Probabilistic structural analysis using a general purpose finite element program
NASA Astrophysics Data System (ADS)
Riha, D. S.; Millwater, H. R.; Thacker, B. H.
1992-07-01
This paper presents an accurate and efficient method to predict the probabilistic response for structural response quantities, such as stress, displacement, natural frequencies, and buckling loads, by combining the capabilities of MSC/NASTRAN, including design sensitivity analysis and fast probability integration. Two probabilistic structural analysis examples have been performed and verified by comparison with Monte Carlo simulation of the analytical solution. The first example consists of a cantilevered plate with several point loads. The second example is a probabilistic buckling analysis of a simply supported composite plate under in-plane loading. The coupling of MSC/NASTRAN and fast probability integration is shown to be orders of magnitude more efficient than Monte Carlo simulation with excellent accuracy.
Probabilistic structural analysis methods for select space propulsion system components
NASA Technical Reports Server (NTRS)
Millwater, H. R.; Cruse, T. A.
1989-01-01
The Probabilistic Structural Analysis Methods (PSAM) project developed at the Southwest Research Institute integrates state-of-the-art structural analysis techniques with probability theory for the design and analysis of complex large-scale engineering structures. An advanced efficient software system (NESSUS) capable of performing complex probabilistic analysis has been developed. NESSUS contains a number of software components to perform probabilistic analysis of structures. These components include: an expert system, a probabilistic finite element code, a probabilistic boundary element code and a fast probability integrator. The NESSUS software system is shown. An expert system is included to capture and utilize PSAM knowledge and experience. NESSUS/EXPERT is an interactive menu-driven expert system that provides information to assist in the use of the probabilistic finite element code NESSUS/FEM and the fast probability integrator (FPI). The expert system menu structure is summarized. The NESSUS system contains a state-of-the-art nonlinear probabilistic finite element code, NESSUS/FEM, to determine the structural response and sensitivities. A broad range of analysis capabilities and an extensive element library is present.
Advanced reliability methods for structural evaluation
NASA Technical Reports Server (NTRS)
Wirsching, P. H.; Wu, Y.-T.
1985-01-01
Fast probability integration (FPI) methods, which can yield approximate solutions to such general structural reliability problems as the computation of the probabilities of complicated functions of random variables, are known to require one-tenth the computer time of Monte Carlo methods for a probability level of 0.001; lower probabilities yield even more dramatic differences. A strategy is presented in which a computer routine is run k times with selected perturbed values of the variables to obtain k solutions for a response variable Y. An approximating polynomial is fit to the k 'data' sets, and FPI methods are employed for this explicit form.
A fast and high performance multiple data integration algorithm for identifying human disease genes
2015-01-01
Background Integrating multiple data sources is indispensable in improving disease gene identification. It is not only due to the fact that disease genes associated with similar genetic diseases tend to lie close with each other in various biological networks, but also due to the fact that gene-disease associations are complex. Although various algorithms have been proposed to identify disease genes, their prediction performances and the computational time still should be further improved. Results In this study, we propose a fast and high performance multiple data integration algorithm for identifying human disease genes. A posterior probability of each candidate gene associated with individual diseases is calculated by using a Bayesian analysis method and a binary logistic regression model. Two prior probability estimation strategies and two feature vector construction methods are developed to test the performance of the proposed algorithm. Conclusions The proposed algorithm is not only generated predictions with high AUC scores, but also runs very fast. When only a single PPI network is employed, the AUC score is 0.769 by using F2 as feature vectors. The average running time for each leave-one-out experiment is only around 1.5 seconds. When three biological networks are integrated, the AUC score using F3 as feature vectors increases to 0.830, and the average running time for each leave-one-out experiment takes only about 12.54 seconds. It is better than many existing algorithms. PMID:26399620
Probabilistic structural analysis methods and applications
NASA Technical Reports Server (NTRS)
Cruse, T. A.; Wu, Y.-T.; Dias, B.; Rajagopal, K. R.
1988-01-01
An advanced algorithm for simulating the probabilistic distribution of structural responses due to statistical uncertainties in loads, geometry, material properties, and boundary conditions is reported. The method effectively combines an advanced algorithm for calculating probability levels for multivariate problems (fast probability integration) together with a general-purpose finite-element code for stress, vibration, and buckling analysis. Application is made to a space propulsion system turbine blade for which the geometry and material properties are treated as random variables.
Probabilistic methods for rotordynamics analysis
NASA Technical Reports Server (NTRS)
Wu, Y.-T.; Torng, T. Y.; Millwater, H. R.; Fossum, A. F.; Rheinfurth, M. H.
1991-01-01
This paper summarizes the development of the methods and a computer program to compute the probability of instability of dynamic systems that can be represented by a system of second-order ordinary linear differential equations. Two instability criteria based upon the eigenvalues or Routh-Hurwitz test functions are investigated. Computational methods based on a fast probability integration concept and an efficient adaptive importance sampling method are proposed to perform efficient probabilistic analysis. A numerical example is provided to demonstrate the methods.
Probability techniques for reliability analysis of composite materials
NASA Technical Reports Server (NTRS)
Wetherhold, Robert C.; Ucci, Anthony M.
1994-01-01
Traditional design approaches for composite materials have employed deterministic criteria for failure analysis. New approaches are required to predict the reliability of composite structures since strengths and stresses may be random variables. This report will examine and compare methods used to evaluate the reliability of composite laminae. The two types of methods that will be evaluated are fast probability integration (FPI) methods and Monte Carlo methods. In these methods, reliability is formulated as the probability that an explicit function of random variables is less than a given constant. Using failure criteria developed for composite materials, a function of design variables can be generated which defines a 'failure surface' in probability space. A number of methods are available to evaluate the integration over the probability space bounded by this surface; this integration delivers the required reliability. The methods which will be evaluated are: the first order, second moment FPI methods; second order, second moment FPI methods; the simple Monte Carlo; and an advanced Monte Carlo technique which utilizes importance sampling. The methods are compared for accuracy, efficiency, and for the conservativism of the reliability estimation. The methodology involved in determining the sensitivity of the reliability estimate to the design variables (strength distributions) and importance factors is also presented.
Structural reliability methods: Code development status
NASA Astrophysics Data System (ADS)
Millwater, Harry R.; Thacker, Ben H.; Wu, Y.-T.; Cruse, T. A.
1991-05-01
The Probabilistic Structures Analysis Method (PSAM) program integrates state of the art probabilistic algorithms with structural analysis methods in order to quantify the behavior of Space Shuttle Main Engine structures subject to uncertain loadings, boundary conditions, material parameters, and geometric conditions. An advanced, efficient probabilistic structural analysis software program, NESSUS (Numerical Evaluation of Stochastic Structures Under Stress) was developed as a deliverable. NESSUS contains a number of integrated software components to perform probabilistic analysis of complex structures. A nonlinear finite element module NESSUS/FEM is used to model the structure and obtain structural sensitivities. Some of the capabilities of NESSUS/FEM are shown. A Fast Probability Integration module NESSUS/FPI estimates the probability given the structural sensitivities. A driver module, PFEM, couples the FEM and FPI. NESSUS, version 5.0, addresses component reliability, resistance, and risk.
Structural reliability methods: Code development status
NASA Technical Reports Server (NTRS)
Millwater, Harry R.; Thacker, Ben H.; Wu, Y.-T.; Cruse, T. A.
1991-01-01
The Probabilistic Structures Analysis Method (PSAM) program integrates state of the art probabilistic algorithms with structural analysis methods in order to quantify the behavior of Space Shuttle Main Engine structures subject to uncertain loadings, boundary conditions, material parameters, and geometric conditions. An advanced, efficient probabilistic structural analysis software program, NESSUS (Numerical Evaluation of Stochastic Structures Under Stress) was developed as a deliverable. NESSUS contains a number of integrated software components to perform probabilistic analysis of complex structures. A nonlinear finite element module NESSUS/FEM is used to model the structure and obtain structural sensitivities. Some of the capabilities of NESSUS/FEM are shown. A Fast Probability Integration module NESSUS/FPI estimates the probability given the structural sensitivities. A driver module, PFEM, couples the FEM and FPI. NESSUS, version 5.0, addresses component reliability, resistance, and risk.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jadaan, O.M.; Powers, L.M.; Nemeth, N.N.
1995-08-01
A probabilistic design methodology which predicts the fast fracture and time-dependent failure behavior of thermomechanically loaded ceramic components is discussed using the CARES/LIFE integrated design computer program. Slow crack growth (SCG) is assumed to be the mechanism responsible for delayed failure behavior. Inert strength and dynamic fatigue data obtained from testing coupon specimens (O-ring and C-ring specimens) are initially used to calculate the fast fracture and SCG material parameters as a function of temperature using the parameter estimation techniques available with the CARES/LIFE code. Finite element analysis (FEA) is used to compute the stress distributions for the tube as amore » function of applied pressure. Knowing the stress and temperature distributions and the fast fracture and SCG material parameters, the life time for a given tube can be computed. A stress-failure probability-time to failure (SPT) diagram is subsequently constructed for these tubes. Such a diagram can be used by design engineers to estimate the time to failure at a given failure probability level for a component subjected to a given thermomechanical load.« less
Fujioka, Shinsuke; Johzaki, Tomoyuki; Arikawa, Yasunobu; Zhang, Zhe; Morace, Alessio; Ikenouchi, Takahito; Ozaki, Tetsuo; Nagai, Takahiro; Abe, Yuki; Kojima, Sadaoki; Sakata, Shohei; Inoue, Hiroaki; Utsugi, Masaru; Hattori, Shoji; Hosoda, Tatsuya; Lee, Seung Ho; Shigemori, Keisuke; Hironaka, Youichiro; Sunahara, Atsushi; Sakagami, Hitoshi; Mima, Kunioki; Fujimoto, Yasushi; Yamanoi, Kohei; Norimatsu, Takayoshi; Tokita, Shigeki; Nakata, Yoshiki; Kawanaka, Junji; Jitsuno, Takahisa; Miyanaga, Noriaki; Nakai, Mitsuo; Nishimura, Hiroaki; Shiraga, Hiroyuki; Nagatomo, Hideo; Azechi, Hiroshi
2015-06-01
A series of experiments were carried out to evaluate the energy-coupling efficiency from heating laser to a fuel core in the fast-ignition scheme of laser-driven inertial confinement fusion. Although the efficiency is determined by a wide variety of complex physics, from intense laser plasma interactions to the properties of high-energy density plasmas and the transport of relativistic electron beams (REB), here we simplify the physics by breaking down the efficiency into three measurable parameters: (i) energy conversion ratio from laser to REB, (ii) probability of collision between the REB and the fusion fuel core, and (iii) fraction of energy deposited in the fuel core from the REB. These three parameters were measured with the newly developed experimental platform designed for mimicking the plasma conditions of a realistic integrated fast-ignition experiment. The experimental results indicate that the high-energy tail of REB must be suppressed to heat the fuel core efficiently.
NASA Astrophysics Data System (ADS)
Ballestra, Luca Vincenzo; Pacelli, Graziella; Radi, Davide
2016-12-01
We propose a numerical method to compute the first-passage probability density function in a time-changed Brownian model. In particular, we derive an integral representation of such a density function in which the integrand functions must be obtained solving a system of Volterra equations of the first kind. In addition, we develop an ad-hoc numerical procedure to regularize and solve this system of integral equations. The proposed method is tested on three application problems of interest in mathematical finance, namely the calculation of the survival probability of an indebted firm, the pricing of a single-knock-out put option and the pricing of a double-knock-out put option. The results obtained reveal that the novel approach is extremely accurate and fast, and performs significantly better than the finite difference method.
A Numerical Round Robin for the Reliability Prediction of Structural Ceramics
NASA Technical Reports Server (NTRS)
Powers, Lynn M.; Janosik, Lesley A.
1993-01-01
A round robin has been conducted on integrated fast fracture design programs for brittle materials. An informal working group (WELFEP-WEakest Link failure probability prediction by Finite Element Postprocessors) was formed to discuss and evaluate the implementation of the programs examined in the study. Results from the study have provided insight on the differences between the various programs examined. Conclusions from the study have shown that when brittle materials are used in design, analysis must understand how to apply the concepts presented herein to failure probability analysis.
Integrated Data Analysis for Fusion: A Bayesian Tutorial for Fusion Diagnosticians
NASA Astrophysics Data System (ADS)
Dinklage, Andreas; Dreier, Heiko; Fischer, Rainer; Gori, Silvio; Preuss, Roland; Toussaint, Udo von
2008-03-01
Integrated Data Analysis (IDA) offers a unified way of combining information relevant to fusion experiments. Thereby, IDA meets with typical issues arising in fusion data analysis. In IDA, all information is consistently formulated as probability density functions quantifying uncertainties in the analysis within the Bayesian probability theory. For a single diagnostic, IDA allows the identification of faulty measurements and improvements in the setup. For a set of diagnostics, IDA gives joint error distributions allowing the comparison and integration of different diagnostics results. Validation of physics models can be performed by model comparison techniques. Typical data analysis applications benefit from IDA capabilities of nonlinear error propagation, the inclusion of systematic effects and the comparison of different physics models. Applications range from outlier detection, background discrimination, model assessment and design of diagnostics. In order to cope with next step fusion device requirements, appropriate techniques are explored for fast analysis applications.
Analytical approach to an integrate-and-fire model with spike-triggered adaptation
NASA Astrophysics Data System (ADS)
Schwalger, Tilo; Lindner, Benjamin
2015-12-01
The calculation of the steady-state probability density for multidimensional stochastic systems that do not obey detailed balance is a difficult problem. Here we present the analytical derivation of the stationary joint and various marginal probability densities for a stochastic neuron model with adaptation current. Our approach assumes weak noise but is valid for arbitrary adaptation strength and time scale. The theory predicts several effects of adaptation on the statistics of the membrane potential of a tonically firing neuron: (i) a membrane potential distribution with a convex shape, (ii) a strongly increased probability of hyperpolarized membrane potentials induced by strong and fast adaptation, and (iii) a maximized variability associated with the adaptation current at a finite adaptation time scale.
Probabilistic Structural Analysis Methods (PSAM) for select space propulsion system components
NASA Technical Reports Server (NTRS)
1991-01-01
This annual report summarizes the work completed during the third year of technical effort on the referenced contract. Principal developments continue to focus on the Probabilistic Finite Element Method (PFEM) which has been under development for three years. Essentially all of the linear capabilities within the PFEM code are in place. Major progress in the application or verifications phase was achieved. An EXPERT module architecture was designed and partially implemented. EXPERT is a user interface module which incorporates an expert system shell for the implementation of a rule-based interface utilizing the experience and expertise of the user community. The Fast Probability Integration (FPI) Algorithm continues to demonstrate outstanding performance characteristics for the integration of probability density functions for multiple variables. Additionally, an enhanced Monte Carlo simulation algorithm was developed and demonstrated for a variety of numerical strategies.
MacRoy-Higgins, Michelle; Dalton, Kevin Patrick
2015-12-01
The purpose of this study was to examine the influence of phonotactic probability on sublexical (phonological) and lexical representations in 3-year-olds who had a history of being late talkers in comparison with their peers with typical language development. Ten 3-year-olds who were late talkers and 10 age-matched typically developing controls completed nonword repetition and fast mapping tasks; stimuli for both experimental procedures differed in phonotactic probability. Both participant groups repeated nonwords containing high phonotactic probability sequences more accurately than nonwords containing low phonotactic probability sequences. Participants with typical language showed an early advantage for fast mapping high phonotactic probability words; children who were late talkers required more exposures to the novel words to show the same advantage for fast mapping high phonotactic probability words. Children who were late talkers showed similar sensitivities to phonotactic probability in nonword repetition and word learning when compared with their peers with no history of language delay. However, word learning in children who were late talkers appeared to be slower when compared with their peers.
NESSUS/EXPERT - An expert system for probabilistic structural analysis methods
NASA Technical Reports Server (NTRS)
Millwater, H.; Palmer, K.; Fink, P.
1988-01-01
An expert system (NESSUS/EXPERT) is presented which provides assistance in using probabilistic structural analysis methods. NESSUS/EXPERT is an interactive menu-driven expert system that provides information to assist in the use of the probabilistic finite element code NESSUS/FEM and the fast probability integrator. NESSUS/EXPERT was developed with a combination of FORTRAN and CLIPS, a C language expert system tool, to exploit the strengths of each language.
Patterns of fast synaptic cholinergic activation of neurons in the celiac ganglia of cats.
Niel, J P; Clerc, N; Jule, Y
1988-12-01
Fast nicotinic transmission was studied in vitro in neurons of isolated cat celiac ganglia. In the absence of nerve stimulation, neurons could be classified into three types: silent neurons, synaptically activated neurons, and spontaneously discharging neurons. In all three types, fast synaptic activation could be obtained in single neurons by stimulating with a single pulse both the splanchnic nerves or one of the peripheral nerves connected to the ganglia. During repetitive nerve stimulation, a gradual depression of the central and peripheral fast nicotinic activation occurred, which was not affected by phentolamine plus propranolol, domperidone, atropine, or naloxone. Repetitive nerve stimulation was followed by a long lasting discharge of excitatory postsynaptic potentials and action potentials that decreased gradually with time. This discharge, which was probably due to presynaptic or prejunctional facilitation of acetylcholine release from cholinergic terminals, was reduced by the application of phentolamine plus propranolol, domperidone, or atropine and increased with naloxone. The existence of the mechanisms described in this study reflects the complexity of the integrative processes at work in neurons of the cat celiac ganglia that involve fast synaptic cholinergic activation.
NASA Technical Reports Server (NTRS)
1981-01-01
A hardware integrated convolutional coding/symbol interleaving and integrated symbol deinterleaving/Viterbi decoding simulation system is described. Validation on the system of the performance of the TDRSS S-band return link with BPSK modulation, operating in a pulsed RFI environment is included. The system consists of three components, the Fast Linkabit Error Rate Tester (FLERT), the Transition Probability Generator (TPG), and a modified LV7017B which includes rate 1/3 capability as well as a periodic interleaver/deinterleaver. Operating and maintenance manuals for each of these units are included.
Fast numerics for the spin orbit equation with realistic tidal dissipation and constant eccentricity
NASA Astrophysics Data System (ADS)
Bartuccelli, Michele; Deane, Jonathan; Gentile, Guido
2017-08-01
We present an algorithm for the rapid numerical integration of a time-periodic ODE with a small dissipation term that is C^1 in the velocity. Such an ODE arises as a model of spin-orbit coupling in a star/planet system, and the motivation for devising a fast algorithm for its solution comes from the desire to estimate probability of capture in various solutions, via Monte Carlo simulation: the integration times are very long, since we are interested in phenomena occurring on timescales of the order of 10^6-10^7 years. The proposed algorithm is based on the high-order Euler method which was described in Bartuccelli et al. (Celest Mech Dyn Astron 121(3):233-260, 2015), and it requires computer algebra to set up the code for its implementation. The payoff is an overall increase in speed by a factor of about 7.5 compared to standard numerical methods. Means for accelerating the purely numerical computation are also discussed.
Least-rattling feedback from strong time-scale separation
NASA Astrophysics Data System (ADS)
Chvykov, Pavel; England, Jeremy
2018-03-01
In most interacting many-body systems associated with some "emergent phenomena," we can identify subgroups of degrees of freedom that relax on dramatically different time scales. Time-scale separation of this kind is particularly helpful in nonequilibrium systems where only the fast variables are subjected to external driving; in such a case, it may be shown through elimination of fast variables that the slow coordinates effectively experience a thermal bath of spatially varying temperature. In this paper, we investigate how such a temperature landscape arises according to how the slow variables affect the character of the driven quasisteady state reached by the fast variables. Brownian motion in the presence of spatial temperature gradients is known to lead to the accumulation of probability density in low-temperature regions. Here, we focus on the implications of attraction to low effective temperature for the long-term evolution of slow variables. After quantitatively deriving the temperature landscape for a general class of overdamped systems using a path-integral technique, we then illustrate in a simple dynamical system how the attraction to low effective temperature has a fine-tuning effect on the slow variable, selecting configurations that bring about exceptionally low force fluctuation in the fast-variable steady state. We furthermore demonstrate that a particularly strong effect of this kind can take place when the slow variable is tuned to bring about orderly, integrable motion in the fast dynamics that avoids thermalizing energy absorbed from the drive. We thus point to a potentially general feedback mechanism in multi-time-scale active systems, that leads to the exploration of slow variable space, as if in search of fine tuning for a "least-rattling" response in the fast coordinates.
Acquired prior knowledge modulates audiovisual integration.
Van Wanrooij, Marc M; Bremen, Peter; John Van Opstal, A
2010-05-01
Orienting responses to audiovisual events in the environment can benefit markedly by the integration of visual and auditory spatial information. However, logically, audiovisual integration would only be considered successful for stimuli that are spatially and temporally aligned, as these would be emitted by a single object in space-time. As humans do not have prior knowledge about whether novel auditory and visual events do indeed emanate from the same object, such information needs to be extracted from a variety of sources. For example, expectation about alignment or misalignment could modulate the strength of multisensory integration. If evidence from previous trials would repeatedly favour aligned audiovisual inputs, the internal state might also assume alignment for the next trial, and hence react to a new audiovisual event as if it were aligned. To test for such a strategy, subjects oriented a head-fixed pointer as fast as possible to a visual flash that was consistently paired, though not always spatially aligned, with a co-occurring broadband sound. We varied the probability of audiovisual alignment between experiments. Reaction times were consistently lower in blocks containing only aligned audiovisual stimuli than in blocks also containing pseudorandomly presented spatially disparate stimuli. Results demonstrate dynamic updating of the subject's prior expectation of audiovisual congruency. We discuss a model of prior probability estimation to explain the results.
Open-loop frequency acquisition for suppressed-carrier biphase signals using one-pole arm filters
NASA Technical Reports Server (NTRS)
Shah, B.; Holmes, J. K.
1991-01-01
Open loop frequency acquisition performance is discussed for suppressed carrier binary phase shift keyed signals in terms of the probability of detecting the carrier frequency offset when the arms of the Costas loop detector have one pole filters. The approach, which does not require symbol timing, uses fast Fourier transforms (FFTs) to detect the carrier frequency offset. The detection probability, which depends on both the 3 dB arm filter bandwidth and the received symbol signal to noise ratio, is derived and is shown to be independent of symbol timing. It is shown that the performance of this technique is slightly better that other open loop acquisition techniques which use integrators in the arms and whose detection performance varies with symbol timing.
Space Object Collision Probability via Monte Carlo on the Graphics Processing Unit
NASA Astrophysics Data System (ADS)
Vittaldev, Vivek; Russell, Ryan P.
2017-09-01
Fast and accurate collision probability computations are essential for protecting space assets. Monte Carlo (MC) simulation is the most accurate but computationally intensive method. A Graphics Processing Unit (GPU) is used to parallelize the computation and reduce the overall runtime. Using MC techniques to compute the collision probability is common in literature as the benchmark. An optimized implementation on the GPU, however, is a challenging problem and is the main focus of the current work. The MC simulation takes samples from the uncertainty distributions of the Resident Space Objects (RSOs) at any time during a time window of interest and outputs the separations at closest approach. Therefore, any uncertainty propagation method may be used and the collision probability is automatically computed as a function of RSO collision radii. Integration using a fixed time step and a quartic interpolation after every Runge Kutta step ensures that no close approaches are missed. Two orders of magnitude speedups over a serial CPU implementation are shown, and speedups improve moderately with higher fidelity dynamics. The tool makes the MC approach tractable on a single workstation, and can be used as a final product, or for verifying surrogate and analytical collision probability methods.
Noorbaloochi, Sharareh; Sharon, Dahlia; McClelland, James L
2015-08-05
We used electroencephalography (EEG) and behavior to examine the role of payoff bias in a difficult two-alternative perceptual decision under deadline pressure in humans. The findings suggest that a fast guess process, biased by payoff and triggered by stimulus onset, occurred on a subset of trials and raced with an evidence accumulation process informed by stimulus information. On each trial, the participant judged whether a rectangle was shifted to the right or left and responded by squeezing a right- or left-hand dynamometer. The payoff for each alternative (which could be biased or unbiased) was signaled 1.5 s before stimulus onset. The choice response was assigned to the first hand reaching a squeeze force criterion and reaction time was defined as time to criterion. Consistent with a fast guess account, fast responses were strongly biased toward the higher-paying alternative and the EEG exhibited an abrupt rise in the lateralized readiness potential (LRP) on a subset of biased payoff trials contralateral to the higher-paying alternative ∼ 150 ms after stimulus onset and 50 ms before stimulus information influenced the LRP. This rise was associated with poststimulus dynamometer activity favoring the higher-paying alternative and predicted choice and response time. Quantitative modeling supported the fast guess account over accounts of payoff effects supported in other studies. Our findings, taken with previous studies, support the idea that payoff and prior probability manipulations produce flexible adaptations to task structure and do not reflect a fixed policy for the integration of payoff and stimulus information. Humans and other animals often face situations in which they must make choices based on uncertain sensory information together with information about expected outcomes (gains or losses) about each choice. We investigated how differences in payoffs between available alternatives affect neural activity, overt choice, and the timing of choice responses. In our experiment, in which participants were under strong time pressure, neural and behavioral findings together with model fitting suggested that our human participants often made a fast guess toward the higher reward rather than integrating stimulus and payoff information. Our findings, taken with findings from other studies, support the idea that payoff and prior probability manipulations produce flexible adaptations to task structure and do not reflect a fixed policy. Copyright © 2015 the authors 0270-6474/15/3510989-23$15.00/0.
Ceramic component reliability with the restructured NASA/CARES computer program
NASA Technical Reports Server (NTRS)
Powers, Lynn M.; Starlinger, Alois; Gyekenyesi, John P.
1992-01-01
The Ceramics Analysis and Reliability Evaluation of Structures (CARES) integrated design program on statistical fast fracture reliability and monolithic ceramic components is enhanced to include the use of a neutral data base, two-dimensional modeling, and variable problem size. The data base allows for the efficient transfer of element stresses, temperatures, and volumes/areas from the finite element output to the reliability analysis program. Elements are divided to insure a direct correspondence between the subelements and the Gaussian integration points. Two-dimensional modeling is accomplished by assessing the volume flaw reliability with shell elements. To demonstrate the improvements in the algorithm, example problems are selected from a round-robin conducted by WELFEP (WEakest Link failure probability prediction by Finite Element Postprocessors).
Endocrine response to fasting in the overwintering captive raccoon dog (Nyctereutes procyonoides).
Nieminen, Petteri; Saarela, Seppo; Pyykönen, Teija; Asikainen, Juha; Mononen, Jaakko; Mustonen, Anne-Mari
2004-12-01
The raccoon dog (Nyctereutes procyonoides) is an omnivorous canid utilizing the passive wintering strategy in the boreal climate. Farmed raccoon dogs (n=12) were randomly assigned into two study groups on 26 November 2003. Between 3 December 2003 and 27 January 2004, half of the animals were fasted for 8 weeks and plasma weight-regulatory hormone concentrations determined on 26 November and 30 December 2003 and on 27 January 2004. The plasma peptide YY, ghrelin, and growth hormone (GH) concentrations increased due to food deprivation, while the T4 and Acrp30 concentrations decreased. Furthermore, the plasma GH concentrations were higher in the fasted raccoon dogs than in the fed animals, which had higher plasma insulin, glucagon, and T4 concentrations. However, fasting had no effect on the plasma leptin concentrations. The results confirm previous findings with unchanged leptin levels in fasting carnivores. Increased GH levels probably contribute to increased lipolysis and mobilization of fat stores. Ghrelin can also enhance lipolysis by increasing the GH levels. The decreased levels of T4 may reduce the metabolic rate. The plasma dopamine concentrations decreased due to fasting unlike observed previously in rats. Together with the unaffected adrenaline, noradrenaline, and cortisol concentrations, this suggests that food deprivation in winter does not cause stress to the raccoon dog but is an integral part of its natural life history.
Bayesian soft X-ray tomography using non-stationary Gaussian Processes
NASA Astrophysics Data System (ADS)
Li, Dong; Svensson, J.; Thomsen, H.; Medina, F.; Werner, A.; Wolf, R.
2013-08-01
In this study, a Bayesian based non-stationary Gaussian Process (GP) method for the inference of soft X-ray emissivity distribution along with its associated uncertainties has been developed. For the investigation of equilibrium condition and fast magnetohydrodynamic behaviors in nuclear fusion plasmas, it is of importance to infer, especially in the plasma center, spatially resolved soft X-ray profiles from a limited number of noisy line integral measurements. For this ill-posed inversion problem, Bayesian probability theory can provide a posterior probability distribution over all possible solutions under given model assumptions. Specifically, the use of a non-stationary GP to model the emission allows the model to adapt to the varying length scales of the underlying diffusion process. In contrast to other conventional methods, the prior regularization is realized in a probability form which enhances the capability of uncertainty analysis, in consequence, scientists who concern the reliability of their results will benefit from it. Under the assumption of normally distributed noise, the posterior distribution evaluated at a discrete number of points becomes a multivariate normal distribution whose mean and covariance are analytically available, making inversions and calculation of uncertainty fast. Additionally, the hyper-parameters embedded in the model assumption can be optimized through a Bayesian Occam's Razor formalism and thereby automatically adjust the model complexity. This method is shown to produce convincing reconstructions and good agreements with independently calculated results from the Maximum Entropy and Equilibrium-Based Iterative Tomography Algorithm methods.
Bayesian soft X-ray tomography using non-stationary Gaussian Processes.
Li, Dong; Svensson, J; Thomsen, H; Medina, F; Werner, A; Wolf, R
2013-08-01
In this study, a Bayesian based non-stationary Gaussian Process (GP) method for the inference of soft X-ray emissivity distribution along with its associated uncertainties has been developed. For the investigation of equilibrium condition and fast magnetohydrodynamic behaviors in nuclear fusion plasmas, it is of importance to infer, especially in the plasma center, spatially resolved soft X-ray profiles from a limited number of noisy line integral measurements. For this ill-posed inversion problem, Bayesian probability theory can provide a posterior probability distribution over all possible solutions under given model assumptions. Specifically, the use of a non-stationary GP to model the emission allows the model to adapt to the varying length scales of the underlying diffusion process. In contrast to other conventional methods, the prior regularization is realized in a probability form which enhances the capability of uncertainty analysis, in consequence, scientists who concern the reliability of their results will benefit from it. Under the assumption of normally distributed noise, the posterior distribution evaluated at a discrete number of points becomes a multivariate normal distribution whose mean and covariance are analytically available, making inversions and calculation of uncertainty fast. Additionally, the hyper-parameters embedded in the model assumption can be optimized through a Bayesian Occam's Razor formalism and thereby automatically adjust the model complexity. This method is shown to produce convincing reconstructions and good agreements with independently calculated results from the Maximum Entropy and Equilibrium-Based Iterative Tomography Algorithm methods.
NASA Astrophysics Data System (ADS)
Chen, Fan; Huang, Shaoxiong; Ding, Jinjin; Ding, Jinjin; Gao, Bo; Xie, Yuguang; Wang, Xiaoming
2018-01-01
This paper proposes a fast reliability assessing method for distribution grid with distributed renewable energy generation. First, the Weibull distribution and the Beta distribution are used to describe the probability distribution characteristics of wind speed and solar irradiance respectively, and the models of wind farm, solar park and local load are built for reliability assessment. Then based on power system production cost simulation probability discretization and linearization power flow, a optimal power flow objected with minimum cost of conventional power generation is to be resolved. Thus a reliability assessment for distribution grid is implemented fast and accurately. The Loss Of Load Probability (LOLP) and Expected Energy Not Supplied (EENS) are selected as the reliability index, a simulation for IEEE RBTS BUS6 system in MATLAB indicates that the fast reliability assessing method calculates the reliability index much faster with the accuracy ensured when compared with Monte Carlo method.
NASA Technical Reports Server (NTRS)
Townsend, J.; Meyers, C.; Ortega, R.; Peck, J.; Rheinfurth, M.; Weinstock, B.
1993-01-01
Probabilistic structural analyses and design methods are steadily gaining acceptance within the aerospace industry. The safety factor approach to design has long been the industry standard, and it is believed by many to be overly conservative and thus, costly. A probabilistic approach to design may offer substantial cost savings. This report summarizes several probabilistic approaches: the probabilistic failure analysis (PFA) methodology developed by Jet Propulsion Laboratory, fast probability integration (FPI) methods, the NESSUS finite element code, and response surface methods. Example problems are provided to help identify the advantages and disadvantages of each method.
An approximate methods approach to probabilistic structural analysis
NASA Technical Reports Server (NTRS)
Mcclung, R. C.; Millwater, H. R.; Wu, Y.-T.; Thacker, B. H.; Burnside, O. H.
1989-01-01
A probabilistic structural analysis method (PSAM) is described which makes an approximate calculation of the structural response of a system, including the associated probabilistic distributions, with minimal computation time and cost, based on a simplified representation of the geometry, loads, and material. The method employs the fast probability integration (FPI) algorithm of Wu and Wirsching. Typical solution strategies are illustrated by formulations for a representative critical component chosen from the Space Shuttle Main Engine (SSME) as part of a major NASA-sponsored program on PSAM. Typical results are presented to demonstrate the role of the methodology in engineering design and analysis.
NASA Astrophysics Data System (ADS)
Ducci, L.; Sidoli, L.; Paizis, A.
2010-11-01
We performed a systematic analysis of all INTEGRAL observations from 2003 to 2009 of 14 supergiant fast X-ray transients (SFXTs), implying a net exposure time of about 30 Ms. For each source we obtained light curves and spectra (3-100 keV), discovering several new outbursts. We discuss the X-ray behaviour of SFXTs emerging from our analysis in the framework of the clumpy wind accretion mechanism we proposed. We discuss the effect of X-ray photoionization on accretion in close binary systems such as IGR J16479-4514 and IGR J17544-2619. We show that, because of X-ray photoionization, there is a high probability of an accretion disc forming from the capture of angular momentum in IGR J16479-4514, and we suggest that the formation of transient accretion discs could be partly responsible for the flaring activity in SFXTs with narrow orbits. We also propose an alternative way to explain the origin of flares with peculiar shapes observed in our analysis applying the model of Lamb et al., which is based on accretion via the Rayleigh-Taylor instability and was originally proposed to explain Type II bursts.
Quantifying Uncertainties in the Thermo-Mechanical Properties of Particulate Reinforced Composites
NASA Technical Reports Server (NTRS)
Mital, Subodh K.; Murthy, Pappu L. N.
1999-01-01
The present paper reports results from a computational simulation of probabilistic particulate reinforced composite behavior. The approach consists use of simplified micromechanics of particulate reinforced composites together with a Fast Probability Integration (FPI) technique. Sample results are presented for a Al/SiC(sub p)(silicon carbide particles in aluminum matrix) composite. The probability density functions for composite moduli, thermal expansion coefficient and thermal conductivities along with their sensitivity factors are computed. The effect of different assumed distributions and the effect of reducing scatter in constituent properties on the thermal expansion coefficient are also evaluated. The variations in the constituent properties that directly effect these composite properties are accounted for by assumed probabilistic distributions. The results show that the present technique provides valuable information about the scatter in composite properties and sensitivity factors, which are useful to test or design engineers.
Accurate hybrid stochastic simulation of a system of coupled chemical or biochemical reactions.
Salis, Howard; Kaznessis, Yiannis
2005-02-01
The dynamical solution of a well-mixed, nonlinear stochastic chemical kinetic system, described by the Master equation, may be exactly computed using the stochastic simulation algorithm. However, because the computational cost scales with the number of reaction occurrences, systems with one or more "fast" reactions become costly to simulate. This paper describes a hybrid stochastic method that partitions the system into subsets of fast and slow reactions, approximates the fast reactions as a continuous Markov process, using a chemical Langevin equation, and accurately describes the slow dynamics using the integral form of the "Next Reaction" variant of the stochastic simulation algorithm. The key innovation of this method is its mechanism of efficiently monitoring the occurrences of slow, discrete events while simultaneously simulating the dynamics of a continuous, stochastic or deterministic process. In addition, by introducing an approximation in which multiple slow reactions may occur within a time step of the numerical integration of the chemical Langevin equation, the hybrid stochastic method performs much faster with only a marginal decrease in accuracy. Multiple examples, including a biological pulse generator and a large-scale system benchmark, are simulated using the exact and proposed hybrid methods as well as, for comparison, a previous hybrid stochastic method. Probability distributions of the solutions are compared and the weak errors of the first two moments are computed. In general, these hybrid methods may be applied to the simulation of the dynamics of a system described by stochastic differential, ordinary differential, and Master equations.
A rapid local singularity analysis algorithm with applications
NASA Astrophysics Data System (ADS)
Chen, Zhijun; Cheng, Qiuming; Agterberg, Frits
2015-04-01
The local singularity model developed by Cheng is fast gaining popularity in characterizing mineralization and detecting anomalies of geochemical, geophysical and remote sensing data. However in one of the conventional algorithms involving the moving average values with different scales is time-consuming especially while analyzing a large dataset. Summed area table (SAT), also called as integral image, is a fast algorithm used within the Viola-Jones object detection framework in computer vision area. Historically, the principle of SAT is well-known in the study of multi-dimensional probability distribution functions, namely in computing 2D (or ND) probabilities (area under the probability distribution) from the respective cumulative distribution functions. We introduce SAT and it's variation Rotated Summed Area Table in the isotropic, anisotropic or directional local singularity mapping in this study. Once computed using SAT, any one of the rectangular sum can be computed at any scale or location in constant time. The area for any rectangular region in the image can be computed by using only 4 array accesses in constant time independently of the size of the region; effectively reducing the time complexity from O(n) to O(1). New programs using Python, Julia, matlab and C++ are implemented respectively to satisfy different applications, especially to the big data analysis. Several large geochemical and remote sensing datasets are tested. A wide variety of scale changes (linear spacing or log spacing) for non-iterative or iterative approach are adopted to calculate the singularity index values and compare the results. The results indicate that the local singularity analysis with SAT is more robust and superior to traditional approach in identifying anomalies.
Fast and accurate computation of projected two-point functions
NASA Astrophysics Data System (ADS)
Grasshorn Gebhardt, Henry S.; Jeong, Donghui
2018-01-01
We present the two-point function from the fast and accurate spherical Bessel transformation (2-FAST) algorithm
A fast elitism Gaussian estimation of distribution algorithm and application for PID optimization.
Xu, Qingyang; Zhang, Chengjin; Zhang, Li
2014-01-01
Estimation of distribution algorithm (EDA) is an intelligent optimization algorithm based on the probability statistics theory. A fast elitism Gaussian estimation of distribution algorithm (FEGEDA) is proposed in this paper. The Gaussian probability model is used to model the solution distribution. The parameters of Gaussian come from the statistical information of the best individuals by fast learning rule. A fast learning rule is used to enhance the efficiency of the algorithm, and an elitism strategy is used to maintain the convergent performance. The performances of the algorithm are examined based upon several benchmarks. In the simulations, a one-dimensional benchmark is used to visualize the optimization process and probability model learning process during the evolution, and several two-dimensional and higher dimensional benchmarks are used to testify the performance of FEGEDA. The experimental results indicate the capability of FEGEDA, especially in the higher dimensional problems, and the FEGEDA exhibits a better performance than some other algorithms and EDAs. Finally, FEGEDA is used in PID controller optimization of PMSM and compared with the classical-PID and GA.
A Fast Elitism Gaussian Estimation of Distribution Algorithm and Application for PID Optimization
Xu, Qingyang; Zhang, Chengjin; Zhang, Li
2014-01-01
Estimation of distribution algorithm (EDA) is an intelligent optimization algorithm based on the probability statistics theory. A fast elitism Gaussian estimation of distribution algorithm (FEGEDA) is proposed in this paper. The Gaussian probability model is used to model the solution distribution. The parameters of Gaussian come from the statistical information of the best individuals by fast learning rule. A fast learning rule is used to enhance the efficiency of the algorithm, and an elitism strategy is used to maintain the convergent performance. The performances of the algorithm are examined based upon several benchmarks. In the simulations, a one-dimensional benchmark is used to visualize the optimization process and probability model learning process during the evolution, and several two-dimensional and higher dimensional benchmarks are used to testify the performance of FEGEDA. The experimental results indicate the capability of FEGEDA, especially in the higher dimensional problems, and the FEGEDA exhibits a better performance than some other algorithms and EDAs. Finally, FEGEDA is used in PID controller optimization of PMSM and compared with the classical-PID and GA. PMID:24892059
Powell, Lisa M; Auld, M Christopher; Chaloupka, Frank J; O'Malley, Patrick M; Johnston, Lloyd D
2007-01-01
We examine the extent to which food prices and restaurant outlet density are associated with adolescent fruit and vegetable consumption, body mass index (BMI), and the probability of overweight. We use repeated cross-sections of individual-level data on adolescents from the Monitoring the Future Surveys from 1997 to 2003 combined with fast food and fruit and vegetable prices obtained from the American Chamber of Commerce Researchers Association and fast food and full-service restaurant outlet density measures obtained from Dun & Bradstreet. The results suggest that the price of a fast food meal is an important determinant of adolescents' body weight and eating habits: a 10% increase in the price of a fast food meal leads to a 3.0% increase in the probability of frequent fruit and vegetable consumption, a 0.4% decrease in BMI, and a 5.9% decrease in probability of overweight. The price of fruits and vegetables and restaurant outlet density are less important determinants, although these variables typically have the expected sign and are often statistically associated with our outcome measures. Despite these findings, changes in all observed economic and socio-demographic characteristics together only explain roughly one-quarter of the change in mean BMI and one-fifth of the change in overweight over the 1997-2003 sampling period.
Dynamics of a Landau-Zener non-dissipative system with fluctuating energy levels
NASA Astrophysics Data System (ADS)
Fai, L. C.; Diffo, J. T.; Ateuafack, M. E.; Tchoffo, M.; Fouokeng, G. C.
2014-12-01
This paper considers a Landau-Zener (two-level) system influenced by a three-dimensional Gaussian and non-Gaussian coloured noise and finds a general form of the time dependent diabatic quantum bit (qubit) flip transition probabilities in the fast, intermediate and slow noise limits. The qubit flip probability is observed to mimic (for low-frequencies noise) that of the standard LZ problem. The qubit flip probability is also observed to be the measure of quantum coherence of states. The transition probability is observed to be tailored by non-Gaussian low-frequency noise and otherwise by Gaussian low-frequency coloured noise. Intermediate and fast noise limits are observed to alter the memory of the system in time and found to improve and control quantum information processing.
... back or joint pain; widening of the pupils; irritability; anxiety; weakness; stomach cramps; difficulty falling asleep or staying asleep; nausea; loss of appetite; vomiting; diarrhea; fast breathing; or fast heartbeat. Your doctor will probably decrease your dose gradually.
... back or joint pain; widening of the pupils; irritability; anxiety; weakness; stomach cramps; difficulty falling asleep or staying asleep; nausea; loss of appetite; vomiting; diarrhea; fast breathing; or fast heartbeat. Your doctor will probably decrease your dose gradually.
The framing effect and skin conductance responses.
Ring, Patrick
2015-01-01
Individuals often rely on simple heuristics when they face complex choice situations under uncertainty. Traditionally, it has been proposed that cognitive processes are the main driver to evaluate different choice options and to finally reach a decision. Growing evidence, however, highlights a strong interrelation between judgment and decision-making (JDM) on the one hand, and emotional processes on the other hand. This also seems to apply to judgmental heuristics, i.e., decision processes that are typically considered to be fast and intuitive. In this study, participants are exposed to different probabilities of receiving an unpleasant electric shock. Information about electric shock probabilities is either positively or negatively framed. Integrated skin conductance responses (ISCRs) while waiting for electric shock realization are used as an indicator for participants' emotional arousal. This measure is compared to objective probabilities. I find evidence for a relation between emotional body reactions measured by ISCRs and the framing effect. Under negative frames, participants show significantly higher ISCRs while waiting for an electric shock to be delivered than under positive frames. This result might contribute to a better understanding of the psychological processes underlying JDM. Further studies are necessary to reveal the causality underlying this finding, i.e., whether emotional processes influence JDM or vice versa.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wixson, J. R.
Root cause analysis (RCA) is an important methodology that can be integrated with the VE Job Plan to generate superior results from the VE Methodology. The point at which RCA is most appropriate is after the function analysis and FAST Model have been built and functions for improvement have been chosen. These functions are then subjected to a simple, but, rigorous RCA to get to the root cause of their deficiencies, whether it is high cost/poor value, poor quality, or poor reliability. Once the most probable causes for these problems have been arrived at, better solutions for improvement can bemore » developed in the creativity phase because the team better understands the problems associated with these functions.« less
A New Method for Generating Probability Tables in the Unresolved Resonance Region
Holcomb, Andrew M.; Leal, Luiz C.; Rahnema, Farzad; ...
2017-04-18
One new method for constructing probability tables in the unresolved resonance region (URR) has been developed. This new methodology is an extensive modification of the single-level Breit-Wigner (SLBW) pseudo-resonance pair sequence method commonly used to generate probability tables in the URR. The new method uses a Monte Carlo process to generate many pseudo-resonance sequences by first sampling the average resonance parameter data in the URR and then converting the sampled resonance parameters to the more robust R-matrix limited (RML) format. Furthermore, for each sampled set of pseudo-resonance sequences, the temperature-dependent cross sections are reconstructed on a small grid around themore » energy of reference using the Reich-Moore formalism and the Leal-Hwang Doppler broadening methodology. We then use the effective cross sections calculated at the energies of reference to construct probability tables in the URR. The RML cross-section reconstruction algorithm has been rigorously tested for a variety of isotopes, including 16O, 19F, 35Cl, 56Fe, 63Cu, and 65Cu. The new URR method also produced normalized cross-section factor probability tables for 238U that were found to be in agreement with current standards. The modified 238U probability tables were shown to produce results in excellent agreement with several standard benchmarks, including the IEU-MET-FAST-007 (BIG TEN), IEU-MET-FAST-003, and IEU-COMP-FAST-004 benchmarks.« less
DOT National Transportation Integrated Search
1999-02-12
This report focuses on the system integration aspect of the Transportation Information Management System (TIMS), a component of the FAST-TRAC project. The TIMS functions as the center of a communications network which integrates advanced traffic cont...
Response statistics of rotating shaft with non-linear elastic restoring forces by path integration
NASA Astrophysics Data System (ADS)
Gaidai, Oleg; Naess, Arvid; Dimentberg, Michael
2017-07-01
Extreme statistics of random vibrations is studied for a Jeffcott rotor under uniaxial white noise excitation. Restoring force is modelled as elastic non-linear; comparison is done with linearized restoring force to see the force non-linearity effect on the response statistics. While for the linear model analytical solutions and stability conditions are available, it is not generally the case for non-linear system except for some special cases. The statistics of non-linear case is studied by applying path integration (PI) method, which is based on the Markov property of the coupled dynamic system. The Jeffcott rotor response statistics can be obtained by solving the Fokker-Planck (FP) equation of the 4D dynamic system. An efficient implementation of PI algorithm is applied, namely fast Fourier transform (FFT) is used to simulate dynamic system additive noise. The latter allows significantly reduce computational time, compared to the classical PI. Excitation is modelled as Gaussian white noise, however any kind distributed white noise can be implemented with the same PI technique. Also multidirectional Markov noise can be modelled with PI in the same way as unidirectional. PI is accelerated by using Monte Carlo (MC) estimated joint probability density function (PDF) as initial input. Symmetry of dynamic system was utilized to afford higher mesh resolution. Both internal (rotating) and external damping are included in mechanical model of the rotor. The main advantage of using PI rather than MC is that PI offers high accuracy in the probability distribution tail. The latter is of critical importance for e.g. extreme value statistics, system reliability, and first passage probability.
On forward inferences of fast and slow readers. An eye movement study
Hawelka, Stefan; Schuster, Sarah; Gagl, Benjamin; Hutzler, Florian
2015-01-01
Unimpaired readers process words incredibly fast and hence it was assumed that top-down processing, such as predicting upcoming words, would be too slow to play an appreciable role in reading. This runs counter the major postulate of the predictive coding framework that our brain continually predicts probable upcoming sensory events. This means, it may generate predictions about the probable upcoming word during reading (dubbed forward inferences). Trying to asses these contradictory assumptions, we evaluated the effect of the predictability of words in sentences on eye movement control during silent reading. Participants were a group of fluent (i.e., fast) and a group of speed-impaired (i.e., slow) readers. The findings indicate that fast readers generate forward inferences, whereas speed-impaired readers do so to a reduced extent - indicating a significant role of predictive coding for fluent reading. PMID:25678030
DOE Office of Scientific and Technical Information (OSTI.GOV)
Libenson, B. N., E-mail: libenson-b@yandex.ru
2011-10-15
The probability of single characteristic energy loss of a fast electron in a reflection experiment has been calculated. Unlike many works concerning this subject, the bremsstrahlung of bulk plasmons in the non- Cherenkov ranges of frequencies and wavevectors of a plasmon has been taken into account. The contributions to the probability of single loss and to the shape of the spectral line from a quantum correction that is due to the interference of elastic and inelastic electron scattering events have been determined. The probability has been calculated in the kinetic approximation for the relative permittivity, where the short-wavelength range ofmore » the plasmon spectrum is correctly taken into account. In view of these circumstances, the expression for the mean free path of the electron with respect to the emission of a bulk plasmon that was obtained by Pines [D. Pines, Elementary Excitations in Solids (Benjamin, New York, 1963)] has been refined. The coherence length of the fast electron in the medium-energy range under consideration has been estimated. The shape of the spectral line of energy losses in the non-Cherenkov frequency range has been determined. It has been shown that the probability of the single emission of the bulk plasmon incompletely corresponds to the Poisson statistics.« less
Probabilistic Simulation of Progressive Fracture in Bolted-Joint Composite Laminates
NASA Technical Reports Server (NTRS)
Minnetyan, L.; Singhal, S. N.; Chamis, C. C.
1996-01-01
This report describes computational methods to probabilistically simulate fracture in bolted composite structures. An innovative approach that is independent of stress intensity factors and fracture toughness was used to simulate progressive fracture. The effect of design variable uncertainties on structural damage was also quantified. A fast probability integrator assessed the scatter in the composite structure response before and after damage. Then the sensitivity of the response to design variables was computed. General-purpose methods, which are applicable to bolted joints in all types of structures and in all fracture processes-from damage initiation to unstable propagation and global structure collapse-were used. These methods were demonstrated for a bolted joint of a polymer matrix composite panel under edge loads. The effects of the fabrication process were included in the simulation of damage in the bolted panel. Results showed that the most effective way to reduce end displacement at fracture is to control both the load and the ply thickness. The cumulative probability for longitudinal stress in all plies was most sensitive to the load; in the 0 deg. plies it was very sensitive to ply thickness. The cumulative probability for transverse stress was most sensitive to the matrix coefficient of thermal expansion. In addition, fiber volume ratio and fiber transverse modulus both contributed significantly to the cumulative probability for the transverse stresses in all the plies.
Fast rerouting schemes for protected mobile IP over MPLS networks
NASA Astrophysics Data System (ADS)
Wen, Chih-Chao; Chang, Sheng-Yi; Chen, Huan; Chen, Kim-Joan
2005-10-01
Fast rerouting is a critical traffic engineering operation in the MPLS networks. To implement the Mobile IP service over the MPLS network, one can collaborate with the fast rerouting operation to enhance the availability and survivability. MPLS can protect critical LSP tunnel between Home Agent (HA) and Foreign Agent (FA) using the fast rerouting scheme. In this paper, we propose a simple but efficient algorithm to address the triangle routing problem for the Mobile IP over the MPLS networks. We consider this routing issue as a link weighting and capacity assignment (LW-CA) problem. The derived solution is used to plan the fast restoration mechanism to protect the link or node failure. In this paper, we first model the LW-CA problem as a mixed integer optimization problem. Our goal is to minimize the call blocking probability on the most congested working truck for the mobile IP connections. Many existing network topologies are used to evaluate the performance of our scheme. Results show that our proposed scheme can obtain the best performance in terms of the smallest blocking probability compared to other schemes.
How Root Cause Analysis Can Improve the Value Methodology
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wixson, James Robert
2002-05-01
Root cause analysis (RCA) is an important methodology that can be integrated with the VE Job Plan to generate superior results from the VE Methodology. The point at which RCA is most appropriate is after the function analysis and FAST Model have been built and functions for improvement have been chosen. These functions are then subjected to a simple, but, rigorous RCA to get to the root cause of their deficiencies, whether it is high cost/poor value, poor quality, or poor reliability. Once the most probable causes for these problems have been arrived at, better solutions for improvement can bemore » developed in the creativity phase because the team better understands the problems associated with these functions.« less
RS-Forest: A Rapid Density Estimator for Streaming Anomaly Detection.
Wu, Ke; Zhang, Kun; Fan, Wei; Edwards, Andrea; Yu, Philip S
Anomaly detection in streaming data is of high interest in numerous application domains. In this paper, we propose a novel one-class semi-supervised algorithm to detect anomalies in streaming data. Underlying the algorithm is a fast and accurate density estimator implemented by multiple fully randomized space trees (RS-Trees), named RS-Forest. The piecewise constant density estimate of each RS-tree is defined on the tree node into which an instance falls. Each incoming instance in a data stream is scored by the density estimates averaged over all trees in the forest. Two strategies, statistical attribute range estimation of high probability guarantee and dual node profiles for rapid model update, are seamlessly integrated into RS-Forest to systematically address the ever-evolving nature of data streams. We derive the theoretical upper bound for the proposed algorithm and analyze its asymptotic properties via bias-variance decomposition. Empirical comparisons to the state-of-the-art methods on multiple benchmark datasets demonstrate that the proposed method features high detection rate, fast response, and insensitivity to most of the parameter settings. Algorithm implementations and datasets are available upon request.
RS-Forest: A Rapid Density Estimator for Streaming Anomaly Detection
Wu, Ke; Zhang, Kun; Fan, Wei; Edwards, Andrea; Yu, Philip S.
2015-01-01
Anomaly detection in streaming data is of high interest in numerous application domains. In this paper, we propose a novel one-class semi-supervised algorithm to detect anomalies in streaming data. Underlying the algorithm is a fast and accurate density estimator implemented by multiple fully randomized space trees (RS-Trees), named RS-Forest. The piecewise constant density estimate of each RS-tree is defined on the tree node into which an instance falls. Each incoming instance in a data stream is scored by the density estimates averaged over all trees in the forest. Two strategies, statistical attribute range estimation of high probability guarantee and dual node profiles for rapid model update, are seamlessly integrated into RS-Forest to systematically address the ever-evolving nature of data streams. We derive the theoretical upper bound for the proposed algorithm and analyze its asymptotic properties via bias-variance decomposition. Empirical comparisons to the state-of-the-art methods on multiple benchmark datasets demonstrate that the proposed method features high detection rate, fast response, and insensitivity to most of the parameter settings. Algorithm implementations and datasets are available upon request. PMID:25685112
The framing effect and skin conductance responses
Ring, Patrick
2015-01-01
Individuals often rely on simple heuristics when they face complex choice situations under uncertainty. Traditionally, it has been proposed that cognitive processes are the main driver to evaluate different choice options and to finally reach a decision. Growing evidence, however, highlights a strong interrelation between judgment and decision-making (JDM) on the one hand, and emotional processes on the other hand. This also seems to apply to judgmental heuristics, i.e., decision processes that are typically considered to be fast and intuitive. In this study, participants are exposed to different probabilities of receiving an unpleasant electric shock. Information about electric shock probabilities is either positively or negatively framed. Integrated skin conductance responses (ISCRs) while waiting for electric shock realization are used as an indicator for participants' emotional arousal. This measure is compared to objective probabilities. I find evidence for a relation between emotional body reactions measured by ISCRs and the framing effect. Under negative frames, participants show significantly higher ISCRs while waiting for an electric shock to be delivered than under positive frames. This result might contribute to a better understanding of the psychological processes underlying JDM. Further studies are necessary to reveal the causality underlying this finding, i.e., whether emotional processes influence JDM or vice versa. PMID:26300747
2016-09-01
par. 4) Based on a RED projected size of 22.16 m, a sample calculation for the unadjusted single shot probability of kill for HELLFIRE missiles is...framework based on intelligent objects (SIMIO) environment to model a fast attack craft/fast inshore attack craft anti-surface warfare expanded kill chain...concept of operation efficiency. Based on the operational environment, low cost and less capable unmanned aircraft provide an alternative to the
FAST-PT: a novel algorithm to calculate convolution integrals in cosmological perturbation theory
DOE Office of Scientific and Technical Information (OSTI.GOV)
McEwen, Joseph E.; Fang, Xiao; Hirata, Christopher M.
2016-09-01
We present a novel algorithm, FAST-PT, for performing convolution or mode-coupling integrals that appear in nonlinear cosmological perturbation theory. The algorithm uses several properties of gravitational structure formation—the locality of the dark matter equations and the scale invariance of the problem—as well as Fast Fourier Transforms to describe the input power spectrum as a superposition of power laws. This yields extremely fast performance, enabling mode-coupling integral computations fast enough to embed in Monte Carlo Markov Chain parameter estimation. We describe the algorithm and demonstrate its application to calculating nonlinear corrections to the matter power spectrum, including one-loop standard perturbation theorymore » and the renormalization group approach. We also describe our public code (in Python) to implement this algorithm. The code, along with a user manual and example implementations, is available at https://github.com/JoeMcEwen/FAST-PT.« less
Recent Results with CVD Diamond Trackers
NASA Astrophysics Data System (ADS)
Adam, W.; Bauer, C.; Berdermann, E.; Bergonzo, P.; Bogani, F.; Borchi, E.; Brambilla, A.; Bruzzi, M.; Colledani, C.; Conway, J.; Dabrowski, W.; Delpierre, P.; Deneuville, A.; Dulinski, W.; van Eijk, B.; Fallou, A.; Fizzotti, F.; Foulon, F.; Friedl, M.; Gan, K. K.; Gheeraert, E.; Grigoriev, E.; Hallewell, G.; Hall-Wilton, R.; Han, S.; Hartjes, F.; Hrubec, J.; Husson, D.; Kagan, H.; Kania, D.; Kaplon, J.; Karl, C.; Kass, R.; Knöpfle, K. T.; Krammer, M.; Logiudice, A.; Lu, R.; Manfredi, P. F.; Manfredotti, C.; Marshall, R. D.; Meier, D.; Mishina, M.; Oh, A.; Pan, L. S.; Palmieri, V. G.; Pernicka, M.; Peitz, A.; Pirollo, S.; Polesello, P.; Pretzl, K.; Procario, M.; Re, V.; Riester, J. L.; Roe, S.; Roff, D.; Rudge, A.; Runolfsson, O.; Russ, J.; Schnetzer, S.; Sciortino, S.; Speziali, V.; Stelzer, H.; Stone, R.; Suter, B.; Tapper, R. J.; Tesarek, R.; Trawick, M.; Trischuk, W.; Vittone, E.; Walsh, A. M.; Wedenig, R.; Weilhammer, P.; White, C.; Ziock, H.; Zoeller, M.; RD42 Collaboration
1999-08-01
We present recent results on the use of Chemical Vapor Deposition (CVD) diamond microstrip detectors for charged particle tracking. A series of detectors was fabricated using 1 x 1 cm 2 diamonds. Good signal-to-noise ratios were observed using both slow and fast readout electronics. For slow readout electronics, 2 μs shaping time, the most probable signal-to-noise ratio was 50 to 1. For fast readout electronics, 25 ns peaking time, the most probable signal-to-noise ratio was 7 to 1. Using the first 2 x 4 cm 2 diamond from a production CVD reactor with slow readout electronics, the most probable signal-to-noise ratio was 23 to 1. The spatial resolution achieved for the detectors was consistent with the digital resolution expected from the detector pitch.
CANCER CONTROL AND POPULATION SCIENCES FAST STATS
Fast Stats links to tables, charts, and graphs of cancer statistics for all major cancer sites by age, sex, race, and geographic area. The statistics include incidence, mortality, prevalence, and the probability of developing or dying from cancer. A large set of statistics is ava...
Efficiency optimization of a fast Poisson solver in beam dynamics simulation
NASA Astrophysics Data System (ADS)
Zheng, Dawei; Pöplau, Gisela; van Rienen, Ursula
2016-01-01
Calculating the solution of Poisson's equation relating to space charge force is still the major time consumption in beam dynamics simulations and calls for further improvement. In this paper, we summarize a classical fast Poisson solver in beam dynamics simulations: the integrated Green's function method. We introduce three optimization steps of the classical Poisson solver routine: using the reduced integrated Green's function instead of the integrated Green's function; using the discrete cosine transform instead of discrete Fourier transform for the Green's function; using a novel fast convolution routine instead of an explicitly zero-padded convolution. The new Poisson solver routine preserves the advantages of fast computation and high accuracy. This provides a fast routine for high performance calculation of the space charge effect in accelerators.
Integrated HI emission in galaxy groups and clusters
NASA Astrophysics Data System (ADS)
Ai, Mei; Zhu, Ming; Fu, Jian
2017-09-01
The integrated HI emission from hierarchical structures such as groups and clusters of galaxies can be detected by FAST at intermediate redshifts. Here we propose to use FAST to study the evolution of the global HI content of clusters and groups over cosmic time by measuring their integrated HI emissions. We use the Virgo Cluster as an example to estimate the detection limit of FAST, and have estimated the integration time to detect a Virgo type cluster at different redshifts (from z = 0.1 to z = 1.5).We have also employed a semi-analytic model (SAM) to simulate the evolution of HI contents in galaxy clusters. Our simulations suggest that the HI mass of a Virgo-like cluster could be 2-3 times higher and the physical size could be more than 50% smaller when redshift increases from z = 0.3 to z = 1. Thus the integration time could be reduced significantly and gas rich clusters at intermediate redshifts can be detected by FAST in less than 2 hours of integration time. For the local Universe, we have also used SAM simulations to create mock catalogs of clusters to predict the outcomes from FAST all sky surveys. Comparing with the optically selected catalogs derived by cross matching the galaxy catalogs from the SDSS survey and the ALFALFA survey, we find that the HI mass distribution of the mock catalog with 20 s of integration time agrees well with that of observations. However, the mock catalog with 120 s of integration time predicts many more groups and clusters that contain a population of low mass HI galaxies not detected by the ALFALFA survey. A future deep HI blind sky survey with FAST would be able to test such prediction and set constraints on the numerical simulation models. The observational strategy and sample selections for future FAST observations of galaxy clusters at high redshifts are also discussed.
Reliability of Radioisotope Stirling Convertor Linear Alternator
NASA Technical Reports Server (NTRS)
Shah, Ashwin; Korovaichuk, Igor; Geng, Steven M.; Schreiber, Jeffrey G.
2006-01-01
Onboard radioisotope power systems being developed and planned for NASA s deep-space missions would require reliable design lifetimes of up to 14 years. Critical components and materials of Stirling convertors have been undergoing extensive testing and evaluation in support of a reliable performance for the specified life span. Of significant importance to the successful development of the Stirling convertor is the design of a lightweight and highly efficient linear alternator. Alternator performance could vary due to small deviations in the permanent magnet properties, operating temperature, and component geometries. Durability prediction and reliability of the alternator may be affected by these deviations from nominal design conditions. Therefore, it is important to evaluate the effect of these uncertainties in predicting the reliability of the linear alternator performance. This paper presents a study in which a reliability-based methodology is used to assess alternator performance. The response surface characterizing the induced open-circuit voltage performance is constructed using 3-D finite element magnetic analysis. Fast probability integration method is used to determine the probability of the desired performance and its sensitivity to the alternator design parameters.
An approximate methods approach to probabilistic structural analysis
NASA Technical Reports Server (NTRS)
Mcclung, R. C.; Millwater, H. R.; Wu, Y.-T.; Thacker, B. H.; Burnside, O. H.
1989-01-01
A major research and technology program in Probabilistic Structural Analysis Methods (PSAM) is currently being sponsored by the NASA Lewis Research Center with Southwest Research Institute as the prime contractor. This program is motivated by the need to accurately predict structural response in an environment where the loadings, the material properties, and even the structure may be considered random. The heart of PSAM is a software package which combines advanced structural analysis codes with a fast probability integration (FPI) algorithm for the efficient calculation of stochastic structural response. The basic idea of PAAM is simple: make an approximate calculation of system response, including calculation of the associated probabilities, with minimal computation time and cost, based on a simplified representation of the geometry, loads, and material. The deterministic solution resulting should give a reasonable and realistic description of performance-limiting system responses, although some error will be inevitable. If the simple model has correctly captured the basic mechanics of the system, however, including the proper functional dependence of stress, frequency, etc. on design parameters, then the response sensitivities calculated may be of significantly higher accuracy.
Probability theory, not the very guide of life.
Juslin, Peter; Nilsson, Håkan; Winman, Anders
2009-10-01
Probability theory has long been taken as the self-evident norm against which to evaluate inductive reasoning, and classical demonstrations of violations of this norm include the conjunction error and base-rate neglect. Many of these phenomena require multiplicative probability integration, whereas people seem more inclined to linear additive integration, in part, at least, because of well-known capacity constraints on controlled thought. In this article, the authors show with computer simulations that when based on approximate knowledge of probabilities, as is routinely the case in natural environments, linear additive integration can yield as accurate estimates, and as good average decision returns, as estimates based on probability theory. It is proposed that in natural environments people have little opportunity or incentive to induce the normative rules of probability theory and, given their cognitive constraints, linear additive integration may often offer superior bounded rationality.
Time Evolving Fission Chain Theory and Fast Neutron and Gamma-Ray Counting Distributions
Kim, K. S.; Nakae, L. F.; Prasad, M. K.; ...
2015-11-01
Here, we solve a simple theoretical model of time evolving fission chains due to Feynman that generalizes and asymptotically approaches the point model theory. The point model theory has been used to analyze thermal neutron counting data. This extension of the theory underlies fast counting data for both neutrons and gamma rays from metal systems. Fast neutron and gamma-ray counting is now possible using liquid scintillator arrays with nanosecond time resolution. For individual fission chains, the differential equations describing three correlated probability distributions are solved: the time-dependent internal neutron population, accumulation of fissions in time, and accumulation of leaked neutronsmore » in time. Explicit analytic formulas are given for correlated moments of the time evolving chain populations. The equations for random time gate fast neutron and gamma-ray counting distributions, due to randomly initiated chains, are presented. Correlated moment equations are given for both random time gate and triggered time gate counting. There are explicit formulas for all correlated moments are given up to triple order, for all combinations of correlated fast neutrons and gamma rays. The nonlinear differential equations for probabilities for time dependent fission chain populations have a remarkably simple Monte Carlo realization. A Monte Carlo code was developed for this theory and is shown to statistically realize the solutions to the fission chain theory probability distributions. Combined with random initiation of chains and detection of external quanta, the Monte Carlo code generates time tagged data for neutron and gamma-ray counting and from these data the counting distributions.« less
Effect of friction on electron transfer: The two reaction coordinate case
NASA Astrophysics Data System (ADS)
Onuchic, José Nelson
1987-04-01
Electron transfer is a very important reaction in many biological processes such as photosynthesis and oxidative phosphorylation. In many of these reactions, most of the interesting dynamics can be included by using two reaction coordinates: one fast (local high frequency vibration modes) and one slow (outersphere modes such as solvent polarization). We report a model to describe this problem, which uses path integral techniques to calculate electron transfer rates, and also to obtain the Fokker-Planck equations associated with this model. Different limiting cases lead to qualitatively different results such as exponential or nonexponential time decay for the donor survival probability. Conditions for the validity of the adiabatic or the nonadiabatic limits will be discussed. Application of this model to real systems is proposed, in particular for a porphyrin rigidly linked to a quinone, which is a very interesting model compound for primary events of photosynthesis. This model can also be used for other multicoordinate biological reactions such as ligand binding to heme proteins. Also, in the concluding part of Sec. III, we discuss the important limit where the fast vibronic mode is much faster than all the other nuclear modes coupled to the problem. In this limit the fast mode ``renormalizes'' the electronic matrix element, and this considerably simplifies the treatment of the problem, reducing it to coupling only to the slow modes.
Limiting fragmentation from scale-invariant merging of fast partons
NASA Astrophysics Data System (ADS)
Bialas, A.; Bzdak, A.; Peschanski, R.
2008-07-01
Exploiting the idea that the fast partons of an energetic projectile can be treated as sources of colour radiation interpreted as wee partons, it is shown that the recently observed property of extended limiting fragmentation implies a scaling law for the rapidity distribution of fast partons. This leads to a picture of a self-similar process where, for fixed total rapidity Y, the sources merge with probability varying as 1 / y.
The propagation of Lamb waves in multilayered plates: phase-velocity measurement
NASA Astrophysics Data System (ADS)
Grondel, Sébastien; Assaad, Jamal; Delebarre, Christophe; Blanquet, Pierrick; Moulin, Emmanuel
1999-05-01
Owing to the dispersive nature and complexity of the Lamb waves generated in a composite plate, the measurement of the phase velocities by using classical methods is complicated. This paper describes a measurement method based upon the spectrum-analysis technique, which allows one to overcome these problems. The technique consists of using the fast Fourier transform to compute the spatial power-density spectrum. Additionally, weighted functions are used to increase the probability of detecting the various propagation modes. Experimental Lamb-wave dispersion curves of multilayered plates are successfully compared with the analytical ones. This technique is expected to be a useful way to design composite parts integrating ultrasonic transducers in the field of health monitoring. Indeed, Lamb waves and particularly their velocities are very sensitive to defects.
Neutronic calculation of fast reactors by the EUCLID/V1 integrated code
NASA Astrophysics Data System (ADS)
Koltashev, D. A.; Stakhanova, A. A.
2017-01-01
This article considers neutronic calculation of a fast-neutron lead-cooled reactor BREST-OD-300 by the EUCLID/V1 integrated code. The main goal of development and application of integrated codes is a nuclear power plant safety justification. EUCLID/V1 is integrated code designed for coupled neutronics, thermomechanical and thermohydraulic fast reactor calculations under normal and abnormal operating conditions. EUCLID/V1 code is being developed in the Nuclear Safety Institute of the Russian Academy of Sciences. The integrated code has a modular structure and consists of three main modules: thermohydraulic module HYDRA-IBRAE/LM/V1, thermomechanical module BERKUT and neutronic module DN3D. In addition, the integrated code includes databases with fuel, coolant and structural materials properties. Neutronic module DN3D provides full-scale simulation of neutronic processes in fast reactors. Heat sources distribution, control rods movement, reactivity level changes and other processes can be simulated. Neutron transport equation in multigroup diffusion approximation is solved. This paper contains some calculations implemented as a part of EUCLID/V1 code validation. A fast-neutron lead-cooled reactor BREST-OD-300 transient simulation (fuel assembly floating, decompression of passive feedback system channel) and cross-validation with MCU-FR code results are presented in this paper. The calculations demonstrate EUCLID/V1 code application for BREST-OD-300 simulating and safety justification.
Factors That Influence Fast Mapping in Children Exposed to Spanish and English
Alt, Mary; Meyers, Christina; Figueroa, Cecilia
2015-01-01
Purpose The purpose of this study was to determine if children exposed to two languages would benefit from the phonotactic probability cues of a single language in the same way as monolingual peers and to determine if cross-linguistic influence would be present in a fast mapping task. Method Two groups of typically-developing children (monolingual English and bilingual Spanish-English) took part in a computer-based fast mapping task which manipulated phonotactic probability. Children were preschool-aged (N = 50) or school-aged (N = 34). Fast mapping was assessed through name identification and naming tasks. Data were analyzed using mixed ANOVAs with post-hoc testing and simple regression. Results Bilingual and monolingual preschoolers showed sensitivity to English phonotactic cues in both tasks, but bilingual preschoolers were less accurate than monolingual peers in the naming task. School-aged bilingual children had nearly identical performance to monolingual peers. Conclusions Knowing that children exposed to two languages can benefit from the statistical cues of a single language can help inform ideas about instruction and assessment for bilingual learners. PMID:23816663
Farinella, Matteo; Ruedt, Daniel T.; Gleeson, Padraig; Lanore, Frederic; Silver, R. Angus
2014-01-01
In vivo, cortical pyramidal cells are bombarded by asynchronous synaptic input arising from ongoing network activity. However, little is known about how such ‘background’ synaptic input interacts with nonlinear dendritic mechanisms. We have modified an existing model of a layer 5 (L5) pyramidal cell to explore how dendritic integration in the apical dendritic tuft could be altered by the levels of network activity observed in vivo. Here we show that asynchronous background excitatory input increases neuronal gain and extends both temporal and spatial integration of stimulus-evoked synaptic input onto the dendritic tuft. Addition of fast and slow inhibitory synaptic conductances, with properties similar to those from dendritic targeting interneurons, that provided a ‘balanced’ background configuration, partially counteracted these effects, suggesting that inhibition can tune spatio-temporal integration in the tuft. Excitatory background input lowered the threshold for NMDA receptor-mediated dendritic spikes, extended their duration and increased the probability of additional regenerative events occurring in neighbouring branches. These effects were also observed in a passive model where all the non-synaptic voltage-gated conductances were removed. Our results show that glutamate-bound NMDA receptors arising from ongoing network activity can provide a powerful spatially distributed nonlinear dendritic conductance. This may enable L5 pyramidal cells to change their integrative properties as a function of local network activity, potentially allowing both clustered and spatially distributed synaptic inputs to be integrated over extended timescales. PMID:24763087
Outcome of Blunt Abdominal Traumas with Stable Hemodynamic and Positive FAST Findings.
Behboodi, Firooz; Mohtasham-Amiri, Zahra; Masjedi, Navid; Shojaie, Reza; Sadri, Peyman
2016-01-01
Focused assessment with sonography for trauma (FAST) is a highly effective first screening tool for initial classification of abdominal trauma patients. The present study was designed to evaluate the outcome of patients with blunt abdominal trauma and positive FAST findings. The present prospective cross-sectional study was done on patients over 7 years old with normal abdominal examination, positive FAST findings, and available abdominopelvic computed tomography (CT) scan findings. The frequency of need for laparotomy as well as its probable risk factors were calculated. 180 patients were enrolled (mean age: 28.0 ± 11.5 years; 76.7% male). FAST findings were confirmed by abdominopelvic CT scan in only 124 (68.9%) cases. Finally, 12 (6.6%) patients needed laparotomy. Mean age of those in need of laparotomy was significantly higher than others (36.75 ± 11.37 versus 27.34 ± 11.37, p = 0.006). Higher grading of spleen (p = 0.001) and hepatic (p = 0.038) ruptures increased the probability of need for laparotomy. 68.9% of the positive FAST findings in patients with blunt abdominal trauma and stable hemodynamics was confirmed by abdominopelvic CT scan and only 6.6% needed laparotomy. Simultaneous presence of free fluid and air in the abdominal area, old age, and higher grading o solid organ injuries were factors that had a significant correlation with need for laparotomy.
A model of systems integration to facilitate ITS deployment : FAST-TRAC phase IV deliverable
DOT National Transportation Integrated Search
2003-04-01
The FAST-TRAC (Faster and Safer Travel through Traffic Routing and Advanced Controls) Intelligent Transportation Systems (ITS) Field Test in Oakland County, an urbanized county in metropolitan Detroit, represents a large and successful systems integr...
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, K. S.; Nakae, L. F.; Prasad, M. K.
Here, we solve a simple theoretical model of time evolving fission chains due to Feynman that generalizes and asymptotically approaches the point model theory. The point model theory has been used to analyze thermal neutron counting data. This extension of the theory underlies fast counting data for both neutrons and gamma rays from metal systems. Fast neutron and gamma-ray counting is now possible using liquid scintillator arrays with nanosecond time resolution. For individual fission chains, the differential equations describing three correlated probability distributions are solved: the time-dependent internal neutron population, accumulation of fissions in time, and accumulation of leaked neutronsmore » in time. Explicit analytic formulas are given for correlated moments of the time evolving chain populations. The equations for random time gate fast neutron and gamma-ray counting distributions, due to randomly initiated chains, are presented. Correlated moment equations are given for both random time gate and triggered time gate counting. There are explicit formulas for all correlated moments are given up to triple order, for all combinations of correlated fast neutrons and gamma rays. The nonlinear differential equations for probabilities for time dependent fission chain populations have a remarkably simple Monte Carlo realization. A Monte Carlo code was developed for this theory and is shown to statistically realize the solutions to the fission chain theory probability distributions. Combined with random initiation of chains and detection of external quanta, the Monte Carlo code generates time tagged data for neutron and gamma-ray counting and from these data the counting distributions.« less
Probabilistic Micromechanics and Macromechanics for Ceramic Matrix Composites
NASA Technical Reports Server (NTRS)
Murthy, Pappu L. N.; Mital, Subodh K.; Shah, Ashwin R.
1997-01-01
The properties of ceramic matrix composites (CMC's) are known to display a considerable amount of scatter due to variations in fiber/matrix properties, interphase properties, interphase bonding, amount of matrix voids, and many geometry- or fabrication-related parameters, such as ply thickness and ply orientation. This paper summarizes preliminary studies in which formal probabilistic descriptions of the material-behavior- and fabrication-related parameters were incorporated into micromechanics and macromechanics for CMC'S. In this process two existing methodologies, namely CMC micromechanics and macromechanics analysis and a fast probability integration (FPI) technique are synergistically coupled to obtain the probabilistic composite behavior or response. Preliminary results in the form of cumulative probability distributions and information on the probability sensitivities of the response to primitive variables for a unidirectional silicon carbide/reaction-bonded silicon nitride (SiC/RBSN) CMC are presented. The cumulative distribution functions are computed for composite moduli, thermal expansion coefficients, thermal conductivities, and longitudinal tensile strength at room temperature. The variations in the constituent properties that directly affect these composite properties are accounted for via assumed probabilistic distributions. Collectively, the results show that the present technique provides valuable information about the composite properties and sensitivity factors, which is useful to design or test engineers. Furthermore, the present methodology is computationally more efficient than a standard Monte-Carlo simulation technique; and the agreement between the two solutions is excellent, as shown via select examples.
An information diffusion technique to assess integrated hazard risks.
Huang, Chongfu; Huang, Yundong
2018-02-01
An integrated risk is a scene in the future associated with some adverse incident caused by multiple hazards. An integrated probability risk is the expected value of disaster. Due to the difficulty of assessing an integrated probability risk with a small sample, weighting methods and copulas are employed to avoid this obstacle. To resolve the problem, in this paper, we develop the information diffusion technique to construct a joint probability distribution and a vulnerability surface. Then, an integrated risk can be directly assessed by using a small sample. A case of an integrated risk caused by flood and earthquake is given to show how the suggested technique is used to assess the integrated risk of annual property loss. Copyright © 2017 Elsevier Inc. All rights reserved.
"Fast" Capitalism and "Fast" Schools: New Realities and New Truths.
ERIC Educational Resources Information Center
Robertson, Susan L.
This paper locates the phenomenon of self-managing schools within the framework of "fast capitalism" and identifies themes of organization central to fast capitalism, which are argued to also underpin the self-managing schools. "Fast capitalism" refers to the rapidly intensified integration of regionalized productive activities into the global…
Sleep-mediated memory consolidation depends on the level of integration at encoding.
Himmer, Lea; Müller, Elias; Gais, Steffen; Schönauer, Monika
2017-01-01
There is robust evidence that sleep facilitates declarative memory consolidation. Integration of newly acquired memories into existing neocortical knowledge networks has been proposed to underlie this effect. Here, we test whether sleep affects memory retention for word-picture associations differently when it was learned explicitly or using a fast mapping strategy. Fast mapping is an incidental form of learning that references new information to existing knowledge and possibly allows neocortical integration already during encoding. If the integration of information into neocortical networks is a main function of sleep-dependent memory consolidation, material learned via fast mapping should therefore benefit less from sleep. Supporting this idea, we find that sleep has a protective effect on explicitly learned associations. In contrast, memory for associations learned by fast mapping does not benefit from sleep and remains stable regardless of whether sleep or wakefulness follows learning. Our results thus indicate that the need for sleep-mediated consolidation depends on the strategy used for learning and might thus be related to the level of integration of newly acquired memory achieved during encoding. Copyright © 2016 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Pankratov, Oleg; Kuvshinov, Alexey
2016-01-01
Despite impressive progress in the development and application of electromagnetic (EM) deterministic inverse schemes to map the 3-D distribution of electrical conductivity within the Earth, there is one question which remains poorly addressed—uncertainty quantification of the recovered conductivity models. Apparently, only an inversion based on a statistical approach provides a systematic framework to quantify such uncertainties. The Metropolis-Hastings (M-H) algorithm is the most popular technique for sampling the posterior probability distribution that describes the solution of the statistical inverse problem. However, all statistical inverse schemes require an enormous amount of forward simulations and thus appear to be extremely demanding computationally, if not prohibitive, if a 3-D set up is invoked. This urges development of fast and scalable 3-D modelling codes which can run large-scale 3-D models of practical interest for fractions of a second on high-performance multi-core platforms. But, even with these codes, the challenge for M-H methods is to construct proposal functions that simultaneously provide a good approximation of the target density function while being inexpensive to be sampled. In this paper we address both of these issues. First we introduce a variant of the M-H method which uses information about the local gradient and Hessian of the penalty function. This, in particular, allows us to exploit adjoint-based machinery that has been instrumental for the fast solution of deterministic inverse problems. We explain why this modification of M-H significantly accelerates sampling of the posterior probability distribution. In addition we show how Hessian handling (inverse, square root) can be made practicable by a low-rank approximation using the Lanczos algorithm. Ultimately we discuss uncertainty analysis based on stochastic inversion results. In addition, we demonstrate how this analysis can be performed within a deterministic approach. In the second part, we summarize modern trends in the development of efficient 3-D EM forward modelling schemes with special emphasis on recent advances in the integral equation approach.
Probabilistic Aeroelastic Analysis Developed for Turbomachinery Components
NASA Technical Reports Server (NTRS)
Reddy, T. S. R.; Mital, Subodh K.; Stefko, George L.; Pai, Shantaram S.
2003-01-01
Aeroelastic analyses for advanced turbomachines are being developed for use at the NASA Glenn Research Center and industry. However, these analyses at present are used for turbomachinery design with uncertainties accounted for by using safety factors. This approach may lead to overly conservative designs, thereby reducing the potential of designing higher efficiency engines. An integration of the deterministic aeroelastic analysis methods with probabilistic analysis methods offers the potential to design efficient engines with fewer aeroelastic problems and to make a quantum leap toward designing safe reliable engines. In this research, probabilistic analysis is integrated with aeroelastic analysis: (1) to determine the parameters that most affect the aeroelastic characteristics (forced response and stability) of a turbomachine component such as a fan, compressor, or turbine and (2) to give the acceptable standard deviation on the design parameters for an aeroelastically stable system. The approach taken is to combine the aeroelastic analysis of the MISER (MIStuned Engine Response) code with the FPI (fast probability integration) code. The role of MISER is to provide the functional relationships that tie the structural and aerodynamic parameters (the primitive variables) to the forced response amplitudes and stability eigenvalues (the response properties). The role of FPI is to perform probabilistic analyses by utilizing the response properties generated by MISER. The results are a probability density function for the response properties. The probabilistic sensitivities of the response variables to uncertainty in primitive variables are obtained as a byproduct of the FPI technique. The combined analysis of aeroelastic and probabilistic analysis is applied to a 12-bladed cascade vibrating in bending and torsion. Out of the total 11 design parameters, 6 are considered as having probabilistic variation. The six parameters are space-to-chord ratio (SBYC), stagger angle (GAMA), elastic axis (ELAXS), Mach number (MACH), mass ratio (MASSR), and frequency ratio (WHWB). The cascade is considered to be in subsonic flow with Mach 0.7. The results of the probabilistic aeroelastic analysis are the probability density function of predicted aerodynamic damping and frequency for flutter and the response amplitudes for forced response.
Failure Analysis by Statistical Techniques (FAST). Volume 1. User’s Manual
1974-10-31
REPORT NUMBER DNA 3336F-1 2. OOVT ACCESSION NO 4. TITLE Cand Sublllle) • FAILURE ANALYSIS BY STATISTICAL TECHNIQUES (FAST) Volume I, User’s...SS2), and t’ a facility ( SS7 ). The other three diagrams break down the three critical subsystems. T le median probability of survival of the
Factors that Influence Fast Mapping in Children Exposed to Spanish and English
ERIC Educational Resources Information Center
Alt, Mary; Meyers, Christina; Figueroa, Cecilia
2013-01-01
Purpose: The purpose of this study was to determine whether children exposed to 2 languages would benefit from the phonotactic probability cues of a single language in the same way as monolingual peers and to determine whether crosslinguistic influence would be present in a fast-mapping task. Method: Two groups of typically developing children…
Fast and High Dynamic Range Imaging with Superconducting Tunnel Junction Detectors
NASA Astrophysics Data System (ADS)
Matsuo, Hiroshi
2014-08-01
We have demonstrated a combined test of the submillimeter-wave SIS photon detectors and GaAs-JFET cryogenic integrated circuits. A relatively large background photo-current can be read out by fast-reset integrating amplifiers. An integration time of 1 ms enables fast frame rate readout and large dynamic range imaging, with an expected dynamic range of 8,000 in 1 ms. Ultimate fast and high dynamic range performance of superconducting tunnel junction detectors (STJ) will be obtained when photon counting capabilities are employed. In the terahertz frequencies, when input photon rate of 100 MHz is measured, the photon bunching gives us enough timing resolution to be used as phase information of intensity fluctuation. Application of photon statistics will be a new tool in the terahertz frequency region. The design parameters of STJ terahertz photon counting detectors are discussed.
Liu, Xuewu; Huang, Yuxiao; Liang, Jiao; Zhang, Shuai; Li, Yinghui; Wang, Jun; Shen, Yan; Xu, Zhikai; Zhao, Ya
2014-11-30
The invasion of red blood cells (RBCs) by malarial parasites is an essential step in the life cycle of Plasmodium falciparum. Human-parasite surface protein interactions play a critical role in this process. Although several interactions between human and parasite proteins have been discovered, the mechanism related to invasion remains poorly understood because numerous human-parasite protein interactions have not yet been identified. High-throughput screening experiments are not feasible for malarial parasites due to difficulty in expressing the parasite proteins. Here, we performed computational prediction of the PPIs involved in malaria parasite invasion to elucidate the mechanism by which invasion occurs. In this study, an expectation maximization algorithm was used to estimate the probabilities of domain-domain interactions (DDIs). Estimates of DDI probabilities were then used to infer PPI probabilities. We found that our prediction performance was better than that based on the information of D. melanogaster alone when information related to the six species was used. Prediction performance was assessed using protein interaction data from S. cerevisiae, indicating that the predicted results were reliable. We then used the estimates of DDI probabilities to infer interactions between 490 parasite and 3,787 human membrane proteins. A small-scale dataset was used to illustrate the usability of our method in predicting interactions between human and parasite proteins. The positive predictive value (PPV) was lower than that observed in S. cerevisiae. We integrated gene expression data to improve prediction accuracy and to reduce false positives. We identified 80 membrane proteins highly expressed in the schizont stage by fast Fourier transform method. Approximately 221 erythrocyte membrane proteins were identified using published mass spectral datasets. A network consisting of 205 interactions was predicted. Results of network analysis suggest that SNARE proteins of parasites and APP of humans may function in the invasion of RBCs by parasites. We predicted a small-scale PPI network that may be involved in parasite invasion of RBCs by integrating DDI information and expression profiles. Experimental studies should be conducted to validate the predicted interactions. The predicted PPIs help elucidate the mechanism of parasite invasion and provide directions for future experimental investigations.
Probabilistic Structural Analysis Methods (PSAM) for select space propulsion system components
NASA Technical Reports Server (NTRS)
1991-01-01
The fourth year of technical developments on the Numerical Evaluation of Stochastic Structures Under Stress (NESSUS) system for Probabilistic Structural Analysis Methods is summarized. The effort focused on the continued expansion of the Probabilistic Finite Element Method (PFEM) code, the implementation of the Probabilistic Boundary Element Method (PBEM), and the implementation of the Probabilistic Approximate Methods (PAppM) code. The principal focus for the PFEM code is the addition of a multilevel structural dynamics capability. The strategy includes probabilistic loads, treatment of material, geometry uncertainty, and full probabilistic variables. Enhancements are included for the Fast Probability Integration (FPI) algorithms and the addition of Monte Carlo simulation as an alternate. Work on the expert system and boundary element developments continues. The enhanced capability in the computer codes is validated by applications to a turbine blade and to an oxidizer duct.
FastMag: Fast micromagnetic simulator for complex magnetic structures (invited)
NASA Astrophysics Data System (ADS)
Chang, R.; Li, S.; Lubarda, M. V.; Livshitz, B.; Lomakin, V.
2011-04-01
A fast micromagnetic simulator (FastMag) for general problems is presented. FastMag solves the Landau-Lifshitz-Gilbert equation and can handle multiscale problems with a high computational efficiency. The simulator derives its high performance from efficient methods for evaluating the effective field and from implementations on massively parallel graphics processing unit (GPU) architectures. FastMag discretizes the computational domain into tetrahedral elements and therefore is highly flexible for general problems. The magnetostatic field is computed via the superposition principle for both volume and surface parts of the computational domain. This is accomplished by implementing efficient quadrature rules and analytical integration for overlapping elements in which the integral kernel is singular. Thus, discretized superposition integrals are computed using a nonuniform grid interpolation method, which evaluates the field from N sources at N collocated observers in O(N) operations. This approach allows handling objects of arbitrary shape, allows easily calculating of the field outside the magnetized domains, does not require solving a linear system of equations, and requires little memory. FastMag is implemented on GPUs with ?> GPU-central processing unit speed-ups of 2 orders of magnitude. Simulations are shown of a large array of magnetic dots and a recording head fully discretized down to the exchange length, with over a hundred million tetrahedral elements on an inexpensive desktop computer.
NASA Technical Reports Server (NTRS)
Sparks, W. M.; Starrfield, S.; Truran, J. W.
1978-01-01
The paper reports use of a Lagrangian implicit hydrodynamics computer code incorporating a full nuclear-reaction network to follow a thermonuclear runaway in the hydrogen-rich envelope of a 1.25 solar-mass white dwarf. In this evolutionary sequence the envelope was assumed to be of normal (solar) composition and the resulting outburst closely resembles that of the slow nova HR Del. In contrast, previous CNO-enhanced models resemble fast nova outbursts. The slow-nova model ejects material by radiation pressure when the high luminosity of the rekindled hydrogen shell source exceeds the local Eddington luminosity of the outer layers. This is in contrast to the fast nova outburst where ejection is caused by the decay of the beta(+)-unstable nuclei. Nevertheless, radiation pressure probably plays a major role in ejecting material from the fast nova remnants. Therefore, the sequence from slow to fast novae can be interpreted as a sequence of white dwarfs with increasing amounts of enhanced CNO nuclei in their hydrogen envelopes, although other parameters such as the white-dwarf mass and accretion rate probably contribute to the observed variation between novae.
The Ongoing Impact of the U.S. Fast Reactor Integral Experiments Program
DOE Office of Scientific and Technical Information (OSTI.GOV)
John D. Bess; Michael A. Pope; Harold F. McFarlane
2012-11-01
The creation of a large database of integral fast reactor physics experiments advanced nuclear science and technology in ways that were unachievable by less capital intensive and operationally challenging approaches. They enabled the compilation of integral physics benchmark data, validated (or not) analytical methods, and provided assurance of future rector designs The integral experiments performed at Argonne National Laboratory (ANL) represent decades of research performed to support fast reactor design and our understanding of neutronics behavior and reactor physics measurements. Experiments began in 1955 with the Zero Power Reactor No. 3 (ZPR-3) and terminated with the Zero Power Physics Reactormore » (ZPPR, originally the Zero Power Plutonium Reactor) in 1990 at the former ANL-West site in Idaho, which is now part of the Idaho National Laboratory (INL). Two additional critical assemblies, ZPR-6 and ZPR-9, operated at the ANL-East site in Illinois. A total of 128 fast reactor assemblies were constructed with these facilities [1]. The infrastructure and measurement capabilities are too expensive to be replicated in the modern era, making the integral database invaluable as the world pushes ahead with development of liquid metal cooled reactors.« less
Humbird, David; Trendewicz, Anna; Braun, Robert; ...
2017-01-12
A biomass fast pyrolysis reactor model with detailed reaction kinetics and one-dimensional fluid dynamics was implemented in an equation-oriented modeling environment (Aspen Custom Modeler). Portions of this work were detailed in previous publications; further modifications have been made here to improve stability and reduce execution time of the model to make it compatible for use in large process flowsheets. The detailed reactor model was integrated into a larger process simulation in Aspen Plus and was stable for different feedstocks over a range of reactor temperatures. Sample results are presented that indicate general agreement with experimental results, but with higher gasmore » losses caused by stripping of the bio-oil by the fluidizing gas in the simulated absorber/condenser. Lastly, this integrated modeling approach can be extended to other well-defined, predictive reactor models for fast pyrolysis, catalytic fast pyrolysis, as well as other processes.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Humbird, David; Trendewicz, Anna; Braun, Robert
A biomass fast pyrolysis reactor model with detailed reaction kinetics and one-dimensional fluid dynamics was implemented in an equation-oriented modeling environment (Aspen Custom Modeler). Portions of this work were detailed in previous publications; further modifications have been made here to improve stability and reduce execution time of the model to make it compatible for use in large process flowsheets. The detailed reactor model was integrated into a larger process simulation in Aspen Plus and was stable for different feedstocks over a range of reactor temperatures. Sample results are presented that indicate general agreement with experimental results, but with higher gasmore » losses caused by stripping of the bio-oil by the fluidizing gas in the simulated absorber/condenser. Lastly, this integrated modeling approach can be extended to other well-defined, predictive reactor models for fast pyrolysis, catalytic fast pyrolysis, as well as other processes.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Smith, R.R.
1986-01-01
This report presents information on the Integral Fast Reactor and its role in the future. Information is presented in the areas of: inherent safety; other virtues of sodium-cooled breeder; and solving LWR fuel cycle problems with IFR technologies. (JDB)
Time Neutron Technique for UXO Discrimination
2010-12-01
mixture of TNT and RDX C-4 CFD Composition 4 military plastic explosive Constant Fraction Discriminator Cps CsI counts per second inorganic...Pdfs Probability Density Functions PET Positron Emission Tomography Pfa Probability of False Alarm PFTNA Pulsed Fast/Thermal Neutron Analysis PMTs...the ordnance type (rocket, mortar , projectile, etc.) and what filler material it contains (inert or empty), practice, HE, illumination, chemical (i.e
Fast and asymptotic computation of the fixation probability for Moran processes on graphs.
Alcalde Cuesta, F; González Sequeiros, P; Lozano Rojo, Á
2015-03-01
Evolutionary dynamics has been classically studied for homogeneous populations, but now there is a growing interest in the non-homogeneous case. One of the most important models has been proposed in Lieberman et al. (2005), adapting to a weighted directed graph the process described in Moran (1958). The Markov chain associated with the graph can be modified by erasing all non-trivial loops in its state space, obtaining the so-called Embedded Markov chain (EMC). The fixation probability remains unchanged, but the expected time to absorption (fixation or extinction) is reduced. In this paper, we shall use this idea to compute asymptotically the average fixation probability for complete bipartite graphs K(n,m). To this end, we firstly review some recent results on evolutionary dynamics on graphs trying to clarify some points. We also revisit the 'Star Theorem' proved in Lieberman et al. (2005) for the star graphs K(1,m). Theoretically, EMC techniques allow fast computation of the fixation probability, but in practice this is not always true. Thus, in the last part of the paper, we compare this algorithm with the standard Monte Carlo method for some kind of complex networks. Copyright © 2015 Elsevier Ireland Ltd. All rights reserved.
A fast and objective multidimensional kernel density estimation method: fastKDE
O'Brien, Travis A.; Kashinath, Karthik; Cavanaugh, Nicholas R.; ...
2016-03-07
Numerous facets of scientific research implicitly or explicitly call for the estimation of probability densities. Histograms and kernel density estimates (KDEs) are two commonly used techniques for estimating such information, with the KDE generally providing a higher fidelity representation of the probability density function (PDF). Both methods require specification of either a bin width or a kernel bandwidth. While techniques exist for choosing the kernel bandwidth optimally and objectively, they are computationally intensive, since they require repeated calculation of the KDE. A solution for objectively and optimally choosing both the kernel shape and width has recently been developed by Bernacchiamore » and Pigolotti (2011). While this solution theoretically applies to multidimensional KDEs, it has not been clear how to practically do so. A method for practically extending the Bernacchia-Pigolotti KDE to multidimensions is introduced. This multidimensional extension is combined with a recently-developed computational improvement to their method that makes it computationally efficient: a 2D KDE on 10 5 samples only takes 1 s on a modern workstation. This fast and objective KDE method, called the fastKDE method, retains the excellent statistical convergence properties that have been demonstrated for univariate samples. The fastKDE method exhibits statistical accuracy that is comparable to state-of-the-science KDE methods publicly available in R, and it produces kernel density estimates several orders of magnitude faster. The fastKDE method does an excellent job of encoding covariance information for bivariate samples. This property allows for direct calculation of conditional PDFs with fastKDE. It is demonstrated how this capability might be leveraged for detecting non-trivial relationships between quantities in physical systems, such as transitional behavior.« less
UKIRT fast guide system improvements
NASA Astrophysics Data System (ADS)
Balius, Al; Rees, Nicholas P.
1997-09-01
The United Kingdom Infra-Red Telescope (UKIRT) has recently undergone the first major upgrade program since its construction. One part of the upgrade program was an adaptive tip-tilt secondary mirror closed with a CCD system collectively called the fast guide system. The installation of the new secondary and associated systems was carried out in the first half of 1996. Initial testing of the fast guide system has shown great improvement in guide accuracy. The initial installation included a fixed integration time CCD. In the first part of 1997 an integration time controller based on computed guide star luminosity was implemented in the fast guide system. Also, a Kalman type estimator was installed in the image tracking loop based on a dynamic model and knowledge of the statistical properties of the guide star position error measurement as a function of computed guide star magnitude and CCD integration time. The new configuration was tested in terms of improved guide performance nd graceful degradation when tracking faint guide stars. This paper describes the modified fast guide system configuration and reports the results of performance tests.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Simpson, M.; Markel, T.
2012-08-01
The growing, though still nascent, plug-in electric vehicle (PEV) market currently operates primarily via level 1 and level 2 charging in the United States. Fast chargers are still a rarity, but offer a confidence boost to oppose 'range anxiety' in consumers making the transition from conventional vehicles to PEVs. Because relatively no real-world usage of fast chargers at scale exists yet, the National Renewable Energy Laboratory developed a simulation to help assess fast charging needs based on real-world travel data. This study documents the data, methods, and results of the simulation run for multiple scenarios, varying fleet sizes, and themore » number of charger ports. The grid impact of this usage is further quantified to assess the opportunity for integration of renewables; specifically, a high frequency of fast charging is found to be in demand during the late afternoons and evenings coinciding with grid peak periods. Proper integration of a solar array and stationary battery thus helps ease the load and reduces the need for new generator construction to meet the demand of a future PEV market.« less
Michiels, Steven; Poels, Kenneth; Crijns, Wouter; Delombaerde, Laurence; De Roover, Robin; Vanstraelen, Bianca; Haustermans, Karin; Nuyts, Sandra; Depuydt, Tom
2018-05-05
Linac improvements in gantry speed, leaf speed and dose rate may increase the time-efficiency of volumetric modulated arc therapy (VMAT) delivery. The plan quality achievable with faster VMAT however remains to be investigated. In this study, a fast-rotating O-ring linac with fast-moving leaves is compared with a C-arm linac in terms of plan quality and delivery time for VMAT of head-and-neck cancer (HNC). For 30 patients with HNC, treatment planning was performed using dual-arc (HA2) and triple-arc (HA3) VMAT on a Halcyon fast-rotating O-ring linac and using dual-arc VMAT on a TrueBeam C-arm linac (TB2). Target coverage metrics and complication probabilities were compared. Plan delivery was verified using 3%/3 mm gamma-index analysis of helical diode array measurements. Volumetric image acquisition and plan delivery times were compared. All studied VMAT-techniques fulfilled the target coverage objectives. D 2% to the boost volume was higher for HA2 (median 103.7%, 1st-3rd quartile [103.5%;104.0%]) and HA3 (103.2% [103.0%;103.7%)] than for TB2 (102.6% [102.3%;103.0%)], resulting in an increased boost target dose heterogeneity for HA2 and HA3. Complication probabilities were comparable between HA2 and TB2, while HA3 showed a xerostomia probability reduction (0.8% [0.2%;1.8%]) and dysphagia probability reduction (1.0% [0.2%;1.8%]) compared with TB2. Gamma-index agreement scores were never below 93.0% for HA2, HA3 and TB2. Volumetric imaging and plan delivery time was shorter for HA2 (1 m 24 s ± 1 s) and HA3 (1 m 54 s ± 1 s) than for TB2 (2 m 47 s ± 1 s). For VMAT of HNC, the fast-rotating O-ring linac at least maintains the plan quality of two arcs on a C-arm linac while reducing the image acquisition and plan delivery time. Copyright © 2018 Elsevier B.V. All rights reserved.
Improved FFT-based numerical inversion of Laplace transforms via fast Hartley transform algorithm
NASA Technical Reports Server (NTRS)
Hwang, Chyi; Lu, Ming-Jeng; Shieh, Leang S.
1991-01-01
The disadvantages of numerical inversion of the Laplace transform via the conventional fast Fourier transform (FFT) are identified and an improved method is presented to remedy them. The improved method is based on introducing a new integration step length Delta(omega) = pi/mT for trapezoidal-rule approximation of the Bromwich integral, in which a new parameter, m, is introduced for controlling the accuracy of the numerical integration. Naturally, this method leads to multiple sets of complex FFT computations. A new inversion formula is derived such that N equally spaced samples of the inverse Laplace transform function can be obtained by (m/2) + 1 sets of N-point complex FFT computations or by m sets of real fast Hartley transform (FHT) computations.
Converting Maturing Nuclear Sites to Integrated Power Production Islands
Solbrig, Charles W.
2011-01-01
Nuclear islands, which are integrated power production sites, could effectively sequester and safeguard the US stockpile of plutonium. A nuclear island, an evolution of the integral fast reactor, utilizes all the Transuranics (Pu plus minor actinides) produced in power production, and it eliminates all spent fuel shipments to and from the site. This latter attribute requires that fuel reprocessing occur on each site and that fast reactors be built on-site to utilize the TRU. All commercial spent fuel shipments could be eliminated by converting all LWR nuclear power sites to nuclear islands. Existing LWR sites have the added advantage ofmore » already possessing a license to produce nuclear power. Each could contribute to an increase in the nuclear power production by adding one or more fast reactors. Both the TRU and the depleted uranium obtained in reprocessing would be used on-site for fast fuel manufacture. Only fission products would be shipped to a repository for storage. The nuclear island concept could be used to alleviate the strain of LWR plant sites currently approaching or exceeding their spent fuel pool storage capacity. Fast reactor breeding ratio could be designed to convert existing sites to all fast reactors, or keep the majority thermal.« less
Reducing the Risk of Human Space Missions with INTEGRITY
NASA Technical Reports Server (NTRS)
Jones, Harry W.; Dillon-Merill, Robin L.; Tri, Terry O.; Henninger, Donald L.
2003-01-01
The INTEGRITY Program will design and operate a test bed facility to help prepare for future beyond-LEO missions. The purpose of INTEGRITY is to enable future missions by developing, testing, and demonstrating advanced human space systems. INTEGRITY will also implement and validate advanced management techniques including risk analysis and mitigation. One important way INTEGRITY will help enable future missions is by reducing their risk. A risk analysis of human space missions is important in defining the steps that INTEGRITY should take to mitigate risk. This paper describes how a Probabilistic Risk Assessment (PRA) of human space missions will help support the planning and development of INTEGRITY to maximize its benefits to future missions. PRA is a systematic methodology to decompose the system into subsystems and components, to quantify the failure risk as a function of the design elements and their corresponding probability of failure. PRA provides a quantitative estimate of the probability of failure of the system, including an assessment and display of the degree of uncertainty surrounding the probability. PRA provides a basis for understanding the impacts of decisions that affect safety, reliability, performance, and cost. Risks with both high probability and high impact are identified as top priority. The PRA of human missions beyond Earth orbit will help indicate how the risk of future human space missions can be reduced by integrating and testing systems in INTEGRITY.
Low Probability Tail Event Analysis and Mitigation in BPA Control Area: Task 2 Report
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lu, Shuai; Makarov, Yuri V.; McKinstry, Craig A.
Task report detailing low probability tail event analysis and mitigation in BPA control area. Tail event refers to the situation in a power system when unfavorable forecast errors of load and wind are superposed onto fast load and wind ramps, or non-wind generators falling short of scheduled output, causing the imbalance between generation and load to become very significant.
Fourier Method for Calculating Fission Chain Neutron Multiplicity Distributions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chambers, David H.; Chandrasekaran, Hema; Walston, Sean E.
Here, a new way of utilizing the fast Fourier transform is developed to compute the probability distribution for a fission chain to create n neutrons. We then extend this technique to compute the probability distributions for detecting n neutrons. Lastly, our technique can be used for fission chains initiated by either a single neutron inducing a fission or by the spontaneous fission of another isotope.
Fourier Method for Calculating Fission Chain Neutron Multiplicity Distributions
Chambers, David H.; Chandrasekaran, Hema; Walston, Sean E.
2017-03-27
Here, a new way of utilizing the fast Fourier transform is developed to compute the probability distribution for a fission chain to create n neutrons. We then extend this technique to compute the probability distributions for detecting n neutrons. Lastly, our technique can be used for fission chains initiated by either a single neutron inducing a fission or by the spontaneous fission of another isotope.
Protein turnover in adipose tissue from fasted or diabetic rats
NASA Technical Reports Server (NTRS)
Tischler, Marc E.; Ost, Alan H.; Coffman, Julia
1986-01-01
Protein synthesis and degradation in vitro were compared in epididymal fat pads from animals deprived of food for 48 h or treated 6 or 12 days prior with streptozotocin to induce diabetes. Although both fasting and diabetes led to depressed (-24 to -57 percent) protein synthesis, the diminution in protein degradation (-63 to -72 percent) was even greater, so that net in vitro protein balance improved dramatically. Insulin failed to inhibit protein degradation in fat pads of these rats as it does for fed animals. Although insulin stimulated protein synthesis in fat pads of fasted and 12 day diabetic rats, the absolute change was much smaller than that seen in the fed state. The inhibition of protein degradation by leucine also seems to be less in fasted animals, probably because leucine catabolism is slower in fasting. These results show that fasting and diabetes may improve protein balance in adipose tissue but diminish the regulatory effects of insulin.
BioStar models of clinical and genomic data for biomedical data warehouse design
Wang, Liangjiang; Ramanathan, Murali
2008-01-01
Biomedical research is now generating large amounts of data, ranging from clinical test results to microarray gene expression profiles. The scale and complexity of these datasets give rise to substantial challenges in data management and analysis. It is highly desirable that data warehousing and online analytical processing technologies can be applied to biomedical data integration and mining. The major difficulty probably lies in the task of capturing and modelling diverse biological objects and their complex relationships. This paper describes multidimensional data modelling for biomedical data warehouse design. Since the conventional models such as star schema appear to be insufficient for modelling clinical and genomic data, we develop a new model called BioStar schema. The new model can capture the rich semantics of biomedical data and provide greater extensibility for the fast evolution of biological research methodologies. PMID:18048122
Extremely Fast Numerical Integration of Ocean Surface Wave Dynamics
2007-09-30
sub-processor must be added as shown in the blue box of Fig. 1. We first consider the Kadomtsev - Petviashvili (KP) equation ηt + coηx +αηηx + βη ...analytic integration of the so-called “soliton equations ,” I have discovered how the GFT can be used to solved higher order equations for which study...analytical study and extremely fast numerical integration of the extended nonlinear Schroedinger equation for fully three dimensional wave motion
Chandrasekar, A; Rakkiyappan, R; Cao, Jinde
2015-10-01
This paper studies the impulsive synchronization of Markovian jumping randomly coupled neural networks with partly unknown transition probabilities via multiple integral approach. The array of neural networks are coupled in a random fashion which is governed by Bernoulli random variable. The aim of this paper is to obtain the synchronization criteria, which is suitable for both exactly known and partly unknown transition probabilities such that the coupled neural network is synchronized with mixed time-delay. The considered impulsive effects can be synchronized at partly unknown transition probabilities. Besides, a multiple integral approach is also proposed to strengthen the Markovian jumping randomly coupled neural networks with partly unknown transition probabilities. By making use of Kronecker product and some useful integral inequalities, a novel Lyapunov-Krasovskii functional was designed for handling the coupled neural network with mixed delay and then impulsive synchronization criteria are solvable in a set of linear matrix inequalities. Finally, numerical examples are presented to illustrate the effectiveness and advantages of the theoretical results. Copyright © 2015 Elsevier Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fan, J; Fan, J; Hu, W
Purpose: To develop a fast automatic algorithm based on the two dimensional kernel density estimation (2D KDE) to predict the dose-volume histogram (DVH) which can be employed for the investigation of radiotherapy quality assurance and automatic treatment planning. Methods: We propose a machine learning method that uses previous treatment plans to predict the DVH. The key to the approach is the framing of DVH in a probabilistic setting. The training consists of estimating, from the patients in the training set, the joint probability distribution of the dose and the predictive features. The joint distribution provides an estimation of the conditionalmore » probability of the dose given the values of the predictive features. For the new patient, the prediction consists of estimating the distribution of the predictive features and marginalizing the conditional probability from the training over this. Integrating the resulting probability distribution for the dose yields an estimation of the DVH. The 2D KDE is implemented to predict the joint probability distribution of the training set and the distribution of the predictive features for the new patient. Two variables, including the signed minimal distance from each OAR (organs at risk) voxel to the target boundary and its opening angle with respect to the origin of voxel coordinate, are considered as the predictive features to represent the OAR-target spatial relationship. The feasibility of our method has been demonstrated with the rectum, breast and head-and-neck cancer cases by comparing the predicted DVHs with the planned ones. Results: The consistent result has been found between these two DVHs for each cancer and the average of relative point-wise differences is about 5% within the clinical acceptable extent. Conclusion: According to the result of this study, our method can be used to predict the clinical acceptable DVH and has ability to evaluate the quality and consistency of the treatment planning.« less
Fast Micromethod: Determination of DNA Integrity in Cell Suspensions and in Solid Tissues.
Bihari, Nevenka
2017-01-01
The Fast Micromethod is a rapid and convenient microplate procedure for the determination of DNA integrity in cell suspensions and in solid tissues. The procedure is based on the ability of fluorochromes to preferentially interact with double-stranded DNA in alkaline conditions. Rapid sample lysis is followed by denaturation at high pH during 15 min. Only 30 ng of DNA from cell suspensions or tissue homogenates per single well are required for the analyses. The whole analysis is performed within 3 h or less (for one 96-well microplate).The Fast Micromethod is broadly used in biology and medicine. Its applications range from environmental pollution tests in marine invertebrates to the analysis of biopsy samples in cancer patients to detect DNA alterations caused by irradiation or chemotherapy.The procedure presented here describes the Fast Micromethod applied for the determination of DNA integrity in cell suspensions (HeLa cells) and solid tissues (mussel gills).
Saleh, Kutaiba; Stucke, Stephan; Uciteli, Alexandr; Faulbrück-Röhr, Sebastian; Neumann, Juliane; Tahar, Kais; Ammon, Danny; Schmidt, Tobias; Neumuth, Thomas; Besting, Andreas; Portheine, Frank; Herre, Heinrich; Kaeding, André; Specht, Martin
2017-01-01
With the growing strain of medical staff and complexity of patient care, the risk of medical errors increases. In this work we present the use of Fast Healthcare Interoperability Resources (FHIR) as communication standard for the integration of an ontology- and agent-based system to identify risks across medical processes in a clinical environment.
Diagnostic Suite for HyperV Coaxial Plasma Gun Development for the PLX- α Project
NASA Astrophysics Data System (ADS)
Case, Andrew; Brockington, Sam; Witherspoon, F. Douglas
2015-11-01
We present the diagnostic suite to be used during development of the coaxial guns HyperV will deliver to LANL in support of the ARPA-E Accelerating Low-Cost Plasma Heating And Assembly (ALPHA) program. For plasma jet diagnostics this includes fast photodiodes for velocimetry, a ballistic pendulum for measuring total plasmoid momentum, interferometry for line integrated plasma density, deflectometry for line integrated perpendicular density gradient measurements, and spectroscopy, both time resolved high resolution spectroscopy using a novel detector developed by HyperV and time integrated survey spectroscopy, for measurements of velocity and temperature as well as impurities. In addition, we plan to use fast pressure probes for stagnation pressure, a Faraday cup for density, fast imaging for plume geometry and time integrated imaging for overall light emission. A novel low resolution long record length camera developed by HyperV will also be used for plume diagnostics. For diagnostics of gun operation, we will use Rogowski coils to measure current, voltage dividers for voltages, B-dot probes for magnetic field, and time resolved fast photodiodes to measure plasmoid velocity inside the accelerator. This work supported by the ARPA-E ALPHA program.
Dependence of core heating properties on heating pulse duration and intensity
NASA Astrophysics Data System (ADS)
Johzaki, Tomoyuki; Nagatomo, Hideo; Sunahara, Atsushi; Cai, Hongbo; Sakagami, Hitoshi; Mima, Kunioki
2009-11-01
In the cone-guiding fast ignition, an imploded core is heated by the energy transport of fast electrons generated by the ultra-intense short-pulse laser at the cone inner surface. The fast core heating (˜800eV) has been demonstrated at integrated experiments with GEKKO-XII+ PW laser systems. As the next step, experiments using more powerful heating laser, FIREX, have been started at ILE, Osaka university. In FIREX-I (phase-I of FIREX), our goal is the demonstration of efficient core heating (Ti ˜ 5keV) using a newly developed 10kJ LFEX laser. In the first integrated experiments, the LFEX laser is operated with low energy mode (˜0.5kJ/4ps) to validate the previous GEKKO+PW experiments. Between the two experiments, though the laser energy is similar (˜0.5kJ), the duration is different; ˜0.5ps in the PW laser and ˜ 4ps in the LFEX laser. In this paper, we evaluate the dependence of core heating properties on the heating pulse duration on the basis of integrated simulations with FI^3 (Fast Ignition Integrated Interconnecting) code system.
Ting, Chih-Chung; Yu, Chia-Chen; Maloney, Laurence T.
2015-01-01
In Bayesian decision theory, knowledge about the probabilities of possible outcomes is captured by a prior distribution and a likelihood function. The prior reflects past knowledge and the likelihood summarizes current sensory information. The two combined (integrated) form a posterior distribution that allows estimation of the probability of different possible outcomes. In this study, we investigated the neural mechanisms underlying Bayesian integration using a novel lottery decision task in which both prior knowledge and likelihood information about reward probability were systematically manipulated on a trial-by-trial basis. Consistent with Bayesian integration, as sample size increased, subjects tended to weigh likelihood information more compared with prior information. Using fMRI in humans, we found that the medial prefrontal cortex (mPFC) correlated with the mean of the posterior distribution, a statistic that reflects the integration of prior knowledge and likelihood of reward probability. Subsequent analysis revealed that both prior and likelihood information were represented in mPFC and that the neural representations of prior and likelihood in mPFC reflected changes in the behaviorally estimated weights assigned to these different sources of information in response to changes in the environment. Together, these results establish the role of mPFC in prior-likelihood integration and highlight its involvement in representing and integrating these distinct sources of information. PMID:25632152
McCune, D.A.
1964-03-17
A method and device for measurement of integrated fast neutron flux in the presence of a large thermal neutron field are described. The device comprises a thorium wire surrounded by a thermal neutron attenuator that is, in turn, enclosed by heat-resistant material. The method consists of irradiating the device in a neutron field whereby neutrons with energies in excess of 1.1 Mev cause fast fissions in the thorium, then removing the thorium wire, separating the cesium-137 fission product by chemical means from the thorium, and finally counting the radioactivity of the cesium to determine the number of fissions which have occurred so that the integrated fast flux may be obtained. (AEC)
Restoring integrity--A grounded theory of coping with a fast track surgery programme.
Jørgensen, Lene Bastrup; Fridlund, Bengt
2016-01-01
The aim of this study was to generate a theory conceptualizing and explaining behavioural processes involved in coping in order to identify the predominant coping types and coping type-specific features. Patients undergoing fast track procedures do not experience a higher risk of complications, readmission, or mortality. However, such programmes presuppose an increasing degree of patient involvement, placing high educational, physical, and mental demands on the patients. There is a lack of knowledge about how patients understand and cope with fast track programmes. The study design used classical grounded theory. The study used a multimodal approach with qualitative and quantitative data sets from 14 patients. Four predominant types of coping, with distinct physiological, cognitive, affective, and psychosocial features, existed among patients going through a fast track total hip replacement programme. These patients' main concern was to restore their physical and psychosocial integrity, which had been compromised by reduced function and mobility in daily life. To restore integrity they economized their mental resources, while striving to fulfil the expectations of the fast track programme. This goal was achieved by being mentally proactive and physically active. Three out of the four predominant types of coping matched the expectations expressed in the fast track programme. The non-matching behaviour was seen among the most nervous patients, who claimed the right to diverge from the programme. In theory, four predominant types of coping with distinct physiological, cognitive, affective, and psychosocial features occur among patients going through a fast track total hip arthroplasty programme.
Rapid consolidation of new knowledge in adulthood via fast mapping.
Coutanche, Marc N; Thompson-Schill, Sharon L
2015-09-01
Rapid word learning, where words are 'fast mapped' onto new concepts, may help build vocabulary during childhood. Recent evidence has suggested that fast mapping might help to rapidly integrate information into memory networks of the adult neocortex. The neural basis for this learning by fast mapping determines key properties of the learned information. Copyright © 2015 Elsevier Ltd. All rights reserved.
Reading strategies of fast and slow readers.
Haberlandt, K F; Graesser, A C; Schneider, N J
1989-09-01
In three subject-paced experiments we evaluated reading patterns at the word, line, and sentence level for fast and slow readers. A moving-window method was used to collect word reading times for natural texts. At the word level, reading times of word N were influenced by features of word N-1 for fast readers but not for slow readers. The lag effect exhibited by fast readers indicates that they continue to process a word when it is no longer in view, thus limiting the notion of immediate processing. Contrary to our initial expectation that fast readers would process only a single new argument from a sentence, whereas slow readers would process several new arguments, we found that both reader groups adopted a many-argument strategy. However, fast and slow readers differed in terms of the text units (lines vs. sentences) defining the new-argument effects: Fast readers exhibited greater new-argument effects relative to lines, whereas slow readers exhibited greater new-argument effects relative to sentences. Specifically, slow readers integrated the new arguments primarily at the end of the sentence, whereas fast readers did so at line boundaries. These results are discussed in terms of a buffer-and-integrate model of reading comprehension.
PCEMCAN - Probabilistic Ceramic Matrix Composites Analyzer: User's Guide, Version 1.0
NASA Technical Reports Server (NTRS)
Shah, Ashwin R.; Mital, Subodh K.; Murthy, Pappu L. N.
1998-01-01
PCEMCAN (Probabalistic CEramic Matrix Composites ANalyzer) is an integrated computer code developed at NASA Lewis Research Center that simulates uncertainties associated with the constituent properties, manufacturing process, and geometric parameters of fiber reinforced ceramic matrix composites and quantifies their random thermomechanical behavior. The PCEMCAN code can perform the deterministic as well as probabilistic analyses to predict thermomechanical properties. This User's guide details the step-by-step procedure to create input file and update/modify the material properties database required to run PCEMCAN computer code. An overview of the geometric conventions, micromechanical unit cell, nonlinear constitutive relationship and probabilistic simulation methodology is also provided in the manual. Fast probability integration as well as Monte-Carlo simulation methods are available for the uncertainty simulation. Various options available in the code to simulate probabilistic material properties and quantify sensitivity of the primitive random variables have been described. The description of deterministic as well as probabilistic results have been described using demonstration problems. For detailed theoretical description of deterministic and probabilistic analyses, the user is referred to the companion documents "Computational Simulation of Continuous Fiber-Reinforced Ceramic Matrix Composite Behavior," NASA TP-3602, 1996 and "Probabilistic Micromechanics and Macromechanics for Ceramic Matrix Composites", NASA TM 4766, June 1997.
... More data FRAX-based Estimates of 10-year Probability of Hip and Major Osteoporotic Fracture Among Adults ... 2016 Content source: CDC/National Center for Health Statistics Email Recommend Tweet YouTube Instagram Listen Watch RSS ...
Stochastic seismic inversion based on an improved local gradual deformation method
NASA Astrophysics Data System (ADS)
Yang, Xiuwei; Zhu, Peimin
2017-12-01
A new stochastic seismic inversion method based on the local gradual deformation method is proposed, which can incorporate seismic data, well data, geology and their spatial correlations into the inversion process. Geological information, such as sedimentary facies and structures, could provide significant a priori information to constrain an inversion and arrive at reasonable solutions. The local a priori conditional cumulative distributions at each node of model to be inverted are first established by indicator cokriging, which integrates well data as hard data and geological information as soft data. Probability field simulation is used to simulate different realizations consistent with the spatial correlations and local conditional cumulative distributions. The corresponding probability field is generated by the fast Fourier transform moving average method. Then, optimization is performed to match the seismic data via an improved local gradual deformation method. Two improved strategies are proposed to be suitable for seismic inversion. The first strategy is that we select and update local areas of bad fitting between synthetic seismic data and real seismic data. The second one is that we divide each seismic trace into several parts and obtain the optimal parameters for each part individually. The applications to a synthetic example and a real case study demonstrate that our approach can effectively find fine-scale acoustic impedance models and provide uncertainty estimations.
ERIC Educational Resources Information Center
MacRoy-Higgins, Michelle; Dalton, Kevin Patrick
2015-01-01
Purpose: The purpose of this study was to examine the influence of phonotactic probability on sublexical (phonological) and lexical representations in 3-year-olds who had a history of being late talkers in comparison with their peers with typical language development. Method: Ten 3-year-olds who were late talkers and 10 age-matched typically…
Retinoic acid concentrations in patients with squamous cell carcinoma of the head and neck.
Wahlberg, P; Fex, G
1996-02-01
The serum concentrations of all-trans (atRA) and 13-cis (13cRA) retinoic acid were determined by high performance liquid chromatography in 27 patients with squamous cell carcinoma of the head and neck and in 80 healthy controls. This investigation seemed relevant as ethanol is an aetiological factor in these cancers and has been suggested to interfere with the synthesis of atRA. Neither the serum concentration of atRA nor that of 13cRA differed between patients and controls. The serum atRA concentration did not differ between fasting and non-fasting patients, but the serum 13cRA concentration was significantly higher in non-fasting than in fasting patients, probably due to the dietary retinoid content.
Exact numerical calculation of fixation probability and time on graphs.
Hindersin, Laura; Möller, Marius; Traulsen, Arne; Bauer, Benedikt
2016-12-01
The Moran process on graphs is a popular model to study the dynamics of evolution in a spatially structured population. Exact analytical solutions for the fixation probability and time of a new mutant have been found for only a few classes of graphs so far. Simulations are time-expensive and many realizations are necessary, as the variance of the fixation times is high. We present an algorithm that numerically computes these quantities for arbitrary small graphs by an approach based on the transition matrix. The advantage over simulations is that the calculation has to be executed only once. Building the transition matrix is automated by our algorithm. This enables a fast and interactive study of different graph structures and their effect on fixation probability and time. We provide a fast implementation in C with this note (Hindersin et al., 2016). Our code is very flexible, as it can handle two different update mechanisms (Birth-death or death-Birth), as well as arbitrary directed or undirected graphs. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
2009-09-14
described for over 20 years in the medical literature. Other less common names include Obesity Dyslipidemia Syndrome, Insulin Resistance Syndrome...abnormalities. The EGIR definition differed from the WHO definition by including treatment for hypertension or dyslipidemia and excluding type 2 DM as...Borderline dyslipidemia and fasting glucose similarly tend to worsen over time.22,23 Therefore, it is again probable that this study actually
NASA Technical Reports Server (NTRS)
Kershaw, David S.; Prasad, Manoj K.; Beason, J. Douglas
1986-01-01
The Klein-Nishina differential cross section averaged over a relativistic Maxwellian electron distribution is analytically reduced to a single integral, which can then be rapidly evaluated in a variety of ways. A particularly fast method for numerically computing this single integral is presented. This is, to the authors' knowledge, the first correct computation of the Compton scattering kernel.
Integrated-Circuit Pseudorandom-Number Generator
NASA Technical Reports Server (NTRS)
Steelman, James E.; Beasley, Jeff; Aragon, Michael; Ramirez, Francisco; Summers, Kenneth L.; Knoebel, Arthur
1992-01-01
Integrated circuit produces 8-bit pseudorandom numbers from specified probability distribution, at rate of 10 MHz. Use of Boolean logic, circuit implements pseudorandom-number-generating algorithm. Circuit includes eight 12-bit pseudorandom-number generators, outputs are uniformly distributed. 8-bit pseudorandom numbers satisfying specified nonuniform probability distribution are generated by processing uniformly distributed outputs of eight 12-bit pseudorandom-number generators through "pipeline" of D flip-flops, comparators, and memories implementing conditional probabilities on zeros and ones.
Silicon photonic integrated circuit for fast and precise dual-comb distance metrology.
Weimann, C; Lauermann, M; Hoeller, F; Freude, W; Koos, C
2017-11-27
We demonstrate an optical distance sensor integrated on a silicon photonic chip with a footprint of well below 1 mm 2 . The integrated system comprises a heterodyne receiver structure with tunable power splitting ratio and on-chip photodetectors. The functionality of the device is demonstrated in a synthetic-wavelength interferometry experiment using frequency combs as optical sources. We obtain accurate and fast distance measurements with an unambiguity range of 3.75 mm, a root-mean-square error of 3.4 µm and acquisition times of 14 µs.
Continuous-variable quantum key distribution in uniform fast-fading channels
NASA Astrophysics Data System (ADS)
Papanastasiou, Panagiotis; Weedbrook, Christian; Pirandola, Stefano
2018-03-01
We investigate the performance of several continuous-variable quantum key distribution protocols in the presence of uniform fading channels. These are lossy channels whose transmissivity changes according to a uniform probability distribution. We assume the worst-case scenario where an eavesdropper induces a fast-fading process, where she chooses the instantaneous transmissivity while the remote parties may only detect the mean statistical effect. We analyze coherent-state protocols in various configurations, including the one-way switching protocol in reverse reconciliation, the measurement-device-independent protocol in the symmetric configuration, and its extension to a three-party network. We show that, regardless of the advantage given to the eavesdropper (control of the fading), these protocols can still achieve high rates under realistic attacks, within reasonable values for the variance of the probability distribution associated with the fading process.
Small violations of Bell inequalities for multipartite pure random states
NASA Astrophysics Data System (ADS)
Drumond, Raphael C.; Duarte, Cristhiano; Oliveira, Roberto I.
2018-05-01
For any finite number of parts, measurements, and outcomes in a Bell scenario, we estimate the probability of random N-qudit pure states to substantially violate any Bell inequality with uniformly bounded coefficients. We prove that under some conditions on the local dimension, the probability to find any significant amount of violation goes to zero exponentially fast as the number of parts goes to infinity. In addition, we also prove that if the number of parts is at least 3, this probability also goes to zero as the local Hilbert space dimension goes to infinity.
Tahirbegi, Islam Bogachan; Ehgartner, Josef; Sulzer, Philipp; Zieger, Silvia; Kasjanow, Alice; Paradiso, Mirco; Strobl, Martin; Bouwes, Dominique; Mayr, Torsten
2017-02-15
The necessities of developing fast, portable, cheap and easy to handle pesticide detection platforms are getting attention of scientific and industrial communities. Although there are some approaches to develop microchip based pesticide detection platforms, there is no compact microfluidic device for the complementary, fast, cheap, reusable and reliable analysis of different pesticides. In this work, a microfluidic device is developed for in-situ analysis of pesticide concentration detected via metabolism/photosynthesis of Chlamydomonas reinhardtii algal cells (algae) in tap water. Algae are grown in glass based microfluidic chip, which contains integrated optical pH and oxygen sensors in a portable system for on-site detection. In addition, intrinsic algal fluorescence is detected to analyze the pesticide concentration in parallel to pH and oxygen sensors with integrated fluorescence detectors. The response of the algae under the effect of different concentrations of pesticides is evaluated and complementary inhibition effects depending on the pesticide concentration are demonstrated. The three different sensors allow the determination of various pesticide concentrations in the nanomolar concentration range. The miniaturized system provides the fast quantification of pesticides in less than 10min and enables the study of toxic effects of different pesticides on Chlamydomonas reinhardtii green algae. Consequently, the microfluidic device described here provides fast and complementary detection of different pesticides with algae in a novel glass based microfluidic device with integrated optical pH, oxygen sensors and algal fluorescence. Copyright © 2016 Elsevier B.V. All rights reserved.
On-the-fly Doppler broadening of unresolved resonance region cross sections
Walsh, Jonathan A.; Forget, Benoit; Smith, Kord S.; ...
2017-07-29
In this paper, two methods for computing temperature-dependent unresolved resonance region cross sections on-the-fly within continuous-energy Monte Carlo neutron transport simulations are presented. The first method calculates Doppler broadened cross sections directly from zero-temperature average resonance parameters. In a simulation, at each event that requires cross section values, a realization of unresolved resonance parameters is generated about the desired energy and temperature-dependent single-level Breit-Wigner resonance cross sections are computed directly via the analytical Ψ-x Doppler integrals. The second method relies on the generation of equiprobable cross section magnitude bands on an energy-temperature mesh. Within a simulation, the bands are sampledmore » and interpolated in energy and temperature to obtain cross section values on-the-fly. Both of the methods, as well as their underlying calculation procedures, are verified numerically in extensive code-to-code comparisons. Energy-dependent pointwise cross sections calculated with the newly-implemented procedures are shown to be in excellent agreement with those calculated by a widely-used nuclear data processing code. Relative differences at or below 0.1% are observed. Integral criticality benchmark results computed with the proposed methods are shown to reproduce those computed with a state-of-the-art processed nuclear data library very well. In simulations of fast spectrum systems which are highly-sensitive to the representation of cross section data in the unresolved region, k-eigenvalue and neutron flux spectra differences of <10 pcm and <1.0% are observed, respectively. The direct method is demonstrated to be well-suited to the calculation of reference solutions — against which results obtained with a discretized representation may be assessed — as a result of its treatment of the energy, temperature, and cross section magnitude variables as continuous. Also, because there is no pre-processed data to store (only temperature-independent average resonance parameters) the direct method is very memory-efficient. Typically, only a few kB of memory are needed to store all required unresolved region data for a single nuclide. However, depending on the details of a particular simulation, performing URR cross section calculations on-the-fly can significantly increase simulation times. Alternatively, the method of interpolating equiprobable probability bands is demonstrated to produce results that are as accurate as the direct reference solutions, to within arbitrary precision, with high computational efficiency in terms of memory requirements and simulation time. Analyses of a fast spectrum system show that interpolation on a coarse energy-temperature mesh can be used to reproduce reference k-eigenvalue results obtained with cross sections calculated continuously in energy and directly at an exact temperature to within <10 pcm. Probability band data on a mesh encompassing the range of temperatures relevant to reactor analysis usually require around 100 kB of memory per nuclide. Finally, relative to the case in which probability table data generated at a single, desired temperature are used, minor increases in simulation times are observed when probability band interpolation is employed.« less
Some New Twists to Problems Involving the Gaussian Probability Integral
NASA Technical Reports Server (NTRS)
Simon, Marvin K.; Divsalar, Dariush
1997-01-01
Using an alternate form of the Gaussian probability integral discovered a number of years ago, it is shown that the solution to a number of previously considered communication problems can be simplified and in some cases made more accurate(i.e., exact rather than bounded).
Neural Encoding and Integration of Learned Probabilistic Sequences in Avian Sensory-Motor Circuitry
Brainard, Michael S.
2013-01-01
Many complex behaviors, such as human speech and birdsong, reflect a set of categorical actions that can be flexibly organized into variable sequences. However, little is known about how the brain encodes the probabilities of such sequences. Behavioral sequences are typically characterized by the probability of transitioning from a given action to any subsequent action (which we term “divergence probability”). In contrast, we hypothesized that neural circuits might encode the probability of transitioning to a given action from any preceding action (which we term “convergence probability”). The convergence probability of repeatedly experienced sequences could naturally become encoded by Hebbian plasticity operating on the patterns of neural activity associated with those sequences. To determine whether convergence probability is encoded in the nervous system, we investigated how auditory-motor neurons in vocal premotor nucleus HVC of songbirds encode different probabilistic characterizations of produced syllable sequences. We recorded responses to auditory playback of pseudorandomly sequenced syllables from the bird's repertoire, and found that variations in responses to a given syllable could be explained by a positive linear dependence on the convergence probability of preceding sequences. Furthermore, convergence probability accounted for more response variation than other probabilistic characterizations, including divergence probability. Finally, we found that responses integrated over >7–10 syllables (∼700–1000 ms) with the sign, gain, and temporal extent of integration depending on convergence probability. Our results demonstrate that convergence probability is encoded in sensory-motor circuitry of the song-system, and suggest that encoding of convergence probability is a general feature of sensory-motor circuits. PMID:24198363
Generalized Wishart Mixtures for Unsupervised Classification of PolSAR Data
NASA Astrophysics Data System (ADS)
Li, Lan; Chen, Erxue; Li, Zengyuan
2013-01-01
This paper presents an unsupervised clustering algorithm based upon the expectation maximization (EM) algorithm for finite mixture modelling, using the complex wishart probability density function (PDF) for the probabilities. The mixture model enables to consider heterogeneous thematic classes which could not be better fitted by the unimodal wishart distribution. In order to make it fast and robust to calculate, we use the recently proposed generalized gamma distribution (GΓD) for the single polarization intensity data to make the initial partition. Then we use the wishart probability density function for the corresponding sample covariance matrix to calculate the posterior class probabilities for each pixel. The posterior class probabilities are used for the prior probability estimates of each class and weights for all class parameter updates. The proposed method is evaluated and compared with the wishart H-Alpha-A classification. Preliminary results show that the proposed method has better performance.
Fast molecular shocks. II - Emission from fast dissociative shocks
NASA Technical Reports Server (NTRS)
Neufeld, David A.; Dalgarno, A.
1989-01-01
The line radiations emitted in the cooling gas behind a fast dissociative shock are studied. The intensities emitted in high rotational transitions of the neutral molecules CO, SiO, HCN, CN, NO, and SO are estimated, as well as in rovibrational transitions of the molecular ions HeH(+) and OH(+) in radio recombination lines of atomic hydrogen and in fine-structure transitions of C, C(+), O, and Si(+). The predictions are compared with the observed intensities of line emission from the Orion-KL region. For Orion-KL the observations do not exclude, but probably do not require, the presence of a fast dissociative shock. Emission from SiO in high-J rotational states and from vibrationally excited OH(+), HeH(+), HeH(+), and SO(+) may be detectable from dissociative shocks under suitable conditions of preshock density and shock velocity; such emission may prove to be a useful diagnostic probe of fast shock activity.
Gravitational Instabilities in Disks: From Polytropes to Protoplanets?
NASA Astrophysics Data System (ADS)
Durisen, R. H.
2004-12-01
Gravitational instabilities (GI's) probably occur in disks around young stellar objects during their early embedded phase. This paper reviews what is known about the nonlinear consequences of GI's for planet formation and disk evolution. All researchers agree that, for sufficiently fast cooling, disks fragment into dense clumps or arclike structures, but there is no universal agreement about whether fast enough cooling to cause fragmentation ever occurs and, if it does, whether any clumps that form will become bound protoplanets.
NASA Astrophysics Data System (ADS)
Solodov, A. A.; Theobald, W.; Anderson, K. S.; Shvydky, A.; Epstein, R.; Betti, R.; Myatt, J. F.; Stoeckl, C.; Jarrott, L. C.; McGuffey, C.; Qiao, B.; Beg, F. N.; Wei, M. S.; Stephens, R. B.
2013-10-01
Integrated fast-ignition experiments on OMEGA benefit from improved performance of the OMEGA EP laser, including higher contrast, higher energy, and a smaller focus. Recent 8-keV, Cu-Kα flash radiography of cone-in-shell implosions and cone-tip breakout measurements showed good agreement with the 2-D radiation-hydrodynamic simulations using the code DRACO. DRACO simulations show that the fuel assembly can be further improved by optimizing the compression laser pulse, evacuating air from the shell, and by adjusting the material of the cone tip. This is found to delay the cone-tip breakout by ~220 ps and increase the core areal density from ~80 mg/cm2 in the current experiments to ~500 mg/cm2 at the time of the OMEGA EP beam arrival before the cone-tip breakout. Simulations using the code LSP of fast-electron transport in the recent integrated OMEGA experiments with Cu-doped shells will be presented. Cu-doping is added to probe the transport of fast electrons via their induced Cu K-shell fluorescent emission. This material is based upon work supported by the Department of Energy National Nuclear Security Administration DE-NA0001944 and the Office of Science under DE-FC02-04ER54789.
FAST-TRAC evaluation : evaluation summary report
DOT National Transportation Integrated Search
FAST-TRAC is an Intelligent Transportation System (ITS) that integrates advanced traffic control with a variety of advanced traffic information systems through centralized collection, processing, and dissemination of traffic data. The Road Commission...
Reinforcement Probability Modulates Temporal Memory Selection and Integration Processes
Matell, Matthew S.; Kurti, Allison N.
2013-01-01
We have previously shown that rats trained in a mixed-interval peak procedure (tone = 4s, light = 12s) respond in a scalar manner at a time in between the trained peak times when presented with the stimulus compound (Swanton & Matell, 2011). In our previous work, the two component cues were reinforced with different probabilities (short = 20%, long = 80%) to equate response rates, and we found that the compound peak time was biased toward the cue with the higher reinforcement probability. Here, we examined the influence that different reinforcement probabilities have on the temporal location and shape of the compound response function. We found that the time of peak responding shifted as a function of the relative reinforcement probability of the component cues, becoming earlier as the relative likelihood of reinforcement associated with the short cue increased. However, as the relative probabilities of the component cues grew dissimilar, the compound peak became non-scalar, suggesting that the temporal control of behavior shifted from a process of integration to one of selection. As our previous work has utilized durations and reinforcement probabilities more discrepant than those used here, these data suggest that the processes underlying the integration/selection decision for time are based on cue value. PMID:23896560
A Tomographic Method for the Reconstruction of Local Probability Density Functions
NASA Technical Reports Server (NTRS)
Sivathanu, Y. R.; Gore, J. P.
1993-01-01
A method of obtaining the probability density function (PDF) of local properties from path integrated measurements is described. The approach uses a discrete probability function (DPF) method to infer the PDF of the local extinction coefficient from measurements of the PDFs of the path integrated transmittance. The local PDFs obtained using the method are compared with those obtained from direct intrusive measurements in propylene/air and ethylene/air diffusion flames. The results of this comparison are good.
NASA Technical Reports Server (NTRS)
Young, D. P.; Woo, A. C.; Bussoletti, J. E.; Johnson, F. T.
1986-01-01
A general method is developed combining fast direct methods and boundary integral equation methods to solve Poisson's equation on irregular exterior regions. The method requires O(N log N) operations where N is the number of grid points. Error estimates are given that hold for regions with corners and other boundary irregularities. Computational results are given in the context of computational aerodynamics for a two-dimensional lifting airfoil. Solutions of boundary integral equations for lifting and nonlifting aerodynamic configurations using preconditioned conjugate gradient are examined for varying degrees of thinness.
Kaushik, Aniruddha M; Hsieh, Kuangwen; Chen, Liben; Shin, Dong Jin; Liao, Joseph C; Wang, Tza-Huei
2017-11-15
There remains an urgent need for rapid diagnostic methods that can evaluate antibiotic resistance for pathogenic bacteria in order to deliver targeted antibiotic treatments. Toward this end, we present a rapid and integrated single-cell biosensing platform, termed dropFAST, for bacterial growth detection and antimicrobial susceptibility assessment. DropFAST utilizes a rapid resazurin-based fluorescent growth assay coupled with stochastic confinement of bacteria in 20 pL droplets to detect signal from growing bacteria after 1h incubation, equivalent to 2-3 bacterial replications. Full integration of droplet generation, incubation, and detection into a single, uninterrupted stream also renders this platform uniquely suitable for in-line bacterial phenotypic growth assessment. To illustrate the concept of rapid digital antimicrobial susceptibility assessment, we employ the dropFAST platform to evaluate the antibacterial effect of gentamicin on E. coli growth. Copyright © 2017 Elsevier B.V. All rights reserved.
Performance of a fast digital integrator in on-field magnetic measurements for particle accelerators
NASA Astrophysics Data System (ADS)
Arpaia, P.; Bottura, L.; Fiscarelli, L.; Walckiers, L.
2012-02-01
The fast digital integrator has been conceived to face most demanding magnet test requirements with a resolution of 10 ppm, a signal-to-noise ratio of 105 dB at 20 kHz, a time resolution of 50 ns, an offset of 10 ppm, and on-line processing. In this paper, the on-field achievements of the fast digital integrator are assessed by a specific measurement campaign at the European Organization for Nuclear Research (CERN). At first, the architecture and the metrological specifications of the instrument are reported. Then, the recent on-field achievements of (i) ±10 ppm of uncertainty in the measurement of the main field for superconducting magnets characterization, (ii) ±0.02 % of field uncertainty in quality assessment of small-aperture permanent magnets, and (iii) ±0.15 % of drift, in an excitation current measurement of 600 s under cryogenic conditions, are presented and discussed.
Integrated Life-Cycle Framework for Maintenance, Monitoring and Reliability of Naval Ship Structures
2012-08-15
number of times, a fast and accurate method for analyzing the ship hull is required. In order to obtain this required computational speed and accuracy...Naval Engineers Fleet Maintenance & Modernization Symposium (FMMS 2011) [8] and the Eleventh International Conference on Fast Sea Transportation ( FAST ...probabilistic strength of the ship hull. First, a novel deterministic method for the fast and accurate calculation of the strength of the ship hull is
Effects of stop-signal probability in the stop-signal paradigm: the N2/P3 complex further validated.
Ramautar, J R; Kok, A; Ridderinkhof, K R
2004-11-01
The aim of this study was to examine the effects of frequency of occurrence of stop signals in the stop-signal paradigm. Presenting stop signals less frequently resulted in faster reaction times to the go stimulus and a lower probability of inhibition. Also, go stimuli elicited larger and somewhat earlier P3 responses when stop signals occurred less frequently. Since the amplitude effect was more pronounced on trials when go signals were followed by fast than slow reactions, it probably reflected a stronger set to produce fast responses. N2 and P3 components to stop signals were observed to be larger and of longer latency when stop signals occurred less frequently. The amplitude enhancement of these N2 and P3 components were more pronounced for unsuccessful than for successful stop-signal trials. Moreover, the successfully inhibited stop trials elicited a frontocentral P3 whereas unsuccessfully inhibited stop trials elicited a more posterior P3 that resembled the classical P3b. P3 amplitude in the unsuccessfully inhibited condition also differed between waveforms synchronized with the stop signal and waveforms synchronized with response onset whereas N2 amplitude did not. Taken together these findings suggest that N2 reflected a greater significance of failed inhibitions after low probability stop signals while P3 reflected continued processing of the erroneous response after response execution.
Approximating Integrals Using Probability
ERIC Educational Resources Information Center
Maruszewski, Richard F., Jr.; Caudle, Kyle A.
2005-01-01
As part of a discussion on Monte Carlo methods, which outlines how to use probability expectations to approximate the value of a definite integral. The purpose of this paper is to elaborate on this technique and then to show several examples using visual basic as a programming tool. It is an interesting method because it combines two branches of…
Huerta, Miguel; Ortiz-Mesina, Mónica; Trujillo, Xóchitl; Sánchez-Pastor, Enrique; Vásquez, Clemente; Castro, Elena; Velasco, Raymundo; Montoya-Pérez, Rocío; Onetti, Carlos
2009-05-01
The effect of cannabinoids on caffeine contractures was investigated in slow and fast skeletal muscle fibers using isometric tension recording. In slow muscle fibers, WIN 55,212-2 (10 and 5 microM) caused a decrease in tension. These doses reduced maximum tension to 67.43 +/- 8.07% (P = 0.02, n = 5) and 79.4 +/- 14.11% (P = 0.007, n = 5) compared to control, respectively. Tension-time integral was reduced to 58.37 +/- 7.17% and 75.10 +/- 3.60% (P = 0.002, n = 5), respectively. Using the CB(1) cannabinoid receptor agonist ACPA (1 microM) reduced the maximum tension of caffeine contractures by 68.70 +/- 11.63% (P = 0.01, n = 5); tension-time integral was reduced by 66.82 +/- 6.89% (P = 0.02, n = 5) compared to controls. When the CB(1) receptor antagonist AM281 was coapplied with ACPA, it reversed the effect of ACPA on caffeine-evoked tension. In slow and fast muscle fibers incubated with the pertussis toxin, ACPA had no effect on tension evoked by caffeine. In fast muscle fibers, ACPA (1 microM) also decreased tension; the maximum tension was reduced by 56.48 +/- 3.4% (P = 0.001, n = 4), and tension-time integral was reduced by 57.81 +/- 2.6% (P = 0.006, n = 4). This ACPA effect was not statistically significant with respect to the reduction in tension in slow muscle fibers. Moreover, we detected the presence of mRNA for the cannabinoid CB(1) receptor on fast and slow skeletal muscle fibers, which was significantly higher in fast compared to slow muscle fiber expression. In conclusion, our results suggest that in the slow and fast muscle fibers of the frog cannabinoids diminish caffeine-evoked tension through a receptor-mediated mechanism.
The fast multipole method and point dipole moment polarizable force fields.
Coles, Jonathan P; Masella, Michel
2015-01-14
We present an implementation of the fast multipole method for computing Coulombic electrostatic and polarization forces from polarizable force-fields based on induced point dipole moments. We demonstrate the expected O(N) scaling of that approach by performing single energy point calculations on hexamer protein subunits of the mature HIV-1 capsid. We also show the long time energy conservation in molecular dynamics at the nanosecond scale by performing simulations of a protein complex embedded in a coarse-grained solvent using a standard integrator and a multiple time step integrator. Our tests show the applicability of fast multipole method combined with state-of-the-art chemical models in molecular dynamical systems.
Statistical analysis of native contact formation in the folding of designed model proteins
NASA Astrophysics Data System (ADS)
Tiana, Guido; Broglia, Ricardo A.
2001-02-01
The time evolution of the formation probability of native bonds has been studied for designed sequences which fold fast into the native conformation. From this analysis a clear hierarchy of bonds emerge: (a) local, fast forming highly stable native bonds built by some of the most strongly interacting amino acids of the protein; (b) nonlocal bonds formed late in the folding process, in coincidence with the folding nucleus, and involving essentially the same strongly interacting amino acids already participating in the fast bonds; (c) the rest of the native bonds whose behavior is subordinated, to a large extent, to that of the strong local and nonlocal native contacts.
Combining endoscopic ultrasound with Time-Of-Flight PET: The EndoTOFPET-US Project
NASA Astrophysics Data System (ADS)
Frisch, Benjamin
2013-12-01
The EndoTOFPET-US collaboration develops a multimodal imaging technique for endoscopic exams of the pancreas or the prostate. It combines the benefits of high resolution metabolic imaging with Time-Of-Flight Positron Emission Tomography (TOF PET) and anatomical imaging with ultrasound (US). EndoTOFPET-US consists of a PET head extension for a commercial US endoscope and a PET plate outside the body in coincidence with the head. The high level of miniaturization and integration creates challenges in fields such as scintillating crystals, ultra-fast photo-detection, highly integrated electronics, system integration and image reconstruction. Amongst the developments, fast scintillators as well as fast and compact digital SiPMs with single SPAD readout are used to obtain the best coincidence time resolution (CTR). Highly integrated ASICs and DAQ electronics contribute to the timing performances of EndoTOFPET. In view of the targeted resolution of around 1 mm in the reconstructed image, we present a prototype detector system with a CTR better than 240 ps FWHM. We discuss the challenges in simulating such a system and introduce reconstruction algorithms based on graphics processing units (GPU).
Connectivity is a Poor Indicator of Fast Quantum Search
NASA Astrophysics Data System (ADS)
Meyer, David A.; Wong, Thomas G.
2015-03-01
A randomly walking quantum particle evolving by Schrödinger's equation searches on d -dimensional cubic lattices in O (√{N }) time when d ≥5 , and with progressively slower runtime as d decreases. This suggests that graph connectivity (including vertex, edge, algebraic, and normalized algebraic connectivities) is an indicator of fast quantum search, a belief supported by fast quantum search on complete graphs, strongly regular graphs, and hypercubes, all of which are highly connected. In this Letter, we show this intuition to be false by giving two examples of graphs for which the opposite holds true: one with low connectivity but fast search, and one with high connectivity but slow search. The second example is a novel two-stage quantum walk algorithm in which the walking rate must be adjusted to yield high search probability.
Atir-Sharon, Tali; Gilboa, Asaf; Hazan, Hananel; Koilis, Ester; Manevitz, Larry M
2015-01-01
Neocortical structures typically only support slow acquisition of declarative memory; however, learning through fast mapping may facilitate rapid learning-induced cortical plasticity and hippocampal-independent integration of novel associations into existing semantic networks. During fast mapping the meaning of new words and concepts is inferred, and durable novel associations are incidentally formed, a process thought to support early childhood's exuberant learning. The anterior temporal lobe, a cortical semantic memory hub, may critically support such learning. We investigated encoding of semantic associations through fast mapping using fMRI and multivoxel pattern analysis. Subsequent memory performance following fast mapping was more efficiently predicted using anterior temporal lobe than hippocampal voxels, while standard explicit encoding was best predicted by hippocampal activity. Searchlight algorithms revealed additional activity patterns that predicted successful fast mapping semantic learning located in lateral occipitotemporal and parietotemporal neocortex and ventrolateral prefrontal cortex. By contrast, successful explicit encoding could be classified by activity in medial and dorsolateral prefrontal and parahippocampal cortices. We propose that fast mapping promotes incidental rapid integration of new associations into existing neocortical semantic networks by activating related, nonoverlapping conceptual knowledge. In healthy adults, this is better captured by unique anterior and lateral temporal lobe activity patterns, while hippocampal involvement is less predictive of this kind of learning.
Invariance of separability probability over reduced states in 4 × 4 bipartite systems
NASA Astrophysics Data System (ADS)
Lovas, Attila; Andai, Attila
2017-07-01
The geometric separability probability of composite quantum systems has been extensively studied in the recent decades. One of the simplest but strikingly difficult problem is to compute the separability probability of qubit-qubit and rebit-rebit quantum states with respect to the Hilbert-Schmidt measure. A lot of numerical simulations confirm the P{rebit - rebit}=\\frac{29}{64} and P{qubit-qubit}=\\frac{8}{33} conjectured probabilities. We provide a rigorous proof for the separability probability in the real case and we give explicit integral formulas for the complex and quaternionic case. Milz and Strunz studied the separability probability with respect to given subsystems. They conjectured that the separability probability of qubit-qubit (and qubit-qutrit) states of the form of ≤ft(\\begin{array}{@{}cc@{}} D1 & C \\ C* & D2 \\end{array}\\right) depends on D=D1+D2 (on single qubit subsystems), moreover it depends only on the Bloch radii (r) of D and it is constant in r. Using the Peres-Horodecki criterion for separability we give a mathematical proof for the \\frac{29}{64} probability and we present an integral formula for the complex case which hopefully will help to prove the \\frac{8}{33} probability, too. We prove Milz and Strunz’s conjecture for rebit-rebit and qubit-qubit states. The case, when the state space is endowed with the volume form generated by the operator monotone function f(x)=\\sqrt{x} is also studied in detail. We show that even in this setting Milz and Strunz’s conjecture holds true and we give an integral formula for separability probability according to this measure.
The Priority Heuristic: Making Choices Without Trade-Offs
Brandstätter, Eduard; Gigerenzer, Gerd; Hertwig, Ralph
2010-01-01
Bernoulli's framework of expected utility serves as a model for various psychological processes, including motivation, moral sense, attitudes, and decision making. To account for evidence at variance with expected utility, we generalize the framework of fast and frugal heuristics from inferences to preferences. The priority heuristic predicts (i) Allais' paradox, (ii) risk aversion for gains if probabilities are high, (iii) risk seeking for gains if probabilities are low (lottery tickets), (iv) risk aversion for losses if probabilities are low (buying insurance), (v) risk seeking for losses if probabilities are high, (vi) certainty effect, (vii) possibility effect, and (viii) intransitivities. We test how accurately the heuristic predicts people's choices, compared to previously proposed heuristics and three modifications of expected utility theory: security-potential/aspiration theory, transfer-of-attention-exchange model, and cumulative prospect theory. PMID:16637767
DOE Office of Scientific and Technical Information (OSTI.GOV)
O'Brien, Travis A.; Kashinath, Karthik; Cavanaugh, Nicholas R.
Numerous facets of scientific research implicitly or explicitly call for the estimation of probability densities. Histograms and kernel density estimates (KDEs) are two commonly used techniques for estimating such information, with the KDE generally providing a higher fidelity representation of the probability density function (PDF). Both methods require specification of either a bin width or a kernel bandwidth. While techniques exist for choosing the kernel bandwidth optimally and objectively, they are computationally intensive, since they require repeated calculation of the KDE. A solution for objectively and optimally choosing both the kernel shape and width has recently been developed by Bernacchiamore » and Pigolotti (2011). While this solution theoretically applies to multidimensional KDEs, it has not been clear how to practically do so. A method for practically extending the Bernacchia-Pigolotti KDE to multidimensions is introduced. This multidimensional extension is combined with a recently-developed computational improvement to their method that makes it computationally efficient: a 2D KDE on 10 5 samples only takes 1 s on a modern workstation. This fast and objective KDE method, called the fastKDE method, retains the excellent statistical convergence properties that have been demonstrated for univariate samples. The fastKDE method exhibits statistical accuracy that is comparable to state-of-the-science KDE methods publicly available in R, and it produces kernel density estimates several orders of magnitude faster. The fastKDE method does an excellent job of encoding covariance information for bivariate samples. This property allows for direct calculation of conditional PDFs with fastKDE. It is demonstrated how this capability might be leveraged for detecting non-trivial relationships between quantities in physical systems, such as transitional behavior.« less
Structural Information from Single-molecule FRET Experiments Using the Fast Nano-positioning System
Röcker, Carlheinz; Nagy, Julia; Michaelis, Jens
2017-01-01
Single-molecule Förster Resonance Energy Transfer (smFRET) can be used to obtain structural information on biomolecular complexes in real-time. Thereby, multiple smFRET measurements are used to localize an unknown dye position inside a protein complex by means of trilateration. In order to obtain quantitative information, the Nano-Positioning System (NPS) uses probabilistic data analysis to combine structural information from X-ray crystallography with single-molecule fluorescence data to calculate not only the most probable position but the complete three-dimensional probability distribution, termed posterior, which indicates the experimental uncertainty. The concept was generalized for the analysis of smFRET networks containing numerous dye molecules. The latest version of NPS, Fast-NPS, features a new algorithm using Bayesian parameter estimation based on Markov Chain Monte Carlo sampling and parallel tempering that allows for the analysis of large smFRET networks in a comparably short time. Moreover, Fast-NPS allows the calculation of the posterior by choosing one of five different models for each dye, that account for the different spatial and orientational behavior exhibited by the dye molecules due to their local environment. Here we present a detailed protocol for obtaining smFRET data and applying the Fast-NPS. We provide detailed instructions for the acquisition of the three input parameters of Fast-NPS: the smFRET values, as well as the quantum yield and anisotropy of the dye molecules. Recently, the NPS has been used to elucidate the architecture of an archaeal open promotor complex. This data is used to demonstrate the influence of the five different dye models on the posterior distribution. PMID:28287526
Structural Information from Single-molecule FRET Experiments Using the Fast Nano-positioning System.
Dörfler, Thilo; Eilert, Tobias; Röcker, Carlheinz; Nagy, Julia; Michaelis, Jens
2017-02-09
Single-molecule Förster Resonance Energy Transfer (smFRET) can be used to obtain structural information on biomolecular complexes in real-time. Thereby, multiple smFRET measurements are used to localize an unknown dye position inside a protein complex by means of trilateration. In order to obtain quantitative information, the Nano-Positioning System (NPS) uses probabilistic data analysis to combine structural information from X-ray crystallography with single-molecule fluorescence data to calculate not only the most probable position but the complete three-dimensional probability distribution, termed posterior, which indicates the experimental uncertainty. The concept was generalized for the analysis of smFRET networks containing numerous dye molecules. The latest version of NPS, Fast-NPS, features a new algorithm using Bayesian parameter estimation based on Markov Chain Monte Carlo sampling and parallel tempering that allows for the analysis of large smFRET networks in a comparably short time. Moreover, Fast-NPS allows the calculation of the posterior by choosing one of five different models for each dye, that account for the different spatial and orientational behavior exhibited by the dye molecules due to their local environment. Here we present a detailed protocol for obtaining smFRET data and applying the Fast-NPS. We provide detailed instructions for the acquisition of the three input parameters of Fast-NPS: the smFRET values, as well as the quantum yield and anisotropy of the dye molecules. Recently, the NPS has been used to elucidate the architecture of an archaeal open promotor complex. This data is used to demonstrate the influence of the five different dye models on the posterior distribution.
Liu, S X; Zou, M S
2018-03-01
The radiation loading on a vibratory finite cylindrical shell is conventionally evaluated through the direct numerical integration (DNI) method. An alternative strategy via the fast Fourier transform algorithm is put forward in this work based on the general expression of radiation impedance. To check the feasibility and efficiency of the proposed method, a comparison with DNI is presented through numerical cases. The results obtained using the present method agree well with those calculated by DNI. More importantly, the proposed calculating strategy can significantly save the time cost compared with the conventional approach of straightforward numerical integration.
Probabilistic structural analysis methods for improving Space Shuttle engine reliability
NASA Technical Reports Server (NTRS)
Boyce, L.
1989-01-01
Probabilistic structural analysis methods are particularly useful in the design and analysis of critical structural components and systems that operate in very severe and uncertain environments. These methods have recently found application in space propulsion systems to improve the structural reliability of Space Shuttle Main Engine (SSME) components. A computer program, NESSUS, based on a deterministic finite-element program and a method of probabilistic analysis (fast probability integration) provides probabilistic structural analysis for selected SSME components. While computationally efficient, it considers both correlated and nonnormal random variables as well as an implicit functional relationship between independent and dependent variables. The program is used to determine the response of a nickel-based superalloy SSME turbopump blade. Results include blade tip displacement statistics due to the variability in blade thickness, modulus of elasticity, Poisson's ratio or density. Modulus of elasticity significantly contributed to blade tip variability while Poisson's ratio did not. Thus, a rational method for choosing parameters to be modeled as random is provided.
Probabilistic Assessment of Fracture Progression in Composite Structures
NASA Technical Reports Server (NTRS)
Chamis, Christos C.; Minnetyan, Levon; Mauget, Bertrand; Huang, Dade; Addi, Frank
1999-01-01
This report describes methods and corresponding computer codes that are used to evaluate progressive damage and fracture and to perform probabilistic assessment in built-up composite structures. Structural response is assessed probabilistically, during progressive fracture. The effects of design variable uncertainties on structural fracture progression are quantified. The fast probability integrator (FPI) is used to assess the response scatter in the composite structure at damage initiation. The sensitivity of the damage response to design variables is computed. The methods are general purpose and are applicable to stitched and unstitched composites in all types of structures and fracture processes starting from damage initiation to unstable propagation and to global structure collapse. The methods are demonstrated for a polymer matrix composite stiffened panel subjected to pressure. The results indicated that composite constituent properties, fabrication parameters, and respective uncertainties have a significant effect on structural durability and reliability. Design implications with regard to damage progression, damage tolerance, and reliability of composite structures are examined.
The bacterial envelope as a target for novel anti-MRSA antibiotics.
Van Bambeke, Françoise; Mingeot-Leclercq, Marie-Paule; Struelens, Marc J; Tulkens, Paul M
2008-03-01
Methicillin-resistant Staphylococcus aureus (MRSA) and vancomycin-intermediate S. aureus (VISA) are spreading worldwide, making the search for antibiotics directed against new targets a high priority. Drugs that anchor in the bacterial membrane (e.g. ceragenins and lipopeptides) or that target the bacterial membrane and proteic (lipoglycopeptides) or lipidic (glycodepsipeptides) cell wall precursors seem to have the most potential because they show a fast and extensive bactericidal effect and are probably less prone to select for resistance owing to the difficulty in modifying their targets in a way that is compatible with bacterial survival. The efficacy of lipopeptides and lipoglycopeptides has been demonstrated in the treatment of skin and skin structure infections, and bacteremia caused by resistant S. aureus. Ceragenins and glycodepsipeptides are restricted to topical applications because of their unsatisfactory safety profile. The mode of action, pharmacological and microbiological properties and target indications of these anti-MRSA agents, which function by disturbing membrane integrity, are reviewed in this article.
Multidisciplinary System Reliability Analysis
NASA Technical Reports Server (NTRS)
Mahadevan, Sankaran; Han, Song; Chamis, Christos C. (Technical Monitor)
2001-01-01
The objective of this study is to develop a new methodology for estimating the reliability of engineering systems that encompass multiple disciplines. The methodology is formulated in the context of the NESSUS probabilistic structural analysis code, developed under the leadership of NASA Glenn Research Center. The NESSUS code has been successfully applied to the reliability estimation of a variety of structural engineering systems. This study examines whether the features of NESSUS could be used to investigate the reliability of systems in other disciplines such as heat transfer, fluid mechanics, electrical circuits etc., without considerable programming effort specific to each discipline. In this study, the mechanical equivalence between system behavior models in different disciplines are investigated to achieve this objective. A new methodology is presented for the analysis of heat transfer, fluid flow, and electrical circuit problems using the structural analysis routines within NESSUS, by utilizing the equivalence between the computational quantities in different disciplines. This technique is integrated with the fast probability integration and system reliability techniques within the NESSUS code, to successfully compute the system reliability of multidisciplinary systems. Traditional as well as progressive failure analysis methods for system reliability estimation are demonstrated, through a numerical example of a heat exchanger system involving failure modes in structural, heat transfer and fluid flow disciplines.
Multi-Disciplinary System Reliability Analysis
NASA Technical Reports Server (NTRS)
Mahadevan, Sankaran; Han, Song
1997-01-01
The objective of this study is to develop a new methodology for estimating the reliability of engineering systems that encompass multiple disciplines. The methodology is formulated in the context of the NESSUS probabilistic structural analysis code developed under the leadership of NASA Lewis Research Center. The NESSUS code has been successfully applied to the reliability estimation of a variety of structural engineering systems. This study examines whether the features of NESSUS could be used to investigate the reliability of systems in other disciplines such as heat transfer, fluid mechanics, electrical circuits etc., without considerable programming effort specific to each discipline. In this study, the mechanical equivalence between system behavior models in different disciplines are investigated to achieve this objective. A new methodology is presented for the analysis of heat transfer, fluid flow, and electrical circuit problems using the structural analysis routines within NESSUS, by utilizing the equivalence between the computational quantities in different disciplines. This technique is integrated with the fast probability integration and system reliability techniques within the NESSUS code, to successfully compute the system reliability of multi-disciplinary systems. Traditional as well as progressive failure analysis methods for system reliability estimation are demonstrated, through a numerical example of a heat exchanger system involving failure modes in structural, heat transfer and fluid flow disciplines.
IVHS Institutional and Legal Issues Program: Review of the FAST-TRAC Operational Test
DOT National Transportation Integrated Search
1994-06-01
The FAST-TRAC (Faster and Safer Travel through Traffic Routing and Advanced Controls) operational test attempted to integrate advanced traveler information systems (ATIS) and advanced traffic management system (ATMS) technologies in Oakland County, M...
Pulse propagation in discrete excitatory networks of integrate-and-fire neurons.
Badel, Laurent; Tonnelier, Arnaud
2004-07-01
We study the propagation of solitary waves in a discrete excitatory network of integrate-and-fire neurons. We show the existence and the stability of a fast wave and a family of slow waves. Fast waves are similar to those already described in continuum networks. Stable slow waves have not been previously reported in purely excitatory networks and their propagation is particular to the discrete nature of the network. The robustness of our results is studied in the presence of noise.
Azari, Mansour R; Sadighzadeh, Asghar; Bayatian, Majid
2018-06-19
Accidents have happened in the chemical industries all over the world with serious consequences for the adjacent heavily populated areas. In this study, the impact of the probable hypothetical event, releasing considerable amounts of hydrogen fluoride (HF) as a strong irritant into the atmosphere over the city of Isfahan from a strategic chemical plant, was simulated by computational fluid dynamics (CFD). In this model, the meteorological parameters were integrated into time and space, and dispersion of the pollutants was estimated based on a probable accidental release of HF. According to the hypothetical results of the simulation model in this study, HF clouds reached Isfahan in 20 min and exposed 80% of the general public to HF concentration in the range of 0-34 ppm. Then, they dissipated 240 min after the time of the incident. Supposing the uniform population density within the proximity of the city of Isfahan with the population of 1.75 million, 5% of the population (87,500 people) could be exposed for a few minutes to a HF concentration as high as 34 ppm. This concentration is higher than a very hazardous concentration described as the Immediate Danger to Life and Health (30 ppm). This hypothetical risk evaluation of environmental exposure to HF with the potential of health risks was very instrumental for the general public of Isfahan in terms of risk management. Similar studies based on probable accidental scenarios along with the application of a simulation model for computation of dispersed pollutants are recommended for risk evaluation and management of cities in the developing countries with a fast pace of urbanization around the industrial sites.
Fusion of Scores in a Detection Context Based on Alpha Integration.
Soriano, Antonio; Vergara, Luis; Ahmed, Bouziane; Salazar, Addisson
2015-09-01
We present a new method for fusing scores corresponding to different detectors (two-hypotheses case). It is based on alpha integration, which we have adapted to the detection context. Three optimization methods are presented: least mean square error, maximization of the area under the ROC curve, and minimization of the probability of error. Gradient algorithms are proposed for the three methods. Different experiments with simulated and real data are included. Simulated data consider the two-detector case to illustrate the factors influencing alpha integration and demonstrate the improvements obtained by score fusion with respect to individual detector performance. Two real data cases have been considered. In the first, multimodal biometric data have been processed. This case is representative of scenarios in which the probability of detection is to be maximized for a given probability of false alarm. The second case is the automatic analysis of electroencephalogram and electrocardiogram records with the aim of reproducing the medical expert detections of arousal during sleeping. This case is representative of scenarios in which probability of error is to be minimized. The general superior performance of alpha integration verifies the interest of optimizing the fusing parameters.
Predicting the probability of slip in gait: methodology and distribution study.
Gragg, Jared; Yang, James
2016-01-01
The likelihood of a slip is related to the available and required friction for a certain activity, here gait. Classical slip and fall analysis presumed that a walking surface was safe if the difference between the mean available and required friction coefficients exceeded a certain threshold. Previous research was dedicated to reformulating the classical slip and fall theory to include the stochastic variation of the available and required friction when predicting the probability of slip in gait. However, when predicting the probability of a slip, previous researchers have either ignored the variation in the required friction or assumed the available and required friction to be normally distributed. Also, there are no published results that actually give the probability of slip for various combinations of required and available frictions. This study proposes a modification to the equation for predicting the probability of slip, reducing the previous equation from a double-integral to a more convenient single-integral form. Also, a simple numerical integration technique is provided to predict the probability of slip in gait: the trapezoidal method. The effect of the random variable distributions on the probability of slip is also studied. It is shown that both the required and available friction distributions cannot automatically be assumed as being normally distributed. The proposed methods allow for any combination of distributions for the available and required friction, and numerical results are compared to analytical solutions for an error analysis. The trapezoidal method is shown to be highly accurate and efficient. The probability of slip is also shown to be sensitive to the input distributions of the required and available friction. Lastly, a critical value for the probability of slip is proposed based on the number of steps taken by an average person in a single day.
Techno-economic analysis Process model development for existing and conceptual processes Detailed heat integration Economic analysis of integrated processes Integration of process simulation learnings into control ;Conceptual Process Design and Techno-Economic Assessment of Ex Situ Catalytic Fast Pyrolysis of Biomass: A
A fast low-power optical memory based on coupled micro-ring lasers
NASA Astrophysics Data System (ADS)
Hill, Martin T.; Dorren, Harmen J. S.; de Vries, Tjibbe; Leijtens, Xaveer J. M.; den Besten, Jan Hendrik; Smalbrugge, Barry; Oei, Yok-Siang; Binsma, Hans; Khoe, Giok-Djan; Smit, Meint K.
2004-11-01
The increasing speed of fibre-optic-based telecommunications has focused attention on high-speed optical processing of digital information. Complex optical processing requires a high-density, high-speed, low-power optical memory that can be integrated with planar semiconductor technology for buffering of decisions and telecommunication data. Recently, ring lasers with extremely small size and low operating power have been made, and we demonstrate here a memory element constructed by interconnecting these microscopic lasers. Our device occupies an area of 18 × 40µm2 on an InP/InGaAsP photonic integrated circuit, and switches within 20ps with 5.5fJ optical switching energy. Simulations show that the element has the potential for much smaller dimensions and switching times. Large numbers of such memory elements can be densely integrated and interconnected on a photonic integrated circuit: fast digital optical information processing systems employing large-scale integration should now be viable.
On fast X-ray rotators with long-term periodicities
NASA Technical Reports Server (NTRS)
Naranan, S.; Elsner, R. F.; Darbro, W.; Ramsey, B. D.; Leahy, D. A.; Weisskopf, M. C.; Williams, A. C.; Hardee, P. E.; Sutherland, P. G.; Grindlay, J. E.
1985-01-01
The support of previous SAS 3 spacecraft observations by new data gathered by the Monitor Proportional Counter aboard the HEAO 2 spacecraft indicates that the pulse period history of the 13.5 sec-pulsing X-ray source LMC X-4 is consistent with standard accretion and torque models only if LMC X-4 is a fast rotator for which the accretion torques nearly cancel. This result leads to a neutron star magnetic field strength estimate of about 1.2 x 10 to the 13th G. Strong evidence is noted for Her X-1's status as a fast rotator, while SMC X-1 is probably an intermediate-to-fast rotator. In the context of slaved disk models for these objects, it is noted that the precession periods expected for the companion stars are significantly longer than the observed 1-2 month time scales; slaved disk models are thereby undermined.
Theory of the stopping power of fast multicharged ions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Yudin, G.L.
1991-12-01
The processes of Coulomb excitation and ionization of atoms by a fast charged particle moving along a classical trajectory are studied. The target electrons are described by the Dirac equation, while the field of the incident particle is described by the Lienard-Wiechert potential. The theory is formulated in the form most convenient for investigation of various characteristics of semiclassical atomic collisions. The theory of sudden perturbations, which is valid at high enough velocities for a high projectile charge, is employed to obtain probabilities and cross sections of the Coulomb excitation and ionization of atomic hydrogen by fast multiply charged ions.more » Based on the semiclassical sudden Born approximation, the ionization cross section and the average electronic energy loss of a fast ion in a single collision with an atom are investigated over a wide specific energy range from 500 keV/amu to 50 MeV/amu.« less
High Quality 3D Photonics using Nano Imprint Lithography of Fast Sol-gel Materials.
Bar-On, Ofer; Brenner, Philipp; Siegle, Tobias; Gvishi, Raz; Kalt, Heinz; Lemmer, Uli; Scheuer, Jacob
2018-05-18
A method for the realization of low-loss integrated optical components is proposed and demonstrated. This approach is simple, fast, inexpensive, scalable for mass production, and compatible with both 2D and 3D geometries. The process is based on a novel dual-step soft nano imprint lithography process for producing devices with smooth surfaces, combined with fast sol-gel technology providing highly transparent materials. As a concrete example, this approach is demonstrated on a micro ring resonator made by direct laser writing (DLW) to achieve a quality factor improvement from one hundred thousand to more than 3 million. To the best of our knowledge this also sets a Q-factor record for UV-curable integrated micro-ring resonators. The process supports the integration of many types of materials such as light-emitting, electro-optic, piezo-electric, and can be readily applied to a wide variety of devices such as waveguides, lenses, diffractive elements and more.
Option volatility and the acceleration Lagrangian
NASA Astrophysics Data System (ADS)
Baaquie, Belal E.; Cao, Yang
2014-01-01
This paper develops a volatility formula for option on an asset from an acceleration Lagrangian model and the formula is calibrated with market data. The Black-Scholes model is a simpler case that has a velocity dependent Lagrangian. The acceleration Lagrangian is defined, and the classical solution of the system in Euclidean time is solved by choosing proper boundary conditions. The conditional probability distribution of final position given the initial position is obtained from the transition amplitude. The volatility is the standard deviation of the conditional probability distribution. Using the conditional probability and the path integral method, the martingale condition is applied, and one of the parameters in the Lagrangian is fixed. The call option price is obtained using the conditional probability and the path integral method.
Metabolic Panel: MedlinePlus Health Topic
... as tests of your cholesterol , protein levels, and liver function . You probably need to fast (not eat any ... What's this? GO National Institutes of Health ... government agencies. MedlinePlus also links to health information from non-government Web sites. See our disclaimer about external ...
The whistle and the rattle: the design of sound producing muscles.
Rome, L C; Syme, D A; Hollingworth, S; Lindstedt, S L; Baylor, S M
1996-01-01
Vertebrate sound producing muscles often operate at frequencies exceeding 100 Hz, making them the fastest vertebrate muscles. Like other vertebrate muscle, these sonic muscles are "synchronous," necessitating that calcium be released and resequestered by the sarcoplasmic reticulum during each contraction cycle. Thus to operate at such high frequencies, vertebrate sonic muscles require extreme adaptations. We have found that to generate the "boatwhistle" mating call (approximately 200 Hz), the swimbladder muscle fibers of toadfish have evolved (i) a large and very fast calcium transient, (ii) a fast crossbridge detachment rate, and (iii) probably a fast kinetic off-rate of Ca2+ from troponin. The fibers of the shaker muscle of rattlesnakes have independently evolved similar traits, permitting tail rattling at approximately 90 Hz. PMID:8755609
Note: Development of fast heating inert gas annealing apparatus operated at atmospheric pressure
NASA Astrophysics Data System (ADS)
Das, S. C.; Majumdar, A.; Shripathi, T.; Hippler, R.
2012-04-01
Here, we report the development of a simple, small, fast heating, and portable, homemade, inert gas (Ar) atmospheric annealing setup. Instead of using a conventional heating element, a commercial soldering rod having an encapsulated fast heating heater is used here. The sample holder is made of a block of stainless steel. It takes 200 s to reach 700 °C, and 10 min to cool down. The probability of oxidation or surface contamination has been examined by means of x ray photoelectron spectroscopy of virgin Cu sample after annealing at 600 °C. In addition, we compare the annealing of a hydrogenated carbon nitride film (HCNx) in both a conventional vacuum and our newly developed ambient Ar atmosphere setup.
Variation across individuals and items determine learning outcomes from fast mapping.
Coutanche, Marc N; Koch, Griffin E
2017-11-01
An approach to learning words known as "fast mapping" has been linked to unique neurobiological and behavioral markers in adult humans, including rapid lexical integration. However, the mechanisms supporting fast mapping are still not known. In this study, we sought to help change this by examining factors that modulate learning outcomes. In 90 subjects, we systematically manipulated the typicality of the items used to support fast mapping (foils), and quantified learners' inclination to employ semantic, episodic, and spatial memory through the Survey of Autobiographical Memory (SAM). We asked how these factors affect lexical competition and recognition performance, and then asked how foil typicality and lexical competition are related in an independent dataset. We find that both the typicality of fast mapping foils, and individual differences in how different memory systems are employed, influence lexical competition effects after fast mapping, but not after other learning approaches. Specifically, learning a word through fast mapping with an atypical foil led to lexical competition, while a typical foil led to lexical facilitation. This effect was particularly evident in individuals with a strong tendency to employ semantic memory. We further replicated the relationship between continuous foil atypicality and lexical competition in an independent dataset. These findings suggest that semantic properties of the foils that support fast mapping can influence the degree and nature of subsequent lexical integration. Further, the effects of foils differ based on an individual's tendency to draw-on the semantic memory system. Copyright © 2017 Elsevier Ltd. All rights reserved.
Quantifying Spatially Integrated Floodplain and Wetland Systems for the Conterminous US
Wetlands interact with other waters across a variable connectivity continuum, from permanent to transient, from fast to slow, and from primarily surface water to exclusively groundwater flows. Floodplain wetlands typically experience fast and frequent surface and near-surface gro...
VizieR Online Data Catalog: ynogkm: code for calculating time-like geodesics (Yang+, 2014)
NASA Astrophysics Data System (ADS)
Yang, X.-L.; Wang, J.-C.
2013-11-01
Here we present the source file for a new public code named ynogkm, aim on calculating the time-like geodesics in a Kerr-Newmann spacetime fast. In the code the four Boyer-Lindquis coordinates and proper time are expressed as functions of a parameter p semi-analytically, i.e., r(p), μ(p), φ(p), t(p), and σ(p), by using the Weiers- trass' and Jacobi's elliptic functions and integrals. All of the ellip- tic integrals are computed by Carlson's elliptic integral method, which guarantees the fast speed of the code.The source Fortran file ynogkm.f90 contains three modules: constants, rootfind, ellfunction, and blcoordinates. (3 data files).
Storm-based Cloud-to-Ground Lightning Probabilities and Warnings
NASA Astrophysics Data System (ADS)
Calhoun, K. M.; Meyer, T.; Kingfield, D.
2017-12-01
A new cloud-to-ground (CG) lightning probability algorithm has been developed using machine-learning methods. With storm-based inputs of Earth Networks' in-cloud lightning, Vaisala's CG lightning, multi-radar/multi-sensor (MRMS) radar derived products including the Maximum Expected Size of Hail (MESH) and Vertically Integrated Liquid (VIL), and near storm environmental data including lapse rate and CAPE, a random forest algorithm was trained to produce probabilities of CG lightning up to one-hour in advance. As part of the Prototype Probabilistic Hazard Information experiment in the Hazardous Weather Testbed in 2016 and 2017, National Weather Service forecasters were asked to use this CG lightning probability guidance to create rapidly updating probability grids and warnings for the threat of CG lightning for 0-60 minutes. The output from forecasters was shared with end-users, including emergency managers and broadcast meteorologists, as part of an integrated warning team.
Mustonen, Anne-Mari; Pyykönen, Teija; Puukka, Matti; Asikainen, Juha; Hänninen, Sari; Mononen, Jaakko; Nieminen, Petteri
2006-01-01
This study investigated the physiological adaptations to fasting using the farmed blue fox (Alopex lagopus) as a model for the endangered wild arctic fox. Sixteen blue foxes were fed throughout the winter and 32 blue foxes were fasted for 22 d in Nov-Dec 2002. Half of the fasted blue foxes were food-deprived again for 22 d in Jan-Feb 2003. The farmed blue fox lost weight at a slower rate (0.97-1.02% body mass d(-1)) than observed previously in the arctic fox, possibly due to its higher initial body fat content. The animals experienced occasional fasting-induced hypoglycaemia, but their locomotor activity was not affected. The plasma triacylglycerol and glycerol concentrations were elevated during phase II of fasting indicating stimulated lipolysis, probably induced by the high growth hormone concentrations. The total cholesterol, HDL- and LDL-cholesterol, urea, uric acid and total protein levels and the urea:creatinine ratio decreased during fasting. Although the plasma levels of some essential amino acids increased, the blue foxes did not enter phase III of starvation characterized by stimulated proteolysis during either of the 22-d fasting procedures. Instead of excessive protein catabolism, it is liver dysfunction, indicated by the increased plasma bilirubin levels and alkaline phosphatase, alanine aminotransferase and aspartate aminotransferase activities, that may limit the duration of fasting in the species.
Ho, Daniel W H; Sze, Karen M F; Ng, Irene O L
2015-08-28
Viral integration into the human genome upon infection is an important risk factor for various human malignancies. We developed viral integration site detection tool called Virus-Clip, which makes use of information extracted from soft-clipped sequencing reads to identify exact positions of human and virus breakpoints of integration events. With initial read alignment to virus reference genome and streamlined procedures, Virus-Clip delivers a simple, fast and memory-efficient solution to viral integration site detection. Moreover, it can also automatically annotate the integration events with the corresponding affected human genes. Virus-Clip has been verified using whole-transcriptome sequencing data and its detection was validated to have satisfactory sensitivity and specificity. Marked advancement in performance was detected, compared to existing tools. It is applicable to versatile types of data including whole-genome sequencing, whole-transcriptome sequencing, and targeted sequencing. Virus-Clip is available at http://web.hku.hk/~dwhho/Virus-Clip.zip.
Improving biomass pyrolysis economics by integrating vapor and liquid phase upgrading
Iisa, Kristiina; Robichaud, David J.; Watson, Michael J.; ...
2017-11-24
Partial deoxygenation of bio-oil by catalytic fast pyrolysis with subsequent coupling and hydrotreating can lead to improved economics and will aid commercial deployment of pyrolytic conversion of biomass technologies. Biomass pyrolysis efficiently depolymerizes and deconstructs solid plant matter into carbonaceous molecules that, upon catalytic upgrading, can be used for fuels and chemicals. Upgrading strategies include catalytic deoxygenation of the vapors before they are condensed (in situ and ex situ catalytic fast pyrolysis), or hydrotreating following condensation of the bio-oil. In general, deoxygenation carbon efficiencies, one of the most important cost drivers, are typically higher for hydrotreating when compared to catalyticmore » fast pyrolysis alone. However, using catalytic fast pyrolysis as the primary conversion step can benefit the entire process chain by: (1) reducing the reactivity of the bio-oil, thereby mitigating issues with aging and transport and eliminating need for multi-stage hydroprocessing configurations; (2) producing a bio-oil that can be fractionated through distillation, which could lead to more efficient use of hydrogen during hydrotreating and facilitate integration in existing petroleum refineries; and (3) allowing for the separation of the aqueous phase. In this perspective, we investigate in detail a combination of these approaches, where some oxygen is removed during catalytic fast pyrolysis and the remainder removed by downstream hydrotreating, accompanied by carbon–carbon coupling reactions in either the vapor or liquid phase to maximize carbon efficiency toward value-driven products (e.g. fuels or chemicals). The economic impact of partial deoxygenation by catalytic fast pyrolysis will be explored in the context of an integrated two-stage process. In conclusion, improving the overall pyrolysis-based biorefinery economics by inclusion of production of high-value co-products will be examined.« less
Improving biomass pyrolysis economics by integrating vapor and liquid phase upgrading
DOE Office of Scientific and Technical Information (OSTI.GOV)
Iisa, Kristiina; Robichaud, David J.; Watson, Michael J.
Partial deoxygenation of bio-oil by catalytic fast pyrolysis with subsequent coupling and hydrotreating can lead to improved economics and will aid commercial deployment of pyrolytic conversion of biomass technologies. Biomass pyrolysis efficiently depolymerizes and deconstructs solid plant matter into carbonaceous molecules that, upon catalytic upgrading, can be used for fuels and chemicals. Upgrading strategies include catalytic deoxygenation of the vapors before they are condensed (in situ and ex situ catalytic fast pyrolysis), or hydrotreating following condensation of the bio-oil. In general, deoxygenation carbon efficiencies, one of the most important cost drivers, are typically higher for hydrotreating when compared to catalyticmore » fast pyrolysis alone. However, using catalytic fast pyrolysis as the primary conversion step can benefit the entire process chain by: (1) reducing the reactivity of the bio-oil, thereby mitigating issues with aging and transport and eliminating need for multi-stage hydroprocessing configurations; (2) producing a bio-oil that can be fractionated through distillation, which could lead to more efficient use of hydrogen during hydrotreating and facilitate integration in existing petroleum refineries; and (3) allowing for the separation of the aqueous phase. In this perspective, we investigate in detail a combination of these approaches, where some oxygen is removed during catalytic fast pyrolysis and the remainder removed by downstream hydrotreating, accompanied by carbon–carbon coupling reactions in either the vapor or liquid phase to maximize carbon efficiency toward value-driven products (e.g. fuels or chemicals). The economic impact of partial deoxygenation by catalytic fast pyrolysis will be explored in the context of an integrated two-stage process. In conclusion, improving the overall pyrolysis-based biorefinery economics by inclusion of production of high-value co-products will be examined.« less
Don’t make cache too complex: A simple probability-based cache management scheme for SSDs
Cho, Sangyeun; Choi, Jongmoo
2017-01-01
Solid-state drives (SSDs) have recently become a common storage component in computer systems, and they are fueled by continued bit cost reductions achieved with smaller feature sizes and multiple-level cell technologies. However, as the flash memory stores more bits per cell, the performance and reliability of the flash memory degrade substantially. To solve this problem, a fast non-volatile memory (NVM-)based cache has been employed within SSDs to reduce the long latency required to write data. Absorbing small writes in a fast NVM cache can also reduce the number of flash memory erase operations. To maximize the benefits of an NVM cache, it is important to increase the NVM cache utilization. In this paper, we propose and study ProCache, a simple NVM cache management scheme, that makes cache-entrance decisions based on random probability testing. Our scheme is motivated by the observation that frequently written hot data will eventually enter the cache with a high probability, and that infrequently accessed cold data will not enter the cache easily. Owing to its simplicity, ProCache is easy to implement at a substantially smaller cost than similar previously studied techniques. We evaluate ProCache and conclude that it achieves comparable performance compared to a more complex reference counter-based cache-management scheme. PMID:28358897
Don't make cache too complex: A simple probability-based cache management scheme for SSDs.
Baek, Seungjae; Cho, Sangyeun; Choi, Jongmoo
2017-01-01
Solid-state drives (SSDs) have recently become a common storage component in computer systems, and they are fueled by continued bit cost reductions achieved with smaller feature sizes and multiple-level cell technologies. However, as the flash memory stores more bits per cell, the performance and reliability of the flash memory degrade substantially. To solve this problem, a fast non-volatile memory (NVM-)based cache has been employed within SSDs to reduce the long latency required to write data. Absorbing small writes in a fast NVM cache can also reduce the number of flash memory erase operations. To maximize the benefits of an NVM cache, it is important to increase the NVM cache utilization. In this paper, we propose and study ProCache, a simple NVM cache management scheme, that makes cache-entrance decisions based on random probability testing. Our scheme is motivated by the observation that frequently written hot data will eventually enter the cache with a high probability, and that infrequently accessed cold data will not enter the cache easily. Owing to its simplicity, ProCache is easy to implement at a substantially smaller cost than similar previously studied techniques. We evaluate ProCache and conclude that it achieves comparable performance compared to a more complex reference counter-based cache-management scheme.
Identification of two new HMXBs in the LMC: an ˜2013 s pulsar and a probable SFXT
NASA Astrophysics Data System (ADS)
Vasilopoulos, G.; Maitra, C.; Haberl, F.; Hatzidimitriou, D.; Petropoulou, M.
2018-03-01
We report on the X-ray and optical properties of two high-mass X-ray binary systems located in the Large Magellanic Cloud (LMC). Based on the obtained optical spectra, we classify the massive companion as a supergiant star in both systems. Timing analysis of the X-ray events collected by XMM-Newton revealed the presence of coherent pulsations (spin period ˜2013 s) for XMMU J053108.3-690923 and fast flaring behaviour for XMMU J053320.8-684122. The X-ray spectra of both systems can be modelled sufficiently well by an absorbed power law, yielding hard spectra and high intrinsic absorption from the environment of the systems. Due to their combined X-ray and optical properties, we classify both systems as SgXRBs: the 19th confirmed X-ray pulsar and a probable supergiant fast X-ray transient in the LMC, the second such candidate outside our Galaxy.
Integrals for IBS and beam cooling
DOE Office of Scientific and Technical Information (OSTI.GOV)
Burov, A.; /Fermilab
Simulation of beam cooling usually requires performing certain integral transformations every time step or so, which is a significant burden on the CPU. Examples are the dispersion integrals (Hilbert transforms) in the stochastic cooling, wake fields and IBS integrals. An original method is suggested for fast and sufficiently accurate computation of the integrals. This method is applied for the dispersion integral. Some methodical aspects of the IBS analysis are discussed.
Integrals for IBS and Beam Cooling
DOE Office of Scientific and Technical Information (OSTI.GOV)
Burov, A.
Simulation of beam cooling usually requires performing certain integral transformations every time step or so, which is a significant burden on the CPU. Examples are the dispersion integrals (Hilbert transforms) in the stochastic cooling, wake fields and IBS integrals. An original method is suggested for fast and sufficiently accurate computation of the integrals. This method is applied for the dispersion integral. Some methodical aspects of the IBS analysis are discussed.
The new ATLAS Fast Calorimeter Simulation
NASA Astrophysics Data System (ADS)
Schaarschmidt, J.; ATLAS Collaboration
2017-10-01
Current and future need for large scale simulated samples motivate the development of reliable fast simulation techniques. The new Fast Calorimeter Simulation is an improved parameterized response of single particles in the ATLAS calorimeter that aims to accurately emulate the key features of the detailed calorimeter response as simulated with Geant4, yet approximately ten times faster. Principal component analysis and machine learning techniques are used to improve the performance and decrease the memory need compared to the current version of the ATLAS Fast Calorimeter Simulation. A prototype of this new Fast Calorimeter Simulation is in development and its integration into the ATLAS simulation infrastructure is ongoing.
Using Bitmap Indexing Technology for Combined Numerical and TextQueries
DOE Office of Scientific and Technical Information (OSTI.GOV)
Stockinger, Kurt; Cieslewicz, John; Wu, Kesheng
2006-10-16
In this paper, we describe a strategy of using compressedbitmap indices to speed up queries on both numerical data and textdocuments. By using an efficient compression algorithm, these compressedbitmap indices are compact even for indices with millions of distinctterms. Moreover, bitmap indices can be used very efficiently to answerBoolean queries over text documents involving multiple query terms.Existing inverted indices for text searches are usually inefficient forcorpora with a very large number of terms as well as for queriesinvolving a large number of hits. We demonstrate that our compressedbitmap index technology overcomes both of those short-comings. In aperformance comparison against amore » commonly used database system, ourindices answer queries 30 times faster on average. To provide full SQLsupport, we integrated our indexing software, called FastBit, withMonetDB. The integrated system MonetDB/FastBit provides not onlyefficient searches on a single table as FastBit does, but also answersjoin queries efficiently. Furthermore, MonetDB/FastBit also provides avery efficient retrieval mechanism of result records.« less
Simulating the inception of pulsed discharges near positive electrodes
NASA Astrophysics Data System (ADS)
Teunissen, Jannis; Ebert, Ute
2013-09-01
With 3D particle simulations we study the inception of pulsed discharges near positive electrodes. In different geometries, we first determine the breakdown voltage. Then we study the probability of inception for a fast voltage pulse. This probability mostly depends on the availability of seed electrons to generate the initial electron avalanches. These results are compared with experimental observations. Then we investigate how the shape of a starting discharge affects its further development. In particular, we discuss the formation of so-called ``inception clouds.'' JT was supported by STW-project 10755.
Fast Nonparametric Machine Learning Algorithms for High-Dimensional Massive Data and Applications
2006-03-01
know the probability of that from Lemma 2. Using the union bound, we know that for any query q, the probability that i-am-feeling-lucky search algorithm...and each point in a d-dimensional space, a naive k-NN search needs to do a linear scan of T for every single query q, and thus the computational time...algorithm based on partition trees with priority search , and give an expected query time O((1/)d log n). But the constant in the O((1/)d log n
A fast direct solver for boundary value problems on locally perturbed geometries
NASA Astrophysics Data System (ADS)
Zhang, Yabin; Gillman, Adrianna
2018-03-01
Many applications including optimal design and adaptive discretization techniques involve solving several boundary value problems on geometries that are local perturbations of an original geometry. This manuscript presents a fast direct solver for boundary value problems that are recast as boundary integral equations. The idea is to write the discretized boundary integral equation on a new geometry as a low rank update to the discretized problem on the original geometry. Using the Sherman-Morrison formula, the inverse can be expressed in terms of the inverse of the original system applied to the low rank factors and the right hand side. Numerical results illustrate for problems where perturbation is localized the fast direct solver is three times faster than building a new solver from scratch.
A "Virtual Spin" on the Teaching of Probability
ERIC Educational Resources Information Center
Beck, Shari A.; Huse, Vanessa E.
2007-01-01
This article, which describes integrating virtual manipulatives with the teaching of probability at the elementary level, puts a "virtual spin" on the teaching of probability to provide more opportunities for students to experience successful learning. The traditional use of concrete manipulatives is enhanced with virtual coins and spinners from…
Probability Theory, Not the Very Guide of Life
ERIC Educational Resources Information Center
Juslin, Peter; Nilsson, Hakan; Winman, Anders
2009-01-01
Probability theory has long been taken as the self-evident norm against which to evaluate inductive reasoning, and classical demonstrations of violations of this norm include the conjunction error and base-rate neglect. Many of these phenomena require multiplicative probability integration, whereas people seem more inclined to linear additive…
Laplace Transforms without Integration
ERIC Educational Resources Information Center
Robertson, Robert L.
2017-01-01
Calculating Laplace transforms from the definition often requires tedious integrations. This paper provides an integration-free technique for calculating Laplace transforms of many familiar functions. It also shows how the technique can be applied to probability theory.
NASA Astrophysics Data System (ADS)
Vidybida, Alexander; Shchur, Olha
We consider a class of spiking neuronal models, defined by a set of conditions typical for basic threshold-type models, such as the leaky integrate-and-fire or the binding neuron model and also for some artificial neurons. A neuron is fed with a Poisson process. Each output impulse is applied to the neuron itself after a finite delay Δ. This impulse acts as being delivered through a fast Cl-type inhibitory synapse. We derive a general relation which allows calculating exactly the probability density function (pdf) p(t) of output interspike intervals of a neuron with feedback based on known pdf p0(t) for the same neuron without feedback and on the properties of the feedback line (the Δ value). Similar relations between corresponding moments are derived. Furthermore, we prove that the initial segment of pdf p0(t) for a neuron with a fixed threshold level is the same for any neuron satisfying the imposed conditions and is completely determined by the input stream. For the Poisson input stream, we calculate that initial segment exactly and, based on it, obtain exactly the initial segment of pdf p(t) for a neuron with feedback. That is the initial segment of p(t) is model-independent as well. The obtained expressions are checked by means of Monte Carlo simulation. The course of p(t) has a pronounced peculiarity, which makes it impossible to approximate p(t) by Poisson or another simple stochastic process.
Persistence Probabilities of Two-Sided (Integrated) Sums of Correlated Stationary Gaussian Sequences
NASA Astrophysics Data System (ADS)
Aurzada, Frank; Buck, Micha
2018-02-01
We study the persistence probability for some two-sided, discrete-time Gaussian sequences that are discrete-time analogues of fractional Brownian motion and integrated fractional Brownian motion, respectively. Our results extend the corresponding ones in continuous time in Molchan (Commun Math Phys 205(1):97-111, 1999) and Molchan (J Stat Phys 167(6):1546-1554, 2017) to a wide class of discrete-time processes.
NASA Astrophysics Data System (ADS)
Strauch, R. L.; Istanbulluoglu, E.
2017-12-01
We develop a landslide hazard modeling approach that integrates a data-driven statistical model and a probabilistic process-based shallow landslide model for mapping probability of landslide initiation, transport, and deposition at regional scales. The empirical model integrates the influence of seven site attribute (SA) classes: elevation, slope, curvature, aspect, land use-land cover, lithology, and topographic wetness index, on over 1,600 observed landslides using a frequency ratio (FR) approach. A susceptibility index is calculated by adding FRs for each SA on a grid-cell basis. Using landslide observations we relate susceptibility index to an empirically-derived probability of landslide impact. This probability is combined with results from a physically-based model to produce an integrated probabilistic map. Slope was key in landslide initiation while deposition was linked to lithology and elevation. Vegetation transition from forest to alpine vegetation and barren land cover with lower root cohesion leads to higher frequency of initiation. Aspect effects are likely linked to differences in root cohesion and moisture controlled by solar insulation and snow. We demonstrate the model in the North Cascades of Washington, USA and identify locations of high and low probability of landslide impacts that can be used by land managers in their design, planning, and maintenance.
A fast collocation method for a variable-coefficient nonlocal diffusion model
NASA Astrophysics Data System (ADS)
Wang, Che; Wang, Hong
2017-02-01
We develop a fast collocation scheme for a variable-coefficient nonlocal diffusion model, for which a numerical discretization would yield a dense stiffness matrix. The development of the fast method is achieved by carefully handling the variable coefficients appearing inside the singular integral operator and exploiting the structure of the dense stiffness matrix. The resulting fast method reduces the computational work from O (N3) required by a commonly used direct solver to O (Nlog N) per iteration and the memory requirement from O (N2) to O (N). Furthermore, the fast method reduces the computational work of assembling the stiffness matrix from O (N2) to O (N). Numerical results are presented to show the utility of the fast method.
Fast mapping rapidly integrates information into existing memory networks.
Coutanche, Marc N; Thompson-Schill, Sharon L
2014-12-01
Successful learning involves integrating new material into existing memory networks. A learning procedure known as fast mapping (FM), thought to simulate the word-learning environment of children, has recently been linked to distinct neuroanatomical substrates in adults. This idea has suggested the (never-before tested) hypothesis that FM may promote rapid incorporation into cortical memory networks. We test this hypothesis here in 2 experiments. In our 1st experiment, we introduced 50 participants to 16 unfamiliar animals and names through FM or explicit encoding (EE) and tested participants on the training day, and again after sleep. Learning through EE produced strong declarative memories, without immediate lexical competition, as expected from slow-consolidation models. Learning through FM, however, led to almost immediate lexical competition, which continued to the next day. Additionally, the learned words began to prime related concepts on the day following FM (but not EE) training. In a 2nd experiment, we replicated the lexical integration results and determined that presenting an already-known item during learning was crucial for rapid integration through FM. The findings presented here indicate that learned items can be integrated into cortical memory networks at an accelerated rate through fast mapping. The retrieval of a related known concept, in order to infer the target of the FM question, is critical for this effect. PsycINFO Database Record (c) 2014 APA, all rights reserved.
A point implicit time integration technique for slow transient flow problems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kadioglu, Samet Y.; Berry, Ray A.; Martineau, Richard C.
2015-05-01
We introduce a point implicit time integration technique for slow transient flow problems. The method treats the solution variables of interest (that can be located at cell centers, cell edges, or cell nodes) implicitly and the rest of the information related to same or other variables are handled explicitly. The method does not require implicit iteration; instead it time advances the solutions in a similar spirit to explicit methods, except it involves a few additional function(s) evaluation steps. Moreover, the method is unconditionally stable, as a fully implicit method would be. This new approach exhibits the simplicity of implementation ofmore » explicit methods and the stability of implicit methods. It is specifically designed for slow transient flow problems of long duration wherein one would like to perform time integrations with very large time steps. Because the method can be time inaccurate for fast transient problems, particularly with larger time steps, an appropriate solution strategy for a problem that evolves from a fast to a slow transient would be to integrate the fast transient with an explicit or semi-implicit technique and then switch to this point implicit method as soon as the time variation slows sufficiently. We have solved several test problems that result from scalar or systems of flow equations. Our findings indicate the new method can integrate slow transient problems very efficiently; and its implementation is very robust.« less
Completion of the Edward Air Force Base Statistical Guidance Wind Tool
NASA Technical Reports Server (NTRS)
Dreher, Joseph G.
2008-01-01
The goal of this task was to develop a GUI using EAFB wind tower data similar to the KSC SLF peak wind tool that is already in operations at SMG. In 2004, MSFC personnel began work to replicate the KSC SLF tool using several wind towers at EAFB. They completed the analysis and QC of the data, but due to higher priority work did not start development of the GUI. MSFC personnel calculated wind climatologies and probabilities of 10-minute peak wind occurrence based on the 2-minute average wind speed for several EAFB wind towers. Once the data were QC'ed and analyzed the climatologies were calculated following the methodology outlined in Lambert (2003). The climatologies were calculated for each tower and month, and then were stratified by hour, direction (10" sectors), and direction (45" sectors)/hour. For all climatologies, MSFC calculated the mean, standard deviation and observation counts of the Zminute average and 10-minute peak wind speeds. MSFC personnel also calculated empirical and modeled probabilities of meeting or exceeding specific 10- minute peak wind speeds using PDFs. The empirical PDFs were asymmetrical and bounded on the left by the 2- minute average wind speed. They calculated the parametric PDFs by fitting the GEV distribution to the empirical distributions. Parametric PDFs were calculated in order to smooth and interpolate over variations in the observed values due to possible under-sampling of certain peak winds and to estimate probabilities associated with average winds outside the observed range. MSFC calculated the individual probabilities of meeting or exceeding specific 10- minute peak wind speeds by integrating the area under each curve. The probabilities assist SMG forecasters in assessing the shuttle FR for various Zminute average wind speeds. The A M ' obtained the processed EAFB data from Dr. Lee Bums of MSFC and reformatted them for input to Excel PivotTables, which allow users to display different values with point-click-drag techniques. The GUI was created from the PivotTables using VBA code. It is run through a macro within Excel and allows forecasters to quickly display and interpret peak wind climatology and probabilities in a fast-paced operational environment. The GUI was designed to look and operate exactly the same as the KSC SLF tool since SMG forecasters were already familiar with that product. SMG feedback was continually incorporated into the GUI ensuring the end product met their needs. The final version of the GUI along with all climatologies, PDFs, and probabilities has been delivered to SMG and will be put into operational use.
Computer Simulation Results for the Two-Point Probability Function of Composite Media
NASA Astrophysics Data System (ADS)
Smith, P.; Torquato, S.
1988-05-01
Computer simulation results are reported for the two-point matrix probability function S2 of two-phase random media composed of disks distributed with an arbitrary degree of impenetrability λ. The novel technique employed to sample S2( r) (which gives the probability of finding the endpoints of a line segment of length r in the matrix) is very accurate and has a fast execution time. Results for the limiting cases λ = 0 (fully penetrable disks) and λ = 1 (hard disks), respectively, compare very favorably with theoretical predictions made by Torquato and Beasley and by Torquato and Lado. Results are also reported for several values of λ. that lie between these two extremes: cases which heretofore have not been examined.
Analytic barrage attack model. Final report, January 1986-January 1989
DOE Office of Scientific and Technical Information (OSTI.GOV)
St Ledger, J.W.; Naegeli, R.E.; Dowden, N.A.
An analytic model is developed for a nuclear barrage attack, assuming weapons with no aiming error and a cookie-cutter damage function. The model is then extended with approximations for the effects of aiming error and distance damage sigma. The final result is a fast running model which calculates probability of damage for a barrage attack. The probability of damage is accurate to within seven percent or better, for weapon reliabilities of 50 to 100 percent, distance damage sigmas of 0.5 or less, and zero to very large circular error probabilities. FORTRAN 77 coding is included in the report for themore » analytic model and for a numerical model used to check the analytic results.« less
2011-01-01
Background Laparoscopy-assisted surgery, fast-track perioperative treatment are both increasingly used in colorectal cancer treatment, for their short-time benefits of enhanced recovery and short hospital stays. However, the benefits of the integration of the Laparoscopy-assisted surgery, fast-track perioperative treatment, and even with the Xelox chemotherapy, are still unknown. In this study, the three treatments integration is defined as "Fast Track Multi-Discipline Treatment Model" for colorectal cancer and this model extends the benefits to the whole treatment process of colorectal cancer. The main purpose of the study is to explore the feasibility of "Fast Track Multi-Discipline Treatment" model in treatment of colorectal cancer. Methods The trial is a prospective randomized controlled study with 2 × 2 balanced factorial design. Patients eligible for the study will be randomized to 4 groups: (I) Laparoscopic surgery with fast track perioperative treatment and Xelox chemotherapy; (II) Open surgery with fast track perioperative treatment and Xelox chemotherapy; (III) Laparoscopic surgery with conventional perioperative treatment and mFolfox6 chemotherapy; (IV) Open surgery with conventional perioperative treatment and mFolfox6 chemotherapy. The primary endpoint of this study is the hospital stays. The secondary endpoints are the quality of life, chemotherapy related adverse events, surgical complications and hospitalization costs. Totally, 340 patients will be enrolled with 85 patients in each group. Conclusions The study initiates a new treatment model "Fast Track Multi-Discipline Treatment" for colorectal cancer, and will provide feasibility evidence on the new model "Fast Track Multi-Discipline Treatment" for patients with colorectal cancer. Trial registration ClinicalTrials.gov: NCT01080547 PMID:22111914
Avilés-Santa, M Larissa; Schneiderman, Neil; Savage, Peter J; Kaplan, Robert C; Teng, Yanping; Pérez, Cynthia M; Suárez, Erick L; Cai, Jianwen; Giachello, Aida L; Talavera, Gregory A; Cowie, Catherine C
2016-10-01
The aim of this study was to compare the ability of American Diabetes Association (ADA) diagnostic criteria to identify U.S. Hispanics/Latinos from diverse heritage groups with probable diabetes mellitus and assess cardiovascular risk factor correlates of those criteria. Cross-sectional analysis of data from 15,507 adults from 6 Hispanic/Latino heritage groups, enrolled in the Hispanic Community Health Study/Study of Latinos. The prevalence of probable diabetes mellitus was estimated using individual or combinations of ADA-defined cut points. The sensitivity and specificity of these criteria at identifying diabetes mellitus from ADA-defined prediabetes and normoglycemia were evaluated. Prevalence ratios of hypertension, abnormal lipids, and elevated urinary albumin-creatinine ratio for unrecognized diabetes mellitus-versus prediabetes and normoglycemia-were calculated. Among Hispanics/Latinos (mean age, 43 years) with diabetes mellitus, 39.4% met laboratory test criteria for probable diabetes, and the prevalence varied by heritage group. Using the oral glucose tolerance test as the gold standard, the sensitivity of fasting plasma glucose (FPG) and hemoglobin A1c-alone or in combination-was low (18, 23, and 33%, respectively) at identifying probable diabetes mellitus. Individuals who met any criterion for probable diabetes mellitus had significantly higher (P<.05) prevalence of most cardiovascular risk factors than those with normoglycemia or prediabetes, and this association was not modified by Hispanic/Latino heritage group. FPG and hemoglobin A1c are not sensitive (but are highly specific) at detecting probable diabetes mellitus among Hispanics/Latinos, independent of heritage group. Assessing cardiovascular risk factors at diagnosis might prompt multitarget interventions and reduce health complications in this young population. 2hPG = 2-hour post-glucose load plasma glucose ADA = American Diabetes Association BMI = body mass index CV = cardiovascular FPG = fasting plasma glucose HbA1c = hemoglobin A1c HCHS/SOL = Hispanic Community Health Study/Study of Latinos HDL-C = high-density-lipoprotein cholesterol NGT = normal glucose tolerance NHANES = National Health and Nutrition Examination Survey OGTT = oral glucose tolerance test TG = triglyceride UACR = urine albumin-creatinine ratio.
Probability distribution for the Gaussian curvature of the zero level surface of a random function
NASA Astrophysics Data System (ADS)
Hannay, J. H.
2018-04-01
A rather natural construction for a smooth random surface in space is the level surface of value zero, or ‘nodal’ surface f(x,y,z) = 0, of a (real) random function f; the interface between positive and negative regions of the function. A physically significant local attribute at a point of a curved surface is its Gaussian curvature (the product of its principal curvatures) because, when integrated over the surface it gives the Euler characteristic. Here the probability distribution for the Gaussian curvature at a random point on the nodal surface f = 0 is calculated for a statistically homogeneous (‘stationary’) and isotropic zero mean Gaussian random function f. Capitalizing on the isotropy, a ‘fixer’ device for axes supplies the probability distribution directly as a multiple integral. Its evaluation yields an explicit algebraic function with a simple average. Indeed, this average Gaussian curvature has long been known. For a non-zero level surface instead of the nodal one, the probability distribution is not fully tractable, but is supplied as an integral expression.
Tinkum, Kelsey L; Stemler, Kristina M; White, Lynn S; Loza, Andrew J; Jeter-Jones, Sabrina; Michalski, Basia M; Kuzmicki, Catherine; Pless, Robert; Stappenbeck, Thaddeus S; Piwnica-Worms, David; Piwnica-Worms, Helen
2015-12-22
Short-term fasting protects mice from lethal doses of chemotherapy through undetermined mechanisms. Herein, we demonstrate that fasting preserves small intestinal (SI) architecture by maintaining SI stem cell viability and SI barrier function following exposure to high-dose etoposide. Nearly all SI stem cells were lost in fed mice, whereas fasting promoted sufficient SI stem cell survival to preserve SI integrity after etoposide treatment. Lineage tracing demonstrated that multiple SI stem cell populations, marked by Lgr5, Bmi1, or HopX expression, contributed to fasting-induced survival. DNA repair and DNA damage response genes were elevated in SI stem/progenitor cells of fasted etoposide-treated mice, which importantly correlated with faster resolution of DNA double-strand breaks and less apoptosis. Thus, fasting preserved SI stem cell viability as well as SI architecture and barrier function suggesting that fasting may reduce host toxicity in patients undergoing dose intensive chemotherapy.
Note: development of fast heating inert gas annealing apparatus operated at atmospheric pressure.
Das, S C; Majumdar, A; Shripathi, T; Hippler, R
2012-04-01
Here, we report the development of a simple, small, fast heating, and portable, homemade, inert gas (Ar) atmospheric annealing setup. Instead of using a conventional heating element, a commercial soldering rod having an encapsulated fast heating heater is used here. The sample holder is made of a block of stainless steel. It takes 200 s to reach 700 °C, and 10 min to cool down. The probability of oxidation or surface contamination has been examined by means of x ray photoelectron spectroscopy of virgin Cu sample after annealing at 600 °C. In addition, we compare the annealing of a hydrogenated carbon nitride film (HCN(x)) in both a conventional vacuum and our newly developed ambient Ar atmosphere setup. © 2012 American Institute of Physics
The priority heuristic: making choices without trade-offs.
Brandstätter, Eduard; Gigerenzer, Gerd; Hertwig, Ralph
2006-04-01
Bernoulli's framework of expected utility serves as a model for various psychological processes, including motivation, moral sense, attitudes, and decision making. To account for evidence at variance with expected utility, the authors generalize the framework of fast and frugal heuristics from inferences to preferences. The priority heuristic predicts (a) the Allais paradox, (b) risk aversion for gains if probabilities are high, (c) risk seeking for gains if probabilities are low (e.g., lottery tickets), (d) risk aversion for losses if probabilities are low (e.g., buying insurance), (e) risk seeking for losses if probabilities are high, (f) the certainty effect, (g) the possibility effect, and (h) intransitivities. The authors test how accurately the heuristic predicts people's choices, compared with previously proposed heuristics and 3 modifications of expected utility theory: security-potential/aspiration theory, transfer-of-attention-exchange model, and cumulative prospect theory. ((c) 2006 APA, all rights reserved).
NASA Astrophysics Data System (ADS)
Kneringer, Philipp; Dietz, Sebastian J.; Mayr, Georg J.; Zeileis, Achim
2018-04-01
Airport operations are sensitive to visibility conditions. Low-visibility events may lead to capacity reduction, delays and economic losses. Different levels of low-visibility procedures (lvp) are enacted to ensure aviation safety. A nowcast of the probabilities for each of the lvp categories helps decision makers to optimally schedule their operations. An ordered logistic regression (OLR) model is used to forecast these probabilities directly. It is applied to cold season forecasts at Vienna International Airport for lead times of 30-min out to 2 h. Model inputs are standard meteorological measurements. The skill of the forecasts is accessed by the ranked probability score. OLR outperforms persistence, which is a strong contender at the shortest lead times. The ranked probability score of the OLR is even better than the one of nowcasts from human forecasters. The OLR-based nowcasting system is computationally fast and can be updated instantaneously when new data become available.
Sundh, Joakim; Juslin, Peter
2018-02-01
In this study, we explore how people integrate risks of assets in a simulated financial market into a judgment of the conjunctive risk that all assets decrease in value, both when assets are independent and when there is a systematic risk present affecting all assets. Simulations indicate that while mental calculation according to naïve application of probability theory is best when the assets are independent, additive or exemplar-based algorithms perform better when systematic risk is high. Considering that people tend to intuitively approach compound probability tasks using additive heuristics, we expected the participants to find it easiest to master tasks with high systematic risk - the most complex tasks from the standpoint of probability theory - while they should shift to probability theory or exemplar memory with independence between the assets. The results from 3 experiments confirm that participants shift between strategies depending on the task, starting off with the default of additive integration. In contrast to results in similar multiple cue judgment tasks, there is little evidence for use of exemplar memory. The additive heuristics also appear to be surprisingly context-sensitive, with limited generalization across formally very similar tasks. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Zhang, Feng-Liang; Ni, Yan-Chun; Au, Siu-Kui; Lam, Heung-Fai
2016-03-01
The identification of modal properties from field testing of civil engineering structures is becoming economically viable, thanks to the advent of modern sensor and data acquisition technology. Its demand is driven by innovative structural designs and increased performance requirements of dynamic-prone structures that call for a close cross-checking or monitoring of their dynamic properties and responses. Existing instrumentation capabilities and modal identification techniques allow structures to be tested under free vibration, forced vibration (known input) or ambient vibration (unknown broadband loading). These tests can be considered complementary rather than competing as they are based on different modeling assumptions in the identification model and have different implications on costs and benefits. Uncertainty arises naturally in the dynamic testing of structures due to measurement noise, sensor alignment error, modeling error, etc. This is especially relevant in field vibration tests because the test condition in the field environment can hardly be controlled. In this work, a Bayesian statistical approach is developed for modal identification using the free vibration response of structures. A frequency domain formulation is proposed that makes statistical inference based on the Fast Fourier Transform (FFT) of the data in a selected frequency band. This significantly simplifies the identification model because only the modes dominating the frequency band need to be included. It also legitimately ignores the information in the excluded frequency bands that are either irrelevant or difficult to model, thereby significantly reducing modeling error risk. The posterior probability density function (PDF) of the modal parameters is derived rigorously from modeling assumptions and Bayesian probability logic. Computational difficulties associated with calculating the posterior statistics, including the most probable value (MPV) and the posterior covariance matrix, are addressed. Fast computational algorithms for determining the MPV are proposed so that the method can be practically implemented. In the companion paper (Part II), analytical formulae are derived for the posterior covariance matrix so that it can be evaluated without resorting to finite difference method. The proposed method is verified using synthetic data. It is also applied to modal identification of full-scale field structures.
Carvalho-Filho, M A; Carvalho, B M; Oliveira, A G; Guadagnini, D; Ueno, M; Dias, M M; Tsukumo, D M; Hirabara, S M; Reis, L F; Curi, R; Carvalheira, J B C; Saad, Mario J A
2012-11-01
The molecular integration of nutrient- and pathogen-sensing pathways has become of great interest in understanding the mechanisms of insulin resistance in obesity. The double-stranded RNA-dependent protein kinase (PKR) is one candidate molecule that may provide cross talk between inflammatory and metabolic signaling. The present study was performed to determine, first, the role of PKR in modulating insulin action and glucose metabolism in physiological situations, and second, the role of PKR in insulin resistance in obese mice. We used Pkr(-/-) and Pkr(+/+) mice to investigate the role of PKR in modulating insulin sensitivity, glucose metabolism, and insulin signaling in liver, muscle, and adipose tissue in response to a high-fat diet. Our data show that in lean Pkr(-/-) mice, there is an improvement in insulin sensitivity, and in glucose tolerance, and a reduction in fasting blood glucose, probably related to a decrease in protein phosphatase 2A activity and a parallel increase in insulin-induced thymoma viral oncogene-1 (Akt) phosphorylation. PKR is activated in tissues of obese mice and can induce insulin resistance by directly binding to and inducing insulin receptor substrate (IRS)-1 serine307 phosphorylation or indirectly through modulation of c-Jun N-terminal kinase and inhibitor of κB kinase β. Pkr(-/-) mice were protected from high-fat diet-induced insulin resistance and glucose intolerance and showed improved insulin signaling associated with a reduction in c-Jun N-terminal kinase and inhibitor of κB kinase β phosphorylation in insulin-sensitive tissues. PKR may have a role in insulin sensitivity under normal physiological conditions, probably by modulating protein phosphatase 2A activity and serine-threonine kinase phosphorylation, and certainly, this kinase may represent a central mechanism for the integration of pathogen response and innate immunity with insulin action and metabolic pathways that are critical in obesity.
NASA Astrophysics Data System (ADS)
Caffo, Michele; Czyż, Henryk; Gunia, Michał; Remiddi, Ettore
2009-03-01
We present the program BOKASUN for fast and precise evaluation of the Master Integrals of the two-loop self-mass sunrise diagram for arbitrary values of the internal masses and the external four-momentum. We use a combination of two methods: a Bernoulli accelerated series expansion and a Runge-Kutta numerical solution of a system of linear differential equations. Program summaryProgram title: BOKASUN Catalogue identifier: AECG_v1_0 Program summary URL:http://cpc.cs.qub.ac.uk/summaries/AECG_v1_0.html Program obtainable from: CPC Program Library, Queen's University, Belfast, N. Ireland Licensing provisions: Standard CPC licence, http://cpc.cs.qub.ac.uk/licence/licence.html No. of lines in distributed program, including test data, etc.: 9404 No. of bytes in distributed program, including test data, etc.: 104 123 Distribution format: tar.gz Programming language: FORTRAN77 Computer: Any computer with a Fortran compiler accepting FORTRAN77 standard. Tested on various PC's with LINUX Operating system: LINUX RAM: 120 kbytes Classification: 4.4 Nature of problem: Any integral arising in the evaluation of the two-loop sunrise Feynman diagram can be expressed in terms of a given set of Master Integrals, which should be calculated numerically. The program provides a fast and precise evaluation method of the Master Integrals for arbitrary (but not vanishing) masses and arbitrary value of the external momentum. Solution method: The integrals depend on three internal masses and the external momentum squared p. The method is a combination of an accelerated expansion in 1/p in its (pretty large!) region of fast convergence and of a Runge-Kutta numerical solution of a system of linear differential equations. Running time: To obtain 4 Master Integrals on PC with 2 GHz processor it takes 3 μs for series expansion with pre-calculated coefficients, 80 μs for series expansion without pre-calculated coefficients, from a few seconds up to a few minutes for Runge-Kutta method (depending on the required accuracy and the values of the physical parameters).
Electrical signals as mechanism of photosynthesis regulation in plants.
Sukhov, Vladimir
2016-12-01
This review summarizes current works concerning the effects of electrical signals (ESs) on photosynthesis, the mechanisms of the effects, and its physiological role in plants. Local irritations of plants induce various photosynthetic responses in intact leaves, including fast and long-term inactivation of photosynthesis, and its activation. Irritation-induced ESs, including action potential, variation potential, and system potential, probably causes the photosynthetic responses in intact leaves. Probable mechanisms of induction of fast inactivation of photosynthesis are associated with Ca 2+ - and (or) H + -influxes during ESs generation; long-term inactivation of photosynthesis might be caused by Ca 2+ - and (or) H + -influxes, production of abscisic and jasmonic acids, and inactivation of phloem H + -sucrose symporters. It is probable that subsequent development of inactivation of photosynthesis is mainly associated with decreased CO 2 influx and inactivation of the photosynthetic dark reactions, which induces decreased photochemical quantum yields of photosystems I and II and increased non-photochemical quenching of photosystem II fluorescence and cyclic electron flow around photosystem I. However, other pathways of the ESs influence on the photosynthetic light reactions are also possible. One of them might be associated with ES-connected acidification of chloroplast stroma inducing ferredoxin-NADP + reductase accumulation at the thylakoids in Tic62 and TROL complexes. Mechanisms of ES-induced activation of photosynthesis require further investigation. The probable ultimate effect of ES-induced photosynthetic responses in plant life is the increased photosynthetic machinery resistance to stressors, including high and low temperatures, and enhanced whole-plant resistance to environmental factors at least during 1 h after irritation.
Characterization of fast-pyrolysis bio-oil distillation residues and their potential applications
USDA-ARS?s Scientific Manuscript database
A typical petroleum refinery makes use of the vacuum gas oil by cracking the large molecular weight compounds into light fuel hydrocarbons. For various types of fast pyrolysis bio-oil, successful analogous methods for processing heavy fractions could expedite integration into a petroleum refinery fo...
Fast Faraday Cup With High Bandwidth
Deibele, Craig E [Knoxville, TN
2006-03-14
A circuit card stripline Fast Faraday cup quantitatively measures the picosecond time structure of a charged particle beam. The stripline configuration maintains signal integrity, and stitching of the stripline increases the bandwidth. A calibration procedure ensures the measurement of the absolute charge and time structure of the charged particle beam.
Assessing the Chances of Success: Naive Statistics versus Kind Experience
ERIC Educational Resources Information Center
Hogarth, Robin M.; Mukherjee, Kanchan; Soyer, Emre
2013-01-01
Additive integration of information is ubiquitous in judgment and has been shown to be effective even when multiplicative rules of probability theory are prescribed. We explore the generality of these findings in the context of estimating probabilities of success in contests. We first define a normative model of these probabilities that takes…
Pump dependence of the dynamics of quantum dot based waveguide absorbers
NASA Astrophysics Data System (ADS)
Viktorov, Evgeny A.; Erneux, Thomas; Piwonski, Tomasz; Pulka, Jaroslaw; Huyet, Guillaume; Houlihan, John
2012-06-01
The nonlinear two stage recovery of quantum dot based reverse-biased waveguide absorbers is investigated experimentally and analytically as a function of the initial ground state occupation probability of the dot. The latter is controlled experimentally by the pump pulse power. The slow stage of the recovery is exponential and its basic timescale is independent of pump power. The fast stage of the recovery is a logistic function which we analyze in detail. The relative strength of slow to fast components is highlighted and the importance of higher order absorption processes at the highest pump level is demonstrated.
Senter, Evan; Sheikh, Saad; Dotu, Ivan; Ponty, Yann; Clote, Peter
2012-01-01
Using complex roots of unity and the Fast Fourier Transform, we design a new thermodynamics-based algorithm, FFTbor, that computes the Boltzmann probability that secondary structures differ by base pairs from an arbitrary initial structure of a given RNA sequence. The algorithm, which runs in quartic time and quadratic space , is used to determine the correlation between kinetic folding speed and the ruggedness of the energy landscape, and to predict the location of riboswitch expression platform candidates. A web server is available at http://bioinformatics.bc.edu/clotelab/FFTbor/. PMID:23284639
Base Rates: Both Neglected and Intuitive
ERIC Educational Resources Information Center
Pennycook, Gordon; Trippas, Dries; Handley, Simon J.; Thompson, Valerie A.
2014-01-01
Base-rate neglect refers to the tendency for people to underweight base-rate probabilities in favor of diagnostic information. It is commonly held that base-rate neglect occurs because effortful (Type 2) reasoning is required to process base-rate information, whereas diagnostic information is accessible to fast, intuitive (Type 1) processing…
A novel bit-wise adaptable entropy coding technique
NASA Technical Reports Server (NTRS)
Kiely, A.; Klimesh, M.
2001-01-01
We present a novel entropy coding technique which is adaptable in that each bit to be encoded may have an associated probability esitmate which depends on previously encoded bits. The technique may have advantages over arithmetic coding. The technique can achieve arbitrarily small redundancy and admits a simple and fast decoder.
Integrated simulation of magnetic-field-assist fast ignition laser fusion
NASA Astrophysics Data System (ADS)
Johzaki, T.; Nagatomo, H.; Sunahara, A.; Sentoku, Y.; Sakagami, H.; Hata, M.; Taguchi, T.; Mima, K.; Kai, Y.; Ajimi, D.; Isoda, T.; Endo, T.; Yogo, A.; Arikawa, Y.; Fujioka, S.; Shiraga, H.; Azechi, H.
2017-01-01
To enhance the core heating efficiency in fast ignition laser fusion, the concept of relativistic electron beam guiding by external magnetic fields was evaluated by integrated simulations for FIREX class targets. For the cone-attached shell target case, the core heating performance deteriorates by applying magnetic fields since the core is considerably deformed and most of the fast electrons are reflected due to the magnetic mirror formed through the implosion. On the other hand, in the case of a cone-attached solid ball target, the implosion is more stable under the kilo-tesla-class magnetic field. In addition, feasible magnetic field configuration is formed through the implosion. As a result, the core heating efficiency doubles by magnetic guiding. The dependence of core heating properties on the heating pulse shot timing was also investigated for the solid ball target.
Diagnostics and results from coaxial plasma gun development for the PLX- α project
NASA Astrophysics Data System (ADS)
Case, A.; Brockington, S.; Cruz, E.; Witherspoon, F. D.
2016-10-01
We present results from the diagnostics used during development of the contoured gap coaxial plasma guns for the PLX- α project at LANL. Plasma-jet diagnostics include fast photodiodes for velocimetry, a ballistic pendulum for total plasmoid momentum, and interferometry for line integrated density. Deflectometry will be used for line integrated perpendicular density gradients. Time-resolved high-resolution spectroscopy using a novel detector and time-integrated survey spectroscopy are used for measurements of velocity and temperature, as well as impurities. We will also use a Faraday cup for density, fast imaging for plume geometry, and time-integrated imaging for overall light emission. Experimental results are compared to the desired target parameters for the plasma jets (up to n 2 ×1016cm-3 , v 50km / s , mass 5gm , radius = 4cm , and length 10cm). This work supported by the ARPA-E ALPHA Program.
Tunable electromagnetically induced transparency in integrated silicon photonics circuit.
Li, Ang; Bogaerts, Wim
2017-12-11
We comprehensively simulate and experimentally demonstrate a novel approach to generate tunable electromagnetically induced transparency (EIT) in a fully integrated silicon photonics circuit. It can also generate tunable fast and slow light. The circuit is a single ring resonator with two integrated tunable reflectors inside, which form an embedded Fabry-Perot (FP) cavity inside the ring cavity. The mode of the FP cavity can be controlled by tuning the reflections using integrated thermo-optic tuners. Under correct tuning conditions, the interaction of the FP mode and the ring resonance mode will generate a Fano resonance and an EIT response. The extinction ratio and bandwidth of the EIT can be tuned by controlling the reflectors. Measured group delay proves that both fast light and slow light can be generated under different tuning conditions. A maximum group delay of 1100 ps is observed because of EIT. Pulse advance around 1200 ps is also demonstrated.
NASA Astrophysics Data System (ADS)
Privas, E.; Archier, P.; Bernard, D.; De Saint Jean, C.; Destouche, C.; Leconte, P.; Noguère, G.; Peneliau, Y.; Capote, R.
2016-02-01
A new IAEA Coordinated Research Project (CRP) aims to test, validate and improve the IRDF library. Among the isotopes of interest, the modelisation of the 238U capture and fission cross sections represents a challenging task. A new description of the 238U neutrons induced reactions in the fast energy range is within progress in the frame of an IAEA evaluation consortium. The Nuclear Data group of Cadarache participates in this effort utilizing the 238U spectral indices measurements and Post Irradiated Experiments (PIE) carried out in the fast reactors MASURCA (CEA Cadarache) and PHENIX (CEA Marcoule). Such a collection of experimental results provides reliable integral information on the (n,γ) and (n,f) cross sections. This paper presents the Integral Data Assimilation (IDA) technique of the CONRAD code used to propagate the uncertainties of the integral data on the 238U cross sections of interest for dosimetry applications.
Flow Modulation and Force Control in Insect Fast Maneuver
NASA Astrophysics Data System (ADS)
Li, Chengyu; Dong, Haibo; Zhang, Wen; Gai, Kuo
2012-11-01
In this work, an integrated study combining high-speed photogrammetry and direct numerical simulation (DNS) is used to study free flying insects in fast maneuver. Quantitative measurement has shown the significant differences between quad-winged flyers such as dragonfly and damselfly and two-winged flyers such as cicada. Comparisons of unsteady 3D vortex formation and associated aerodynamic force production reveal the different mechanisms used by insects in fast turn. This work is supported by NSF CBET-1055949.
Fast packet switch architectures for broadband integrated services digital networks
NASA Technical Reports Server (NTRS)
Tobagi, Fouad A.
1990-01-01
Background information on networking and switching is provided, and the various architectures that have been considered for fast packet switches are described. The focus is solely on switches designed to be implemented electronically. A set of definitions and a brief description of the functionality required of fast packet switches are given. Three basic types of packet switches are identified: the shared-memory, shared-medium, and space-division types. Each of these is described, and examples are given.
Raghuram, Jayaram; Miller, David J; Kesidis, George
2014-07-01
We propose a method for detecting anomalous domain names, with focus on algorithmically generated domain names which are frequently associated with malicious activities such as fast flux service networks, particularly for bot networks (or botnets), malware, and phishing. Our method is based on learning a (null hypothesis) probability model based on a large set of domain names that have been white listed by some reliable authority. Since these names are mostly assigned by humans, they are pronounceable, and tend to have a distribution of characters, words, word lengths, and number of words that are typical of some language (mostly English), and often consist of words drawn from a known lexicon. On the other hand, in the present day scenario, algorithmically generated domain names typically have distributions that are quite different from that of human-created domain names. We propose a fully generative model for the probability distribution of benign (white listed) domain names which can be used in an anomaly detection setting for identifying putative algorithmically generated domain names. Unlike other methods, our approach can make detections without considering any additional (latency producing) information sources, often used to detect fast flux activity. Experiments on a publicly available, large data set of domain names associated with fast flux service networks show encouraging results, relative to several baseline methods, with higher detection rates and low false positive rates.
Raghuram, Jayaram; Miller, David J.; Kesidis, George
2014-01-01
We propose a method for detecting anomalous domain names, with focus on algorithmically generated domain names which are frequently associated with malicious activities such as fast flux service networks, particularly for bot networks (or botnets), malware, and phishing. Our method is based on learning a (null hypothesis) probability model based on a large set of domain names that have been white listed by some reliable authority. Since these names are mostly assigned by humans, they are pronounceable, and tend to have a distribution of characters, words, word lengths, and number of words that are typical of some language (mostly English), and often consist of words drawn from a known lexicon. On the other hand, in the present day scenario, algorithmically generated domain names typically have distributions that are quite different from that of human-created domain names. We propose a fully generative model for the probability distribution of benign (white listed) domain names which can be used in an anomaly detection setting for identifying putative algorithmically generated domain names. Unlike other methods, our approach can make detections without considering any additional (latency producing) information sources, often used to detect fast flux activity. Experiments on a publicly available, large data set of domain names associated with fast flux service networks show encouraging results, relative to several baseline methods, with higher detection rates and low false positive rates. PMID:25685511
FastSKAT: Sequence kernel association tests for very large sets of markers.
Lumley, Thomas; Brody, Jennifer; Peloso, Gina; Morrison, Alanna; Rice, Kenneth
2018-06-22
The sequence kernel association test (SKAT) is widely used to test for associations between a phenotype and a set of genetic variants that are usually rare. Evaluating tail probabilities or quantiles of the null distribution for SKAT requires computing the eigenvalues of a matrix related to the genotype covariance between markers. Extracting the full set of eigenvalues of this matrix (an n×n matrix, for n subjects) has computational complexity proportional to n 3 . As SKAT is often used when n>104, this step becomes a major bottleneck in its use in practice. We therefore propose fastSKAT, a new computationally inexpensive but accurate approximations to the tail probabilities, in which the k largest eigenvalues of a weighted genotype covariance matrix or the largest singular values of a weighted genotype matrix are extracted, and a single term based on the Satterthwaite approximation is used for the remaining eigenvalues. While the method is not particularly sensitive to the choice of k, we also describe how to choose its value, and show how fastSKAT can automatically alert users to the rare cases where the choice may affect results. As well as providing faster implementation of SKAT, the new method also enables entirely new applications of SKAT that were not possible before; we give examples grouping variants by topologically associating domains, and comparing chromosome-wide association by class of histone marker. © 2018 WILEY PERIODICALS, INC.
Willems, Janske G. P.; Wadman, Wytse J.
2018-01-01
Abstract The perirhinal (PER) and lateral entorhinal (LEC) cortex form an anatomical link between the neocortex and the hippocampus. However, neocortical activity is transmitted through the PER and LEC to the hippocampus with a low probability, suggesting the involvement of the inhibitory network. This study explored the role of interneuron mediated inhibition, activated by electrical stimulation in the agranular insular cortex (AiP), in the deep layers of the PER and LEC. Activated synaptic input by AiP stimulation rarely evoked action potentials in the PER‐LEC deep layer excitatory principal neurons, most probably because the evoked synaptic response consisted of a small excitatory and large inhibitory conductance. Furthermore, parvalbumin positive (PV) interneurons—a subset of interneurons projecting onto the axo‐somatic region of principal neurons—received synaptic input earlier than principal neurons, suggesting recruitment of feedforward inhibition. This synaptic input in PV interneurons evoked varying trains of action potentials, explaining the fast rising, long lasting synaptic inhibition received by deep layer principal neurons. Altogether, the excitatory input from the AiP onto deep layer principal neurons is overruled by strong feedforward inhibition. PV interneurons, with their fast, extensive stimulus‐evoked firing, are able to deliver this fast evoked inhibition in principal neurons. This indicates an essential role for PV interneurons in the gating mechanism of the PER‐LEC network. PMID:29341361
NASA Astrophysics Data System (ADS)
Mosunova, N. A.
2018-05-01
The article describes the basic models included in the EUCLID/V1 integrated code intended for safety analysis of liquid metal (sodium, lead, and lead-bismuth) cooled fast reactors using fuel rods with a gas gap and pellet dioxide, mixed oxide or nitride uranium-plutonium fuel under normal operation, under anticipated operational occurrences and accident conditions by carrying out interconnected thermal-hydraulic, neutronics, and thermal-mechanical calculations. Information about the Russian and foreign analogs of the EUCLID/V1 integrated code is given. Modeled objects, equation systems in differential form solved in each module of the EUCLID/V1 integrated code (the thermal-hydraulic, neutronics, fuel rod analysis module, and the burnup and decay heat calculation modules), the main calculated quantities, and also the limitations on application of the code are presented. The article also gives data on the scope of functions performed by the integrated code's thermal-hydraulic module, using which it is possible to describe both one- and twophase processes occurring in the coolant. It is shown that, owing to the availability of the fuel rod analysis module in the integrated code, it becomes possible to estimate the performance of fuel rods in different regimes of the reactor operation. It is also shown that the models implemented in the code for calculating neutron-physical processes make it possible to take into account the neutron field distribution over the fuel assembly cross section as well as other features important for the safety assessment of fast reactors.
NASA Astrophysics Data System (ADS)
Zhang, Guoqing; Lina, Liu
2018-02-01
An ultra-fast photon counting method is proposed based on the charge integration of output electrical pulses of passive quenching silicon photomultipliers (SiPMs). The results of the numerical analysis with actual parameters of SiPMs show that the maximum photon counting rate of a state-of-art passive quenching SiPM can reach ~THz levels which is much larger than that of the existing photon counting devices. The experimental procedure is proposed based on this method. This photon counting regime of SiPMs is promising in many fields such as large dynamic light power detection.
The Need for Optical Means as an Alternative for Electronic Computing
NASA Technical Reports Server (NTRS)
Adbeldayem, Hossin; Frazier, Donald; Witherow, William; Paley, Steve; Penn, Benjamin; Bank, Curtis; Whitaker, Ann F. (Technical Monitor)
2001-01-01
An increasing demand for faster computers is rapidly growing to encounter the fast growing rate of Internet, space communication, and robotic industry. Unfortunately, the Very Large Scale Integration technology is approaching its fundamental limits beyond which the device will be unreliable. Optical interconnections and optical integrated circuits are strongly believed to provide the way out of the extreme limitations imposed on the growth of speed and complexity of nowadays computations by conventional electronics. This paper demonstrates two ultra-fast, all-optical logic gates and a high-density storage medium, which are essential components in building the future optical computer.
Integral Fast Reactor fuel pin processor
DOE Office of Scientific and Technical Information (OSTI.GOV)
Levinskas, D.
1993-01-01
This report discusses the pin processor which receives metal alloy pins cast from recycled Integral Fast Reactor (IFR) fuel and prepares them for assembly into new IFR fuel elements. Either full length as-cast or precut pins are fed to the machine from a magazine, cut if necessary, and measured for length, weight, diameter and deviation from straightness. Accepted pins are loaded into cladding jackets located in a magazine, while rejects and cutting scraps are separated into trays. The magazines, trays, and the individual modules that perform the different machine functions are assembled and removed using remote manipulators and master-slaves.
Integral Fast Reactor fuel pin processor
DOE Office of Scientific and Technical Information (OSTI.GOV)
Levinskas, D.
1993-03-01
This report discusses the pin processor which receives metal alloy pins cast from recycled Integral Fast Reactor (IFR) fuel and prepares them for assembly into new IFR fuel elements. Either full length as-cast or precut pins are fed to the machine from a magazine, cut if necessary, and measured for length, weight, diameter and deviation from straightness. Accepted pins are loaded into cladding jackets located in a magazine, while rejects and cutting scraps are separated into trays. The magazines, trays, and the individual modules that perform the different machine functions are assembled and removed using remote manipulators and master-slaves.
How Can Histograms Be Useful for Introducing Continuous Probability Distributions?
ERIC Educational Resources Information Center
Derouet, Charlotte; Parzysz, Bernard
2016-01-01
The teaching of probability has changed a great deal since the end of the last century. The development of technologies is indeed part of this evolution. In France, continuous probability distributions began to be studied in 2002 by scientific 12th graders, but this subject was marginal and appeared only as an application of integral calculus.…
Energy efficient engine: Propulsion system-aircraft integration evaluation
NASA Technical Reports Server (NTRS)
Owens, R. E.
1979-01-01
Flight performance and operating economics of future commercial transports utilizing the energy efficient engine were assessed as well as the probability of meeting NASA's goals for TSFC, DOC, noise, and emissions. Results of the initial propulsion systems aircraft integration evaluation presented include estimates of engine performance, predictions of fuel burns, operating costs of the flight propulsion system installed in seven selected advanced study commercial transports, estimates of noise and emissions, considerations of thrust growth, and the achievement-probability analysis.
NASA Astrophysics Data System (ADS)
Giraud, O.; Thain, A.; Hannay, J. H.
2004-02-01
The shrunk loop theorem proved here is an integral identity which facilitates the calculation of the relative probability (or probability amplitude) of any given topology that a free, closed Brownian (or Feynman) path of a given 'duration' might have on the twice punctured plane (plane with two marked points). The result is expressed as a 'scattering' series of integrals of increasing dimensionality based on the maximally shrunk version of the path. Physically, this applies in different contexts: (i) the topology probability of a closed ideal polymer chain on a plane with two impassable points, (ii) the trace of the Schrödinger Green function, and thence spectral information, in the presence of two Aharonov-Bohm fluxes and (iii) the same with two branch points of a Riemann surface instead of fluxes. Our theorem starts from the Stovicek scattering expansion for the Green function in the presence of two Aharonov-Bohm flux lines, which itself is based on the famous Sommerfeld one puncture point solution of 1896 (the one puncture case has much easier topology, just one winding number). Stovicek's expansion itself can supply the results at the expense of choosing a base point on the loop and then integrating it away. The shrunk loop theorem eliminates this extra two-dimensional integration, distilling the topology from the geometry.
Hemizygosity Enhances Purifying Selection: Lack of Fast-Z Evolution in Two Satyrine Butterflies.
Rousselle, Marjolaine; Faivre, Nicolas; Ballenghien, Marion; Galtier, Nicolas; Nabholz, Benoit
2016-10-23
The fixation probability of a recessive beneficial mutation is increased on the X or Z chromosome, relative to autosomes, because recessive alleles carried by X or Z are exposed to selection in the heterogametic sex. This leads to an increased dN/dS ratio on sex chromosomes relative to autosomes, a pattern called the "fast-X" or "fast-Z" effect. Besides positive selection, the strength of genetic drift and the efficacy of purifying selection, which affect the rate of molecular evolution, might differ between sex chromosomes and autosomes. Disentangling the complex effects of these distinct forces requires the genome-wide analysis of polymorphism, divergence and gene expression data in a variety of taxa. Here we study the influence of hemizygosity of the Z chromosome in Maniola jurtina and Pyronia tithonus, two species of butterflies (Lepidoptera, Nymphalidae, Satyrinae). Using transcriptome data, we compare the strength of positive and negative selection between Z and autosomes accounting for sex-specific gene expression. We show that M. jurtina and P. tithonus do not experience a faster, but rather a slightly slower evolutionary rate on the Z than on autosomes. Our analysis failed to detect a significant difference in adaptive evolutionary rate between Z and autosomes, but comparison of male-biased, unbiased and female-biased Z-linked genes revealed an increased efficacy of purifying selection against recessive deleterious mutations in female-biased Z-linked genes. This probably contributes to the lack of fast-Z evolution of satyrines. We suggest that the effect of hemizygosity on the fate of recessive deleterious mutations should be taken into account when interpreting patterns of molecular evolution in sex chromosomes vs. autosomes. © The Author 2016. Published by Oxford University Press on behalf of the Society for Molecular Biology and Evolution.
Quantitative 13C NMR characterization of fast pyrolysis oils
Happs, Renee M.; Lisa, Kristina; Ferrell, III, Jack R.
2016-10-20
Quantitative 13C NMR analysis of model catalytic fast pyrolysis (CFP) oils following literature procedures showed poor agreement for aromatic hydrocarbons between NMR measured concentrations and actual composition. Furthermore, modifying integration regions based on DEPT analysis for aromatic carbons resulted in better agreement. Solvent effects were also investigated for hydrotreated CFP oil.
Common and Specific Factors Approaches to Home-Based Treatment: I-FAST and MST
ERIC Educational Resources Information Center
Lee, Mo Yee; Greene, Gilbert J.; Fraser, J. Scott; Edwards, Shivani G.; Grove, David; Solovey, Andrew D.; Scott, Pamela
2013-01-01
Objectives: This study examined the treatment outcomes of integrated families and systems treatment (I-FAST), a moderated common factors approach, in reference to multisystemic therapy (MST), an established specific factor approach, for treating at risk children and adolescents and their families in an intensive community-based setting. Method:…
Quantitative 13C NMR characterization of fast pyrolysis oils
DOE Office of Scientific and Technical Information (OSTI.GOV)
Happs, Renee M.; Lisa, Kristina; Ferrell, III, Jack R.
Quantitative 13C NMR analysis of model catalytic fast pyrolysis (CFP) oils following literature procedures showed poor agreement for aromatic hydrocarbons between NMR measured concentrations and actual composition. Furthermore, modifying integration regions based on DEPT analysis for aromatic carbons resulted in better agreement. Solvent effects were also investigated for hydrotreated CFP oil.
NASA Astrophysics Data System (ADS)
Zheng, Chang-Jun; Chen, Hai-Bo; Chen, Lei-Lei
2013-04-01
This paper presents a novel wideband fast multipole boundary element approach to 3D half-space/plane-symmetric acoustic wave problems. The half-space fundamental solution is employed in the boundary integral equations so that the tree structure required in the fast multipole algorithm is constructed for the boundary elements in the real domain only. Moreover, a set of symmetric relations between the multipole expansion coefficients of the real and image domains are derived, and the half-space fundamental solution is modified for the purpose of applying such relations to avoid calculating, translating and saving the multipole/local expansion coefficients of the image domain. The wideband adaptive multilevel fast multipole algorithm associated with the iterative solver GMRES is employed so that the present method is accurate and efficient for both lowand high-frequency acoustic wave problems. As for exterior acoustic problems, the Burton-Miller method is adopted to tackle the fictitious eigenfrequency problem involved in the conventional boundary integral equation method. Details on the implementation of the present method are described, and numerical examples are given to demonstrate its accuracy and efficiency.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shiraga, H.; Nagatomo, H.; Theobald, W.
Here, integrated fast ignition experiments were performed at ILE, Osaka, and LLE, Rochester, in which a nanosecond driver laser implodes a deuterated plastic shell in front of the tip of a hollow metal cone and an intense ultrashort-pulse laser is injected through the cone to heat the compressed plasma. Based on the initial successful results of fast electron heating of cone-in-shell targets, large-energy short-pulse laser beam lines were constructed and became operational: OMEGA-EP at Rochester and LFEX at Osaka. Neutron enhancement due to heating with a ~kJ short-pulse laser has been demonstrated in the integrated experiments at Osaka and Rochester.more » The neutron yields are being analyzed by comparing the experimental results with simulations. Details of the fast electron beam transport and the electron energy deposition in the imploded fuel plasma are complicated and further studies are imperative. The hydrodynamics of the implosion was studied including the interaction of the imploded core plasma with the cone tip. Theory and simulation studies are presented on the hydrodynamics of a high-gain target for a fast ignition point design.« less
Intense Ion Pulses for Radiation Effects Research
2017-04-01
station; here, the time dependent beam current can be measured with a fast Faraday cup (ə ns time resolution). The transverse distribution of the...focused into a spot with a diameter of about 2 mm [8]. The helium ion current and the integrated charge ver- sus time, measured with the fast Faraday cup...target measured with the fast Faraday cup. The sharp peak in the current measurement shows the beam pulse compression from 1 μs to a few ns. The full
Probabilistic inference in discrete spaces can be implemented into networks of LIF neurons.
Probst, Dimitri; Petrovici, Mihai A; Bytschok, Ilja; Bill, Johannes; Pecevski, Dejan; Schemmel, Johannes; Meier, Karlheinz
2015-01-01
The means by which cortical neural networks are able to efficiently solve inference problems remains an open question in computational neuroscience. Recently, abstract models of Bayesian computation in neural circuits have been proposed, but they lack a mechanistic interpretation at the single-cell level. In this article, we describe a complete theoretical framework for building networks of leaky integrate-and-fire neurons that can sample from arbitrary probability distributions over binary random variables. We test our framework for a model inference task based on a psychophysical phenomenon (the Knill-Kersten optical illusion) and further assess its performance when applied to randomly generated distributions. As the local computations performed by the network strongly depend on the interaction between neurons, we compare several types of couplings mediated by either single synapses or interneuron chains. Due to its robustness to substrate imperfections such as parameter noise and background noise correlations, our model is particularly interesting for implementation on novel, neuro-inspired computing architectures, which can thereby serve as a fast, low-power substrate for solving real-world inference problems.
Chapman-Enskog expansion for the Vicsek model of self-propelled particles
NASA Astrophysics Data System (ADS)
Ihle, Thomas
2016-08-01
Using the standard Vicsek model, I show how the macroscopic transport equations can be systematically derived from microscopic collision rules. The approach starts with the exact evolution equation for the N-particle probability distribution and, after making the mean-field assumption of molecular chaos, leads to a multi-particle Enskog-type equation. This equation is treated by a non-standard Chapman-Enskog expansion to extract the macroscopic behavior. The expansion includes terms up to third order in a formal expansion parameter ɛ, and involves a fast time scale. A self-consistent closure of the moment equations is presented that leads to a continuity equation for the particle density and a Navier-Stokes-like equation for the momentum density. Expressions for all transport coefficients in these macroscopic equations are given explicitly in terms of microscopic parameters of the model. The transport coefficients depend on specific angular integrals which are evaluated asymptotically in the limit of infinitely many collision partners, using an analogy to a random walk. The consistency of the Chapman-Enskog approach is checked by an independent calculation of the shear viscosity using a Green-Kubo relation.
Geobiological constraints on Earth system sensitivity to CO₂ during the Cretaceous and Cenozoic.
Royer, D L; Pagani, M; Beerling, D J
2012-07-01
Earth system climate sensitivity (ESS) is the long-term (>10³ year) response of global surface temperature to doubled CO₂ that integrates fast and slow climate feedbacks. ESS has energy policy implications because global temperatures are not expected to decline appreciably for at least 10³ year, even if anthropogenic greenhouse gas emissions drop to zero. We report provisional ESS estimates of 3 °C or higher for some of the Cretaceous and Cenozoic based on paleo-reconstructions of CO₂ and temperature. These estimates are generally higher than climate sensitivities simulated from global climate models for the same ancient periods (approximately 3 °C). Climate models probably do not capture the full suite of positive climate feedbacks that amplify global temperatures during some globally warm periods, as well as other characteristic features of warm climates such as low meridional temperature gradients. These absent feedbacks may be related to clouds, trace greenhouse gases (GHGs), seasonal snow cover, and/or vegetation, especially in polar regions. Better characterization and quantification of these feedbacks is a priority given the current accumulation of atmospheric GHGs. © 2012 Blackwell Publishing Ltd.
Forcing of the Coupled Ionosphere-Thermosphere (IT) System During Magnetic Storms
NASA Technical Reports Server (NTRS)
Huang, Cheryl; Huang, Yanshi; Su, Yi-Jiun; Sutton, Eric; Hairston, Marc; Coley, W. Robin; Doornbos, Eelco; Zhang, Yongliang
2014-01-01
Poynting flux shows peaks around auroral zone AND inside polar cap. Energy enters IT system at all local times in polar cap. Track-integrated flux at DMSP often peaks at polar latitudes- probably due to increased area of polar cap during storm main phases. center dot lon temperatures at DMSP show large increases in polar region at all local times; cusp and auroral zones do not show distinctively high Ti. center dot I on temperatures in the polar cap are higher than in the auroral zones during quiet times. center dot Neutral densities at GRACE and GOCE show maxima at polar latitudes without clear auroral signatures. Response is fast, minutes from onset to density peaks. center dot GUVI observations of O/N2 ratio during storms show similar response as direct measurements of ion and neutral densities, i.e. high temperatures in polar cap during prestorm quiet period, heating proceeding from polar cap to lower latitudes during storm main phase. center dot Discrepancy between maps of Poynting flux and of ion temperatures/neutral densities suggests that connection between Poynting flux and Joule heating is not simple.
Probabilistic inference in discrete spaces can be implemented into networks of LIF neurons
Probst, Dimitri; Petrovici, Mihai A.; Bytschok, Ilja; Bill, Johannes; Pecevski, Dejan; Schemmel, Johannes; Meier, Karlheinz
2015-01-01
The means by which cortical neural networks are able to efficiently solve inference problems remains an open question in computational neuroscience. Recently, abstract models of Bayesian computation in neural circuits have been proposed, but they lack a mechanistic interpretation at the single-cell level. In this article, we describe a complete theoretical framework for building networks of leaky integrate-and-fire neurons that can sample from arbitrary probability distributions over binary random variables. We test our framework for a model inference task based on a psychophysical phenomenon (the Knill-Kersten optical illusion) and further assess its performance when applied to randomly generated distributions. As the local computations performed by the network strongly depend on the interaction between neurons, we compare several types of couplings mediated by either single synapses or interneuron chains. Due to its robustness to substrate imperfections such as parameter noise and background noise correlations, our model is particularly interesting for implementation on novel, neuro-inspired computing architectures, which can thereby serve as a fast, low-power substrate for solving real-world inference problems. PMID:25729361
BeiDou Signal Acquisition with Neumann–Hoffman Code Modulation in a Degraded Channel
Zhao, Lin; Liu, Aimeng; Ding, Jicheng; Wang, Jing
2017-01-01
With the modernization of global navigation satellite systems (GNSS), secondary codes, also known as the Neumann–Hoffman (NH) codes, are modulated on the satellite signal to obtain a better positioning performance. However, this leads to an attenuation of the acquisition sensitivity of classic integration algorithms because of the frequent bit transitions that refer to the NH codes. Taking weak BeiDou navigation satellite system (BDS) signals as objects, the present study analyzes the side effect of NH codes on acquisition in detail and derives a straightforward formula, which indicates that bit transitions decrease the frequency accuracy. To meet the requirement of carrier-tracking loop initialization, a frequency recalculation algorithm is proposed based on verified fast Fourier transform (FFT) to mitigate the effect, meanwhile, the starting point of NH codes is found. Then, a differential correction is utilized to improve the acquisition accuracy of code phase. Monte Carlo simulations and real BDS data tests demonstrate that the new structure is superior to the conventional algorithms both in detection probability and frequency accuracy in a degraded channel. PMID:28208776
Hostetter, Nathan J.; Evans, Allen F.; Cramer, Bradley M.; Collis, Ken; Lyons, Donald E.; Roby, Daniel D.
2015-01-01
Accurate assessment of specific mortality factors is vital to prioritize recovery actions for threatened and endangered species. For decades, tag recovery methods have been used to estimate fish mortality due to avian predation. Predation probabilities derived from fish tag recoveries on piscivorous waterbird colonies typically reflect minimum estimates of predation due to an unknown and unaccounted-for fraction of tags that are consumed but not deposited on-colony (i.e., deposition probability). We applied an integrated tag recovery modeling approach in a Bayesian context to estimate predation probabilities that accounted for predator-specific tag detection and deposition probabilities in a multiple-predator system. Studies of PIT tag deposition were conducted across three bird species nesting at seven different colonies in the Columbia River basin, USA. Tag deposition probabilities differed significantly among predator species (Caspian ternsHydroprogne caspia: deposition probability = 0.71, 95% credible interval [CRI] = 0.51–0.89; double-crested cormorants Phalacrocorax auritus: 0.51, 95% CRI = 0.34–0.70; California gulls Larus californicus: 0.15, 95% CRI = 0.11–0.21) but showed little variation across trials within a species or across years. Data from a 6-year study (2008–2013) of PIT-tagged juvenile Snake River steelhead Oncorhynchus mykiss (listed as threatened under the Endangered Species Act) indicated that colony-specific predation probabilities ranged from less than 0.01 to 0.17 and varied by predator species, colony location, and year. Integrating the predator-specific deposition probabilities increased the predation probabilities by a factor of approximately 1.4 for Caspian terns, 2.0 for double-crested cormorants, and 6.7 for California gulls compared with traditional minimum predation rate methods, which do not account for deposition probabilities. Results supported previous findings on the high predation impacts from strictly piscivorous waterbirds nesting in the Columbia River estuary (i.e., terns and cormorants), but our findings also revealed greater impacts of a generalist predator species (i.e., California gulls) than were previously documented. Approaches used in this study allow for direct comparisons among multiple fish mortality factors and considerably improve the reliability of tag recovery models for estimating predation probabilities in multiple-predator systems.
Restoring integrity—A grounded theory of coping with a fast track surgery programme
Jørgensen, Lene Bastrup; Fridlund, Bengt
2016-01-01
Aims and objectives The aim of this study was to generate a theory conceptualizing and explaining behavioural processes involved in coping in order to identify the predominant coping types and coping type-specific features. Background Patients undergoing fast track procedures do not experience a higher risk of complications, readmission, or mortality. However, such programmes presuppose an increasing degree of patient involvement, placing high educational, physical, and mental demands on the patients. There is a lack of knowledge about how patients understand and cope with fast track programmes. Design The study design used classical grounded theory. Methods The study used a multimodal approach with qualitative and quantitative data sets from 14 patients. Results Four predominant types of coping, with distinct physiological, cognitive, affective, and psychosocial features, existed among patients going through a fast track total hip replacement programme. These patients’ main concern was to restore their physical and psychosocial integrity, which had been compromised by reduced function and mobility in daily life. To restore integrity they economized their mental resources, while striving to fulfil the expectations of the fast track programme. This goal was achieved by being mentally proactive and physically active. Three out of the four predominant types of coping matched the expectations expressed in the fast track programme. The non-matching behaviour was seen among the most nervous patients, who claimed the right to diverge from the programme. Conclusion In theory, four predominant types of coping with distinct physiological, cognitive, affective, and psychosocial features occur among patients going through a fast track total hip arthroplasty programme. PMID:26751199
Effect of inaction on function of fast and slow muscle spindles
NASA Technical Reports Server (NTRS)
Arutyunyan, R. S.
1980-01-01
There is no data on the comparative effect of tenotomy on the function of the muscle spindles of fast and slow muscles. This study covers this question. The experiments were conducted on cats. The musuculus extensor digitorum longus (m. EDL) was selected as the fast muscle, and the musculus soleus (m. Sol.) as the slow. In a comparison of the spontaneous activity of primary and secondary endings of the fast and slow muscle spindles (i.e., the activity with complete relaxation of the muscles) normally no difference between them was successfully found. The authors recorded the integrative, and not the individual activity, and secondly, under conditions of such recording technique, those slight changes that are observed in the fast muscle receptors could remain unnoticed.
Singh, Rabindarjeet; Hwa, Ooi Cheong; Roy, Jolly; Jin, Chai Wen; Ismail, Siti Musyrifah; Lan, Mohamad Faizal; Hiong, Loo Lean; Aziz, Abdul-Rashid
2011-09-01
To examine the subjective perception of daily acute fasting on sports performance, training, sleep and dietary patterns of Muslim athletes during the Ramadan month. Seven hundred and thirty-four (411 male and 323 female) Malaysian Junior-level Muslim athletes (mean age 16.3 ± 2.6 y) participated in the survey which was designed to establish the personal perception of their sport performance, sleep pattern, food and fluid intake during Ramadan fasting. The survey was conducted during and immediately after the month of Ramadan in 2009. Twenty-four percent of the athletes perceived that there was an adverse effect of the Ramadan fast on their sporting performance and 29.3% reported that quality of training during Ramadan was also negatively influenced. Majority (48.2%) of the athletes stated that Ramadan fasting did not affect their normal sleep pattern but 66.6% of them complained of sleepiness during the daytime. Half of the athletes (41.4%) maintained the caloric intake during Ramadan as they normally would with the majority of them (76.2%) reporting that they consumed more fluids during Ramadan. Overall, Malaysian Junior-level Muslim athletes showed diverse views in their perception of changes in their training, sleep and dietary patterns during Ramadan fast. These individual differences probably indicate differences in the athletes' adaptability and coping strategies during fasting and training in Ramadan.
Influence of fasting during moult on the faecal microbiota of penguins.
Dewar, Meagan L; Arnould, John P Y; Krause, Lutz; Trathan, Phil; Dann, Peter; Smith, Stuart C
2014-01-01
Many seabirds including penguins are adapted to long periods of fasting, particularly during parts of the reproductive cycle and during moult. However, the influence of fasting on the gastrointestinal (GI) microbiota has not been investigated in seabirds. Therefore, the present study aimed to examine the microbial composition and diversity of the GI microbiota of fasting little (Eudyptula minor) and king penguins (Aptenodytes patagonicus) penguins during early and late moult. The results from this study indicated that there was little change in the abundance of the major phyla during moult, except for a significant increase in the level of Proteobacteria in king penguins. In king penguins the abundance of Fusobacteria increases from 1.73% during early moult to 33.6% by late moult, whilst the abundance of Proteobacteria (35.7% to 17.2%) and Bacteroidetes (19.5% to 11%) decrease from early to late moult. In little penguins, a decrease in the abundances of Firmicutes (44% to 29%) and an increase in the abundance of Bacteroidetes (11% to 20%) were observed from early to late moult respectively. The results from this study indicate that the microbial composition of both king and little penguins alters during fasting. However, it appears that the microbial composition of king penguins is more affected by fasting than little penguins with the length of fast the most probable cause for this difference.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bustraan, M.; Coehoorn, J.; Veenema, J.J.
This report is a collection of separate contributions on some aspects of the work done on STEK up to February 1970. A description is given of STEK together with the philosophy of its design, i.e. integral measurements of fission product cross sections by a sample oscillator technique in fast reactor spectra. The influences fission products may have on fast breeder reactors are briefly demonstrated by an example. A description of the facility and of the sample oscillator and sample exchange mechanism is given. Some preliminary results of measurements of reactor parameters and the neutron spectrum in the first fast zonemore » in STEK are given. For the use of lead as material for the buffer an argumentation is given. The proposed program for the measurements of the integral fission product cross sections is outlined. The procurement of some, highly active, samples of actual fission products is briefly sketched. (auth)« less
Application and sensitivity investigation of Fourier transforms for microwave radiometric inversions
NASA Technical Reports Server (NTRS)
Holmes, J. J.; Balanis, C. A.
1974-01-01
Existing microwave radiometer technology now provides a suitable method for remote determination of the ocean surface's absolute brightness temperature. To extract the brightness temperature of the water from the antenna temperature equation, an unstable Fredholm integral equation of the first kind was solved. Fast Fourier Transform techniques were used to invert the integral after it is placed into a cross-correlation form. Application and verification of the methods to a two-dimensional modeling of a laboratory wave tank system were included. The instability of the Fredholm equation was then demonstrated and a restoration procedure was included which smooths the resulting oscillations. With the recent availability and advances of Fast Fourier Transform techniques, the method presented becomes very attractive in the evaluation of large quantities of data. Actual radiometric measurements of sea water are inverted using the restoration method, incorporating the advantages of the Fast Fourier Transform algorithm for computations.
Multiscale solvers and systematic upscaling in computational physics
NASA Astrophysics Data System (ADS)
Brandt, A.
2005-07-01
Multiscale algorithms can overcome the scale-born bottlenecks that plague most computations in physics. These algorithms employ separate processing at each scale of the physical space, combined with interscale iterative interactions, in ways which use finer scales very sparingly. Having been developed first and well known as multigrid solvers for partial differential equations, highly efficient multiscale techniques have more recently been developed for many other types of computational tasks, including: inverse PDE problems; highly indefinite (e.g., standing wave) equations; Dirac equations in disordered gauge fields; fast computation and updating of large determinants (as needed in QCD); fast integral transforms; integral equations; astrophysics; molecular dynamics of macromolecules and fluids; many-atom electronic structures; global and discrete-state optimization; practical graph problems; image segmentation and recognition; tomography (medical imaging); fast Monte-Carlo sampling in statistical physics; and general, systematic methods of upscaling (accurate numerical derivation of large-scale equations from microscopic laws).
Fast algorithms for Quadrature by Expansion I: Globally valid expansions
NASA Astrophysics Data System (ADS)
Rachh, Manas; Klöckner, Andreas; O'Neil, Michael
2017-09-01
The use of integral equation methods for the efficient numerical solution of PDE boundary value problems requires two main tools: quadrature rules for the evaluation of layer potential integral operators with singular kernels, and fast algorithms for solving the resulting dense linear systems. Classically, these tools were developed separately. In this work, we present a unified numerical scheme based on coupling Quadrature by Expansion, a recent quadrature method, to a customized Fast Multipole Method (FMM) for the Helmholtz equation in two dimensions. The method allows the evaluation of layer potentials in linear-time complexity, anywhere in space, with a uniform, user-chosen level of accuracy as a black-box computational method. Providing this capability requires geometric and algorithmic considerations beyond the needs of standard FMMs as well as careful consideration of the accuracy of multipole translations. We illustrate the speed and accuracy of our method with various numerical examples.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Korhonen, Marko; Lee, Eunghyun
2014-01-15
We treat the N-particle zero range process whose jumping rates satisfy a certain condition. This condition is required to use the Bethe ansatz and the resulting model is the q-boson model by Sasamoto and Wadati [“Exact results for one-dimensional totally asymmetric diffusion models,” J. Phys. A 31, 6057–6071 (1998)] or the q-totally asymmetric zero range process (TAZRP) by Borodin and Corwin [“Macdonald processes,” Probab. Theory Relat. Fields (to be published)]. We find the explicit formula of the transition probability of the q-TAZRP via the Bethe ansatz. By using the transition probability we find the probability distribution of the left-most particle'smore » position at time t. To find the probability for the left-most particle's position we find a new identity corresponding to identity for the asymmetric simple exclusion process by Tracy and Widom [“Integral formulas for the asymmetric simple exclusion process,” Commun. Math. Phys. 279, 815–844 (2008)]. For the initial state that all particles occupy a single site, the probability distribution of the left-most particle's position at time t is represented by the contour integral of a determinant.« less
ERIC Educational Resources Information Center
Noddings, Nel
2004-01-01
Most teachers have been good students. Some students are fast learners and attain the required knowledge and skills easily; others are obedient, hard workers. In either case, teachers are likely to believe that if students really try, they will do well. Listening to students over many years, the author has learned that this is probably not true.…
2008-09-01
Littoral Combat Ships OTH Over-The-Horizon Pd Probability of Detection PGGF Fast Attack Craft – Missile PIM Plan of Intended Movement Pk...Lucas Naval Postgraduate School Monterey, California 2. Dudley Knox Library Naval Postgraduate School Monterey, California 3. CAPT Doug Otte
NASA Astrophysics Data System (ADS)
Ros, Paul; Leconte, Pierre; Blaise, Patrick; Naymeh, Laurent
2017-09-01
The current knowledge of nuclear data in the fast neutron energy range is not as good as in the thermal range, resulting in larger propagated uncertainties in integral quantities such as critical masses or reactivity effects. This situation makes it difficult to get the full benefit from recent advances in modeling and simulation. Zero power facilities such as the French ZPR MINERVE have already demonstrated that they can contribute to significantly reduce those uncertainties thanks to dedicated experiments. Historically, MINERVE has been mainly dedicated to thermal spectrum studies. However, experiments involving fast-thermal coupled cores were also performed in MINERVE as part of the ERMINE program, in order to improve nuclear data in fast spectra for the two French SFRs: PHENIX and SUPERPHENIX. Some of those experiments have been recently revisited. In particular, a full characterization of ZONA-1 and ZONA-3, two different cores loaded in the ERMINE V campaign, has been done, with much attention paid to possible sources of errors. It includes detailed geometric descriptions, energy profiles of the direct and adjoint fluxes and spectral indices obtained thanks to Monte Carlo calculations and compared to a reference fast core configuration. Sample oscillation experiments of separated fission products such as 103Rh or 99Tc, which were part of the ERMINE V program, have been simulated using recently-developed options in the TRIPOLI-4 code and compared to the experimental values. The present paper describes the corresponding results. The findings motivate in-depth studies for designing optimized coupled-core conditions in ZEPHYR, a new ZPR which will replace MINERVE and will provide integral data to meet the needs of Gen-III and Gen-IV reactors.
Strategic interactions, affective reactions, and fast adaptations.
Kareev, Yaakov; Avrahami, Judith; Fiedler, Klaus
2014-06-01
We studied repeated choices under uncertainty in situations in which the source of uncertainty is the choice of an interaction partner. In 1 experiment the participants engaged in repeated decisions in a mixed motive game; in another experiment the options and outcomes were identical to those in the 1st, but periods of the mixed-motive game alternated with periods of a coordination game, with the change in period not announced. We analyzed choice dynamics-the relationship between an outcome and the choice that followed-and aggregate choice probabilities to gauge the relative merit of reward-based or affect-based accounts (the affects considered being disappointment and regret). In both experiments choice dynamics were essentially identical and were compatible with only the regret-based account. This was true irrespective of the game played or the stage (early or late) of the game. Moreover, the same dynamics explained the very different aggregate probabilities with which the 2 options were chosen in the 2 games and the remarkably fast adaptations to unannounced changes in the game played. PsycINFO Database Record (c) 2014 APA, all rights reserved.
Sub-barrier quasifission in heavy element formation reactions with deformed actinide target nuclei
NASA Astrophysics Data System (ADS)
Hinde, D. J.; Jeung, D. Y.; Prasad, E.; Wakhle, A.; Dasgupta, M.; Evers, M.; Luong, D. H.; du Rietz, R.; Simenel, C.; Simpson, E. C.; Williams, E.
2018-02-01
Background: The formation of superheavy elements (SHEs) by fusion of two massive nuclei is severely inhibited by the competing quasifission process. Low excitation energies favor SHE survival against fusion-fission competition. In "cold" fusion with spherical target nuclei near 208Pb, SHE yields are largest at beam energies significantly below the average capture barrier. In "hot" fusion with statically deformed actinide nuclei, this is not the case. Here the elongated deformation-aligned configurations in sub-barrier capture reactions inhibits fusion (formation of a compact compound nucleus), instead favoring rapid reseparation through quasifission. Purpose: To determine the probabilities of fast and slow quasifission in reactions with prolate statically deformed actinide nuclei, through measurement and quantitative analysis of the dependence of quasifission characteristics at beam energies spanning the average capture barrier energy. Methods: The Australian National University Heavy Ion Accelerator Facility and CUBE fission spectrometer have been used to measure fission and quasifission mass and angle distributions for reactions with projectiles from C to S, bombarding Th and U target nuclei. Results: Mass-asymmetric quasifission occurring on a fast time scale, associated with collisions with the tips of the prolate actinide nuclei, shows a rapid increase in probability with increasing projectile charge, the transition being centered around projectile atomic number ZP=14 . For mass-symmetric fission events, deviations of angular anisotropies from expectations for fusion fission, indicating a component of slower quasifission, suggest a similar transition, but centered around ZP˜8 . Conclusions: Collisions with the tips of statically deformed prolate actinide nuclei show evidence for two distinct quasifission processes of different time scales. Their probabilities both increase rapidly with the projectile charge. The probability of fusion can be severely suppressed by these two quasifission processes, since the sub-barrier heavy element yield is likely to be determined by the product of the probabilities of surviving each quasifission process.
Burst wait time simulation of CALIBAN reactor at delayed super-critical state
DOE Office of Scientific and Technical Information (OSTI.GOV)
Humbert, P.; Authier, N.; Richard, B.
2012-07-01
In the past, the super prompt critical wait time probability distribution was measured on CALIBAN fast burst reactor [4]. Afterwards, these experiments were simulated with a very good agreement by solving the non-extinction probability equation [5]. Recently, the burst wait time probability distribution has been measured at CEA-Valduc on CALIBAN at different delayed super-critical states [6]. However, in the delayed super-critical case the non-extinction probability does not give access to the wait time distribution. In this case it is necessary to compute the time dependent evolution of the full neutron count number probability distribution. In this paper we present themore » point model deterministic method used to calculate the probability distribution of the wait time before a prescribed count level taking into account prompt neutrons and delayed neutron precursors. This method is based on the solution of the time dependent adjoint Kolmogorov master equations for the number of detections using the generating function methodology [8,9,10] and inverse discrete Fourier transforms. The obtained results are then compared to the measurements and Monte-Carlo calculations based on the algorithm presented in [7]. (authors)« less
El-Migdadi, Fayig; Shotar, Ali; El-Akawi, Zeyad; Banihani, Ibrahim; Abudheese, Rola
2004-01-01
To study a possible effect of Ramadan fasting on luteinizing hormone and testosterone in people of the Jordan Valley. A comparative study (n=40) of serum levels of luteinizing hormone (LH) and testosterone (T) between people living in the Jordan Valley (JV), n=20, 360 meters below sea level, and those living in Ramtha City (RC), n=20, 600 meters above sea level, was conducted in December, 1998. A similar study (n=40) was also done during January 1999 in fasting people during the month of Ramadan. Serum levels of LH in non-fasting people of the JV were statistically similar to those in people of RC. There was also no difference in serum levels of T between non-fasting people of the JV and those in RC. Serum levels of LH in fasting people of the JV were statistically indifferent from those fasting in RC. Serum T levels in fasting people of the JV, on the other hand, were higher than those in fasting people of RC (76+/-18.3 ng/ml compared to 62.7+/-24.2 ng/ml). It is probably the environmental factors such as the higher barometric pressure of the JV compared to that at above sea level that play a role in higher serum levels of T in people of the JV. Other factors, such as genetic background and/or the cultural and nutritional characteristics of the people of the JV, may also contribute to this difference in serum T levels.
Two ways to feel the pressure: an endothelial Ca(2+) entry channel with dual mechanosensitivity.
Groschner, Klaus
2002-01-01
One impressive function of the vascular endothelium is its ability to adjust the release of vasoactive mediators such as NO and PGI(2) almost instantaneously to changes in blood flow or blood pressure. Besides this fast feedback response to hemodynamic alterations, the endothelium is subject to long-term adaptations that are crucial for prevention of pathological processes such as atherogenesis. Among the various signals that are sensed by endothelial cells, mechanical forces which arise from pulsatile blood flow are probably most important for fast as well as long-term control of blood vessel function by the endothelium.
Kim, K. S.; Nakae, L. F.; Prasad, M. K.; ...
2017-07-31
We present that fast nanosecond timescale neutron and gamma-ray counting can be performed with a (liquid) scintillator array. Fission chains in metal evolve over a timescale of tens of nanoseconds. If the metal is surrounded by moderator, neutrons leaking from the metal can thermalize and diffuse in the moderator. With finite probability, the diffusing neutrons can return to the metal and restart the fast fission chain. The timescale for this restart process is microseconds. A theory describing time evolving fission chains for metal surrounded by moderator, including this restart process, is presented. Finally, this theory is sufficiently simple for itmore » to be implemented for real-time analysis.« less
The effect of social integration on outcomes after major lower extremity amputation.
Hawkins, Alexander T; Pallangyo, Anthony J; Herman, Ayesiga M; Schaumeier, Maria J; Smith, Ann D; Hevelone, Nathanael D; Crandell, David M; Nguyen, Louis L
2016-01-01
Major lower extremity (MLE) amputation is a common procedure that results in a profound change in a patient's life. We sought to determine the association between social support and outcomes after amputation. We hypothesized that patients with greater social support will have better post amputation outcomes. From November 2011 to May 2013, we conducted a cross-sectional, observational, multicenter study. Social integration was measured by the social integration subset of the Short Form Craig Handicap Assessment and Reporting Technique. Systemic social support was assessed by comparing a United States and Tanzanian population. Walking function was measured using the 6-minute walk test and quality of life (QoL) was measured using the EuroQol-5D. We recruited 102 MLE amputees. Sixty-three patients were enrolled in the United States with a mean age of 58.0. Forty-two (67%) were male. Patients with low social integration were more likely to be unable to ambulate (no walk 39% vs slow walk 23% vs fast walk 10%; P = .01) and those with high social integration were more likely to be fast walkers (no walk 10% vs slow walk 59% vs fast walk 74%; P = .01). This relationship persisted in a multivariable analysis. Increasing social integration scores were also positively associated with increasing QoL scores in a multivariable analysis (β, .002; standard error, 0.0008; P = .02). In comparing the United States population with the Tanzanian cohort (39 subjects), there were no differences between functional or QoL outcomes in the systemic social support analysis. In the United States population, increased social integration is associated with both improved function and QoL outcomes among MLE amputees. Systemic social support, as measured by comparing the United States population with a Tanzanian population, was not associated with improved function or QoL outcomes. In the United States, steps should be taken to identify and aid amputees with poor social integration. Copyright © 2016 Society for Vascular Surgery. Published by Elsevier Inc. All rights reserved.
The Effect of Social Integration on Outcomes after Major Lower Extremity Amputation
Hawkins, Alexander T.; Pallangyo, Anthony J.; Herman, Ayesiga M.; Schaumeier, Maria J.; Smith, Ann D.; Hevelone, Nathanael D.; Crandell, David M.; Nguyen, Louis
2016-01-01
Objective Major lower extremity amputation is a common procedure that results in a profound change in a patient's life. We sought to determine the association between social support and outcomes after amputation. We hypothesized that patients with greater social support will have better post amputation outcomes. Methods From November 2011 to May 2013, we conducted a cross-sectional, observational, multicenter study. Social integration was measured by the social integration subset of the Short Form (Craig Handicap Assessment and Reporting Technique (CHART). Systemic social support was assessed by comparing a US and Tanzanian population. Walking function was measured using the 6MWT and quality of life (QoL) was measured using the EQ-5D. Results 102 major lower extremity amputees were recruited. 63 patients were enrolled in the US with a mean age of 58.0. Forty-two (67%) were male. Patients with low social integration were more likely to be unable to ambulate (no walk 39% vs. slow walk 23% vs. fast walk 10%; P=.01) and those with high social integration were more likely to be fast walkers (no walk 10% vs. slow walk 59% vs. fast walk 74%; P=.01). This relationship persisted in a multivariable analysis. Increasing social integration scores were also positively associated with increasing quality of life scores in a multivariable analysis (β .002; SE .0008; P = .02). In comparing the US population with the Tanzanian cohort (39 subjects), there were no significant differences between functional or quality of life outcomes in the systemic social support analysis. Conclusions In the US population, increased social integration is associated with both improved function and quality of life outcomes among major lower extremity amputees. Systemic social support, as measured by comparing the US population with a Tanzanian population, was not associated with improved function or quality of life outcomes. In the US, steps should be taken to identify and aid amputees with poor social integration. PMID:26474508
Thermochemical Process Integration, Scale-Up, and Piloting Publications |
-Economic Assessment of Ex Situ Catalytic Fast Pyrolysis of Biomass: A Fixed Bed Reactor Implementation Scenario for Future Feasibility, Topics in Catalysis Image of a schematic of hot gas filter and ex situ Research Pathways with In Situ and Ex Situ Upgrading of Fast Pyrolysis Vapors, NREL Technical Report Image
Fast Electromagnetic Solvers for Large-Scale Naval Scattering Problems
2008-09-27
IEEE Trans. Antennas Propag., vol. 52, no. 8, pp. 2141–2146, 2004. [12] R. J. Burkholder and J. F. Lee, “Fast dual-MGS block-factorization algorithm...Golub and C. F. V. Loan, Matrix Computations. Baltimore: The Johns Hopkins University Press, 1996. [20] W. D. Li, W. Hong, and H. X. Zhou, “Integral
A water-based fast integrated mobility spectrometer (WFIMS) with enhanced dynamic size range
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pinterich, Tamara; Spielman, Steven R.; Hering, Susanne
We developed a water-based fast integrated mobility spectrometer (WFIMS) with enhanced dynamic size range. The WFIMS builds on two established technologies: the fast integrated mobility spectrometer and laminar flow water-based condensation methodology. Inside WFIMS, particles of differing electrical mobility are separated in a drift tube and subsequently enlarged through water condensation. Particle size and concentration are measured via digital imaging at a frame rate of 10 Hz. When we measure particles of different mobilities simultaneously, the WFIMS resolves particle diameters ranging from 8 to 580 nm within 1 s or less. The performance of WFIMS was characterized with differential mobilitymore » analyzer (DMA) classified (NH 4) 2SO 2 particles with diameters ranging from 8 to 265 nm. The mean particle diameters measured by WFIMS were found to be in excellent agreement with DMA centroid diameters. Furthermore, detection efficiency of WFIMS was characterized using a condensation particle counter as a reference and is nearly 100% for particles with diameter greater than 8 nm. In general, measured and simulated WFIMS mobility resolutions are in good agreement. But, some deviations are observed at low particle mobilities, likely due to the non-idealities of the WFIMS electric field.« less
A water-based fast integrated mobility spectrometer (WFIMS) with enhanced dynamic size range
Pinterich, Tamara; Spielman, Steven R.; Hering, Susanne; ...
2017-06-08
We developed a water-based fast integrated mobility spectrometer (WFIMS) with enhanced dynamic size range. The WFIMS builds on two established technologies: the fast integrated mobility spectrometer and laminar flow water-based condensation methodology. Inside WFIMS, particles of differing electrical mobility are separated in a drift tube and subsequently enlarged through water condensation. Particle size and concentration are measured via digital imaging at a frame rate of 10 Hz. When we measure particles of different mobilities simultaneously, the WFIMS resolves particle diameters ranging from 8 to 580 nm within 1 s or less. The performance of WFIMS was characterized with differential mobilitymore » analyzer (DMA) classified (NH 4) 2SO 2 particles with diameters ranging from 8 to 265 nm. The mean particle diameters measured by WFIMS were found to be in excellent agreement with DMA centroid diameters. Furthermore, detection efficiency of WFIMS was characterized using a condensation particle counter as a reference and is nearly 100% for particles with diameter greater than 8 nm. In general, measured and simulated WFIMS mobility resolutions are in good agreement. But, some deviations are observed at low particle mobilities, likely due to the non-idealities of the WFIMS electric field.« less
Are there common mathematical structures in economics and physics?
NASA Astrophysics Data System (ADS)
Mimkes, Jürgen
2016-12-01
Economics is a field that looks into the future. We may know a few things ahead (ex ante), but most things we only know, afterwards (ex post). How can we work in a field, where much of the important information is missing? Mathematics gives two answers: 1. Probability theory leads to microeconomics: the Lagrange function optimizes utility under constraints of economic terms (like costs). The utility function is the entropy, the logarithm of probability. The optimal result is given by a probability distribution and an integrating factor. 2. Calculus leads to macroeconomics: In economics we have two production factors, capital and labour. This requires two dimensional calculus with exact and not-exact differentials, which represent the "ex ante" and "ex post" terms of economics. An integrating factor turns a not-exact term (like income) into an exact term (entropy, the natural production function). The integrating factor is the same as in microeconomics and turns the not-exact field of economics into an exact physical science.
A hybrid method for transient wave propagation in a multilayered solid
NASA Astrophysics Data System (ADS)
Tian, Jiayong; Xie, Zhoumin
2009-08-01
We present a hybrid method for the evaluation of transient elastic-wave propagation in a multilayered solid, integrating reverberation matrix method with the theory of generalized rays. Adopting reverberation matrix formulation, Laplace-Fourier domain solutions of elastic waves in the multilayered solid are expanded into the sum of a series of generalized-ray group integrals. Each generalized-ray group integral containing Kth power of reverberation matrix R represents the set of K-times reflections and refractions of source waves arriving at receivers in the multilayered solid, which was computed by fast inverse Laplace transform (FILT) and fast Fourier transform (FFT) algorithms. However, the calculation burden and low precision of FILT-FFT algorithm limit the application of reverberation matrix method. In this paper, we expand each of generalized-ray group integrals into the sum of a series of generalized-ray integrals, each of which is accurately evaluated by Cagniard-De Hoop method in the theory of generalized ray. The numerical examples demonstrate that the proposed method makes it possible to calculate the early-time transient response in the complex multilayered-solid configuration efficiently.
NASA Astrophysics Data System (ADS)
Mohammadian-Behbahani, Mohammad-Reza; Saramad, Shahyar
2018-04-01
Model based analysis methods are relatively new approaches for processing the output data of radiation detectors in nuclear medicine imaging and spectroscopy. A class of such methods requires fast algorithms for fitting pulse models to experimental data. In order to apply integral-equation based methods for processing the preamplifier output pulses, this article proposes a fast and simple method for estimating the parameters of the well-known bi-exponential pulse model by solving an integral equation. The proposed method needs samples from only three points of the recorded pulse as well as its first and second order integrals. After optimizing the sampling points, the estimation results were calculated and compared with two traditional integration-based methods. Different noise levels (signal-to-noise ratios from 10 to 3000) were simulated for testing the functionality of the proposed method, then it was applied to a set of experimental pulses. Finally, the effect of quantization noise was assessed by studying different sampling rates. Promising results by the proposed method endorse it for future real-time applications.
Switching probability of all-perpendicular spin valve nanopillars
NASA Astrophysics Data System (ADS)
Tzoufras, M.
2018-05-01
In all-perpendicular spin valve nanopillars the probability density of the free-layer magnetization is independent of the azimuthal angle and its evolution equation simplifies considerably compared to the general, nonaxisymmetric geometry. Expansion of the time-dependent probability density to Legendre polynomials enables analytical integration of the evolution equation and yields a compact expression for the practically relevant switching probability. This approach is valid when the free layer behaves as a single-domain magnetic particle and it can be readily applied to fitting experimental data.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Howe, Daniel T.; Westover, Tyler; Carpenter, Daniel
2015-05-21
Feedstock composition can affect final fuel yields and quality for the fast pyrolysis and hydrotreatment upgrading pathway. However, previous studies have focused on individual unit operations rather than the integrated system. In this study, a suite of six pure lignocellulosic feedstocks (clean pine, whole pine, tulip poplar, hybrid poplar, switchgrass, and corn stover) and two blends (equal weight percentages whole pine/tulip poplar/switchgrass and whole pine/clean pine/hybrid poplar) were prepared and characterized at Idaho National Laboratory. These blends then underwent fast pyrolysis at the National Renewable Energy Laboratory and hydrotreatment at Pacific Northwest National Laboratory. Although some feedstocks showed a highmore » fast pyrolysis bio-oil yield such as tulip poplar at 57%, high yields in the hydrotreater were not always observed. Results showed overall fuel yields of 15% (switchgrass), 18% (corn stover), 23% (tulip poplar, Blend 1, Blend 2), 24% (whole pine, hybrid poplar) and 27% (clean pine). Simulated distillation of the upgraded oils indicated that the gasoline fraction varied from 39% (clean pine) to 51% (corn stover), while the diesel fraction ranged from 40% (corn stover) to 46% (tulip poplar). Little variation was seen in the jet fuel fraction at 11 to 12%. Hydrogen consumption during hydrotreating, a major factor in the economic feasibility of the integrated process, ranged from 0.051 g/g dry feed (tulip poplar) to 0.070 g/g dry feed (clean pine).« less
Effects of heterogeneous traffic with speed limit zone on the car accidents
NASA Astrophysics Data System (ADS)
Marzoug, R.; Lakouari, N.; Bentaleb, K.; Ez-Zahraouy, H.; Benyoussef, A.
2016-06-01
Using the extended Nagel-Schreckenberg (NS) model, we numerically study the impact of the heterogeneity of traffic with speed limit zone (SLZ) on the probability of occurrence of car accidents (Pac). SLZ in the heterogeneous traffic has an important effect, typically in the mixture velocities case. In the deterministic case, SLZ leads to the appearance of car accidents even in the low densities, in this region Pac increases with increasing of fraction of fast vehicles (Ff). In the nondeterministic case, SLZ decreases the effect of braking probability Pb in the low densities. Furthermore, the impact of multi-SLZ on the probability Pac is also studied. In contrast with the homogeneous case [X. Li, H. Kuang, Y. Fan and G. Zhang, Int. J. Mod. Phys. C 25 (2014) 1450036], it is found that in the low densities the probability Pac without SLZ (n = 0) is low than Pac with multi-SLZ (n > 0). However, the existence of multi-SLZ in the road decreases the risk of collision in the congestion phase.
Reducing the two-loop large-scale structure power spectrum to low-dimensional, radial integrals
Schmittfull, Marcel; Vlah, Zvonimir
2016-11-28
Modeling the large-scale structure of the universe on nonlinear scales has the potential to substantially increase the science return of upcoming surveys by increasing the number of modes available for model comparisons. One way to achieve this is to model nonlinear scales perturbatively. Unfortunately, this involves high-dimensional loop integrals that are cumbersome to evaluate. Here, trying to simplify this, we show how two-loop (next-to-next-to-leading order) corrections to the density power spectrum can be reduced to low-dimensional, radial integrals. Many of those can be evaluated with a one-dimensional fast Fourier transform, which is significantly faster than the five-dimensional Monte-Carlo integrals thatmore » are needed otherwise. The general idea of this fast fourier transform perturbation theory method is to switch between Fourier and position space to avoid convolutions and integrate over orientations, leaving only radial integrals. This reformulation is independent of the underlying shape of the initial linear density power spectrum and should easily accommodate features such as those from baryonic acoustic oscillations. We also discuss how to account for halo bias and redshift space distortions.« less
Spiesberger, John L
2013-02-01
The hypothesis tested is that internal gravity waves limit the coherent integration time of sound at 1346 km in the Pacific ocean at 133 Hz and a pulse resolution of 0.06 s. Six months of continuous transmissions at about 18 min intervals are examined. The source and receiver are mounted on the bottom of the ocean with timing governed by atomic clocks. Measured variability is only due to fluctuations in the ocean. A model for the propagation of sound through fluctuating internal waves is run without any tuning with data. Excellent resemblance is found between the model and data's probability distributions of integration time up to five hours.
Maximizing Information Diffusion in the Cyber-physical Integrated Network †
Lu, Hongliang; Lv, Shaohe; Jiao, Xianlong; Wang, Xiaodong; Liu, Juan
2015-01-01
Nowadays, our living environment has been embedded with smart objects, such as smart sensors, smart watches and smart phones. They make cyberspace and physical space integrated by their abundant abilities of sensing, communication and computation, forming a cyber-physical integrated network. In order to maximize information diffusion in such a network, a group of objects are selected as the forwarding points. To optimize the selection, a minimum connected dominating set (CDS) strategy is adopted. However, existing approaches focus on minimizing the size of the CDS, neglecting an important factor: the weight of links. In this paper, we propose a distributed maximizing the probability of information diffusion (DMPID) algorithm in the cyber-physical integrated network. Unlike previous approaches that only consider the size of CDS selection, DMPID also considers the information spread probability that depends on the weight of links. To weaken the effects of excessively-weighted links, we also present an optimization strategy that can properly balance the two factors. The results of extensive simulation show that DMPID can nearly double the information diffusion probability, while keeping a reasonable size of selection with low overhead in different distributed networks. PMID:26569254
2014-01-01
The survival rate of lung cancer can be significantly improved by monitoring biomarkers in exhaled air that indicate diseases in early stage, so it is very important to develop micro analytical systems which can offer a fast, on-site, real-time detecting biomarkers in exhaled air. In this paper, a mini-gas chromatography (GC)-photo-ionization detector (PID) system integrated with a micro GC column and a micro pre-concentrator was developed for forming an inexpensive, fast, and non-invasive diagnostic tool for lung cancer. This system has very strong concentrate ability owing to its integrated micro pre-concentrator, which make the detection of trace components in exhaled air very easy. In addition, the integrated micro GC column can separate complex mixtures, which overcome low resolution and poor anti-interference ability of other instruments. The results indicated that the mini-GC-PID system can effectively separate and detect the biomarkers at parts-per-billion (ppb) level. PMID:25339856
1981-06-01
for a de- tection probability of PD and associated false alarm probability PFA (in dB). 21 - - - II V. REFERENCE MODEL A. INTRODUCTION In order to...space for which to choose HI . PFA = P (wI 0o)dw = Q(---) (26) j 0 Similarity, the miss probability=l-detection probability is obtained by integrating...31) = 2 (1+ (22 [()BT z] ~Z The input signal-to-noise ratio: S/N(input) - a2 (32) The probability of false alarm: PFA = Q[ tB(j-I) 1 (33) The
Analysis techniques for diagnosing runaway ion distributions in the reversed field pinch
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, J., E-mail: jkim536@wisc.edu; Anderson, J. K.; Capecchi, W.
2016-11-15
An advanced neutral particle analyzer (ANPA) on the Madison Symmetric Torus measures deuterium ions of energy ranges 8-45 keV with an energy resolution of 2-4 keV and time resolution of 10 μs. Three different experimental configurations measure distinct portions of the naturally occurring fast ion distributions: fast ions moving parallel, anti-parallel, or perpendicular to the plasma current. On a radial-facing port, fast ions moving perpendicular to the current have the necessary pitch to be measured by the ANPA. With the diagnostic positioned on a tangent line through the plasma core, a chord integration over fast ion density, background neutral density,more » and local appropriate pitch defines the measured sample. The plasma current can be reversed to measure anti-parallel fast ions in the same configuration. Comparisons of energy distributions for the three configurations show an anisotropic fast ion distribution favoring high pitch ions.« less
Probability distributions for multimeric systems.
Albert, Jaroslav; Rooman, Marianne
2016-01-01
We propose a fast and accurate method of obtaining the equilibrium mono-modal joint probability distributions for multimeric systems. The method necessitates only two assumptions: the copy number of all species of molecule may be treated as continuous; and, the probability density functions (pdf) are well-approximated by multivariate skew normal distributions (MSND). Starting from the master equation, we convert the problem into a set of equations for the statistical moments which are then expressed in terms of the parameters intrinsic to the MSND. Using an optimization package on Mathematica, we minimize a Euclidian distance function comprising of a sum of the squared difference between the left and the right hand sides of these equations. Comparison of results obtained via our method with those rendered by the Gillespie algorithm demonstrates our method to be highly accurate as well as efficient.
Reversing a tree regeneration crisis in an endangered ecoregion.
Fischer, Joern; Stott, Jenny; Zerger, Andre; Warren, Garth; Sherren, Kate; Forrester, Robert I
2009-06-23
Global food demand is growing rapidly. Livestock grazing can provide a valuable source of protein, but conventional grazing is often unsustainable. We studied an 800,000-ha section of a threatened ecoregion in southeastern Australia. Conventional management in the region involves continuous livestock grazing with few rest periods and regular fertilizer application. By using remotely sensed data on tree cover and extensive field data on livestock grazing regimes, soil chemistry, tree diameters, and tree regeneration, we show that the region is facing a tree regeneration crisis. Under conventional management, across the region, millions of hectares of land currently supporting tens of millions of trees will be treeless within decades from now. This would have severe negative ramifications for biodiversity and key ecosystem services, including water infiltration and shade provision for livestock. However, we identified an unexpected win-win solution for tree regeneration and commercial grazing. A relatively new practice in the region is fast-rotational grazing, characterized by prolonged rest periods in between short, intensive grazing events. The probability of regeneration under fast-rotational grazing was up to 4-fold higher than under conventional grazing, and it did not differ significantly from the probability of regeneration in ungrazed areas. In addition, trees were more likely to regenerate where soil nutrient levels were low. These findings suggest that the tree regeneration crisis can be reversed by applying low-input, fast-rotational grazing. New policy settings supporting these practices could signal a turning point for the region, from ecological decline to ecological recovery.
Reversing a tree regeneration crisis in an endangered ecoregion
Fischer, Joern; Stott, Jenny; Zerger, Andre; Warren, Garth; Sherren, Kate; Forrester, Robert I.
2009-01-01
Global food demand is growing rapidly. Livestock grazing can provide a valuable source of protein, but conventional grazing is often unsustainable. We studied an 800,000-ha section of a threatened ecoregion in southeastern Australia. Conventional management in the region involves continuous livestock grazing with few rest periods and regular fertilizer application. By using remotely sensed data on tree cover and extensive field data on livestock grazing regimes, soil chemistry, tree diameters, and tree regeneration, we show that the region is facing a tree regeneration crisis. Under conventional management, across the region, millions of hectares of land currently supporting tens of millions of trees will be treeless within decades from now. This would have severe negative ramifications for biodiversity and key ecosystem services, including water infiltration and shade provision for livestock. However, we identified an unexpected win–win solution for tree regeneration and commercial grazing. A relatively new practice in the region is fast-rotational grazing, characterized by prolonged rest periods in between short, intensive grazing events. The probability of regeneration under fast-rotational grazing was up to 4-fold higher than under conventional grazing, and it did not differ significantly from the probability of regeneration in ungrazed areas. In addition, trees were more likely to regenerate where soil nutrient levels were low. These findings suggest that the tree regeneration crisis can be reversed by applying low-input, fast-rotational grazing. New policy settings supporting these practices could signal a turning point for the region, from ecological decline to ecological recovery. PMID:19497886
Evaluation of Mean and Variance Integrals without Integration
ERIC Educational Resources Information Center
Joarder, A. H.; Omar, M. H.
2007-01-01
The mean and variance of some continuous distributions, in particular the exponentially decreasing probability distribution and the normal distribution, are considered. Since they involve integration by parts, many students do not feel comfortable. In this note, a technique is demonstrated for deriving mean and variance through differential…
Melissa's Year in Sixth Grade: A Technology Integration Vignette.
ERIC Educational Resources Information Center
Hemmer, Jeanie
1998-01-01
In 1995, rather than require seventh-grade computer literacy classes, Texas allowed school districts to integrate technology skills into curricula. This article, the first of three, describes technology integration for sixth grade. Includes unit ideas on nations; the Holocaust; Olympic diving; Christmas; probability; organisms; Antarctica;…
Robust object tacking based on self-adaptive search area
NASA Astrophysics Data System (ADS)
Dong, Taihang; Zhong, Sheng
2018-02-01
Discriminative correlation filter (DCF) based trackers have recently achieved excellent performance with great computational efficiency. However, DCF based trackers suffer boundary effects, which result in the unstable performance in challenging situations exhibiting fast motion. In this paper, we propose a novel method to mitigate this side-effect in DCF based trackers. We change the search area according to the prediction of target motion. When the object moves fast, broad search area could alleviate boundary effects and reserve the probability of locating object. When the object moves slowly, narrow search area could prevent effect of useless background information and improve computational efficiency to attain real-time performance. This strategy can impressively soothe boundary effects in situations exhibiting fast motion and motion blur, and it can be used in almost all DCF based trackers. The experiments on OTB benchmark show that the proposed framework improves the performance compared with the baseline trackers.
Photodissociation of ethylbenzene and n-propylbenzene in a molecular beam
NASA Astrophysics Data System (ADS)
Huang, Cheng-Liang; Jiang, Jyh-Chiang; Lee, Yuan T.; Ni, Chi-Kung
2002-10-01
The photodissociation of jet-cooled ethylbenzene and n-propylbenzene at both 193 and 248 nm was studied using vacuum ultraviolet photoionization/multimass ion imaging techniques. The photofragment translational energy distributions from both the molecules obtained at 193 nm show that the probability of portioning energy to product translational energy decreases monotonically with increasing translational energy. They indicate that the dissociation occurs from the ground electronic state. However, the photofragment translational energy distributions from both molecules obtained at 248 nm contain a fast and a slow component. 75% of ethylbenzene and 80% of n-propylbenzene following the 248 nm photoexcitation dissociate from electronic excited state, resulting in the fast component. The remaining 25% of ethylbenzene and 20% of n-propylbenzene dissociate through the ground electronic state, giving rise to the slow component. A comparison with an ab initio calculation suggests that the dissociation from the first triplet state corresponds to the fast component in translational energy distribution.
Perceptions and practices of fasting in Ramadan during pregnancy in Pakistan
Mubeen, Syed M.; Mansoor, Salman; Hussain, Asad; Qadir, Shayan
2012-01-01
Background: Islam clearly exempts fasting in Ramadan during sickness, pregnancy, and breastfeeding. Yet, some women prefer to fast despite realizing their increase in nutritional demand during pregnancy. Nurses and other healthcare providers face a difficult task of providing appropriate medical advice to Muslim women regarding the safety and impact of their fasting. The present study was conducted to examine the concepts and practices of maternal fasting during pregnancy in Pakistani Muslim women. Materials and Methods: A cross-sectional study was done in 2008 on 353 females from gynecology departments of hospitals in major cities of two provinces of Pakistan. Non-probability convenient sampling technique was used and a questionnaire was used to collect data from females who had experienced pregnancy during Ramadan at least once in life. Questions related to perceptions and practices of fasting in pregnancy along with demographic data were asked. Analysis was done on SPSS version 16. Chi-square test was used to assess associations and P < 0.05 was considered statistically significant. All ethical considerations were taken into account for the respondents. Results: Out of the total, 309 (87.5%) fasted during their pregnancies while 43 (12.5%) did not fast at all. Among those who fasted, 42.5% (n = 150) fasted for the whole month, 23.8% (n = 84) fasted on alternate days, 10.5% (n = 37) fasted only on weekends/holidays, and 10.8% (n = 38) fasted on particular special days. The majority perceived no harm and risk caused by fasting to them or to fetus during pregnancy. Although adverse effects during fasting were frequently reported, only 24% break their fast. A significant association was found between gravida and education with fasting (P w< 0.05). A total of 88% of women believed that fasting during pregnancy (in good health) is “obligatory” while 12% thought otherwise. Conclusion: Pakistani Muslim women showed a strong compassion for fasting in pregnancy despite overall decrease in maternal health indicators. There is a need for nurses and other healthcare providers in all Muslim countries and especially in non-Muslim countries with cultural diversity to be fully aware of the effects of fasting in pregnancy so that they can act promptly ensuring safe and healthy delivery. PMID:23922591
Jiang, Liqun; Wu, Nannan; Zheng, Anqing; Zhao, Zengli; He, Fang; Li, Haibin
2016-01-01
Fermentable sugars are important intermediates in the biological conversion of biomass. Hemicellulose and amorphous cellulose are easily hydrolyzed to fermentable sugars in dilute acid, whereas crystalline cellulose is more difficult to be hydrolyzed. Cellulose fast pyrolysis is an alternative method to liberate valuable fermentable sugars from biomass. The amount of levoglucosan generated from lignocellulose by fast pyrolysis is usually lower than the theoretical yield based on the cellulose fraction. Pretreatment is a promising route to improve the yield of levoglucosan from lignocellulose. The integration of dilute sulfuric acid hydrolysis and fast pyrolysis to obtain fermentable sugars was evaluated in this study. Dilute sulfuric acid hydrolysis could remove more than 95.1 and 93.4 % of xylan (the main component of hemicellulose) from sugarcane bagasse and corncob with high yield of xylose. On the other hand, dilute sulfuric acid hydrolysis was also an effective pretreatment to enhance levoglucosan yield from lignocellulose. Dilute acid hydrolysis could accumulate glucan (the component of cellulose) and remove most of the alkali and alkaline earth metals which were powerful catalysts during fast pyrolysis. Further increase in dilute acid concentration (from 0 to 2 %) in pretreatment could promote the yield of levoglucosan in fast pyrolysis. The acid pretreated sugarcane bagasse and corncob gave levoglucosan yields of 43.8 and 35.2 % which were obvious higher than those of raw sugarcane bagasse (12.0 %) and corncob (7.0 %). Obtaining fermentable sugars by combination dilute acid hydrolysis of xylan and fast pyrolysis of glucan could make full utilization of biomass, and get fermentable sugars economically from biomass for bio-refinery.
Optimal temperature ladders in replica exchange simulations
NASA Astrophysics Data System (ADS)
Denschlag, Robert; Lingenheil, Martin; Tavan, Paul
2009-04-01
In replica exchange simulations, a temperature ladder with N rungs spans a given temperature interval. Considering systems with heat capacities independent of the temperature, here we address the question of how large N should be chosen for an optimally fast diffusion of the replicas through the temperature space. Using a simple example we show that choosing average acceptance probabilities of about 45% and computing N accordingly maximizes the round trip rates r across the given temperature range. This result differs from previous analyses which suggested smaller average acceptance probabilities of about 23%. We show that the latter choice maximizes the ratio r/N instead of r.
The Secondary Development of ABAQUS by using Python and the Application of the Advanced GA
NASA Astrophysics Data System (ADS)
Luo, Lilong; Zhao, Meiying
Realizing the secondary development of the ABAQUS based on the manual of ABAQUS. In order to overcome the prematurity and the worse convergence of the Simple Genetic Algorithm (SGA), a new strategy how to improve the efficiency of the SGA has been put forward. In the new GA, the selection probability and the mutation probability are self-adaptive. Taking the stability of the composite laminates as the target, the optimized laminates sequences and radius of the hatch are analyzed with the help of ABAQUS. Compared with the SGA, the new GA method shows a good consistency, fast convergence and practical feasibility.
Microwaves in chemistry: Another way of heating reaction mixtures
NASA Astrophysics Data System (ADS)
Berlan, J.
1995-04-01
The question of a possible "microwave activation" of chemical reaction is discussed. In fact two cases should be distinguished: homogeneous or heterogeneous reaction mixtures. In homogeneous mixtures there are no (or very low) rate enhancements compared to a conventional heating, but some influence on chemioselectivity has been observed. These effects derive from fast and mass heating of microwaves, and probably, especially under reflux, from different boiling rates and/or overheating. With heterogeneous mixtures non conventional effects probably derive from mass heating and selective overheating. This is illustrated with several reactions: Diels-Alder, naphthalene sulphonation, preparation of cyanuric acid, hydrolysis of nitriles, transposition reaction on solid support.
QUEST - A Bayesian adaptive psychometric method
NASA Technical Reports Server (NTRS)
Watson, A. B.; Pelli, D. G.
1983-01-01
An adaptive psychometric procedure that places each trial at the current most probable Bayesian estimate of threshold is described. The procedure takes advantage of the common finding that the human psychometric function is invariant in form when expressed as a function of log intensity. The procedure is simple, fast, and efficient, and may be easily implemented on any computer.
The Shape of Things to Come: A Photo Essay
ERIC Educational Resources Information Center
Connection: The Journal of the New England Board of Higher Education, 2002
2002-01-01
Think of New England higher education and one probably thinks of ivy-covered brick or perhaps 1950s concrete, but the environment of New England's college campuses is far more diverse, more complex, more daring, and more educational than the stereotype. It is changing fast, thanks to profound forces such as consumerization among students,…
Gaming the Law of Large Numbers
ERIC Educational Resources Information Center
Hoffman, Thomas R.; Snapp, Bart
2012-01-01
Many view mathematics as a rich and wonderfully elaborate game. In turn, games can be used to illustrate mathematical ideas. Fibber's Dice, an adaptation of the game Liar's Dice, is a fast-paced game that rewards gutsy moves and favors the underdog. It also brings to life concepts arising in the study of probability. In particular, Fibber's Dice…
Origin of the enhancement of tunneling probability in the nearly integrable system
NASA Astrophysics Data System (ADS)
Hanada, Yasutaka; Shudo, Akira; Ikeda, Kensuke S.
2015-04-01
The enhancement of tunneling probability in the nearly integrable system is closely examined, focusing on tunneling splittings plotted as a function of the inverse of the Planck's constant. On the basis of the analysis using the absorber which efficiently suppresses the coupling, creating spikes in the plot, we found that the splitting curve should be viewed as the staircase-shaped skeleton accompanied by spikes. We further introduce renormalized integrable Hamiltonians and explore the origin of such a staircase structure by investigating the nature of eigenfunctions closely. It is found that the origin of the staircase structure could trace back to the anomalous structure in tunneling tail which manifests itself in the representation using renormalized action bases. This also explains the reason why the staircase does not appear in the completely integrable system.
Fast transform decoding of nonsystematic Reed-Solomon codes
NASA Technical Reports Server (NTRS)
Truong, T. K.; Cheung, K.-M.; Reed, I. S.; Shiozaki, A.
1989-01-01
A Reed-Solomon (RS) code is considered to be a special case of a redundant residue polynomial (RRP) code, and a fast transform decoding algorithm to correct both errors and erasures is presented. This decoding scheme is an improvement of the decoding algorithm for the RRP code suggested by Shiozaki and Nishida, and can be realized readily on very large scale integration chips.
Diabetes mellitus and Ramadan in Algeria
Chentli, Farida; Azzoug, Said; Amani, Mohammed El Amine; Elgradechi, Aldjia
2013-01-01
Ramadan, one of the five pillars of Islam, is a holy month in Algeria where diabetes mellitus (DM) is more frequent in urban areas with a frequency which varies from 8 to 16%. DM complications are broadly as frequent as in developed countries, except for neuropathy which seems more frequent. Despite contraindications which are regularly explained to our patients and despite the flexible side of Islam toward chronic diseases, most Algerian people with DM insist on fasting. Not fasting is considered a sin and shameful. There are also other reasons put forward by diabetic persons, such as very strong religious faith, habit of fasting together with the whole family since an early age, solidarity with the family, friends, and neighbors, and finally and probably because of the desire to appear “normal” and share a festive and a spiritual atmosphere of Ramadan. As in other Muslim countries, severe hypoglycemia the main motive of hospitalizations during the holy month, ketoacidosis, dehydration, orthostatic hypotension and thrombosis are some of the complications which Algerian people with DM are exposed to when fasting. PMID:24251192
NASA Technical Reports Server (NTRS)
Mavris, Dimitri N.
1998-01-01
Over the past few years, modem aircraft design has experienced a paradigm shift from designing for performance to designing for affordability. This report contains a probabilistic approach that will allow traditional deterministic design methods to be extended to account for disciplinary, economic, and technological uncertainty. The probabilistic approach was facilitated by the Fast Probability Integration (FPI) technique; a technique which allows the designer to gather valuable information about the vehicle's behavior in the design space. This technique is efficient for assessing multi-attribute, multi-constraint problems in a more realistic fashion. For implementation purposes, this technique is applied to illustrate how both economic and technological uncertainty associated with a Very Large Transport aircraft concept may be assessed. The assessment is evaluated with the FPI technique to determine the cumulative probability distributions of the design space, as bound by economic objectives and performance constraints. These distributions were compared to established targets for a comparable large capacity aircraft, similar in size to the Boeing 747-400. The conventional baseline configuration design space was determined to be unfeasible and marginally viable, motivating the infusion of advanced technologies, including reductions in drag, specific fuel consumption, wing weight, and Research, Development, Testing, and Evaluation costs. The resulting system design space was qualitatively assessed with technology metric "k" factors. The infusion of technologies shifted the VLT design into regions of feasibility and greater viability. The study also demonstrated a method and relationship by which the impact of new technologies may be assessed in a more system focused approach.
Kasiri, Keyvan; Kazemi, Kamran; Dehghani, Mohammad Javad; Helfroush, Mohammad Sadegh
2013-01-01
In this paper, we present a new semi-automatic brain tissue segmentation method based on a hybrid hierarchical approach that combines a brain atlas as a priori information and a least-square support vector machine (LS-SVM). The method consists of three steps. In the first two steps, the skull is removed and the cerebrospinal fluid (CSF) is extracted. These two steps are performed using the toolbox FMRIB's automated segmentation tool integrated in the FSL software (FSL-FAST) developed in Oxford Centre for functional MRI of the brain (FMRIB). Then, in the third step, the LS-SVM is used to segment grey matter (GM) and white matter (WM). The training samples for LS-SVM are selected from the registered brain atlas. The voxel intensities and spatial positions are selected as the two feature groups for training and test. SVM as a powerful discriminator is able to handle nonlinear classification problems; however, it cannot provide posterior probability. Thus, we use a sigmoid function to map the SVM output into probabilities. The proposed method is used to segment CSF, GM and WM from the simulated magnetic resonance imaging (MRI) using Brainweb MRI simulator and real data provided by Internet Brain Segmentation Repository. The semi-automatically segmented brain tissues were evaluated by comparing to the corresponding ground truth. The Dice and Jaccard similarity coefficients, sensitivity and specificity were calculated for the quantitative validation of the results. The quantitative results show that the proposed method segments brain tissues accurately with respect to corresponding ground truth. PMID:24696800
A tool for selecting SNPs for association studies based on observed linkage disequilibrium patterns.
De La Vega, Francisco M; Isaac, Hadar I; Scafe, Charles R
2006-01-01
The design of genetic association studies using single-nucleotide polymorphisms (SNPs) requires the selection of subsets of the variants providing high statistical power at a reasonable cost. SNPs must be selected to maximize the probability that a causative mutation is in linkage disequilibrium (LD) with at least one marker genotyped in the study. The HapMap project performed a genome-wide survey of genetic variation with about a million SNPs typed in four populations, providing a rich resource to inform the design of association studies. A number of strategies have been proposed for the selection of SNPs based on observed LD, including construction of metric LD maps and the selection of haplotype tagging SNPs. Power calculations are important at the study design stage to ensure successful results. Integrating these methods and annotations can be challenging: the algorithms required to implement these methods are complex to deploy, and all the necessary data and annotations are deposited in disparate databases. Here, we present the SNPbrowser Software, a freely available tool to assist in the LD-based selection of markers for association studies. This stand-alone application provides fast query capabilities and swift visualization of SNPs, gene annotations, power, haplotype blocks, and LD map coordinates. Wizards implement several common SNP selection workflows including the selection of optimal subsets of SNPs (e.g. tagging SNPs). Selected SNPs are screened for their conversion potential to either TaqMan SNP Genotyping Assays or the SNPlex Genotyping System, two commercially available genotyping platforms, expediting the set-up of genetic studies with an increased probability of success.
Reliability and Creep/Fatigue Analysis of a CMC Component
NASA Technical Reports Server (NTRS)
Murthy, Pappu L. N.; Mital, Subodh K.; Gyekenyesi, John Z.; Gyekenyesi, John P.
2007-01-01
High temperature ceramic matrix composites (CMC) are being explored as viable candidate materials for hot section gas turbine components. These advanced composites can potentially lead to reduced weight and enable higher operating temperatures requiring less cooling; thus leading to increased engine efficiencies. There is a need for convenient design tools that can accommodate various loading conditions and material data with their associated uncertainties to estimate the minimum predicted life as well as the failure probabilities of a structural component. This paper presents a review of the life prediction and probabilistic analyses performed for a CMC turbine stator vane. A computer code, NASALife, is used to predict the life of a 2-D woven silicon carbide fiber reinforced silicon carbide matrix (SiC/SiC) turbine stator vane due to a mission cycle which induces low cycle fatigue and creep. The output from this program includes damage from creep loading, damage due to cyclic loading and the combined damage due to the given loading cycle. Results indicate that the trends predicted by NASALife are as expected for the loading conditions used for this study. In addition, a combination of woven composite micromechanics, finite element structural analysis and Fast Probability Integration (FPI) techniques has been used to evaluate the maximum stress and its probabilistic distribution in a CMC turbine stator vane. Input variables causing scatter are identified and ranked based upon their sensitivity magnitude. Results indicate that reducing the scatter in proportional limit strength of the vane material has the greatest effect in improving the overall reliability of the CMC vane.
Choudhary, P; Davies, C; Emery, C J; Heller, S R
2013-08-01
The Somogyi effect postulates that nocturnal hypoglycaemia causes fasting hyperglycaemia attributable to counter-regulatory hormone release. Although most published evidence has failed to support this hypothesis, this concept remains firmly embedded in clinical practice and often prevents patients and professionals from optimizing overnight insulin. Previous observational data found lower fasting glucose was associated with nocturnal hypoglycaemia, but did not assess the probability of infrequent individual episodes of rebound hypoglycaemia. We analysed continuous glucose monitoring data to explore its prevalence. We analysed data from 89 patients with Type 1 diabetes who participated in the UK Hypoglycaemia study. We compared fasting capillary glucose following nights with and without nocturnal hypoglycaemia (sensor glucose < 3.5 mmol/l). Fasting capillary blood glucose was lower after nights with hypoglycaemia than without [5.5 (3.0) vs. 14.5 (4.5) mmol/l, P < 0.0001], and was lower on nights with more severe nocturnal hypoglycaemia [5.5 (3.0) vs. 8.2 (2.3) mmol/l; P = 0.018 on nights with nadir sensor glucose of < 2.2 mmol/l vs. 3.5 mmol/l]. There were only two instances of fasting capillary blood glucose > 10 mmol/l after nocturnal hypoglycaemia, both after likely treatment of the episode. When fasting capillary blood glucose is < 5 mmol/l, there was evidence of nocturnal hypoglycaemia on 94% of nights. Our data indicate that, in clinical practice, the Somogyi effect is rare. Fasting capillary blood glucose ≤ 5 mmol/l appears an important indicator of preceding silent nocturnal hypoglycaemia. © 2013 The Authors. Diabetic Medicine © 2013 Diabetes UK.
Mölle, Matthias; Bergmann, Til O; Marshall, Lisa; Born, Jan
2011-10-01
Thalamo-cortical spindles driven by the up-state of neocortical slow (< 1 Hz) oscillations (SOs) represent a candidate mechanism of memory consolidation during sleep. We examined interactions between SOs and spindles in human slow wave sleep, focusing on the presumed existence of 2 kinds of spindles, i.e., slow frontocortical and fast centro-parietal spindles. Two experiments were performed in healthy humans (24.5 ± 0.9 y) investigating undisturbed sleep (Experiment I) and the effects of prior learning (word paired associates) vs. non-learning (Experiment II) on multichannel EEG recordings during sleep. Only fast spindles (12-15 Hz) were synchronized to the depolarizing SO up-state. Slow spindles (9-12 Hz) occurred preferentially at the transition into the SO down-state, i.e., during waning depolarization. Slow spindles also revealed a higher probability to follow rather than precede fast spindles. For sequences of individual SOs, fast spindle activity was largest for "initial" SOs, whereas SO amplitude and slow spindle activity were largest for succeeding SOs. Prior learning enhanced this pattern. The finding that fast and slow spindles occur at different times of the SO cycle points to disparate generating mechanisms for the 2 kinds of spindles. The reported temporal relationships during SO sequences suggest that fast spindles, driven by the SO up-state feed back to enhance the likelihood of succeeding SOs together with slow spindles. By enforcing such SO-spindle cycles, particularly after prior learning, fast spindles possibly play a key role in sleep-dependent memory processing.
Singh, Rabindarjeet; Hwa, Ooi Cheong; Roy, Jolly; Jin, Chai Wen; Ismail, Siti Musyrifah; Lan, Mohamad Faizal; Hiong, Loo Lean; Aziz, Abdul-Rashid
2011-01-01
Purpose To examine the subjective perception of daily acute fasting on sports performance, training, sleep and dietary patterns of Muslim athletes during the Ramadan month. Methods Seven hundred and thirty-four (411 male and 323 female) Malaysian Junior-level Muslim athletes (mean age 16.3 ± 2.6 y) participated in the survey which was designed to establish the personal perception of their sport performance, sleep pattern, food and fluid intake during Ramadan fasting. The survey was conducted during and immediately after the month of Ramadan in 2009. Results Twenty-four percent of the athletes perceived that there was an adverse effect of the Ramadan fast on their sporting performance and 29.3% reported that quality of training during Ramadan was also negatively influenced. Majority (48.2%) of the athletes stated that Ramadan fasting did not affect their normal sleep pattern but 66.6% of them complained of sleepiness during the daytime. Half of the athletes (41.4%) maintained the caloric intake during Ramadan as they normally would with the majority of them (76.2%) reporting that they consumed more fluids during Ramadan. Conclusions Overall, Malaysian Junior-level Muslim athletes showed diverse views in their perception of changes in their training, sleep and dietary patterns during Ramadan fast. These individual differences probably indicate differences in the athletes’ adaptability and coping strategies during fasting and training in Ramadan. PMID:22375236
Probability Elicitation Under Severe Time Pressure: A Rank-Based Method.
Jaspersen, Johannes G; Montibeller, Gilberto
2015-07-01
Probability elicitation protocols are used to assess and incorporate subjective probabilities in risk and decision analysis. While most of these protocols use methods that have focused on the precision of the elicited probabilities, the speed of the elicitation process has often been neglected. However, speed is also important, particularly when experts need to examine a large number of events on a recurrent basis. Furthermore, most existing elicitation methods are numerical in nature, but there are various reasons why an expert would refuse to give such precise ratio-scale estimates, even if highly numerate. This may occur, for instance, when there is lack of sufficient hard evidence, when assessing very uncertain events (such as emergent threats), or when dealing with politicized topics (such as terrorism or disease outbreaks). In this article, we adopt an ordinal ranking approach from multicriteria decision analysis to provide a fast and nonnumerical probability elicitation process. Probabilities are subsequently approximated from the ranking by an algorithm based on the principle of maximum entropy, a rule compatible with the ordinal information provided by the expert. The method can elicit probabilities for a wide range of different event types, including new ways of eliciting probabilities for stochastically independent events and low-probability events. We use a Monte Carlo simulation to test the accuracy of the approximated probabilities and try the method in practice, applying it to a real-world risk analysis recently conducted for DEFRA (the U.K. Department for the Environment, Farming and Rural Affairs): the prioritization of animal health threats. © 2015 Society for Risk Analysis.
Mölle, Matthias; Bergmann, Til O.; Marshall, Lisa; Born, Jan
2011-01-01
Study Objectives: Thalamo-cortical spindles driven by the up-state of neocortical slow (< 1 Hz) oscillations (SOs) represent a candidate mechanism of memory consolidation during sleep. We examined interactions between SOs and spindles in human slow wave sleep, focusing on the presumed existence of 2 kinds of spindles, i.e., slow frontocortical and fast centro-parietal spindles. Design: Two experiments were performed in healthy humans (24.5 ± 0.9 y) investigating undisturbed sleep (Experiment I) and the effects of prior learning (word paired associates) vs. non-learning (Experiment II) on multichannel EEG recordings during sleep. Measurements and Results: Only fast spindles (12-15 Hz) were synchronized to the depolarizing SO up-state. Slow spindles (9-12 Hz) occurred preferentially at the transition into the SO down-state, i.e., during waning depolarization. Slow spindles also revealed a higher probability to follow rather than precede fast spindles. For sequences of individual SOs, fast spindle activity was largest for “initial” SOs, whereas SO amplitude and slow spindle activity were largest for succeeding SOs. Prior learning enhanced this pattern. Conclusions: The finding that fast and slow spindles occur at different times of the SO cycle points to disparate generating mechanisms for the 2 kinds of spindles. The reported temporal relationships during SO sequences suggest that fast spindles, driven by the SO up-state feed back to enhance the likelihood of succeeding SOs together with slow spindles. By enforcing such SO-spindle cycles, particularly after prior learning, fast spindles possibly play a key role in sleep-dependent memory processing. Citation: Mölle M; Bergmann TO; Marshall L; Born J. Fast and slow spindles during the sleep slow oscillation: disparate coalescence and engagement in memory processing. SLEEP 2011;34(10):1411–1421. PMID:21966073
Wu, Ying; Zou, Meng; Raulerson, Chelsea K.; Jackson, Kayla; Yuan, Wentao; Wang, Haifeng; Shou, Weihua; Wang, Ying; Luo, Jingchun; Lange, Leslie A.; Lange, Ethan M.; Gordon-Larsen, Penny; Du, Shufa; Huang, Wei; Mohlke, Karen L.
2018-01-01
To identify genetic contributions to type 2 diabetes (T2D) and related glycemic traits (fasting glucose, fasting insulin, and HbA1c), we conducted genome-wide association analyses (GWAS) in up to 7,178 Chinese subjects from nine provinces in the China Health and Nutrition Survey (CHNS). We examined patterns of population structure within CHNS and found that allele frequencies differed across provinces, consistent with genetic drift and population substructure. We further validated 32 previously described T2D- and glycemic trait-loci, including G6PC2 and SIX3-SIX2 associated with fasting glucose. At G6PC2, we replicated a known fasting glucose-associated variant (rs34177044) and identified a second signal (rs2232326), a low-frequency (4%), probably damaging missense variant (S324P). A variant within the lead fasting glucose-associated signal at SIX3-SIX2 co-localized with pancreatic islet expression quantitative trait loci (eQTL) for SIX3, SIX2, and three noncoding transcripts. To identify variants functionally responsible for the fasting glucose association at SIX3-SIX2, we tested five candidate variants for allelic differences in regulatory function. The rs12712928-C allele, associated with higher fasting glucose and lower transcript expression level, showed lower transcriptional activity in reporter assays and increased binding to GABP compared to the rs12712928-G, suggesting that rs12712928-C contributes to elevated fasting glucose levels by disrupting an islet enhancer, resulting in reduced gene expression. Taken together, these analyses identified multiple loci associated with glycemic traits across China, and suggest a regulatory mechanism at the SIX3-SIX2 fasting glucose GWAS locus. PMID:29621232
Zhu, Lin; Dai, Zhenxue; Gong, Huili; ...
2015-06-12
Understanding the heterogeneity arising from the complex architecture of sedimentary sequences in alluvial fans is challenging. This study develops a statistical inverse framework in a multi-zone transition probability approach for characterizing the heterogeneity in alluvial fans. An analytical solution of the transition probability matrix is used to define the statistical relationships among different hydrofacies and their mean lengths, integral scales, and volumetric proportions. A statistical inversion is conducted to identify the multi-zone transition probability models and estimate the optimal statistical parameters using the modified Gauss–Newton–Levenberg–Marquardt method. The Jacobian matrix is computed by the sensitivity equation method, which results in anmore » accurate inverse solution with quantification of parameter uncertainty. We use the Chaobai River alluvial fan in the Beijing Plain, China, as an example for elucidating the methodology of alluvial fan characterization. The alluvial fan is divided into three sediment zones. In each zone, the explicit mathematical formulations of the transition probability models are constructed with optimized different integral scales and volumetric proportions. The hydrofacies distributions in the three zones are simulated sequentially by the multi-zone transition probability-based indicator simulations. Finally, the result of this study provides the heterogeneous structure of the alluvial fan for further study of flow and transport simulations.« less
Robust approaches to quantification of margin and uncertainty for sparse data
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hund, Lauren; Schroeder, Benjamin B.; Rumsey, Kelin
Characterizing the tails of probability distributions plays a key role in quantification of margins and uncertainties (QMU), where the goal is characterization of low probability, high consequence events based on continuous measures of performance. When data are collected using physical experimentation, probability distributions are typically fit using statistical methods based on the collected data, and these parametric distributional assumptions are often used to extrapolate about the extreme tail behavior of the underlying probability distribution. In this project, we character- ize the risk associated with such tail extrapolation. Specifically, we conducted a scaling study to demonstrate the large magnitude of themore » risk; then, we developed new methods for communicat- ing risk associated with tail extrapolation from unvalidated statistical models; lastly, we proposed a Bayesian data-integration framework to mitigate tail extrapolation risk through integrating ad- ditional information. We conclude that decision-making using QMU is a complex process that cannot be achieved using statistical analyses alone.« less
NASA Astrophysics Data System (ADS)
Gilani, Seyed-Omid; Sattarvand, Javad
2016-02-01
Meeting production targets in terms of ore quantity and quality is critical for a successful mining operation. In-situ grade uncertainty causes both deviations from production targets and general financial deficits. A new stochastic optimization algorithm based on ant colony optimization (ACO) approach is developed herein to integrate geological uncertainty described through a series of the simulated ore bodies. Two different strategies were developed based on a single predefined probability value (Prob) and multiple probability values (Pro bnt) , respectively in order to improve the initial solutions that created by deterministic ACO procedure. Application at the Sungun copper mine in the northwest of Iran demonstrate the abilities of the stochastic approach to create a single schedule and control the risk of deviating from production targets over time and also increase the project value. A comparison between two strategies and traditional approach illustrates that the multiple probability strategy is able to produce better schedules, however, the single predefined probability is more practical in projects requiring of high flexibility degree.
Sullivan, David; Csicsvari, Jozsef; Mizuseki, Kenji; Montgomery, Sean; Diba, Kamran; Buzsáki, György
2011-01-01
Summary Hippocampal sharp waves (SPW) and associated fast (‘ripple’) oscillations in the CA1 region are among the most synchronous physiological patterns in the mammalian brain. Using two-dimensional arrays of electrodes for recording local field potentials and unit discharges in freely moving rats, we studied the emergence of ripple oscillations (140–220 Hz) and compared their origin and cellular-synaptic mechanisms with fast gamma oscillations (90–140 Hz). We show that (a) hippocampal SPW-Rs and fast gamma oscillations are quantitatively distinct patterns but involve the same networks and share similar mechanisms, (b) both the frequency and magnitude of fast oscillations is positively correlated with the magnitude of SPWs, (c) during both ripples and fast gamma oscillations the frequency of network oscillation is higher in CA1 than in CA3, (d) SPWs and associated firing of neurons are synchronous in the dorsal hippocampus and dorso-medial entorhinal cortex but ripples are confined to the CA1 pyramidal layer and its downstream targets and (e) the emergence of CA3 population bursts, a prerequisite for SPW-ripples, is biased by activity patterns in the dentate gyrus and entorhinal cortex, with highest probability of ripples associated with an ‘optimum’ level of dentate gamma power. We hypothesize that each hippocampal subnetwork possesses distinct resonant properties, tuned by the magnitude of the excitatory drive. PMID:21653864
An Integrated Nonlinear Analysis library - (INA) for solar system plasma turbulence
NASA Astrophysics Data System (ADS)
Munteanu, Costel; Kovacs, Peter; Echim, Marius; Koppan, Andras
2014-05-01
We present an integrated software library dedicated to the analysis of time series recorded in space and adapted to investigate turbulence, intermittency and multifractals. The library is written in MATLAB and provides a graphical user interface (GUI) customized for the analysis of space physics data available online like: Coordinated Data Analysis Web (CDAWeb), Automated Multi Dataset Analysis system (AMDA), Planetary Science Archive (PSA), World Data Center Kyoto (WDC), Ulysses Final Archive (UFA) and Cluster Active Archive (CAA). Three main modules are already implemented in INA : the Power Spectral Density (PSD) Analysis, the Wavelet and Intemittency Analysis and the Probability Density Functions (PDF) analysis.The layered structure of the software allows the user to easily switch between different modules/methods while retaining the same time interval for the analysis. The wavelet analysis module includes algorithms to compute and analyse the PSD, the Scalogram, the Local Intermittency Measure (LIM) or the Flatness parameter. The PDF analysis module includes algorithms for computing the PDFs for a range of scales and parameters fully customizable by the user; it also computes the Flatness parameter and enables fast comparison with standard PDF profiles like, for instance, the Gaussian PDF. The library has been already tested on Cluster and Venus Express data and we will show relevant examples. Research supported by the European Community's Seventh Framework Programme (FP7/2007-2013) under grant agreement no 313038/STORM, and a grant of the Romanian Ministry of National Education, CNCS UEFISCDI, project number PN-II-ID PCE-2012-4-0418.
Plasmonic phased array feeder enabling ultra-fast beam steering at millimeter waves.
Bonjour, R; Burla, M; Abrecht, F C; Welschen, S; Hoessbacher, C; Heni, W; Gebrewold, S A; Baeuerle, B; Josten, A; Salamin, Y; Haffner, C; Johnston, P V; Elder, D L; Leuchtmann, P; Hillerkuss, D; Fedoryshyn, Y; Dalton, L R; Hafner, C; Leuthold, J
2016-10-31
In this paper, we demonstrate an integrated microwave phoneeded for beamtonics phased array antenna feeder at 60 GHz with a record-low footprint. Our design is based on ultra-compact plasmonic phase modulators (active area <2.5µm2) that not only provide small size but also ultra-fast tuning speed. In our design, the integrated circuit footprint is in fact only limited by the contact pads of the electrodes and by the optical feeding waveguides. Using the high speed of the plasmonic modulators, we demonstrate beam steering with less than 1 ns reconfiguration time, i.e. the beam direction is reconfigured in-between 1 GBd transmitted symbols.
On the studies of thermodynamics properties of fast neutron irradiated (LixK1-x)2SO4 crystals
NASA Astrophysics Data System (ADS)
El-Khatib, A. M.; Kassem, M. E.; Gomaa, N. G.; Mahmoud, S. A.
The effect of fast neutron irradiation on the thermodynamic properties of (LixK1-x)2SO4, (x = 0.1, 0.2,˙˙˙˙˙˙˙˙0.5) has been studied. The measurements were carried out in the vicinity of phase transition. The study reveals that as the lithium content decreases the first high temperature phase Tc = 705 K disappears, while the second one is shifted to lower temperature. It is observed also that the specific heat, Cp, decreases sharply with neutron integrated fluence φ and increases once more. Both entropy and enthalpy changes increase with the increase of neutron integrated fluence.
NASA Astrophysics Data System (ADS)
Bays, Samuel Eugene
2008-10-01
In the past several years there has been a renewed interest in sodium fast reactor (SFR) technology for the purpose of destroying transuranic waste (TRU) produced by light water reactors (LWR). The utility of SFRs as waste burners is due to the fact that higher neutron energies allow all of the actinides, including the minor actinides (MA), to contribute to fission. It is well understood that many of the design issues of LWR spent nuclear fuel (SNF) disposal in a geologic repository are linked to MAs. Because the probability of fission for essentially all the "non-fissile" MAs is nearly zero at low neutron energies, these isotopes act as a neutron capture sink in most thermal reactor systems. Furthermore, because most of the isotopes produced by these capture reactions are also non-fissile, they too are neutron sinks in most thermal reactor systems. Conversely, with high neutron energies, the MAs can produce neutrons by fast fission. Additionally, capture reactions transmute the MAs into mostly plutonium isotopes, which can fission more readily at any energy. The transmutation of non-fissile into fissile atoms is the premise of the plutonium breeder reactor. In a breeder reactor, not only does the non-fissile "fertile" U-238 atom contribute fast fission neutrons, but also transmutes into fissile Pu-239. The fissile value of the plutonium produced by MA transmutation can only be realized in fast neutron spectra. This is due to the fact that the predominate isotope produced by MA transmutation, Pu-238, is itself not fissile. However, the Pu-238 fission cross section is significantly larger than the original transmutation parent, predominately: Np-237 and Am-241, in the fast energy range. Also, Pu-238's fission cross section and fission-to-capture ratio is almost as high as that of fissile Pu-239 in the fast neutron spectrum. It is also important to note that a neutron absorption in Pu-238, that does not cause fission, will instead produce fissile Pu-239. Given this fast fissile quality and also the fact that Pu-238 is transmuted from Np-237 and Am-241, these MAs are regarded as fertile material in the SFR design proposed by this dissertation. This dissertation demonstrates a SFR design which is dedicated to plutonium breeding by targeting Am-241 transmutation. This SFR design uses a moderated axial transmutation target that functions primarily as a pseudo-blanket fuel, which is reprocessed with the active driver fuel in an integrated recycling strategy. This work demonstrates the cost and feasibility advantages of plutonium breeding via MA transmutation by adopting reactor, reprocessing and fuel technologies previously demonstrated for traditional breeder reactors. The fuel cycle proposed seeks to find a harmony between the waste management advantages of transuranic burning SFRs and the resource sustainability of traditional plutonium breeder SFRs. As a result, the enhanced plutonium conversion from MAs decreases the burner SFR's fuel costs, by extracting more fissile value from the initial TRU purchased through SNF reprocessing.
Exact transition probabilities in a 6-state Landau–Zener system with path interference
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sinitsyn, Nikolai A.
2015-04-23
In this paper, we identify a nontrivial multistate Landau–Zener (LZ) model for which transition probabilities between any pair of diabatic states can be determined analytically and exactly. In the semiclassical picture, this model features the possibility of interference of different trajectories that connect the same initial and final states. Hence, transition probabilities are generally not described by the incoherent successive application of the LZ formula. Finally, we discuss reasons for integrability of this system and provide numerical tests of the suggested expression for the transition probability matrix.
Fast camera observations of injected and intrinsic dust in TEXTOR
NASA Astrophysics Data System (ADS)
Shalpegin, A.; Vignitchouk, L.; Erofeev, I.; Brochard, F.; Litnovsky, A.; Bozhenkov, S.; Bykov, I.; den Harder, N.; Sergienko, G.
2015-12-01
Stereoscopic fast camera observations of pre-characterized carbon and tungsten dust injection in TEXTOR are reported, along with the modelling of tungsten particle trajectories with MIGRAINe. Particle tracking analysis of the video data showed significant differences in dust dynamics: while carbon flakes were prone to agglomeration and explosive destruction, spherical tungsten particles followed quasi-inertial trajectories. Although this inertial nature prevented any validation of the force models used in MIGRAINe, comparisons between the experimental and simulated lifetimes provide a direct evidence of dust temperature overestimation in dust dynamics codes. Furthermore, wide-view observations of the TEXTOR interior revealed the main production mechanism of intrinsic carbon dust, as well as the location of probable dust remobilization sites.
Prospect evaluation as a function of numeracy and probability denominator.
Millroth, Philip; Juslin, Peter
2015-05-01
This study examines how numeracy and probability denominator (a direct-ratio probability, a relative frequency with denominator 100, a relative frequency with denominator 10,000) affect the evaluation of prospects in an expected-value based pricing task. We expected that numeracy would affect the results due to differences in the linearity of number perception and the susceptibility to denominator neglect with different probability formats. An analysis with functional measurement verified that participants integrated value and probability into an expected value. However, a significant interaction between numeracy and probability format and subsequent analyses of the parameters of cumulative prospect theory showed that the manipulation of probability denominator changed participants' psychophysical response to probability and value. Standard methods in decision research may thus confound people's genuine risk attitude with their numerical capacities and the probability format used. Copyright © 2015 Elsevier B.V. All rights reserved.
Edwards, James P; Gerber, Urs; Schubert, Christian; Trejo, Maria Anabel; Weber, Axel
2018-04-01
We introduce two integral transforms of the quantum mechanical transition kernel that represent physical information about the path integral. These transforms can be interpreted as probability distributions on particle trajectories measuring respectively the relative contribution to the path integral from paths crossing a given spatial point (the hit function) and the likelihood of values of the line integral of the potential along a path in the ensemble (the path-averaged potential).
NASA Astrophysics Data System (ADS)
Edwards, James P.; Gerber, Urs; Schubert, Christian; Trejo, Maria Anabel; Weber, Axel
2018-04-01
We introduce two integral transforms of the quantum mechanical transition kernel that represent physical information about the path integral. These transforms can be interpreted as probability distributions on particle trajectories measuring respectively the relative contribution to the path integral from paths crossing a given spatial point (the hit function) and the likelihood of values of the line integral of the potential along a path in the ensemble (the path-averaged potential).
MM&T for VHSIC Multichip Packages
1989-09-20
broader the network bandwidth must be in order to preserve the integrity of the rising and falling edges of the signal. However, for very fast ...inch of path length adds approximately 100-300 ps of delay to the signal. System designers of fast ECL type devices must incorporate these delays into... intermittently in low temperature areas. Permanent changes in operating characteristics and physical damage produced during temperature shock occur
Analysis of blocking probability for OFDM-based variable bandwidth optical network
NASA Astrophysics Data System (ADS)
Gong, Lei; Zhang, Jie; Zhao, Yongli; Lin, Xuefeng; Wu, Yuyao; Gu, Wanyi
2011-12-01
Orthogonal Frequency Division Multiplexing (OFDM) has recently been proposed as a modulation technique. For optical networks, because of its good spectral efficiency, flexibility, and tolerance to impairments, optical OFDM is much more flexible compared to traditional WDM systems, enabling elastic bandwidth transmissions, and optical networking is the future trend of development. In OFDM-based optical network the research of blocking rate has very important significance for network assessment. Current research for WDM network is basically based on a fixed bandwidth, in order to accommodate the future business and the fast-changing development of optical network, our study is based on variable bandwidth OFDM-based optical networks. We apply the mathematical analysis and theoretical derivation, based on the existing theory and algorithms, research blocking probability of the variable bandwidth of optical network, and then we will build a model for blocking probability.
NASA Astrophysics Data System (ADS)
Tian, F.; Lu, Y.
2017-12-01
Based on socioeconomic and hydrological data in three arid inland basins and error analysis, the dynamics of human water consumption (HWC) are analyzed to be asymmetric, i.e., HWC increase rapidly in wet periods while maintain or decrease slightly in dry periods. Besides the qualitative analysis that in wet periods great water availability inspires HWC to grow fast but the now expanded economy is managed to sustain by over-exploitation in dry periods, two quantitative models are established and tested, based on expected utility theory (EUT) and prospect theory (PT) respectively. EUT states that humans make decisions based on the total expected utility, namely the sum of utility function multiplied by probability of each result, while PT states that the utility function is defined over gains and losses separately, and probability should be replaced by probability weighting function.
NASA Astrophysics Data System (ADS)
Shao, Lin; Peng, Luohan
2009-12-01
Although multiple scattering theories have been well developed, numerical calculation is complicated and only tabulated values have been available, which has caused inconvenience in practical use. We have found that a Pearson VII distribution function can be used to fit Lugujjo and Mayer's probability curves in describing the dechanneling phenomenon in backscattering analysis, over a wide range of disorder levels. Differentiation of the obtained function gives another function to calculate angular dispersion of the beam in the frameworks by Sigmund and Winterbon. The present work provides an easy calculation of both dechanneling probability and angular dispersion for any arbitrary combination of beam and target having a reduced thickness ⩾0.6, which can be implemented in modeling of channeling spectra. Furthermore, we used a Monte Carlo simulation program to calculate the deflection probability and compared them with previously tabulated data. A good agreement was reached.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zhu, Lin; Dai, Zhenxue; Gong, Huili
Understanding the heterogeneity arising from the complex architecture of sedimentary sequences in alluvial fans is challenging. This study develops a statistical inverse framework in a multi-zone transition probability approach for characterizing the heterogeneity in alluvial fans. An analytical solution of the transition probability matrix is used to define the statistical relationships among different hydrofacies and their mean lengths, integral scales, and volumetric proportions. A statistical inversion is conducted to identify the multi-zone transition probability models and estimate the optimal statistical parameters using the modified Gauss–Newton–Levenberg–Marquardt method. The Jacobian matrix is computed by the sensitivity equation method, which results in anmore » accurate inverse solution with quantification of parameter uncertainty. We use the Chaobai River alluvial fan in the Beijing Plain, China, as an example for elucidating the methodology of alluvial fan characterization. The alluvial fan is divided into three sediment zones. In each zone, the explicit mathematical formulations of the transition probability models are constructed with optimized different integral scales and volumetric proportions. The hydrofacies distributions in the three zones are simulated sequentially by the multi-zone transition probability-based indicator simulations. Finally, the result of this study provides the heterogeneous structure of the alluvial fan for further study of flow and transport simulations.« less
NASA Astrophysics Data System (ADS)
Wang, Yu; Jiang, Wenchun; Luo, Yun; Zhang, Yucai; Tu, Shan-Tung
2017-12-01
The reduction and re-oxidation of anode have significant effects on the integrity of the solid oxide fuel cell (SOFC) sealed by the glass-ceramic (GC). The mechanical failure is mainly controlled by the stress distribution. Therefore, a three dimensional model of SOFC is established to investigate the stress evolution during the reduction and re-oxidation by finite element method (FEM) in this paper, and the failure probability is calculated using the Weibull method. The results demonstrate that the reduction of anode can decrease the thermal stresses and reduce the failure probability due to the volumetric contraction and porosity increasing. The re-oxidation can result in a remarkable increase of the thermal stresses, and the failure probabilities of anode, cathode, electrolyte and GC all increase to 1, which is mainly due to the large linear strain rather than the porosity decreasing. The cathode and electrolyte fail as soon as the linear strains are about 0.03% and 0.07%. Therefore, the re-oxidation should be controlled to ensure the integrity, and a lower re-oxidation temperature can decrease the stress and failure probability.
NASA Astrophysics Data System (ADS)
Zhang, Shunli; Zhang, Dinghua; Gong, Hao; Ghasemalizadeh, Omid; Wang, Ge; Cao, Guohua
2014-11-01
Iterative algorithms, such as the algebraic reconstruction technique (ART), are popular for image reconstruction. For iterative reconstruction, the area integral model (AIM) is more accurate for better reconstruction quality than the line integral model (LIM). However, the computation of the system matrix for AIM is more complex and time-consuming than that for LIM. Here, we propose a fast and accurate method to compute the system matrix for AIM. First, we calculate the intersection of each boundary line of a narrow fan-beam with pixels in a recursive and efficient manner. Then, by grouping the beam-pixel intersection area into six types according to the slopes of the two boundary lines, we analytically compute the intersection area of the narrow fan-beam with the pixels in a simple algebraic fashion. Overall, experimental results show that our method is about three times faster than the Siddon algorithm and about two times faster than the distance-driven model (DDM) in computation of the system matrix. The reconstruction speed of our AIM-based ART is also faster than the LIM-based ART that uses the Siddon algorithm and DDM-based ART, for one iteration. The fast reconstruction speed of our method was accomplished without compromising the image quality.
Short bouts of vocalization induce long lasting fast gamma oscillations in a sensorimotor nucleus
Lewandowski, Brian; Schmidt, Marc
2011-01-01
Performance evaluation is a critical feature of motor learning. In the vocal system, it requires the integration of auditory feedback signals with vocal motor commands. The network activity that supports such integration is unknown, but it has been proposed that vocal performance evaluation occurs offline. Recording from NIf, a sensorimotor structure in the avian song system, we show that short bouts of singing in adult male zebra finches (Taeniopygia guttata) induce persistent increases in firing activity and coherent oscillations in the fast gamma range (90–150 Hz). Single units are strongly phase-locked to these oscillations, which can last up to 30 s, often outlasting vocal activity by an order of magnitude. In other systems, oscillations often are triggered by events or behavioral tasks but rarely outlast the event that triggered them by more than 1 second. The present observations are the longest reported gamma oscillations triggered by an isolated behavioral event. In mammals, gamma oscillations have been associated with memory consolidation and are hypothesized to facilitate communication between brain regions. We suggest that the timing and persistent nature of NIf’s fast gamma oscillations make them well suited to facilitate the integration of auditory and vocal motor traces associated with vocal performance evaluation. PMID:21957255
Sensory augmentation: integration of an auditory compass signal into human perception of space
Schumann, Frank; O’Regan, J. Kevin
2017-01-01
Bio-mimetic approaches to restoring sensory function show great promise in that they rapidly produce perceptual experience, but have the disadvantage of being invasive. In contrast, sensory substitution approaches are non-invasive, but may lead to cognitive rather than perceptual experience. Here we introduce a new non-invasive approach that leads to fast and truly perceptual experience like bio-mimetic techniques. Instead of building on existing circuits at the neural level as done in bio-mimetics, we piggy-back on sensorimotor contingencies at the stimulus level. We convey head orientation to geomagnetic North, a reliable spatial relation not normally sensed by humans, by mimicking sensorimotor contingencies of distal sounds via head-related transfer functions. We demonstrate rapid and long-lasting integration into the perception of self-rotation. Short training with amplified or reduced rotation gain in the magnetic signal can expand or compress the perceived extent of vestibular self-rotation, even with the magnetic signal absent in the test. We argue that it is the reliability of the magnetic signal that allows vestibular spatial recalibration, and the coding scheme mimicking sensorimotor contingencies of distal sounds that permits fast integration. Hence we propose that contingency-mimetic feedback has great potential for creating sensory augmentation devices that achieve fast and genuinely perceptual experiences. PMID:28195187
The syndromic deafness mutation G12R impairs fast and slow gating in Cx26 hemichannels.
García, Isaac E; Villanelo, Felipe; Contreras, Gustavo F; Pupo, Amaury; Pinto, Bernardo I; Contreras, Jorge E; Pérez-Acle, Tomás; Alvarez, Osvaldo; Latorre, Ramon; Martínez, Agustín D; González, Carlos
2018-05-07
Mutations in connexin 26 (Cx26) hemichannels can lead to syndromic deafness that affects the cochlea and skin. These mutations lead to gain-of-function hemichannel phenotypes by unknown molecular mechanisms. In this study, we investigate the biophysical properties of the syndromic mutant Cx26G12R (G12R). Unlike wild-type Cx26, G12R macroscopic hemichannel currents do not saturate upon depolarization, and deactivation is faster during hyperpolarization, suggesting that these channels have impaired fast and slow gating. Single G12R hemichannels show a large increase in open probability, and transitions to the subconductance state are rare and short-lived, demonstrating an inoperative fast gating mechanism. Molecular dynamics simulations indicate that G12R causes a displacement of the N terminus toward the cytoplasm, favoring an interaction between R12 in the N terminus and R99 in the intracellular loop. Disruption of this interaction recovers the fast and slow voltage-dependent gating mechanisms. These results suggest that the mechanisms of fast and slow gating in connexin hemichannels are coupled and provide a molecular mechanism for the gain-of-function phenotype displayed by the syndromic G12R mutation. © 2018 García et al.
Never Say No … How the Brain Interprets the Pregnant Pause in Conversation
Bögels, Sara; Kendrick, Kobin H.; Levinson, Stephen C.
2015-01-01
In conversation, negative responses to invitations, requests, offers, and the like are more likely to occur with a delay–conversation analysts talk of them as dispreferred. Here we examine the contrastive cognitive load ‘yes’ and ‘no’ responses make, either when relatively fast (300 ms after question offset) or delayed (1000 ms). Participants heard short dialogues contrasting in speed and valence of response while having their EEG recorded. We found that a fast ‘no’ evokes an N400-effect relative to a fast ‘yes’; however, this contrast disappeared in the delayed responses. 'No' responses, however, elicited a late frontal positivity both if they were fast and if they were delayed. We interpret these results as follows: a fast ‘no’ evoked an N400 because an immediate response is expected to be positive–this effect disappears as the response time lengthens because now in ordinary conversation the probability of a ‘no’ has increased. However, regardless of the latency of response, a ‘no’ response is associated with a late positivity, since a negative response is always dispreferred. Together these results show that negative responses to social actions exact a higher cognitive load, but especially when least expected, in immediate response. PMID:26699335
Factors Predicting Staying in School to Eat Lunch
ERIC Educational Resources Information Center
Beaulieu, Dominique; Godin, Gaston
2011-01-01
Purpose: Easy access to fast-food restaurants in the immediate environment of a high school is such that a high proportion of students do not remain in school for lunch. Hence, the probability that they will eat a healthy meal is reduced. The aim of this study is to identify the behavioral determinants of "staying in school to eat lunch" among…
ERIC Educational Resources Information Center
McKean, Cristina; Letts, Carolyn; Howard, David
2013-01-01
Neighbourhood Density (ND) and Phonotactic Probability (PP) influence word learning in children. This influence appears to change over development but the separate developmental trajectories of influence of PP and ND on word learning have not previously been mapped. This study examined the cross-sectional developmental trajectories of influence of…
Bradford M. Sanders; David H. van Lear
1988-01-01
Paired photographs show fuel conditions before and after burning in recently clearcut stands of mixed pine-hardwoods in the Southern Appalachians. Comparison with the photos permits fast assessment of fuel loading and probable burning success. Information with each photo includes measured weights, volumes, and other residue data, information about the timber stand and...
Mark J. Ducey; Jeffrey H. Gove; Harry T. Valentine
2008-01-01
Perpendicular distance sampling (PDS) is a fast probability-proportional-to-size method for inventory of downed wood. However, previous development of PDS had limited the method to estimating only one variable (such as volume per hectare, or surface area per hectare) at a time. Here, we develop a general design-unbiased estimator for PDS. We then show how that...
Preparation and Characterization of Ato Nanoparticles by Coprecipitation with Modified Drying Method
NASA Astrophysics Data System (ADS)
Liu, Shimin; Liang, Dongdong; Liu, Jindong; Jiang, Weiwei; Liu, Chaoqian; Ding, Wanyu; Wang, Hualin; Wang, Nan
Antimony-doped tin oxide (ATO) nanoparticles were prepared by coprecipitation by packing drying and traditional direct drying (for comparison) methods. The as-prepared ATO nanoparticles were characterized by TG, XRD, EDS, TEM, HRTEM, BET, bulk density and electrical resistivity measurements. Results indicated that the ATO nanoparticles obtained by coprecipitation with direct drying method featured hard-agglomerated morphology, high bulk density, low surface area and low electrical resistivity, probably due to the direct liquid evaporation during drying, the fast shrinkage of the precipitate, the poor removal efficiency of liquid molecules and the hard agglomerate formation after calcination. Very differently, the ATO product obtained by the packing and drying method featured free-agglomerated morphology, low bulk density, high surface area and high electrical resistivity ascribed probably to the formed vapor cyclone environment and liquid evaporation-resistance, avoiding fast liquid removal and improving the removal efficiency of liquid molecules. The intrinsic formation mechanism of ATO nanoparticles from different drying methods was illustrated based on the dehydration process of ATO precipitates. Additionally, the packing and drying time played key roles in determining the bulk density, morphology and electrical conductivity of ATO nanoparticles.
Fast-crawling cell types migrate to avoid the direction of periodic substratum stretching
Okimura, Chika; Ueda, Kazuki; Sakumura, Yuichi; Iwadate, Yoshiaki
2016-01-01
ABSTRACT To investigate the relationship between mechanical stimuli from substrata and related cell functions, one of the most useful techniques is the application of mechanical stimuli via periodic stretching of elastic substrata. In response to this stimulus, Dictyostelium discoideum cells migrate in a direction perpendicular to the stretching direction. The origins of directional migration, higher migration velocity in the direction perpendicular to the stretching direction or the higher probability of a switch of migration direction to perpendicular to the stretching direction, however, remain unknown. In this study, we applied periodic stretching stimuli to neutrophil-like differentiated HL-60 cells, which migrate perpendicular to the direction of stretch. Detailed analysis of the trajectories of HL-60 cells and Dictyostelium cells obtained in a previous study revealed that the higher probability of a switch of migration direction to that perpendicular to the direction of stretching was the main cause of such directional migration. This directional migration appears to be a strategy adopted by fast-crawling cells in which they do not migrate faster in the direction they want to go, but migrate to avoid a direction they do not want to go. PMID:26980079
Improving detection of low SNR targets using moment-based detection
NASA Astrophysics Data System (ADS)
Young, Shannon R.; Steward, Bryan J.; Hawks, Michael; Gross, Kevin C.
2016-05-01
Increases in the number of cameras deployed, frame rate, and detector array sizes have led to a dramatic increase in the volume of motion imagery data that is collected. Without a corresponding increase in analytical manpower, much of the data is not analyzed to full potential. This creates a need for fast, automated, and robust methods for detecting signals of interest. Current approaches fall into two categories: detect-before-track (DBT), which are fast but often poor at detecting dim targets, and track-before-detect (TBD) methods which can offer better performance but are typically much slower. This research seeks to contribute to the near real time detection of low SNR, unresolved moving targets through an extension of earlier work on higher order moments anomaly detection, a method that exploits both spatial and temporal information but is still computationally efficient and massively parallelizable. It was found that intelligent selection of parameters can improve probability of detection by as much as 25% compared to earlier work with higherorder moments. The present method can reduce detection thresholds by 40% compared to the Reed-Xiaoli anomaly detector for low SNR targets (for a given probability of detection and false alarm).
DNA integrity determination in marine invertebrates by Fast Micromethod.
Jaksić, Zeljko; Batel, Renato
2003-12-10
This study was focused toward the adaptation of the previously developed Fast Micromethod for DNA damage determination to marine invertebrates for the establishment of biomonitoring assessment. The Fast Micromethod detects DNA damage (strand breaks, alkali-labile sites and incomplete excision repair) and determines DNA integrity in cell suspensions or tissue homogenates in single microplates. The procedure is based on the ability of the specific fluorochrome dye PicoGreen to preferentially interact with high integrity DNA molecules, dsDNA, in the presence of ssDNA and proteins in high alkaline medium, thereby allowing direct fluorometric measurements of dsDNA denaturation without sample handling and stepwise DNA separations. The results presented herein describe the influence of the DNA amount and the pH of the denaturation media on slopes of the kinetic denaturation curves and calculated strand scission factors (SSFs). The optimal amount of DNA in Mytilus galloprovincialis gills homogenate was found to be 100 ng ml(-1) and the greatest differences in DNA unwinding kinetics (slopes and SSF values) were reached at pH 11.5. The induction of DNA damage and loss of DNA integrity was measured in native DNA isolated from cotton-spinner Holothuria tubulosa, marine sponge Suberites domuncula cells and mussel M. galloprovincialis gills homogenate. DNA damage and loss of DNA integrity were detected after induction by different doses of (gamma-rays, generated by 137Cs 1800 Ci; 0-500 rad in marine sponge S. domuncula cells up to SSFx(-1) values 0.082 +/- 0.012 for the highest radiation dose). Analysis by chemical xenobiotics based on the in vitro action of bleomycin (bleomycin-Fe(II) complex 0-50 or 0-83 microg ml(-1) (microM)) with native DNA from cotton-spinner H. tubulosa and mussel M. galloprovincialis gills homogenate yielded values of 0.537 +/- 0.072 and 0.130 +/- 0.018, respectively. In vivo experiments with mussel M. galloprovincialis gills homogenate by 4-nitroquinoline-N-oxide (NQO; 0-1 microg g(-1) NQO mussel) and benzo[a]pyrene (B[a]P; 0-20 microg g(-1) B[a]P mussel) indicated SSFx(-1) values of 0.121 +/- 0.016 and 0.090 +/- 0.007, respectively, for the highest applied doses of chemical xenobiotics. The analytical technique described here allows simple and fast analysis of DNA integrity, requires very short time for multiple analyses (less than 3 h) and even less than 100 ng DNA per single well (50 ng DNA isolated from cotton-spinner, 12,500 sponge cells or about 10 mg of mussel gills homogenate) in a microplate. This makes the Fast Micromethod applicable for the measurement of DNA integrity of small samples for genotoxicity assessment (biomonitoring), the effects of genotoxins on lower marine taxa or sessile invertebrates in marine environment (e.g. sponges, mussels) and the estimation of directional changes and harmful effects in the ecosystem.
Evaluation of Fast-Time Wake Vortex Models using Wake Encounter Flight Test Data
NASA Technical Reports Server (NTRS)
Ahmad, Nashat N.; VanValkenburg, Randal L.; Bowles, Roland L.; Limon Duparcmeur, Fanny M.; Gloudesman, Thijs; van Lochem, Sander; Ras, Eelco
2014-01-01
This paper describes a methodology for the integration and evaluation of fast-time wake models with flight data. The National Aeronautics and Space Administration conducted detailed flight tests in 1995 and 1997 under the Aircraft Vortex Spacing System Program to characterize wake vortex decay and wake encounter dynamics. In this study, data collected during Flight 705 were used to evaluate NASA's fast-time wake transport and decay models. Deterministic and Monte-Carlo simulations were conducted to define wake hazard bounds behind the wake generator. The methodology described in this paper can be used for further validation of fast-time wake models using en-route flight data, and for determining wake turbulence constraints in the design of air traffic management concepts.
Fast, high-fidelity readout of multiple qubits
NASA Astrophysics Data System (ADS)
Bronn, N. T.; Abdo, B.; Inoue, K.; Lekuch, S.; Córcoles, A. D.; Hertzberg, J. B.; Takita, M.; Bishop, L. S.; Gambetta, J. M.; Chow, J. M.
2017-05-01
Quantum computing requires a delicate balance between coupling quantum systems to external instruments for control and readout, while providing enough isolation from sources of decoherence. Circuit quantum electrodynamics has been a successful method for protecting superconducting qubits, while maintaining the ability to perform readout [1, 2]. Here, we discuss improvements to this method that allow for fast, high-fidelity readout. Specifically, the integration of a Purcell filter, which allows us to increase the resonator bandwidth for fast readout, the incorporation of a Josephson parametric converter, which enables us to perform high-fidelity readout by amplifying the readout signal while adding the minimum amount of noise required by quantum mechanics, and custom control electronics, which provide us with the capability of fast decision and control.
Flightspeed Integral Image Analysis Toolkit
NASA Technical Reports Server (NTRS)
Thompson, David R.
2009-01-01
The Flightspeed Integral Image Analysis Toolkit (FIIAT) is a C library that provides image analysis functions in a single, portable package. It provides basic low-level filtering, texture analysis, and subwindow descriptor for applications dealing with image interpretation and object recognition. Designed with spaceflight in mind, it addresses: Ease of integration (minimal external dependencies) Fast, real-time operation using integer arithmetic where possible (useful for platforms lacking a dedicated floatingpoint processor) Written entirely in C (easily modified) Mostly static memory allocation 8-bit image data The basic goal of the FIIAT library is to compute meaningful numerical descriptors for images or rectangular image regions. These n-vectors can then be used directly for novelty detection or pattern recognition, or as a feature space for higher-level pattern recognition tasks. The library provides routines for leveraging training data to derive descriptors that are most useful for a specific data set. Its runtime algorithms exploit a structure known as the "integral image." This is a caching method that permits fast summation of values within rectangular regions of an image. This integral frame facilitates a wide range of fast image-processing functions. This toolkit has applicability to a wide range of autonomous image analysis tasks in the space-flight domain, including novelty detection, object and scene classification, target detection for autonomous instrument placement, and science analysis of geomorphology. It makes real-time texture and pattern recognition possible for platforms with severe computational restraints. The software provides an order of magnitude speed increase over alternative software libraries currently in use by the research community. FIIAT can commercially support intelligent video cameras used in intelligent surveillance. It is also useful for object recognition by robots or other autonomous vehicles
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shannon M. Bragg-Sitton; Richard D. Boardman; Robert S. Cherry
2014-03-01
Integration of an advanced, sodium-cooled fast spectrum reactor into nuclear hybrid energy system (NHES) architectures is the focus of the present study. A techno-economic evaluation of several conceptual system designs was performed for the integration of a sodium-cooled Advanced Fast Reactor (AFR) with the electric grid in conjunction with wind-generated electricity. Cases in which excess thermal and electrical energy would be reapportioned within an integrated energy system to a chemical plant are presented. The process applications evaluated include hydrogen production via high temperature steam electrolysis and methanol production via steam methane reforming to produce carbon monoxide and hydrogen which feedmore » a methanol synthesis reactor. Three power cycles were considered for integration with the AFR, including subcritical and supercritical Rankine cycles and a modified supercritical carbon dioxide modified Brayton cycle. The thermal efficiencies of all of the modeled power conversions units were greater than 40%. A thermal efficiency of 42% was adopted in economic studies because two of the cycles either performed at that level or could potentially do so (subcritical Rankine and S-CO2 Brayton). Each of the evaluated hybrid architectures would be technically feasible but would demonstrate a different internal rate of return (IRR) as a function of multiple parameters; all evaluated configurations showed a positive IRR. As expected, integration of an AFR with a chemical plant increases the IRR when “must-take” wind-generated electricity is added to the energy system. Additional dynamic system analyses are recommended to draw detailed conclusions on the feasibility and economic benefits associated with AFR-hybrid energy system operation.« less
Data management software concept for WEST plasma measurement system
NASA Astrophysics Data System (ADS)
Zienkiewicz, P.; Kasprowicz, G.; Byszuk, A.; Wojeński, A.; Kolasinski, P.; Cieszewski, R.; Czarski, T.; Chernyshova, M.; Pozniak, K.; Zabolotny, W.; Juszczyk, B.; Mazon, D.; Malard, P.
2014-11-01
This paper describes the concept of data management software for the multichannel readout system for the GEM detector used in WEST Plasma experiment. The proposed system consists of three separate communication channels: fast data channel, diagnostics channel, slow data channel. Fast data channel is provided by the FPGA with integrated ARM cores providing direct readout data from Analog Front Ends through 10GbE with short, guaranteed intervals. Slow data channel is provided by multiple, fast CPUs after data processing with detailed readout data with use of GNU/Linux OS and appropriate software. Diagnostic channel provides detailed feedback for control purposes.
A Bayesian pick-the-winner design in a randomized phase II clinical trial.
Chen, Dung-Tsa; Huang, Po-Yu; Lin, Hui-Yi; Chiappori, Alberto A; Gabrilovich, Dmitry I; Haura, Eric B; Antonia, Scott J; Gray, Jhanelle E
2017-10-24
Many phase II clinical trials evaluate unique experimental drugs/combinations through multi-arm design to expedite the screening process (early termination of ineffective drugs) and to identify the most effective drug (pick the winner) to warrant a phase III trial. Various statistical approaches have been developed for the pick-the-winner design but have been criticized for lack of objective comparison among the drug agents. We developed a Bayesian pick-the-winner design by integrating a Bayesian posterior probability with Simon two-stage design in a randomized two-arm clinical trial. The Bayesian posterior probability, as the rule to pick the winner, is defined as probability of the response rate in one arm higher than in the other arm. The posterior probability aims to determine the winner when both arms pass the second stage of the Simon two-stage design. When both arms are competitive (i.e., both passing the second stage), the Bayesian posterior probability performs better to correctly identify the winner compared with the Fisher exact test in the simulation study. In comparison to a standard two-arm randomized design, the Bayesian pick-the-winner design has a higher power to determine a clear winner. In application to two studies, the approach is able to perform statistical comparison of two treatment arms and provides a winner probability (Bayesian posterior probability) to statistically justify the winning arm. We developed an integrated design that utilizes Bayesian posterior probability, Simon two-stage design, and randomization into a unique setting. It gives objective comparisons between the arms to determine the winner.
Measures of School Integration: Comparing Coleman's Index to Measures of Species Diversity.
ERIC Educational Resources Information Center
Mercil, Steven Bray; Williams, John Delane
This study used species diversity indices developed in ecology as a measure of socioethnic diversity, and compared them to Coleman's Index of Segregation. The twelve indices were Simpson's Concentration Index ("ell"), Simpson's Index of Diversity, Hurlbert's Probability of Interspecific Encounter (PIE), Simpson's Probability of…
NASA Astrophysics Data System (ADS)
von der Linden, Wolfgang; Dose, Volker; von Toussaint, Udo
2014-06-01
Preface; Part I. Introduction: 1. The meaning of probability; 2. Basic definitions; 3. Bayesian inference; 4. Combinatrics; 5. Random walks; 6. Limit theorems; 7. Continuous distributions; 8. The central limit theorem; 9. Poisson processes and waiting times; Part II. Assigning Probabilities: 10. Transformation invariance; 11. Maximum entropy; 12. Qualified maximum entropy; 13. Global smoothness; Part III. Parameter Estimation: 14. Bayesian parameter estimation; 15. Frequentist parameter estimation; 16. The Cramer-Rao inequality; Part IV. Testing Hypotheses: 17. The Bayesian way; 18. The frequentist way; 19. Sampling distributions; 20. Bayesian vs frequentist hypothesis tests; Part V. Real World Applications: 21. Regression; 22. Inconsistent data; 23. Unrecognized signal contributions; 24. Change point problems; 25. Function estimation; 26. Integral equations; 27. Model selection; 28. Bayesian experimental design; Part VI. Probabilistic Numerical Techniques: 29. Numerical integration; 30. Monte Carlo methods; 31. Nested sampling; Appendixes; References; Index.
REPRESENTATIONS OF WEAK AND STRONG INTEGRALS IN BANACH SPACES
Brooks, James K.
1969-01-01
We establish a representation of the Gelfand-Pettis (weak) integral in terms of unconditionally convergent series. Moreover, absolute convergence of the series is a necessary and sufficient condition in order that the weak integral coincide with the Bochner integral. Two applications of the representation are given. The first is a simplified proof of the countable additivity and absolute continuity of the indefinite weak integral. The second application is to probability theory; we characterize the conditional expectation of a weakly integrable function. PMID:16591755
Master equations and the theory of stochastic path integrals
NASA Astrophysics Data System (ADS)
Weber, Markus F.; Frey, Erwin
2017-04-01
This review provides a pedagogic and self-contained introduction to master equations and to their representation by path integrals. Since the 1930s, master equations have served as a fundamental tool to understand the role of fluctuations in complex biological, chemical, and physical systems. Despite their simple appearance, analyses of master equations most often rely on low-noise approximations such as the Kramers-Moyal or the system size expansion, or require ad-hoc closure schemes for the derivation of low-order moment equations. We focus on numerical and analytical methods going beyond the low-noise limit and provide a unified framework for the study of master equations. After deriving the forward and backward master equations from the Chapman-Kolmogorov equation, we show how the two master equations can be cast into either of four linear partial differential equations (PDEs). Three of these PDEs are discussed in detail. The first PDE governs the time evolution of a generalized probability generating function whose basis depends on the stochastic process under consideration. Spectral methods, WKB approximations, and a variational approach have been proposed for the analysis of the PDE. The second PDE is novel and is obeyed by a distribution that is marginalized over an initial state. It proves useful for the computation of mean extinction times. The third PDE describes the time evolution of a ‘generating functional’, which generalizes the so-called Poisson representation. Subsequently, the solutions of the PDEs are expressed in terms of two path integrals: a ‘forward’ and a ‘backward’ path integral. Combined with inverse transformations, one obtains two distinct path integral representations of the conditional probability distribution solving the master equations. We exemplify both path integrals in analysing elementary chemical reactions. Moreover, we show how a well-known path integral representation of averaged observables can be recovered from them. Upon expanding the forward and the backward path integrals around stationary paths, we then discuss and extend a recent method for the computation of rare event probabilities. Besides, we also derive path integral representations for processes with continuous state spaces whose forward and backward master equations admit Kramers-Moyal expansions. A truncation of the backward expansion at the level of a diffusion approximation recovers a classic path integral representation of the (backward) Fokker-Planck equation. One can rewrite this path integral in terms of an Onsager-Machlup function and, for purely diffusive Brownian motion, it simplifies to the path integral of Wiener. To make this review accessible to a broad community, we have used the language of probability theory rather than quantum (field) theory and do not assume any knowledge of the latter. The probabilistic structures underpinning various technical concepts, such as coherent states, the Doi-shift, and normal-ordered observables, are thereby made explicit.
Master equations and the theory of stochastic path integrals.
Weber, Markus F; Frey, Erwin
2017-04-01
This review provides a pedagogic and self-contained introduction to master equations and to their representation by path integrals. Since the 1930s, master equations have served as a fundamental tool to understand the role of fluctuations in complex biological, chemical, and physical systems. Despite their simple appearance, analyses of master equations most often rely on low-noise approximations such as the Kramers-Moyal or the system size expansion, or require ad-hoc closure schemes for the derivation of low-order moment equations. We focus on numerical and analytical methods going beyond the low-noise limit and provide a unified framework for the study of master equations. After deriving the forward and backward master equations from the Chapman-Kolmogorov equation, we show how the two master equations can be cast into either of four linear partial differential equations (PDEs). Three of these PDEs are discussed in detail. The first PDE governs the time evolution of a generalized probability generating function whose basis depends on the stochastic process under consideration. Spectral methods, WKB approximations, and a variational approach have been proposed for the analysis of the PDE. The second PDE is novel and is obeyed by a distribution that is marginalized over an initial state. It proves useful for the computation of mean extinction times. The third PDE describes the time evolution of a 'generating functional', which generalizes the so-called Poisson representation. Subsequently, the solutions of the PDEs are expressed in terms of two path integrals: a 'forward' and a 'backward' path integral. Combined with inverse transformations, one obtains two distinct path integral representations of the conditional probability distribution solving the master equations. We exemplify both path integrals in analysing elementary chemical reactions. Moreover, we show how a well-known path integral representation of averaged observables can be recovered from them. Upon expanding the forward and the backward path integrals around stationary paths, we then discuss and extend a recent method for the computation of rare event probabilities. Besides, we also derive path integral representations for processes with continuous state spaces whose forward and backward master equations admit Kramers-Moyal expansions. A truncation of the backward expansion at the level of a diffusion approximation recovers a classic path integral representation of the (backward) Fokker-Planck equation. One can rewrite this path integral in terms of an Onsager-Machlup function and, for purely diffusive Brownian motion, it simplifies to the path integral of Wiener. To make this review accessible to a broad community, we have used the language of probability theory rather than quantum (field) theory and do not assume any knowledge of the latter. The probabilistic structures underpinning various technical concepts, such as coherent states, the Doi-shift, and normal-ordered observables, are thereby made explicit.
Extinction times of epidemic outbreaks in networks.
Holme, Petter
2013-01-01
In the Susceptible-Infectious-Recovered (SIR) model of disease spreading, the time to extinction of the epidemics happens at an intermediate value of the per-contact transmission probability. Too contagious infections burn out fast in the population. Infections that are not contagious enough die out before they spread to a large fraction of people. We characterize how the maximal extinction time in SIR simulations on networks depend on the network structure. For example we find that the average distances in isolated components, weighted by the component size, is a good predictor of the maximal time to extinction. Furthermore, the transmission probability giving the longest outbreaks is larger than, but otherwise seemingly independent of, the epidemic threshold.
ERIC Educational Resources Information Center
Roberson, James H.; Hagevik, Rita A.
2008-01-01
Cell phones are fast becoming an integral part of students' everyday lives. They are regarded as important companions and tools for personal expression. School-age children are integrating the cell phone as such, and thus placing a high value on them. Educators endeavor to instill in students a high value for education, but often meet with…
Integrated semiconductor-magnetic random access memory system
NASA Technical Reports Server (NTRS)
Katti, Romney R. (Inventor); Blaes, Brent R. (Inventor)
2001-01-01
The present disclosure describes a non-volatile magnetic random access memory (RAM) system having a semiconductor control circuit and a magnetic array element. The integrated magnetic RAM system uses CMOS control circuit to read and write data magnetoresistively. The system provides a fast access, non-volatile, radiation hard, high density RAM for high speed computing.
Klaseboer, Evert; Sepehrirahnama, Shahrokh; Chan, Derek Y C
2017-08-01
The general space-time evolution of the scattering of an incident acoustic plane wave pulse by an arbitrary configuration of targets is treated by employing a recently developed non-singular boundary integral method to solve the Helmholtz equation in the frequency domain from which the space-time solution of the wave equation is obtained using the fast Fourier transform. The non-singular boundary integral solution can enforce the radiation boundary condition at infinity exactly and can account for multiple scattering effects at all spacings between scatterers without adverse effects on the numerical precision. More generally, the absence of singular kernels in the non-singular integral equation confers high numerical stability and precision for smaller numbers of degrees of freedom. The use of fast Fourier transform to obtain the time dependence is not constrained to discrete time steps and is particularly efficient for studying the response to different incident pulses by the same configuration of scatterers. The precision that can be attained using a smaller number of Fourier components is also quantified.
Dangerous "spin": the probability myth of evidence-based prescribing - a Merleau-Pontyian approach.
Morstyn, Ron
2011-08-01
The aim of this study was to examine logical positivist statistical probability statements used to support and justify "evidence-based" prescribing rules in psychiatry when viewed from the major philosophical theories of probability, and to propose "phenomenological probability" based on Maurice Merleau-Ponty's philosophy of "phenomenological positivism" as a better clinical and ethical basis for psychiatric prescribing. The logical positivist statistical probability statements which are currently used to support "evidence-based" prescribing rules in psychiatry have little clinical or ethical justification when subjected to critical analysis from any of the major theories of probability and represent dangerous "spin" because they necessarily exclude the individual , intersubjective and ambiguous meaning of mental illness. A concept of "phenomenological probability" founded on Merleau-Ponty's philosophy of "phenomenological positivism" overcomes the clinically destructive "objectivist" and "subjectivist" consequences of logical positivist statistical probability and allows psychopharmacological treatments to be appropriately integrated into psychiatric treatment.
Latitude Dependence of Low-Altitude O+ Ion Upflow: Statistical Results From FAST Observations
NASA Astrophysics Data System (ADS)
Zhao, K.; Chen, K. W.; Jiang, Y.; Chen, W. J.; Huang, L. F.; Fu, S.
2017-09-01
We introduce a statistical model to explain the latitudinal dependence of the occurrence rate and energy flux of the ionospheric escaping ions, taking advantage of advances in the spatial coverage and accuracy of FAST observations. We use a weighted piecewise Gaussian function to fit the dependence, because two probability peaks are located in the dayside polar cusp source region and the nightside auroral oval zone source region. The statistical results show that (1) the Gaussian Mixture Model suitably describes the dayside polar cusp upflows, and the dayside and the nightside auroral oval zone upflows. (2) The magnetic latitudes of the ionospheric upflow source regions expand toward the magnetic equator as Kp increases, from 81° magnetic latitude (MLAT) (cusp upflows) and 63° MLAT (auroral oval upflows) during quiet times to 76° MLAT and 61° MLAT, respectively. (3) The dayside polar cusp region provides only 3-5% O+ upflows among all the source regions, which include the dayside auroral oval zone, dayside polar cusp, nightside auroral oval zone, and even the polar cap. However, observations show that more than 70% of upflows occur in the auroral oval zone and that the occurrence probability increases at the altitudes of 3500-4200 km, which is considered to be the lower altitude boundary of ion beams. This observed result suggests that soft electron precipitation and transverse wave heating are the most efficient ion energization/acceleration mechanisms at the altitudes of FAST orbit, and that the parallel acceleration caused by field-aligned potential drops becomes effective above that altitude.
Baser, Gonen; Cengiz, Hakan; Uyar, Murat; Seker Un, Emine
2016-01-01
To investigate the effects of dehydration due to fasting on diurnal changes of intraocular pressure, anterior segment biometrics, and refraction. The intraocular pressures, anterior segment biometrics (axial length: AL; Central corneal thickness: CCT; Lens thickness: LT; Anterior chamber depth: ACD), and refractive measurements of 30 eyes of 15 fasting healthy male volunteers were recorded at 8:00 in the morning and 17:00 in the evening in the Ramadan of 2013 and two months later. The results were compared and the statistical analyses were performed using the Rstudio software version 0.98.501. The variables were investigated using visual (histograms, probability plots) and analytical methods (Kolmogorov-Smirnov/Shapiro-Wilk test) to determine whether or not they were normally distributed. The refractive values remained stable in the fasting as well as in the control period (p = 0.384). The axial length measured slightly shorter in the fasting period (p = 0.001). The corneal thickness presented a diurnal variation, in which the cornea measured thinner in the evening. The difference between the fasting and control period was not statistically significant (p = 0.359). The major differences were observed in the anterior chamber depth and IOP. The ACD was shallower in the evening during the fasting period, where it was deeper in the control period. The diurnal IOP difference was greater in the fasting period than the control period. Both were statistically significant (p = 0.001). The LT remained unchanged in both periods. The major difference was shown in the anterior chamber shallowing in the evening hours and IOP. Our study contributes the hypothesis that the posterior segment of the eye is more responsible for the axial length alterations and normovolemia has a more dominant influence on diurnal IOP changes.
Probabilistic Integrated Assessment of ``Dangerous'' Climate Change
NASA Astrophysics Data System (ADS)
Mastrandrea, Michael D.; Schneider, Stephen H.
2004-04-01
Climate policy decisions are being made despite layers of uncertainty. Such decisions directly influence the potential for ``dangerous anthropogenic interference with the climate system.'' We mapped a metric for this concept, based on Intergovernmental Panel on Climate Change assessment of climate impacts, onto probability distributions of future climate change produced from uncertainty in key parameters of the coupled social-natural system-climate sensitivity, climate damages, and discount rate. Analyses with a simple integrated assessment model found that, under midrange assumptions, endogenously calculated, optimal climate policy controls can reduce the probability of dangerous anthropogenic interference from ~45% under minimal controls to near zero.
NASA Astrophysics Data System (ADS)
Li, Zhenjie; Li, Qiuju; Chang, Jinfan; Ma, Yichao; Liu, Peng; Wang, Zheng; Hu, Michael Y.; Zhao, Jiyong; Alp, E. E.; Xu, Wei; Tao, Ye; Wu, Chaoqun; Zhou, Yangfan
2017-10-01
A four-channel nanosecond time-resolved avalanche-photodiode (APD) detector system is developed at Beijing Synchrotron Radiation. It uses a single module for signal processing and readout. This integrated system provides better reliability and flexibility for custom improvement. The detector system consists of three parts: (i) four APD sensors, (ii) four fast preamplifiers and (iii) a time-digital-converter (TDC) readout electronics. The C30703FH silicon APD chips fabricated by Excelitas are used as the sensors of the detectors. It has an effective light-sensitive area of 10 × 10 mm2 and an absorption layer thickness of 110 μm. A fast preamplifier with a gain of 59 dB and bandwidth of 2 GHz is designed to readout of the weak signal from the C30703FH APD. The TDC is realized by a Spartan-6 field-programmable-gate-array (FPGA) with multiphase method in a resolution of 1ns. The arrival time of all scattering events between two start triggers can be recorded by the TDC. The detector has been used for nuclear resonant scattering study at both Advanced Photon Source and also at Beijing Synchrotron Radiation Facility. For the X-ray energy of 14.4 keV, the time resolution, the full width of half maximum (FWHM) of the detector (APD sensor + fast amplifier) is 0.86 ns, and the whole detector system (APD sensors + fast amplifiers + TDC readout electronics) achieves a time resolution of 1.4 ns.
Simulation of the space station information system in Ada
NASA Technical Reports Server (NTRS)
Spiegel, James R.
1986-01-01
The Flexible Ada Simulation Tool (FAST) is a discrete event simulation language which is written in Ada. FAST has been used to simulate a number of options for ground data distribution of Space Station payload data. The fact that Ada language is used for implementation has allowed a number of useful interactive features to be built into FAST and has facilitated quick enhancement of its capabilities to support new modeling requirements. General simulation concepts are discussed, and how these concepts are implemented in FAST. The FAST design is discussed, and it is pointed out how the used of the Ada language enabled the development of some significant advantages over classical FORTRAN based simulation languages. The advantages discussed are in the areas of efficiency, ease of debugging, and ease of integrating user code. The specific Ada language features which enable these advances are discussed.
Does access to fast food lead to super-sized pregnant women and whopper babies?
Lhila, Aparna
2011-12-01
Rise in the availability of fast-food restaurants has been blamed, at least partly, for the increasing obesity in the U.S. The existing studies of obesity have focused primarily on children, adolescents, and adults, and this paper extends the literature by raising a little-studied question and using nationally representative data to answer it. It examines the relationship between the supply of fast-food restaurants and weight gain of pregnant women and their newborns. I study prenatal weight gain because excessive weight gain has been linked to postpartum overweight/obesity and I study both tails of the birthweight distribution because the origin of obesity may be traced to the prenatal period and both tail outcomes have been associated with obesity later in life. I merge the 1998 and 2004 Natality Detail Files with the Area Resource File, and County Business Patterns, which provide data on the number of fast-food restaurants in the metropolitan area where the mother resides. The empirical model includes an extensive list of MSA characteristics and MSA fixed effects to control for factors that may be correlated with both health outcomes and restaurants' location decision. Results reveal that the fast-food and weight gain relationship is robust to the inclusion of these controls but these controls greatly mitigate the fast food-infant health relationship. Greater access to fast-food restaurants is positively related to mothers' probability of excessive weight gain but it does not share a statistically significant relationship with birthweight. These relationships hold in all the socioeconomic and demographic subgroups studied. 2011 Elsevier B.V. All rights reserved.
Fast algorithm for probabilistic bone edge detection (FAPBED)
NASA Astrophysics Data System (ADS)
Scepanovic, Danilo; Kirshtein, Joshua; Jain, Ameet K.; Taylor, Russell H.
2005-04-01
The registration of preoperative CT to intra-operative reality systems is a crucial step in Computer Assisted Orthopedic Surgery (CAOS). The intra-operative sensors include 3D digitizers, fiducials, X-rays and Ultrasound (US). FAPBED is designed to process CT volumes for registration to tracked US data. Tracked US is advantageous because it is real time, noninvasive, and non-ionizing, but it is also known to have inherent inaccuracies which create the need to develop a framework that is robust to various uncertainties, and can be useful in US-CT registration. Furthermore, conventional registration methods depend on accurate and absolute segmentation. Our proposed probabilistic framework addresses the segmentation-registration duality, wherein exact segmentation is not a prerequisite to achieve accurate registration. In this paper, we develop a method for fast and automatic probabilistic bone surface (edge) detection in CT images. Various features that influence the likelihood of the surface at each spatial coordinate are combined using a simple probabilistic framework, which strikes a fair balance between a high-level understanding of features in an image and the low-level number crunching of standard image processing techniques. The algorithm evaluates different features for detecting the probability of a bone surface at each voxel, and compounds the results of these methods to yield a final, low-noise, probability map of bone surfaces in the volume. Such a probability map can then be used in conjunction with a similar map from tracked intra-operative US to achieve accurate registration. Eight sample pelvic CT scans were used to extract feature parameters and validate the final probability maps. An un-optimized fully automatic Matlab code runs in five minutes per CT volume on average, and was validated by comparison against hand-segmented gold standards. The mean probability assigned to nonzero surface points was 0.8, while nonzero non-surface points had a mean value of 0.38 indicating clear identification of surface points on average. The segmentation was also sufficiently crisp, with a full width at half maximum (FWHM) value of 1.51 voxels.
Nilsson, Håkan; Juslin, Peter; Winman, Anders
2016-01-01
Costello and Watts (2014) present a model assuming that people's knowledge of probabilities adheres to probability theory, but that their probability judgments are perturbed by a random noise in the retrieval from memory. Predictions for the relationships between probability judgments for constituent events and their disjunctions and conjunctions, as well as for sums of such judgments were derived from probability theory. Costello and Watts (2014) report behavioral data showing that subjective probability judgments accord with these predictions. Based on the finding that subjective probability judgments follow probability theory, Costello and Watts (2014) conclude that the results imply that people's probability judgments embody the rules of probability theory and thereby refute theories of heuristic processing. Here, we demonstrate the invalidity of this conclusion by showing that all of the tested predictions follow straightforwardly from an account assuming heuristic probability integration (Nilsson, Winman, Juslin, & Hansson, 2009). We end with a discussion of a number of previous findings that harmonize very poorly with the predictions by the model suggested by Costello and Watts (2014). (c) 2015 APA, all rights reserved).
Compact Single Site Resolution Cold Atom Experiment for Adiabatic Quantum Computing
2016-02-03
goal of our scientific investigation is to demonstrate high fidelity and fast atom-atom entanglement between physically 1. REPORT DATE (DD-MM-YYYY) 4...of our scientific investigation is to demonstrate high fidelity and fast atom-atom entanglement between physically separated and optically addressed...Specifically, we will design and construct a set of compact single atom traps with integrated optics, suitable for heralded entanglement and loophole
Zhang, Lei; Pavlica, Egon; Zhong, Xiaolan; Liscio, Fabiola; Li, Songlin; Bratina, Gvido; Orgiu, Emanuele; Samorì, Paolo
2017-03-01
Crystalline dioctyl-3,4,9,10-perylenedicarboximide nanowires and 6,13-bis(triisopropylsilylethynyl) pentacene microplates are integrated into a vertical-yet-open asymmetrical heterojunction for the realization of a high-performance organic photovoltaic detector, which shows fast photoresponse, ultrahigh signal-to-noise ratio, and high sensitivity to weak light. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Improved Measures of Integrated Information
Tegmark, Max
2016-01-01
Although there is growing interest in measuring integrated information in computational and cognitive systems, current methods for doing so in practice are computationally unfeasible. Existing and novel integration measures are investigated and classified by various desirable properties. A simple taxonomy of Φ-measures is presented where they are each characterized by their choice of factorization method (5 options), choice of probability distributions to compare (3 × 4 options) and choice of measure for comparing probability distributions (7 options). When requiring the Φ-measures to satisfy a minimum of attractive properties, these hundreds of options reduce to a mere handful, some of which turn out to be identical. Useful exact and approximate formulas are derived that can be applied to real-world data from laboratory experiments without posing unreasonable computational demands. PMID:27870846
Probing satellite galaxies in the Local Group by using FAST
NASA Astrophysics Data System (ADS)
Li, Jing; Wang, You-Gang; Kong, Min-Zhi; Wang, Jie; Chen, Xuelei; Guo, Rui
2018-01-01
The abundance of neutral hydrogen (HI) in satellite galaxies in the local group is important for studying the formation history of our local group. In this work, we generated mock HI satellite galaxies in the Local Group using the high mass-resolution hydrodynamic APOSTLE simulation. The simulated HI mass function agrees with the ALFALFA survey very well above 106 M ⊙, although there is a discrepancy below this scale because of the observed flux limit. After carefully checking various systematic elements in the observations, including fitting of line width, sky coverage, integration time and frequency drift due to uncertainty in a galaxy’s distance, we predicted the abundance of HI in galaxies in a future survey that will be conducted by FAST. FAST has a larger aperture and higher sensitivity than the Arecibo telescope. We found that the HI mass function could be estimated well around 105 M ⊙ if the integration time is 40 minutes. Our results indicate that there are 61 HI satellites in the Local Group and 36 in the FAST field above 105 M ⊙. This estimation is one order of magnitude better than the current data, and will put a strong constraint on the formation history of the Local Group. Also more high resolution simulated samples are needed to achieve this target.
Integral luminosities of radio pulsars
NASA Astrophysics Data System (ADS)
Malov, I.; Malov, O.
The integral radio luminosities L for 311 normal pulsars and for 27 ones with the rotation period P<0.1 s (fast pulsars) are calculated using new data on their compilated spectra. The values of L lie in the range 10^27-10^30 erg/s for 88% of the normal pulsars and in the range 10^28-10^31 erg/s for 88% of the fast objects. The high correlation between L and estimates l=S x d^2 from the known catalogues is detected. It is shown that the coefficient K of a transformation of the neutron star rotation energy into radio emission increases when P grows for normal pulsars and falls for fast ones. The mean values of K are -3.73 and -4.85 for normal and fast pulsars, respectively. There are no changes of L with the kinematic age T = z/V, where z is the pulsar height over the Galactic plane and V = 300 km/s is its mean velocity. The correlation between L and the rate of the rotation energy losses E is detected for both pulsar groups under consideration. It is shown that L= A E^(1/3) for the whole sample. The total number of pulsars in the Galaxy and their birth rate are in agreement with data on the rate of supernova explosions.
NASA Astrophysics Data System (ADS)
Seitenzahl, Ivo R.; Vogt, Frédéric P. A.; Terry, Jason P.; Ghavamian, Parviz; Dopita, Michael A.; Ruiter, Ashley J.; Sukhbold, Tuguldur
2018-02-01
We study the optical emission from heavy element ejecta in the oxygen-rich young supernova remnant 1E 0102.2–7219 (1E 0102) in the Small Magellanic Cloud. We have used the Multi-Unit Spectroscopic Explorer optical integral field spectrograph at the Very Large Telescope on Cerro Paranal and the wide field spectrograph (WiFeS) at the ANU 2.3 m telescope at Siding Spring Observatory to obtain deep observations of 1E 0102. Our observations cover the entire extent of the remnant from below 3500 Å to 9350 Å. Our observations unambiguously reveal the presence of fast-moving ejecta emitting in [S II], [S III], [Ar III], and [Cl II]. The sulfur-rich ejecta appear more asymmetrically distributed compared to oxygen or neon, a product of carbon burning. In addition to the forbidden line emission from products of oxygen burning (S, Ar, Cl), we have also discovered Hα and Hβ emission from several knots of low surface brightness, fast-moving ejecta. The presence of fast-moving hydrogen points toward a progenitor that had not entirely shed its hydrogen envelope prior to the supernova. The explosion that gave rise to 1E 0102 is therefore commensurate with a Type IIb supernova.
Brito, Maíra M; Lúcio, Cristina F; Angrimani, Daniel S R; Losano, João Diego A; Dalmazzo, Andressa; Nichi, Marcílio; Vannucchi, Camila I
2017-01-02
In addition to the existence of several cryopreservation protocols, no systematic research has been carried out in order to confirm the suitable protocol for canine sperm. This study aims to assess the effect of adding 5% glycerol during cryopreservation at 37°C (one-step) and 5°C (two-steps), in addition of testing two thawing protocols (37°C for 30 seconds, and 70°C for 8 seconds). We used 12 sperm samples divided into four experimental groups: Single-Step - Slow Thawing Group; Two-Step - Slow Thawing Group; Single-Step - Fast Thawing Group; and Two-Step - Fast Thawing Group. Frozen-thawed samples were submitted to automated analysis of sperm motility, evaluation of plasmatic membrane integrity, acrosomal integrity, mitochondrial activity, sperm morphology, sperm susceptibility to oxidative stress, and sperm binding assay to perivitellinic membrane of chicken egg yolk. Considering the comparison between freezing protocols, no statistical differences were verified for any of the response variables. When comparison between thawing protocols was performed, slow thawing protocol presented higher sperm count bound to perivitelline membrane of chicken egg yolk, compared to fast thawing protocol. Regardless of the freezing process, the slow thawing protocol can be recommended for the large scale cryopreservation of canine semen, since it shows a consistent better functional result.
Fründ, Ingo; Busch, Niko A; Schadow, Jeanette; Körner, Ursula; Herrmann, Christoph S
2007-01-01
Background Phase-locked gamma oscillations have so far mainly been described in relation to perceptual processes such as sensation, attention or memory matching. Due to its very short latency (≈90 ms) such oscillations are a plausible candidate for very rapid integration of sensory and motor processes. Results We measured EEG in 13 healthy participants in a speeded reaction task. Participants had to press a button as fast as possible whenever a visual stimulus was presented. The stimulus was always identical and did not have to be discriminated from other possible stimuli. In trials in which the participants showed a fast response, a slow negative potential over central electrodes starting approximately 800 ms before the response and highly phase-locked gamma oscillations over central and posterior electrodes between 90 and 140 ms after the stimulus were observed. In trials in which the participants showed a slow response, no slow negative potential was observed and phase-locked gamma oscillations were significantly reduced. Furthermore, for slow response trials the phase-locked gamma oscillations were significantly delayed with respect to fast response trials. Conclusion These results indicate the relevance of phase-locked gamma oscillations for very fast (not necessarily detailed) integration processes. PMID:17439642
Fast and fuel efficient? Optimal use of wind by flying albatrosses.
Weimerskirch, H; Guionnet, T; Martin, J; Shaffer, S A; Costa, D P
2000-09-22
The influence of wind patterns on behaviour and effort of free-ranging male wandering albatrosses (Diomedea exulans) was studied with miniaturized external heart-rate recorders in conjunction with satellite transmitters and activity recorders. Heart rate was used as an instantaneous index of energy expenditure. When cruising with favourable tail or side winds, wandering albatrosses can achieve high flight speeds while expending little more energy than birds resting on land. In contrast, heart rate increases concomitantly with increasing head winds, and flight speeds decrease. Our results show that effort is greatest when albatrosses take off from or land on the water. On a larger scale, we show that in order for birds to have the highest probability of experiencing favourable winds, wandering albatrosses use predictable weather systems to engage in a stereotypical flight pattern of large looping tracks. When heading north, albatrosses fly in anticlockwise loops, and to the south, movements are in a clockwise direction. Thus, the capacity to integrate instantaneous eco-physiological measures with records of large-scale flight and wind patterns allows us to understand better the complex interplay between the evolution of morphological, physiological and behavioural adaptations of albatrosses in the windiest place on earth.
NASA Technical Reports Server (NTRS)
Duffy, S. F.; Hu, J.; Hopkins, D. A.
1995-01-01
The article begins by examining the fundamentals of traditional deterministic design philosophy. The initial section outlines the concepts of failure criteria and limit state functions two traditional notions that are embedded in deterministic design philosophy. This is followed by a discussion regarding safety factors (a possible limit state function) and the common utilization of statistical concepts in deterministic engineering design approaches. Next the fundamental aspects of a probabilistic failure analysis are explored and it is shown that deterministic design concepts mentioned in the initial portion of the article are embedded in probabilistic design methods. For components fabricated from ceramic materials (and other similarly brittle materials) the probabilistic design approach yields the widely used Weibull analysis after suitable assumptions are incorporated. The authors point out that Weibull analysis provides the rare instance where closed form solutions are available for a probabilistic failure analysis. Since numerical methods are usually required to evaluate component reliabilities, a section on Monte Carlo methods is included to introduce the concept. The article concludes with a presentation of the technical aspects that support the numerical method known as fast probability integration (FPI). This includes a discussion of the Hasofer-Lind and Rackwitz-Fiessler approximations.
Lotka-Volterra systems in environments with randomly disordered temporal periodicity
NASA Astrophysics Data System (ADS)
Naess, Arvid; Dimentberg, Michael F.; Gaidai, Oleg
2008-08-01
A generalized Lotka-Volterra model for a pair of interacting populations of predators and prey is studied. The model accounts for the prey’s interspecies competition and therefore is asymptotically stable, whereas its oscillatory behavior is induced by temporal variations in environmental conditions simulated by those in the prey’s reproduction rate. Two models of the variations are considered, each of them combining randomness with “hidden” periodicity. The stationary joint probability density function (PDF) of the number of predators and prey is calculated numerically by the path integration (PI) method based on the use of characteristic functions and the fast Fourier transform. The numerical results match those for the asymptotic case of white-noise variations for which an analytical solution is available. Several examples are studied, with calculations of important characteristics of oscillations, for example the expected rate of up-crossings given the level of the predator number. The calculated PDFs may be of predominantly random (unimodal) or predominantly periodic nature (bimodal). Thus, the PI method has been demonstrated to be a powerful tool for studies of the dynamics of predator-prey pairs. The method captures the random oscillations as observed in nature, taking into account potential periodicity in the environmental conditions.
Lotka-Volterra systems in environments with randomly disordered temporal periodicity.
Naess, Arvid; Dimentberg, Michael F; Gaidai, Oleg
2008-08-01
A generalized Lotka-Volterra model for a pair of interacting populations of predators and prey is studied. The model accounts for the prey's interspecies competition and therefore is asymptotically stable, whereas its oscillatory behavior is induced by temporal variations in environmental conditions simulated by those in the prey's reproduction rate. Two models of the variations are considered, each of them combining randomness with "hidden" periodicity. The stationary joint probability density function (PDF) of the number of predators and prey is calculated numerically by the path integration (PI) method based on the use of characteristic functions and the fast Fourier transform. The numerical results match those for the asymptotic case of white-noise variations for which an analytical solution is available. Several examples are studied, with calculations of important characteristics of oscillations, for example the expected rate of up-crossings given the level of the predator number. The calculated PDFs may be of predominantly random (unimodal) or predominantly periodic nature (bimodal). Thus, the PI method has been demonstrated to be a powerful tool for studies of the dynamics of predator-prey pairs. The method captures the random oscillations as observed in nature, taking into account potential periodicity in the environmental conditions.
Foundation Mathematics for the Physical Sciences
NASA Astrophysics Data System (ADS)
Riley, K. F.; Hobson, M. P.
2011-03-01
1. Arithmetic and geometry; 2. Preliminary algebra; 3. Differential calculus; 4. Integral calculus; 5. Complex numbers and hyperbolic functions; 6. Series and limits; 7. Partial differentiation; 8. Multiple integrals; 9. Vector algebra; 10. Matrices and vector spaces; 11. Vector calculus; 12. Line, surface and volume integrals; 13. Laplace transforms; 14. Ordinary differential equations; 15. Elementary probability; Appendices; Index.
Student Solution Manual for Foundation Mathematics for the Physical Sciences
NASA Astrophysics Data System (ADS)
Riley, K. F.; Hobson, M. P.
2011-03-01
1. Arithmetic and geometry; 2. Preliminary algebra; 3. Differential calculus; 4. Integral calculus; 5. Complex numbers and hyperbolic functions; 6. Series and limits; 7. Partial differentiation; 8. Multiple integrals; 9. Vector algebra; 10. Matrices and vector spaces; 11. Vector calculus; 12. Line, surface and volume integrals; 13. Laplace transforms; 14. Ordinary differential equations; 15. Elementary probability; Appendix.
The description of two-photon Rabi oscillations in the path integral approach
NASA Astrophysics Data System (ADS)
Biryukov, A. A.; Degtyareva, Ya. V.; Shleenkov, M. A.
2018-04-01
The probability of quantum transitions of a molecule between its states under the action of an electromagnetic field is represented as an integral over trajectories from a real alternating functional. A method is proposed for computing the integral using recurrence relations. The method is attached to describe the two-photon Rabi oscillations.
NASA Technical Reports Server (NTRS)
Hou, Gene J.-W; Newman, Perry A. (Technical Monitor)
2004-01-01
A major step in a most probable point (MPP)-based method for reliability analysis is to determine the MPP. This is usually accomplished by using an optimization search algorithm. The minimum distance associated with the MPP provides a measurement of safety probability, which can be obtained by approximate probability integration methods such as FORM or SORM. The reliability sensitivity equations are derived first in this paper, based on the derivatives of the optimal solution. Examples are provided later to demonstrate the use of these derivatives for better reliability analysis and reliability-based design optimization (RBDO).
Error analysis for fast scintillator-based inertial confinement fusion burn history measurements
NASA Astrophysics Data System (ADS)
Lerche, R. A.; Ognibene, T. J.
1999-01-01
Plastic scintillator material acts as a neutron-to-light converter in instruments that make inertial confinement fusion burn history measurements. Light output for a detected neutron in current instruments has a fast rise time (<20 ps) and a relatively long decay constant (1.2 ns). For a burst of neutrons whose duration is much shorter than the decay constant, instantaneous light output is approximately proportional to the integral of the neutron interaction rate with the scintillator material. Burn history is obtained by deconvolving the exponential decay from the recorded signal. The error in estimating signal amplitude for these integral measurements is calculated and compared with a direct measurement in which light output is linearly proportional to the interaction rate.
A compact light-sheet microscope for the study of the mammalian central nervous system
Yang, Zhengyi; Haslehurst, Peter; Scott, Suzanne; Emptage, Nigel; Dholakia, Kishan
2016-01-01
Investigation of the transient processes integral to neuronal function demands rapid and high-resolution imaging techniques over a large field of view, which cannot be achieved with conventional scanning microscopes. Here we describe a compact light sheet fluorescence microscope, featuring a 45° inverted geometry and an integrated photolysis laser, that is optimized for applications in neuroscience, in particular fast imaging of sub-neuronal structures in mammalian brain slices. We demonstrate the utility of this design for three-dimensional morphological reconstruction, activation of a single synapse with localized photolysis, and fast imaging of neuronal Ca2+ signalling across a large field of view. The developed system opens up a host of novel applications for the neuroscience community. PMID:27215692
NASA Astrophysics Data System (ADS)
Dumazert, Jonathan; Coulon, Romain; Carrel, Frédérick; Corre, Gwenolé; Normand, Stéphane; Méchin, Laurence; Hamel, Matthieu
2016-08-01
Neutron detection forms a critical branch of nuclear-related issues, currently driven by the search for competitive alternative technologies to neutron counters based on the helium-3 isotope. The deployment of plastic scintillators shows a high potential for efficient detectors, safer and more reliable than liquids, more easily scalable and cost-effective than inorganic. In the meantime, natural gadolinium, through its 155 and mostly 157 isotopes, presents an exceptionally high interaction probability with thermal neutrons. This paper introduces a dual system including a metal gadolinium core inserted at the center of a high-scale plastic scintillator sphere. Incident fast neutrons are thermalized by the scintillator shell and then may be captured with a significant probability by gadolinium 155 and 157 nuclei in the core. The deposition of a sufficient fraction of the capture high-energy prompt gamma signature inside the scintillator shell will then allow discrimination from background radiations by energy threshold, and therefore neutron detection. The scaling of the system with the Monte Carlo MCNPX2.7 code was carried out according to a tradeoff between the moderation of incident fast neutrons and the probability of slow neutron capture by a moderate-cost metal gadolinium core. Based on the parameters extracted from simulation, a first laboratory prototype for the assessment of the detection method principle has been synthetized. The robustness and sensitivity of the neutron detection principle are then assessed by counting measurement experiments. Experimental results confirm the potential for a stable, highly sensitive, transportable and cost-efficient neutron detector and orientate future investigation toward promising axes.
Monternier, Pierre-Axel; Teulier, Loïc; Drai, Jocelyne; Bourguignon, Aurore; Collin-Chavagnac, Delphine; Hervant, Frédéric; Rouanet, Jean-Louis; Roussel, Damien
2017-10-01
Fasted endothermic vertebrates must develop physiological responses to maximize energy conservation and survival. The aim of this study was to determine the effect of 1-wk. fasting in 5-wk. old ducklings (Cairina moschata) from whole-body resting metabolic rate and body temperature to metabolic phenotype of tissues and mitochondrial coupling efficiency. At the level of whole organism, the mass-specific metabolic rate of ducklings was decreased by 40% after 1-wk. of fasting, which was associated with nocturnal Tb declines and shallow diurnal hypothermia during fasting. At the cellular level, fasting induced a large reduction in liver, gastrocnemius (oxidative) and pectoralis (glycolytic) muscle masses together with a fuel selection towards lipid oxidation and ketone body production in liver and a lower glycolytic phenotype in skeletal muscles. At the level of mitochondria, fasting induced a reduction of oxidative phosphorylation activities and an up-regulation of coupling efficiency (+30% on average) in liver and skeletal muscles. The present integrative study shows that energy conservation in fasted ducklings is mainly achieved by an overall reduction in mitochondrial activity and an increase in mitochondrial coupling efficiency, which would, in association with shallow hypothermia, increase the conservation of endogenous fuel stores during fasting. Copyright © 2017 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Wallace, Jon Michael
2003-10-01
Reliability prediction of components operating in complex systems has historically been conducted in a statistically isolated manner. Current physics-based, i.e. mechanistic, component reliability approaches focus more on component-specific attributes and mathematical algorithms and not enough on the influence of the system. The result is that significant error can be introduced into the component reliability assessment process. The objective of this study is the development of a framework that infuses the needs and influence of the system into the process of conducting mechanistic-based component reliability assessments. The formulated framework consists of six primary steps. The first three steps, identification, decomposition, and synthesis, are primarily qualitative in nature and employ system reliability and safety engineering principles to construct an appropriate starting point for the component reliability assessment. The following two steps are the most unique. They involve a step to efficiently characterize and quantify the system-driven local parameter space and a subsequent step using this information to guide the reduction of the component parameter space. The local statistical space quantification step is accomplished using two proposed multivariate probability models: Multi-Response First Order Second Moment and Taylor-Based Inverse Transformation. Where existing joint probability models require preliminary distribution and correlation information of the responses, these models combine statistical information of the input parameters with an efficient sampling of the response analyses to produce the multi-response joint probability distribution. Parameter space reduction is accomplished using Approximate Canonical Correlation Analysis (ACCA) employed as a multi-response screening technique. The novelty of this approach is that each individual local parameter and even subsets of parameters representing entire contributing analyses can now be rank ordered with respect to their contribution to not just one response, but the entire vector of component responses simultaneously. The final step of the framework is the actual probabilistic assessment of the component. Although the same multivariate probability tools employed in the characterization step can be used for the component probability assessment, variations of this final step are given to allow for the utilization of existing probabilistic methods such as response surface Monte Carlo and Fast Probability Integration. The overall framework developed in this study is implemented to assess the finite-element based reliability prediction of a gas turbine airfoil involving several failure responses. Results of this implementation are compared to results generated using the conventional 'isolated' approach as well as a validation approach conducted through large sample Monte Carlo simulations. The framework resulted in a considerable improvement to the accuracy of the part reliability assessment and an improved understanding of the component failure behavior. Considerable statistical complexity in the form of joint non-normal behavior was found and accounted for using the framework. Future applications of the framework elements are discussed.
NASA Astrophysics Data System (ADS)
Christen, Alejandra; Escarate, Pedro; Curé, Michel; Rial, Diego F.; Cassetti, Julia
2016-10-01
Aims: Knowing the distribution of stellar rotational velocities is essential for understanding stellar evolution. Because we measure the projected rotational speed v sin I, we need to solve an ill-posed problem given by a Fredholm integral of the first kind to recover the "true" rotational velocity distribution. Methods: After discretization of the Fredholm integral we apply the Tikhonov regularization method to obtain directly the probability distribution function for stellar rotational velocities. We propose a simple and straightforward procedure to determine the Tikhonov parameter. We applied Monte Carlo simulations to prove that the Tikhonov method is a consistent estimator and asymptotically unbiased. Results: This method is applied to a sample of cluster stars. We obtain confidence intervals using a bootstrap method. Our results are in close agreement with those obtained using the Lucy method for recovering the probability density distribution of rotational velocities. Furthermore, Lucy estimation lies inside our confidence interval. Conclusions: Tikhonov regularization is a highly robust method that deconvolves the rotational velocity probability density function from a sample of v sin I data directly without the need for any convergence criteria.
Quantum return probability of a system of N non-interacting lattice fermions
NASA Astrophysics Data System (ADS)
Krapivsky, P. L.; Luck, J. M.; Mallick, K.
2018-02-01
We consider N non-interacting fermions performing continuous-time quantum walks on a one-dimensional lattice. The system is launched from a most compact configuration where the fermions occupy neighboring sites. We calculate exactly the quantum return probability (sometimes referred to as the Loschmidt echo) of observing the very same compact state at a later time t. Remarkably, this probability depends on the parity of the fermion number—it decays as a power of time for even N, while for odd N it exhibits periodic oscillations modulated by a decaying power law. The exponent also slightly depends on the parity of N, and is roughly twice smaller than what it would be in the continuum limit. We also consider the same problem, and obtain similar results, in the presence of an impenetrable wall at the origin constraining the particles to remain on the positive half-line. We derive closed-form expressions for the amplitudes of the power-law decay of the return probability in all cases. The key point in the derivation is the use of Mehta integrals, which are limiting cases of the Selberg integral.
NASA Astrophysics Data System (ADS)
Keane, E. F.
2016-06-01
In this paper, we consider the fact that the simple criterion used to label fast radio transient events as either fast radio bursts (FRBs, thought to be extragalactic with as yet unknown progenitors) or rotating radio transients (RRATs, thought to be Galactic neutron stars) is uncertain. We identify single pulse events reported in the literature which have never been seen to repeat, and which have been labelled as RRATs, but are potentially mislabelled FRBs. We examine the probability that such `grey area' events are within the Milky Way. The uncertainty in the RRAT/FRB labelling criterion, as well as Galactic-latitude dependent reporting bias may be contributing to the observed latitude dependence of the FRB rate, in addition to effects such as Eddington bias due to scintillation.
Intrinsic Multi-Scale Dynamic Behaviors of Complex Financial Systems.
Ouyang, Fang-Yan; Zheng, Bo; Jiang, Xiong-Fei
2015-01-01
The empirical mode decomposition is applied to analyze the intrinsic multi-scale dynamic behaviors of complex financial systems. In this approach, the time series of the price returns of each stock is decomposed into a small number of intrinsic mode functions, which represent the price motion from high frequency to low frequency. These intrinsic mode functions are then grouped into three modes, i.e., the fast mode, medium mode and slow mode. The probability distribution of returns and auto-correlation of volatilities for the fast and medium modes exhibit similar behaviors as those of the full time series, i.e., these characteristics are rather robust in multi time scale. However, the cross-correlation between individual stocks and the return-volatility correlation are time scale dependent. The structure of business sectors is mainly governed by the fast mode when returns are sampled at a couple of days, while by the medium mode when returns are sampled at dozens of days. More importantly, the leverage and anti-leverage effects are dominated by the medium mode.
Fast and Epsilon-Optimal Discretized Pursuit Learning Automata.
Zhang, JunQi; Wang, Cheng; Zhou, MengChu
2015-10-01
Learning automata (LA) are powerful tools for reinforcement learning. A discretized pursuit LA is the most popular one among them. During an iteration its operation consists of three basic phases: 1) selecting the next action; 2) finding the optimal estimated action; and 3) updating the state probability. However, when the number of actions is large, the learning becomes extremely slow because there are too many updates to be made at each iteration. The increased updates are mostly from phases 1 and 3. A new fast discretized pursuit LA with assured ε -optimality is proposed to perform both phases 1 and 3 with the computational complexity independent of the number of actions. Apart from its low computational complexity, it achieves faster convergence speed than the classical one when operating in stationary environments. This paper can promote the applications of LA toward the large-scale-action oriented area that requires efficient reinforcement learning tools with assured ε -optimality, fast convergence speed, and low computational complexity for each iteration.
Newman, Ian R; Gibb, Maia; Thompson, Valerie A
2017-07-01
It is commonly assumed that belief-based reasoning is fast and automatic, whereas rule-based reasoning is slower and more effortful. Dual-Process theories of reasoning rely on this speed-asymmetry explanation to account for a number of reasoning phenomena, such as base-rate neglect and belief-bias. The goal of the current study was to test this hypothesis about the relative speed of belief-based and rule-based processes. Participants solved base-rate problems (Experiment 1) and conditional inferences (Experiment 2) under a challenging deadline; they then gave a second response in free time. We found that fast responses were informed by rules of probability and logical validity, and that slow responses incorporated belief-based information. Implications for Dual-Process theories and future research options for dissociating Type I and Type II processes are discussed. (PsycINFO Database Record (c) 2017 APA, all rights reserved).
Fast visible imaging of turbulent plasma in TORPEX
DOE Office of Scientific and Technical Information (OSTI.GOV)
Iraji, D.; Diallo, A.; Fasoli, A.
2008-10-15
Fast framing cameras constitute an important recent diagnostic development aimed at monitoring light emission from magnetically confined plasmas, and are now commonly used to study turbulence in plasmas. In the TORPEX toroidal device [A. Fasoli et al., Phys. Plasmas 13, 055902 (2006)], low frequency electrostatic fluctuations associated with drift-interchange waves are routinely measured by means of extensive sets of Langmuir probes. A Photron Ultima APX-RS fast framing camera has recently been acquired to complement Langmuir probe measurements, which allows comparing statistical and spectral properties of visible light and electrostatic fluctuations. A direct imaging system has been developed, which allows viewingmore » the light, emitted from microwave-produced plasmas tangentially and perpendicularly to the toroidal direction. The comparison of the probability density function, power spectral density, and autoconditional average of the camera data to those obtained using a multiple head electrostatic probe covering the plasma cross section shows reasonable agreement in the case of perpendicular view and in the plasma region where interchange modes dominate.« less
Proline Can Have Opposite Effects on Fast and Slow Protein Folding Phases
Osváth, Szabolcs; Gruebele, Martin
2003-01-01
Proline isomerization is well known to cause additional slow phases during protein refolding. We address a new question: does the presence of prolines significantly affect the very fast kinetics that lead to the formation of folding intermediates? We examined both the very slow (10–100 min) and very fast (4 μs–2.5 ms) folding kinetics of the two-domain enzyme yeast phosphoglycerate kinase by temperature-jump relaxation. Phosphoglycerate kinase contains a conserved cis-proline in position 204, in addition to several trans-prolines. Native cis-prolines have the largest effect on folding kinetics because the unfolded state favors trans isomerization, so we compared the kinetics of a P204H mutant with the wild-type as a proof of principle. The presence of Pro-204 causes an additional slow phase upon refolding from the cold denatured state, as reported in the literature. Contrary to this, the fast folding events are sped up in the presence of the cis-proline, probably by restriction of the conformational space accessible to the molecule. The wild-type and Pro204His mutant would be excellent models for off-lattice simulations probing the effects of conformational restriction on short timescales. PMID:12885665
ProbFAST: Probabilistic functional analysis system tool.
Silva, Israel T; Vêncio, Ricardo Z N; Oliveira, Thiago Y K; Molfetta, Greice A; Silva, Wilson A
2010-03-30
The post-genomic era has brought new challenges regarding the understanding of the organization and function of the human genome. Many of these challenges are centered on the meaning of differential gene regulation under distinct biological conditions and can be performed by analyzing the Multiple Differential Expression (MDE) of genes associated with normal and abnormal biological processes. Currently MDE analyses are limited to usual methods of differential expression initially designed for paired analysis. We proposed a web platform named ProbFAST for MDE analysis which uses Bayesian inference to identify key genes that are intuitively prioritized by means of probabilities. A simulated study revealed that our method gives a better performance when compared to other approaches and when applied to public expression data, we demonstrated its flexibility to obtain relevant genes biologically associated with normal and abnormal biological processes. ProbFAST is a free accessible web-based application that enables MDE analysis on a global scale. It offers an efficient methodological approach for MDE analysis of a set of genes that are turned on and off related to functional information during the evolution of a tumor or tissue differentiation. ProbFAST server can be accessed at http://gdm.fmrp.usp.br/probfast.
ProbFAST: Probabilistic Functional Analysis System Tool
2010-01-01
Background The post-genomic era has brought new challenges regarding the understanding of the organization and function of the human genome. Many of these challenges are centered on the meaning of differential gene regulation under distinct biological conditions and can be performed by analyzing the Multiple Differential Expression (MDE) of genes associated with normal and abnormal biological processes. Currently MDE analyses are limited to usual methods of differential expression initially designed for paired analysis. Results We proposed a web platform named ProbFAST for MDE analysis which uses Bayesian inference to identify key genes that are intuitively prioritized by means of probabilities. A simulated study revealed that our method gives a better performance when compared to other approaches and when applied to public expression data, we demonstrated its flexibility to obtain relevant genes biologically associated with normal and abnormal biological processes. Conclusions ProbFAST is a free accessible web-based application that enables MDE analysis on a global scale. It offers an efficient methodological approach for MDE analysis of a set of genes that are turned on and off related to functional information during the evolution of a tumor or tissue differentiation. ProbFAST server can be accessed at http://gdm.fmrp.usp.br/probfast. PMID:20353576
Electron theory of fast and ultrafast dissipative magnetization dynamics.
Fähnle, M; Illg, C
2011-12-14
For metallic magnets we review the experimental and electron-theoretical investigations of fast magnetization dynamics (on a timescale of ns to 100 ps) and of laser-pulse-induced ultrafast dynamics (few hundred fs). It is argued that for both situations the dominant contributions to the dissipative part of the dynamics arise from the excitation of electron-hole pairs and from the subsequent relaxation of these pairs by spin-dependent scattering processes, which transfer angular momentum to the lattice. By effective field theories (generalized breathing and bubbling Fermi-surface models) it is shown that the Gilbert equation of motion, which is often used to describe the fast dissipative magnetization dynamics, must be extended in several aspects. The basic assumptions of the Elliott-Yafet theory, which is often used to describe the ultrafast spin relaxation after laser-pulse irradiation, are discussed very critically. However, it is shown that for Ni this theory probably yields a value for the spin-relaxation time T(1) in good agreement with the experimental value. A relation between the quantity α characterizing the damping of the fast dynamics in simple situations and the time T(1) is derived. © 2011 IOP Publishing Ltd
FAST copper for broadband access
NASA Astrophysics Data System (ADS)
Chiang, Mung; Huang, Jianwei; Cendrillon, Raphael; Tan, Chee Wei; Xu, Dahai
2006-10-01
FAST Copper is a multi-year, U.S. NSF funded project that started in 2004, and is jointly pursued by the research groups of Mung Chiang at Princeton University, John Cioffi at Stanford University, and Alexader Fraser at Fraser Research Lab, and in collaboration with several industrial partners including AT&T. The goal of the FAST Copper Project is to provide ubiquitous, 100 Mbps, fiber/DSL broadband access to everyone in the U.S. with a phone line. This goal will be achieved through two threads of research: dynamic and joint optimization of resources in Frequency, Amplitude, Space, and Time (thus the name 'FAST') to overcome the attenuation and crosstalk bottlenecks, and the integration of communication, networking, computation, modeling, and distributed information management and control for the multi-user twisted pair network.
High-Resolution Strain Analysis of the Human Heart with Fast-DENSE
NASA Astrophysics Data System (ADS)
Aletras, Anthony H.; Balaban, Robert S.; Wen, Han
1999-09-01
Single breath-hold displacement data from the human heart were acquired with fast-DENSE (fast displacement encoding with stimulated echoes) during systolic contraction at 2.5 × 2.5 mm in-plane resolution. Encoding strengths of 0.86-1.60 mm/π were utilized in order to extend the dynamic range of the phase measurements and minimize effects of physiologic and instrument noise. The noise level in strain measurements for both contraction and dilation corresponded to a strain value of 2.8%. In the human heart, strain analysis has sufficient resolution to reveal transmural variation across the left ventricular wall. Data processing required minimal user intervention and provided a rapid quantitative feedback. The intrinsic temporal integration of fast-DENSE achieves high accuracy at the expense of temporal resolution.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, Yang; Pinterich, Tamara; Wang, Jian
We present rapid measurement of submicron particle size distributions enables the characterization of aerosols with fast changing properties, and is often necessary for measurements onboard mobile platforms (e.g., research aircraft). Aerosol mobility size distribution is commonly measured by a scanning mobility particle sizer (SMPS), which relies on voltage scanning or stepping to classify particles of different sizes, and may take up to several minutes to obtain a complete size spectrum of aerosol particles. The recently developed fast integrated mobility spectrometer (FIMS) with enhanced dynamic size range classifies and detects particles from 10 to ~600 nm simultaneously, allowing submicron aerosol mobilitymore » size distributions to be captured at a time resolution of 1 second. In this study, we present a detailed data inversion routine for deriving aerosol size distribution from FIMS measurements. The inversion routine takes into consideration the FIMS transfer function, particle penetration efficiency in the FIMS, and multiple charging of aerosols. The accuracy of the FIMS measurement is demonstrated by comparing parallel FIMS and SMPS measurements of stable aerosols with a wide range of size spectrum shapes, including ambient aerosols and aerosols classified by a differential mobility analyzer (DMA). The FIMS and SMPS-derived size distributions show excellent agreements for all aerosols tested. In addition, total number concentrations of ambient aerosols were integrated from 1 Hz FIMS size distributions, and compared with those directly measured by a condensation particle counter (CPC) operated in parallel. Finally, the integrated and measured total particle concentrations agree well within 5%.« less
Xue, Lu; Lin, Lin; Zhou, Wenbin; Chen, Wendong; Tang, Jun; Sun, Xiujie; Huang, Peiwu; Tian, Ruijun
2018-06-09
Plasma proteome profiling by LC-MS based proteomics has drawn great attention recently for biomarker discovery from blood liquid biopsy. Due to standard multi-step sample preparation could potentially cause plasma protein degradation and analysis variation, integrated proteomics sample preparation technologies became promising solution towards this end. Here, we developed a fully integrated proteomics sample preparation technology for both fast and deep plasma proteome profiling under its native pH. All the sample preparation steps, including protein digestion and two-dimensional fractionation by both mixed-mode ion exchange and high-pH reversed phase mechanism were integrated into one spintip device for the first time. The mixed-mode ion exchange beads design achieved the sample loading at neutral pH and protein digestion within 30 min. Potential sample loss and protein degradation by pH changing could be voided. 1 μL of plasma sample with depletion of high abundant proteins was processed by the developed technology with 12 equally distributed fractions and analyzed with 12 h of LC-MS gradient time, resulting in the identification of 862 proteins. The combination of the Mixed-mode-SISPROT and data-independent MS method achieved fast plasma proteome profiling in 2 h with high identification overlap and quantification precision for a proof-of-concept study of plasma samples from 5 healthy donors. We expect that the Mixed-mode-SISPROT become a generally applicable sample preparation technology for clinical oriented plasma proteome profiling. Copyright © 2018 Elsevier B.V. All rights reserved.
Wang, Yang; Pinterich, Tamara; Wang, Jian
2018-03-30
We present rapid measurement of submicron particle size distributions enables the characterization of aerosols with fast changing properties, and is often necessary for measurements onboard mobile platforms (e.g., research aircraft). Aerosol mobility size distribution is commonly measured by a scanning mobility particle sizer (SMPS), which relies on voltage scanning or stepping to classify particles of different sizes, and may take up to several minutes to obtain a complete size spectrum of aerosol particles. The recently developed fast integrated mobility spectrometer (FIMS) with enhanced dynamic size range classifies and detects particles from 10 to ~600 nm simultaneously, allowing submicron aerosol mobilitymore » size distributions to be captured at a time resolution of 1 second. In this study, we present a detailed data inversion routine for deriving aerosol size distribution from FIMS measurements. The inversion routine takes into consideration the FIMS transfer function, particle penetration efficiency in the FIMS, and multiple charging of aerosols. The accuracy of the FIMS measurement is demonstrated by comparing parallel FIMS and SMPS measurements of stable aerosols with a wide range of size spectrum shapes, including ambient aerosols and aerosols classified by a differential mobility analyzer (DMA). The FIMS and SMPS-derived size distributions show excellent agreements for all aerosols tested. In addition, total number concentrations of ambient aerosols were integrated from 1 Hz FIMS size distributions, and compared with those directly measured by a condensation particle counter (CPC) operated in parallel. Finally, the integrated and measured total particle concentrations agree well within 5%.« less
NASA Astrophysics Data System (ADS)
Heck, Martijn J. R.
2017-01-01
Technologies for efficient generation and fast scanning of narrow free-space laser beams find major applications in three-dimensional (3D) imaging and mapping, like Lidar for remote sensing and navigation, and secure free-space optical communications. The ultimate goal for such a system is to reduce its size, weight, and power consumption, so that it can be mounted on, e.g. drones and autonomous cars. Moreover, beam scanning should ideally be done at video frame rates, something that is beyond the capabilities of current opto-mechanical systems. Photonic integrated circuit (PIC) technology holds the promise of achieving low-cost, compact, robust and energy-efficient complex optical systems. PICs integrate, for example, lasers, modulators, detectors, and filters on a single piece of semiconductor, typically silicon or indium phosphide, much like electronic integrated circuits. This technology is maturing fast, driven by high-bandwidth communications applications, and mature fabrication facilities. State-of-the-art commercial PICs integrate hundreds of elements, and the integration of thousands of elements has been shown in the laboratory. Over the last few years, there has been a considerable research effort to integrate beam steering systems on a PIC, and various beam steering demonstrators based on optical phased arrays have been realized. Arrays of up to thousands of coherent emitters, including their phase and amplitude control, have been integrated, and various applications have been explored. In this review paper, I will present an overview of the state of the art of this technology and its opportunities, illustrated by recent breakthroughs.
Integrating modeling and surveys for more effective assessments
A false dichotomy currently exists in monitoring that pits sample surveys based on probability designs against targeted monitoring of hand-picked sites. We maintain that judicious use of both, when designed to be integrated, produces assessments of greater value than either inde...
Stochastic static fault slip inversion from geodetic data with non-negativity and bound constraints
NASA Astrophysics Data System (ADS)
Nocquet, J.-M.
2018-07-01
Despite surface displacements observed by geodesy are linear combinations of slip at faults in an elastic medium, determining the spatial distribution of fault slip remains a ill-posed inverse problem. A widely used approach to circumvent the illness of the inversion is to add regularization constraints in terms of smoothing and/or damping so that the linear system becomes invertible. However, the choice of regularization parameters is often arbitrary, and sometimes leads to significantly different results. Furthermore, the resolution analysis is usually empirical and cannot be made independently of the regularization. The stochastic approach of inverse problems provides a rigorous framework where the a priori information about the searched parameters is combined with the observations in order to derive posterior probabilities of the unkown parameters. Here, I investigate an approach where the prior probability density function (pdf) is a multivariate Gaussian function, with single truncation to impose positivity of slip or double truncation to impose positivity and upper bounds on slip for interseismic modelling. I show that the joint posterior pdf is similar to the linear untruncated Gaussian case and can be expressed as a truncated multivariate normal (TMVN) distribution. The TMVN form can then be used to obtain semi-analytical formulae for the single, 2-D or n-D marginal pdf. The semi-analytical formula involves the product of a Gaussian by an integral term that can be evaluated using recent developments in TMVN probabilities calculations. Posterior mean and covariance can also be efficiently derived. I show that the maximum posterior (MAP) can be obtained using a non-negative least-squares algorithm for the single truncated case or using the bounded-variable least-squares algorithm for the double truncated case. I show that the case of independent uniform priors can be approximated using TMVN. The numerical equivalence to Bayesian inversions using Monte Carlo Markov chain (MCMC) sampling is shown for a synthetic example and a real case for interseismic modelling in Central Peru. The TMVN method overcomes several limitations of the Bayesian approach using MCMC sampling. First, the need of computer power is largely reduced. Second, unlike Bayesian MCMC-based approach, marginal pdf, mean, variance or covariance are obtained independently one from each other. Third, the probability and cumulative density functions can be obtained with any density of points. Finally, determining the MAP is extremely fast.
Derivation of an eigenvalue probability density function relating to the Poincaré disk
NASA Astrophysics Data System (ADS)
Forrester, Peter J.; Krishnapur, Manjunath
2009-09-01
A result of Zyczkowski and Sommers (2000 J. Phys. A: Math. Gen. 33 2045-57) gives the eigenvalue probability density function for the top N × N sub-block of a Haar distributed matrix from U(N + n). In the case n >= N, we rederive this result, starting from knowledge of the distribution of the sub-blocks, introducing the Schur decomposition and integrating over all variables except the eigenvalues. The integration is done by identifying a recursive structure which reduces the dimension. This approach is inspired by an analogous approach which has been recently applied to determine the eigenvalue probability density function for random matrices A-1B, where A and B are random matrices with entries standard complex normals. We relate the eigenvalue distribution of the sub-blocks to a many-body quantum state, and to the one-component plasma, on the pseudosphere.
Instruction in information structuring improves Bayesian judgment in intelligence analysts.
Mandel, David R
2015-01-01
An experiment was conducted to test the effectiveness of brief instruction in information structuring (i.e., representing and integrating information) for improving the coherence of probability judgments and binary choices among intelligence analysts. Forty-three analysts were presented with comparable sets of Bayesian judgment problems before and immediately after instruction. After instruction, analysts' probability judgments were more coherent (i.e., more additive and compliant with Bayes theorem). Instruction also improved the coherence of binary choices regarding category membership: after instruction, subjects were more likely to invariably choose the category to which they assigned the higher probability of a target's membership. The research provides a rare example of evidence-based validation of effectiveness in instruction to improve the statistical assessment skills of intelligence analysts. Such instruction could also be used to improve the assessment quality of other types of experts who are required to integrate statistical information or make probabilistic assessments.
Emotion and decision-making: affect-driven belief systems in anxiety and depression.
Paulus, Martin P; Yu, Angela J
2012-09-01
Emotion processing and decision-making are integral aspects of daily life. However, our understanding of the interaction between these constructs is limited. In this review, we summarize theoretical approaches that link emotion and decision-making, and focus on research with anxious or depressed individuals to show how emotions can interfere with decision-making. We integrate the emotional framework based on valence and arousal with a Bayesian approach to decision-making in terms of probability and value processing. We discuss how studies of individuals with emotional dysfunctions provide evidence that alterations of decision-making can be viewed in terms of altered probability and value computation. We argue that the probabilistic representation of belief states in the context of partially observable Markov decision processes provides a useful approach to examine alterations in probability and value representation in individuals with anxiety and depression, and outline the broader implications of this approach. Copyright © 2012. Published by Elsevier Ltd.
Emotion and decision-making: affect-driven belief systems in anxiety and depression
Paulus, Martin P.; Yu, Angela J.
2012-01-01
Emotion processing and decision-making are integral aspects of daily life. However, our understanding of the interaction between these constructs is limited. In this review, we summarize theoretical approaches to the link between emotion and decision-making, and focus on research with anxious or depressed individuals that reveals how emotions can interfere with decision-making. We integrate the emotional framework based on valence and arousal with a Bayesian approach to decision-making in terms of probability and value processing. We then discuss how studies of individuals with emotional dysfunctions provide evidence that alterations of decision-making can be viewed in terms of altered probability and value computation. We argue that the probabilistic representation of belief states in the context of partially observable Markov decision processes provides a useful approach to examine alterations in probability and value representation in individuals with anxiety and depression and outline the broader implications of this approach. PMID:22898207
DOE Office of Scientific and Technical Information (OSTI.GOV)
D. Apelian
2007-07-23
The objective of the project is to develop an integrated process for fast, high-temperature carburizing. The new process results in an order of magnitude reduction in cycle time compared to conventional carburizing and represents significant energy savings in addition to a corresponding reduction of scrap associated with distortion free carburizing steels.
Novel Methods for Electromagnetic Simulation and Design
2016-08-03
The resulting discretized integral equations are compatible with fast multipoleaccelerated solvers and will form the basis for high fidelity...expansion”) which are high-order, efficient and easy to use on arbitrarily triangulated surfaces. The resulting discretized integral equations are...created a user interface compatible with both low and high order discretizations , and implemented the generalized Debye approach of [4]. The
NASA Astrophysics Data System (ADS)
Porcar-Castell, A.; Atherton, J.; Rajewicz, P. A.; Riikonen, A.; Gebre, S.; Liu, W.; Aalto, J.; Bendoula, R.; Burkart, A.; Chen, H.; Erkkilä, K. M.; Feret, J. B.; Fernández-Marín, B.; García-Plazaola, J. I.; Hakala, T.; Hartikainen, S.; Honkavaara, E.; Ihalainen, J.; Julitta, T.; Kolari, P.; Kooijmans, L.; Levula, J.; Loponen, M.; Mac Arthur, A.; Magney, T.; Maseyk, K. S.; Mottus, M.; Neimane, S.; Oksa, E.; Osterman, G. B.; Robinson, I.; Robson, M. T.; Sabater, N.; Solanki, T.; Tikkanen, M.; Mäkipää, R.; Aro, E. M.; Rascher, U.; Frankenberg, C.; Kulmala, M. T.; Vesala, T.; Back, J. K.
2017-12-01
The use of solar-induced chlorophyll fluorescence (ChlF) as a tracer of photosynthesis is rapidly expanding with increasing numbers of measurements from towers, drones, aircrafts, or satellites. But how to integrate all the informative potential of these multiscale datasets? The connection between ChlF and photosynthesis takes place via multiple mechansisms that depend on the scale. At the leaf level, diurnal variations in ChlF may indicate changes in photochemical or non-photochemical quenching processes, whereas seasonal variations may indicate changes in the protein structure or pigment composition of the photosynthetic apparatus. At the canopy level, variations in ChlF may also reflect changes in total leaf area, canopy structure, species composition, changes in illumination or sun-target-sensor geometry, background properties, etc. At the pixel level, the dynamics of the atmosphere are also important. It is therefore essential to characterize the impact of factors that control ChlF and photosynthesis at each scale. A combination of multiscale and continuous experimentation and modelling is probably the best option to close the remaining knowledge gaps. The goal of the FAST campaign was to characterize the processes that control the ChlF signal dynamics at each scale, establishing a comprehensive dataset for multiscale hypothesis and model validation. The campaign took place in Hyytiälä (Southern Finland) and lasted for 6 months. Measurements expanded from the molecular to the satellite pixel level and from the picosecond to the seasonal scale, including multiple species, and providing a unique optical and phenomenological record of the multiscale spring recovery of photosynthesis in a boreal forest. Amongst others we measured and registered: leaf ChlF spectra, OJIP kinetics, PSI and PSII activity, photosynthetic gas exchange, carbonyl sulphide (COS), volatile organic compounds (VOCs), total leaf absorption, pigment concentrations, photosynthetic proteins, fluorescence lifetime, canopy SIF, CO2, water, COS, and VOC fluxes, as well as vertical profiles of forest SIF using a drone and target OCO-2 observations at 1x2km pixel resolution. We here present preliminary results from the FAST campaign which emphasize the variability and role of different controls across scales.
Fast Euler solver for transonic airfoils. I - Theory. II - Applications
NASA Technical Reports Server (NTRS)
Dadone, Andrea; Moretti, Gino
1988-01-01
Equations written in terms of generalized Riemann variables are presently integrated by inverting six bidiagonal matrices and two tridiagonal matrices, using an implicit Euler solver that is based on the lambda-formulation. The solution is found on a C-grid whose boundaries are very close to the airfoil. The fast solver is then applied to the computation of several flowfields on a NACA 0012 airfoil at various Mach number and alpha values, yielding results that are primarily concerned with transonic flows. The effects of grid fineness and boundary distances are analyzed; the code is found to be robust and accurate, as well as fast.
Ergül, Özgür
2011-11-01
Fast and accurate solutions of large-scale electromagnetics problems involving homogeneous dielectric objects are considered. Problems are formulated with the electric and magnetic current combined-field integral equation and discretized with the Rao-Wilton-Glisson functions. Solutions are performed iteratively by using the multilevel fast multipole algorithm (MLFMA). For the solution of large-scale problems discretized with millions of unknowns, MLFMA is parallelized on distributed-memory architectures using a rigorous technique, namely, the hierarchical partitioning strategy. Efficiency and accuracy of the developed implementation are demonstrated on very large problems involving as many as 100 million unknowns.
NASA Astrophysics Data System (ADS)
Nurmohammadi, Tofiq; Abbasian, Karim; Yadipour, Reza
2018-05-01
In this work, an ultra-fast all-optical plasmon induced transparency based on a metal–insulator–metal nanoplasmonic waveguide with two Kerr nonlinear ring resonators is studied. Two-dimensional simulations utilizing the finite-difference time-domain method are used to show an obvious optical bistability and significant switching mechanisms of the signal light by varying the pump-light intensity. The proposed all-optical switching based on plasmon induced transparency demonstrates femtosecond-scale feedback time (90 fs), meaning ultra-fast switching can be achieved. The presented all-optical switch may have potential significant applications in integrated optical circuits.
NASA Astrophysics Data System (ADS)
Kawai, Y.; Alton, G. D.; Bilheux, J.-C.
2005-12-01
An inexpensive, fast, and close to universal infiltration coating technique has been developed for fabricating fast diffusion-release ISOL targets. Targets are fabricated by deposition of finely divided (∼1 μm) compound materials in a paint-slurry onto highly permeable, complex structure reticulated-vitreous-carbon-foam (RVCF) matrices, followed by thermal heat treatment. In this article, we describe the coating method and present information on the physical integrity, uniformity of deposition, and matrix adherence of SiC, HfC and UC2 targets, destined for on-line use as targets at the Holifield Radioactive Ion Beam Facility (HRIBF).
NASA Astrophysics Data System (ADS)
Cajiao Vélez, F.; Kamiński, J. Z.; Krajewska, K.
2018-04-01
High-energy photoionization driven by short and circularly-polarized laser pulses is studied in the framework of the relativistic strong-field approximation. The saddle-point analysis of the integrals defining the probability amplitude is used to determine the general properties of the probability distributions. Additionally, an approximate solution to the saddle-point equation is derived. This leads to the concept of the three-dimensional spiral of life in momentum space, around which the ionization probability distribution is maximum. We demonstrate that such spiral is also obtained from a classical treatment.
An Equality in Stochastic Processes
1971-06-21
Neyman have dis- cussed extensively this model in their study of the probabilities of relapse, re- covery, and death for cancer patients [6] (see also Du...negative constant and vp a positive constant. Explicit formulas of the probabilities Pa )(to, t) have been derived in terms of vP.. and vP. (see [2...transitions from Sp occurring during (T, t); the probability of this sequence of events is (52) P’aja’(to 7-)[ vap d7-]P,’’ (TX t)- Integrating (52) from T
Chemical preconcentrator with integral thermal flow sensor
Manginell, Ronald P.; Frye-Mason, Gregory C.
2003-01-01
A chemical preconcentrator with integral thermal flow sensor can be used to accurately measure fluid flow rate in a microanalytical system. The thermal flow sensor can be operated in either constant temperature or constant power mode and variants thereof. The chemical preconcentrator with integral thermal flow sensor can be fabricated with the same MEMS technology as the rest of the microanlaytical system. Because of its low heat capacity, low-loss, and small size, the chemical preconcentrator with integral thermal flow sensor is fast and efficient enough to be used in battery-powered, portable microanalytical systems.
Colonius, Hans; Diederich, Adele
2011-07-01
The concept of a "time window of integration" holds that information from different sensory modalities must not be perceived too far apart in time in order to be integrated into a multisensory perceptual event. Empirical estimates of window width differ widely, however, ranging from 40 to 600 ms depending on context and experimental paradigm. Searching for theoretical derivation of window width, Colonius and Diederich (Front Integr Neurosci 2010) developed a decision-theoretic framework using a decision rule that is based on the prior probability of a common source, the likelihood of temporal disparities between the unimodal signals, and the payoff for making right or wrong decisions. Here, this framework is extended to the focused attention task where subjects are asked to respond to signals from a target modality only. Evoking the framework of the time-window-of-integration (TWIN) model, an explicit expression for optimal window width is obtained. The approach is probed on two published focused attention studies. The first is a saccadic reaction time study assessing the efficiency with which multisensory integration varies as a function of aging. Although the window widths for young and older adults differ by nearly 200 ms, presumably due to their different peripheral processing speeds, neither of them deviates significantly from the optimal values. In the second study, head saccadic reactions times to a perfectly aligned audiovisual stimulus pair had been shown to depend on the prior probability of spatial alignment. Intriguingly, they reflected the magnitude of the time-window widths predicted by our decision-theoretic framework, i.e., a larger time window is associated with a higher prior probability.
Ghavami, Behnam; Raji, Mohsen; Pedram, Hossein
2011-08-26
Carbon nanotube field-effect transistors (CNFETs) show great promise as building blocks of future integrated circuits. However, synthesizing single-walled carbon nanotubes (CNTs) with accurate chirality and exact positioning control has been widely acknowledged as an exceedingly complex task. Indeed, density and chirality variations in CNT growth can compromise the reliability of CNFET-based circuits. In this paper, we present a novel statistical compact model to estimate the failure probability of CNFETs to provide some material and process guidelines for the design of CNFETs in gigascale integrated circuits. We use measured CNT spacing distributions within the framework of detailed failure analysis to demonstrate that both the CNT density and the ratio of metallic to semiconducting CNTs play dominant roles in defining the failure probability of CNFETs. Besides, it is argued that the large-scale integration of these devices within an integrated circuit will be feasible only if a specific range of CNT density with an acceptable ratio of semiconducting to metallic CNTs can be adjusted in a typical synthesis process.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brown, Joseph; Pirrung, Meg; McCue, Lee Ann
FQC is software that facilitates quality control of FASTQ files by carrying out a QC protocol using FastQC, parsing results, and aggregating quality metrics into an interactive dashboard designed to richly summarize individual sequencing runs. The dashboard groups samples in dropdowns for navigation among the data sets, utilizes human-readable configuration files to manipulate the pages and tabs, and is extensible with CSV data.
Brown, Joseph; Pirrung, Meg; McCue, Lee Ann
2017-06-09
FQC is software that facilitates quality control of FASTQ files by carrying out a QC protocol using FastQC, parsing results, and aggregating quality metrics into an interactive dashboard designed to richly summarize individual sequencing runs. The dashboard groups samples in dropdowns for navigation among the data sets, utilizes human-readable configuration files to manipulate the pages and tabs, and is extensible with CSV data.
Superior model for fault tolerance computation in designing nano-sized circuit systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Singh, N. S. S., E-mail: narinderjit@petronas.com.my; Muthuvalu, M. S., E-mail: msmuthuvalu@gmail.com; Asirvadam, V. S., E-mail: vijanth-sagayan@petronas.com.my
2014-10-24
As CMOS technology scales nano-metrically, reliability turns out to be a decisive subject in the design methodology of nano-sized circuit systems. As a result, several computational approaches have been developed to compute and evaluate reliability of desired nano-electronic circuits. The process of computing reliability becomes very troublesome and time consuming as the computational complexity build ups with the desired circuit size. Therefore, being able to measure reliability instantly and superiorly is fast becoming necessary in designing modern logic integrated circuits. For this purpose, the paper firstly looks into the development of an automated reliability evaluation tool based on the generalizationmore » of Probabilistic Gate Model (PGM) and Boolean Difference-based Error Calculator (BDEC) models. The Matlab-based tool allows users to significantly speed-up the task of reliability analysis for very large number of nano-electronic circuits. Secondly, by using the developed automated tool, the paper explores into a comparative study involving reliability computation and evaluation by PGM and, BDEC models for different implementations of same functionality circuits. Based on the reliability analysis, BDEC gives exact and transparent reliability measures, but as the complexity of the same functionality circuits with respect to gate error increases, reliability measure by BDEC tends to be lower than the reliability measure by PGM. The lesser reliability measure by BDEC is well explained in this paper using distribution of different signal input patterns overtime for same functionality circuits. Simulation results conclude that the reliability measure by BDEC depends not only on faulty gates but it also depends on circuit topology, probability of input signals being one or zero and also probability of error on signal lines.« less
Operational NDT simulator, towards human factors integration in simulated probability of detection
NASA Astrophysics Data System (ADS)
Rodat, Damien; Guibert, Frank; Dominguez, Nicolas; Calmon, Pierre
2017-02-01
In the aeronautic industry, the performance demonstration of Non-Destructive Testing (NDT) procedures relies on Probability Of Detection (POD) analyses. This statistical approach measures the ability of the procedure to detect a flaw with regard to one of its characteristic dimensions. The inspection chain is evaluated as a whole, including equipment configuration, probe effciency but also operator manipulations. Traditionally, a POD study requires an expensive campaign during which several operators apply the procedure on a large set of representative samples. Recently, new perspectives for the POD estimation have been introduced using NDT simulation to generate data. However, these approaches do not offer straightforward solutions to take the operator into account. The simulation of human factors, including cognitive aspects, often raises questions. To address these diffculties, we propose a concept of operational NDT simulator [1]. This work presents the first steps in the implementation of such simulator for ultrasound phased array inspection of composite parts containing Flat Bottom Holes (FBHs). The final system will look like a classical ultrasound testing equipment with a single exception: the displayed signals will be synthesized. Our hardware (ultrasound acquisition card, 3D position tracker) and software (position analysis, inspection scenario, synchronization, simulations) environments are developed as a bench to test the meta-modeling techniques able to provide fast-simulated realistic ultra-sound signals. The results presented here are obtained by on-the-fly merging of real and simulated signals. They confirm the feasibility of our approach: the replacement of real signals by purely simulated ones has been unnoticed by operators. We believe this simulator is a great prospect for POD evaluation including human factors, and may also find applications for training or procedure set-up.
Zhang, Yanan; Hu, Guiping; Brown, Robert C
2014-04-01
This study investigates the optimal supply chain design for commodity chemicals (BTX, etc.) production via woody biomass fast pyrolysis and hydroprocessing pathway. The locations and capacities of distributed preprocessing hubs and integrated biorefinery facilities are optimized with a mixed integer linear programming model. In this integrated supply chain system, decisions on the biomass chipping methods (roadside chipping vs. facility chipping) are also explored. The economic objective of the supply chain model is to maximize the profit for a 20-year chemicals production system. In addition to the economic objective, the model also incorporates an environmental objective of minimizing life cycle greenhouse gas emissions, analyzing the trade-off between the economic and environmental considerations. The capital cost, operating cost, and revenues for the biorefinery facilities are based on techno-economic analysis, and the proposed approach is illustrated through a case study of Minnesota, with Minneapolis-St. Paul serving as the chemicals distribution hub. Copyright © 2014 Elsevier Ltd. All rights reserved.
Diagnostic layer integration in FPGA-based pipeline measurement systems for HEP experiments
NASA Astrophysics Data System (ADS)
Pozniak, Krzysztof T.
2007-08-01
Integrated triggering and data acquisition systems for high energy physics experiments may be considered as fast, multichannel, synchronous, distributed, pipeline measurement systems. A considerable extension of functional, technological and monitoring demands, which has recently been imposed on them, forced a common usage of large field-programmable gate array (FPGA), digital signal processing-enhanced matrices and fast optical transmission for their realization. This paper discusses modelling, design, realization and testing of pipeline measurement systems. A distribution of synchronous data stream flows is considered in the network. A general functional structure of a single network node is presented. A suggested, novel block structure of the node model facilitates full implementation in the FPGA chip, circuit standardization and parametrization, as well as integration of functional and diagnostic layers. A general method for pipeline system design was derived. This method is based on a unified model of the synchronous data network node. A few examples of practically realized, FPGA-based, pipeline measurement systems were presented. The described systems were applied in ZEUS and CMS.
NASA Technical Reports Server (NTRS)
Collins, Jeffery D.; Volakis, John L.; Jin, Jian-Ming
1990-01-01
A new technique is presented for computing the scattering by 2-D structures of arbitrary composition. The proposed solution approach combines the usual finite element method with the boundary-integral equation to formulate a discrete system. This is subsequently solved via the conjugate gradient (CG) algorithm. A particular characteristic of the method is the use of rectangular boundaries to enclose the scatterer. Several of the resulting boundary integrals are therefore convolutions and may be evaluated via the fast Fourier transform (FFT) in the implementation of the CG algorithm. The solution approach offers the principal advantage of having O(N) memory demand and employs a 1-D FFT versus a 2-D FFT as required with a traditional implementation of the CGFFT algorithm. The speed of the proposed solution method is compared with that of the traditional CGFFT algorithm, and results for rectangular bodies are given and shown to be in excellent agreement with the moment method.
High resolution, high rate X-ray spectrometer
Goulding, Frederick S.; Landis, Donald A.
1987-01-01
A pulse processing system (10) for use in an X-ray spectrometer in which a ain channel pulse shaper (12) and a fast channel pulse shaper (13) each produce a substantially symmetrical triangular pulse (f, p) for each event detected by the spectrometer, with the pulse width of the pulses being substantially independent of the magnitude of the detected event and with the pulse width of the fast pulses (p) being substantially shorter than the pulse width of the main channel pulses (f). A pile-up rejector circuit (19) allows output pulses to be generated, with amplitudes linearly related to the magnitude of the detected events, whenever the peak of a main channel pulse (f) is not affected by a preceding or succeeding main channel pulse, while inhibiting output pulses wherein peak magnitudes of main channel pulses are affected by adjacent pulses. The substantially symmetrical triangular main channel pulses (f) are generated by the weighted addition (27-31) of successive RC integrations (24, 25, 26) of an RC differentiated step wave (23). The substantially symmetrical triangular fast channel pulses (p) are generated by the RC integration ( 43) of a bipolar pulse (o) in which the amplitude of the second half is 1/e that of the first half, with the RC time constant of integration being equal to one-half the width of the bipolar pulse.
Genetic evidence for role of integration of fast and slow neurotransmission in schizophrenia.
Devor, A; Andreassen, O A; Wang, Y; Mäki-Marttunen, T; Smeland, O B; Fan, C-C; Schork, A J; Holland, D; Thompson, W K; Witoelar, A; Chen, C-H; Desikan, R S; McEvoy, L K; Djurovic, S; Greengard, P; Svenningsson, P; Einevoll, G T; Dale, A M
2017-06-01
The most recent genome-wide association studies (GWAS) of schizophrenia (SCZ) identified hundreds of risk variants potentially implicated in the disease. Further, novel statistical methodology designed for polygenic architecture revealed more potential risk variants. This can provide a link between individual genetic factors and the mechanistic underpinnings of SCZ. Intriguingly, a large number of genes coding for ionotropic and metabotropic receptors for various neurotransmitters-glutamate, γ-aminobutyric acid (GABA), dopamine, serotonin, acetylcholine and opioids-and numerous ion channels were associated with SCZ. Here, we review these findings from the standpoint of classical neurobiological knowledge of neuronal synaptic transmission and regulation of electrical excitability. We show that a substantial proportion of the identified genes are involved in intracellular cascades known to integrate 'slow' (G-protein-coupled receptors) and 'fast' (ionotropic receptors) neurotransmission converging on the protein DARPP-32. Inspection of the Human Brain Transcriptome Project database confirms that that these genes are indeed expressed in the brain, with the expression profile following specific developmental trajectories, underscoring their relevance to brain organization and function. These findings extend the existing pathophysiology hypothesis by suggesting a unifying role of dysregulation in neuronal excitability and synaptic integration in SCZ. This emergent model supports the concept of SCZ as an 'associative' disorder-a breakdown in the communication across different slow and fast neurotransmitter systems through intracellular signaling pathways-and may unify a number of currently competing hypotheses of SCZ pathophysiology.
NASA Astrophysics Data System (ADS)
England, Troy; Curry, Matthew; Carr, Steve; Swartzentruber, Brian; Lilly, Michael; Bishop, Nathan; Carrol, Malcolm
2015-03-01
Fast, low-power quantum state readout is one of many challenges facing quantum information processing. Single electron transistors (SETs) are potentially fast, sensitive detectors for performing spin readout of electrons bound to Si:P donors. From a circuit perspective, however, their output impedance and nonlinear conductance are ill suited to drive the parasitic capacitance typical of coaxial conductors used in cryogenic environments, necessitating a cryogenic amplification stage. We will discuss calibration data, as well as modeling and simulation of cryogenic silicon-germanium (SiGe) heterojunction bipolar transistor (HBT) circuits connected to a silicon SET and operating at 4 K. We find a continuum of solutions from simple, single-HBT amplifiers to more complex, multi-HBT circuits suitable for integration, with varying noise levels and power vs. bandwidth tradeoffs. This work was performed, in part, at the Center for Integrated Nanotechnologies, a U.S. DOE Office of Basic Energy Sciences user facility. Sandia National Laboratories is a multi-program laboratory operated by Sandia Corporation, a Lockheed-Martin Company, for the U. S. Department of Energy under Contract No. DE-AC04-94AL85000.
Latash, M L
1994-01-01
Predictions of three models of single-joint motor control were compared with experimental observations of the changes in electromyographic (EMG) patterns during fast voluntary movements against an unexpectedly reduced inertial load. The subjects performed elbow flexions over 40 degrees "as fast as possible" in two series. During the first series, an approximately 40% decrease in inertia, simulated by a torque-motor, might occur unpredictably on half of the trials (unloaded trials). During the second series, all the trials were unloaded. The major findings are: (1) no differences in the antagonist burst latency in unexpectedly unloaded and unperturbed trials; (2) a decrease in the antagonist latency during expected unloadings; (3) a small, statistically non significant decrease in the first agonist burst EMG integral; and (4) a larger, statistically significant increase in the antagonist burst EMG integral in unexpectedly unloaded trials as compared to unperturbed trials. The data are in good correspondence with a version of the equilibrium-point hypothesis that assumes central programming of the beginning of the antagonist burst and incorporates the possibility of reflex-induced changes in EMG amplitudes.
Valentino, Giovanna; Kramer, Verónica; Bustamante, María José; Casasbellas, Cinthia; Adasme, Marcela; Salazar, Alejandra; Acevedo, Mónica
2015-01-01
Background. Impaired fasting glucose (IFG) through the nondiabetic range (100–125 mg/dL) is not considered in the cardiovascular (CV) risk profile. Aim. To compare the clustering of CV risk factors (RFs) in nondiabetic subjects with normal fasting glucose (NFG) and IFG. Material and Methods. Cross-sectional study in 3739 nondiabetic subjects. Demographics, medical history, and CV risk factors were collected and lipid profile, fasting glucose levels (FBG), C-reactive protein (hsCRP), blood pressure (BP), anthropometric measurements, and aerobic capacity were determined. Results. 559 (15%) subjects had IFG: they had a higher mean age, BMI, waist circumference, non-HDL cholesterol, BP, and hsCRP (p < 0.0001) and lower HDL (p < 0.001) and aerobic capacity (p < 0.001). They also had a higher prevalence of hypertension (34% versus 25%; p < 0.001), dyslipidemia (79% versus 74%; p < 0.001), and obesity (29% versus 16%; p < 0.001) and a higher Framingham risk score (8% versus 6%; p < 0.001). The probability of presenting 3 or more CV RFs adjusted by age and gender was significantly higher in the top quintile of fasting glucose (≥98 mg/dL; OR = 2.02; 1.62–2.51). Conclusions. IFG in the nondiabetic range is associated with increased cardiovascular RF clustering. PMID:26504260
Five months of fasting in king penguin chicks: body mass loss and fuel metabolism.
Cherel, Y; Le Maho, Y
1985-10-01
When king penguin chicks are 3-4 mo old, they enter a season of interrupted growth due to long periods of fasting, because they are irregularly fed in winter. Nine captive chicks [mean initial body mass (m) = 12.5 kg] had fasted an average of 5 mo at the end of the experiment; m was then 4.0 kg, a 68% decrease. They probably could have fasted longer, since chicks of parents delayed in the return to the colony die from starvation at an m of 3.0 kg. The long fast could be divided into three periods based on the changes in rate of decrease of m. The remarkable resistance of king penguin chicks to starvation may be partly explained by their ability to maintain protein sparing for as much as 4 mo, the duration of period II; plasma concentrations of uric acid, urea, and alanine were then minimum, 0.1, 0.4, and 0.4 mmol X l-1 respectively. Particular changes during this period, i.e., progressive increase of beta-hydroxybutyrate and decrease of glucose concentrations, might contribute to the efficiency of protein sparing. Period III was marked by a rise in protein utilization, plasma concentrations of uric acid, urea, and alanine increasing to 0.7, 1.5, and 0.8 mmol X l-1, respectively.
NASA Astrophysics Data System (ADS)
Straus, D. M.
2007-12-01
The probability distribution (pdf) of errors is followed in identical twin studies using the COLA T63 AGCM, integrated with observed SST for 15 recent winters. 30 integrations per winter (for 15 winters) are available with initial errors that are extremely small. The evolution of the pdf is tested for multi-modality, and the results interpreted in terms of clusters / regimes found in: (a) the set of 15x30 integrations mentioned, and (b) a larger ensemble of 55x15 integrations made with the same GCM using the same SSTs. The mapping of pdf evolution and clusters is also carried out for each winter separately, using the clusters found in the 55-member ensemble for the same winter alone. This technique yields information on the change in regimes caused by different boundary forcing (Straus and Molteni, 2004; Straus, Corti and Molteni, 2006). Analysis of the growing errors in terms of baroclinic and barotropic components allows for interpretation of the corresponding instabilities.
Enforcing positivity in intrusive PC-UQ methods for reactive ODE systems
Najm, Habib N.; Valorani, Mauro
2014-04-12
We explore the relation between the development of a non-negligible probability of negative states and the instability of numerical integration of the intrusive Galerkin ordinary differential equation system describing uncertain chemical ignition. To prevent this instability without resorting to either multi-element local polynomial chaos (PC) methods or increasing the order of the PC representation in time, we propose a procedure aimed at modifying the amplitude of the PC modes to bring the probability of negative state values below a user-defined threshold. This modification can be effectively described as a filtering procedure of the spectral PC coefficients, which is applied on-the-flymore » during the numerical integration when the current value of the probability of negative states exceeds the prescribed threshold. We demonstrate the filtering procedure using a simple model of an ignition process in a batch reactor. This is carried out by comparing different observables and error measures as obtained by non-intrusive Monte Carlo and Gauss-quadrature integration and the filtered intrusive procedure. Lastly, the filtering procedure has been shown to effectively stabilize divergent intrusive solutions, and also to improve the accuracy of stable intrusive solutions which are close to the stability limits.« less
Vadillo, Miguel A; Ortega-Castro, Nerea; Barberia, Itxaso; Baker, A G
2014-01-01
Many theories of causal learning and causal induction differ in their assumptions about how people combine the causal impact of several causes presented in compound. Some theories propose that when several causes are present, their joint causal impact is equal to the linear sum of the individual impact of each cause. However, some recent theories propose that the causal impact of several causes needs to be combined by means of a noisy-OR integration rule. In other words, the probability of the effect given several causes would be equal to the sum of the probability of the effect given each cause in isolation minus the overlap between those probabilities. In the present series of experiments, participants were given information about the causal impact of several causes and then they were asked what compounds of those causes they would prefer to use if they wanted to produce the effect. The results of these experiments suggest that participants actually use a variety of strategies, including not only the linear and the noisy-OR integration rules, but also averaging the impact of several causes.
NASA Astrophysics Data System (ADS)
Popov, V. D.; Khamidullina, N. M.
2006-10-01
In developing radio-electronic devices (RED) of spacecraft operating in the fields of ionizing radiation in space, one of the most important problems is the correct estimation of their radiation tolerance. The “weakest link” in the element base of onboard microelectronic devices under radiation effect is the integrated microcircuits (IMC), especially of large scale (LSI) and very large scale (VLSI) degree of integration. The main characteristic of IMC, which is taken into account when making decisions on using some particular type of IMC in the onboard RED, is the probability of non-failure operation (NFO) at the end of the spacecraft’s lifetime. It should be noted that, until now, the NFO has been calculated only from the reliability characteristics, disregarding the radiation effect. This paper presents the so-called “reliability” approach to determination of radiation tolerance of IMC, which allows one to estimate the probability of non-failure operation of various types of IMC with due account of radiation-stimulated dose failures. The described technique is applied to RED onboard the Spektr-R spacecraft to be launched in 2007.
Logic, probability, and human reasoning.
Johnson-Laird, P N; Khemlani, Sangeet S; Goodwin, Geoffrey P
2015-04-01
This review addresses the long-standing puzzle of how logic and probability fit together in human reasoning. Many cognitive scientists argue that conventional logic cannot underlie deductions, because it never requires valid conclusions to be withdrawn - not even if they are false; it treats conditional assertions implausibly; and it yields many vapid, although valid, conclusions. A new paradigm of probability logic allows conclusions to be withdrawn and treats conditionals more plausibly, although it does not address the problem of vapidity. The theory of mental models solves all of these problems. It explains how people reason about probabilities and postulates that the machinery for reasoning is itself probabilistic. Recent investigations accordingly suggest a way to integrate probability and deduction. Copyright © 2015 Elsevier Ltd. All rights reserved.
Community Integration and Quality of Life in Aphasia after Stroke.
Lee, Hyejin; Lee, Yuna; Choi, Hyunsoo; Pyun, Sung-Bom
2015-11-01
To examine community integration and contributing factors in people with aphasia (PWA) following stroke and to investigate the relationship between community integration and quality of life (QOL). Thirty PWA and 42 age-and education-matched control subjects were involved. Main variables were as follows: socioeconomic status, mobility, and activity of daily living (ADL) (Modified Barthel Index), language function [Frenchay Aphasia Screening Test (FAST)], depression [Geriatric Depression Scale (GDS)], Community Integration Questionnaire (CIQ) and Stroke and Aphasia Quality of Life Scale-39 (SAQOL-39). Differences between aphasia and control groups and factors affecting community integration and QOL were analyzed. Home and social integration and productive activity were significantly decreased in the aphasia group compared to the control group; 8.5 and 18.3 points in total CIQ score, respectively. Amount of time spent outside the home and frequency of social contact were also significantly reduced in the aphasia group. Total mean score on the SAQOL-39 was 2.75±0.80 points and was significantly correlated with economic status, gait performance, ADL, depressive mood, and social domain score on the CIQ. Depression score measured by GDS was the single most important factor for the prediction of QOL, but the FAST score was significantly correlated only with the communication domain of the SAQOL-39. Community activities of PWA were very limited, and depression was highly associated with decreased community integration and QOL. Enhancing social participation and reducing emotional distress should be emphasized for rehabilitation of PWA.
Development of a Nonlinear Probability of Collision Tool for the Earth Observing System
NASA Technical Reports Server (NTRS)
McKinley, David P.
2006-01-01
The Earth Observing System (EOS) spacecraft Terra, Aqua, and Aura fly in constellation with several other spacecraft in 705-kilometer mean altitude sun-synchronous orbits. All three spacecraft are operated by the Earth Science Mission Operations (ESMO) Project at Goddard Space Flight Center (GSFC). In 2004, the ESMO project began assessing the probability of collision of the EOS spacecraft with other space objects. In addition to conjunctions with high relative velocities, the collision assessment method for the EOS spacecraft must address conjunctions with low relative velocities during potential collisions between constellation members. Probability of Collision algorithms that are based on assumptions of high relative velocities and linear relative trajectories are not suitable for these situations; therefore an algorithm for handling the nonlinear relative trajectories was developed. This paper describes this algorithm and presents results from its validation for operational use. The probability of collision is typically calculated by integrating a Gaussian probability distribution over the volume swept out by a sphere representing the size of the space objects involved in the conjunction. This sphere is defined as the Hard Body Radius. With the assumption of linear relative trajectories, this volume is a cylinder, which translates into simple limits of integration for the probability calculation. For the case of nonlinear relative trajectories, the volume becomes a complex geometry. However, with an appropriate choice of coordinate systems, the new algorithm breaks down the complex geometry into a series of simple cylinders that have simple limits of integration. This nonlinear algorithm will be discussed in detail in the paper. The nonlinear Probability of Collision algorithm was first verified by showing that, when used in high relative velocity cases, it yields similar answers to existing high relative velocity linear relative trajectory algorithms. The comparison with the existing high velocity/linear theory will also be used to determine at what relative velocity the analysis should use the new nonlinear theory in place of the existing linear theory. The nonlinear algorithm was also compared to a known exact solution for the probability of collision between two objects when the relative motion is strictly circular and the error covariance is spherically symmetric. Figure I shows preliminary results from this comparison by plotting the probabilities calculated from the new algorithm and those from the exact solution versus the Hard Body Radius to Covariance ratio. These results show about 5% error when the Hard Body Radius is equal to one half the spherical covariance magnitude. The algorithm was then combined with a high fidelity orbit state and error covariance propagator into a useful tool for analyzing low relative velocity nonlinear relative trajectories. The high fidelity propagator is capable of using atmospheric drag, central body gravitational, solar radiation, and third body forces to provide accurate prediction of the relative trajectories and covariance evolution. The covariance propagator also includes a process noise model to ensure realistic evolutions of the error covariance. This paper will describe the integration of the nonlinear probability algorithm and the propagators into a useful collision assessment tool. Finally, a hypothetical case study involving a low relative velocity conjunction between members of the Earth Observation System constellation will be presented.
Validation of the SURE Program, phase 1
NASA Technical Reports Server (NTRS)
Dotson, Kelly J.
1987-01-01
Presented are the results of the first phase in the validation of the SURE (Semi-Markov Unreliability Range Evaluator) program. The SURE program gives lower and upper bounds on the death-state probabilities of a semi-Markov model. With these bounds, the reliability of a semi-Markov model of a fault-tolerant computer system can be analyzed. For the first phase in the validation, fifteen semi-Markov models were solved analytically for the exact death-state probabilities and these solutions compared to the corresponding bounds given by SURE. In every case, the SURE bounds covered the exact solution. The bounds, however, had a tendency to separate in cases where the recovery rate was slow or the fault arrival rate was fast.
A classification scheme for edge-localized modes based on their probability distributions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shabbir, A., E-mail: aqsa.shabbir@ugent.be; Max Planck Institute for Plasma Physics, D-85748 Garching; Hornung, G.
We present here an automated classification scheme which is particularly well suited to scenarios where the parameters have significant uncertainties or are stochastic quantities. To this end, the parameters are modeled with probability distributions in a metric space and classification is conducted using the notion of nearest neighbors. The presented framework is then applied to the classification of type I and type III edge-localized modes (ELMs) from a set of carbon-wall plasmas at JET. This provides a fast, standardized classification of ELM types which is expected to significantly reduce the effort of ELM experts in identifying ELM types. Further, themore » classification scheme is general and can be applied to various other plasma phenomena as well.« less
ERIC Educational Resources Information Center
Gibbons, Robert D.; And Others
The probability integral of the multivariate normal distribution (ND) has received considerable attention since W. F. Sheppard's (1900) and K. Pearson's (1901) seminal work on the bivariate ND. This paper evaluates the formula that represents the "n x n" correlation matrix of the "chi(sub i)" and the standardized multivariate…
Federal Register 2010, 2011, 2012, 2013, 2014
2011-08-22
... adversely affect plant safety, and would have no adverse effect on the probability of any accident. For the accidents that involve damage or melting of the fuel in the reactor core, fuel rod integrity has been shown to be unaffected by extended burnup under consideration; therefore, the probability of an accident...
Cordeiro, Aline; de Souza, Luana Lopes; Oliveira, Lorraine Soares; Faustino, Larissa Costa; Santiago, Letícia Aragão; Bloise, Flavia Fonseca; Ortiga-Carvalho, Tania Maria; Almeida, Norma Aparecida Dos Santos; Pazos-Moura, Carmen Cabanelas
2013-02-01
Sirtuin 1 (SIRT1), a NAD(+)-dependent deacetylase, has been connected to beneficial effects elicited by calorie restriction. Physiological adaptation to starvation requires higher activity of SIRT1 and also the suppression of thyroid hormone (TH) action to achieve energy conservation. Here, we tested the hypothesis that those two events are correlated and that TH may be a regulator of SIRT1 expression. Forty-eight-hour fasting mice exhibited reduced serum TH and increased SIRT1 protein content in liver and brown adipose tissue (BAT), and physiological thyroxine replacement prevented or attenuated the increment of SIRT1 in liver and BAT of fasted mice. Hypothyroid mice exhibited increased liver SIRT1 protein, while hyperthyroid ones showed decreased SIRT1 in liver and BAT. In the liver, decreased protein is accompanied by reduced SIRT1 activity and no alteration in its mRNA. Hyperthyroid and hypothyroid mice exhibited increases and decreases in food intake and body weight gain respectively. Food-restricted hyperthyroid animals (pair-fed to euthyroid group) exhibited liver and BAT SIRT1 protein levels intermediary between euthyroid and hyperthyroid mice fed ad libitum. Mice with TH resistance at the liver presented increased hepatic SIRT1 protein and activity, with no alteration in Sirt1 mRNA. These results suggest that TH decreases SIRT1 protein, directly and indirectly, via food ingestion control and, in the liver, this reduction involves TRβ. The SIRT1 reduction induced by TH has important implication to integrated metabolic responses to fasting, as the increase in SIRT1 protein requires the fasting-associated suppression of TH serum levels.
Preparatory Study for Constructing FAST, the World's Largest Single Dish
NASA Astrophysics Data System (ADS)
Peng, Bo; Jin, Chengjin; Wang, Qiming; Zhu, Lichun; Zhu, Wenbai; Zhang, Haiyan; Nan, Rendong
2009-08-01
A 500-m aperture spherical telescope (FAST) was funded by the National Development and Reform Commission of China (NDRC) in July 2007 and will be located in the unique Karst region, a sinkhole-like landform, in Guizhou province. FAST can be seen as a modified "Arecibo" type radio telescope using many innovative techniques, with as much as twice the collecting area and a wider sky coverage. FAST has, first, an active reflector, conforming to a paraboloid of revolution from a sphere in real time through actuated control, which enables the realization of wide bandwidth and full polarization capability by using standard feed design. Secondly, it has a light focus cabin suspension system, integrating optical, mechanical, and electronic technologies, reducing effectively the cost of the support structure and control system. With such a huge collecting area of more than 30 football fields, FAST will become the largest single dish ever built. Here we will summarize the FAST concept and the milestones achieved in experiments on its key technologies, i.e., site exploration, active reflector prototyping, focus cabin driving mechanism, measurement and control techniques, and the receiver layout. The Miyun FAST demonstrator also will be presented.
Preliminary investigation of a water-based method for fast integrating mobility spectrometry
DOE Office of Scientific and Technical Information (OSTI.GOV)
Spielman, Steven R.; Hering, Susanne V.; Kuang, Chongai
A water-based condensational growth channel was developed for imaging mobility-separated particles within a parallel plate separation channel of the Fast Integrated Mobility Spectrometer (FIMS). Reported are initial tests of that system, in which the alcohol condenser of the FIMS was replaced by a water-based condensational growth channel. Tests with monodispersed sodium chloride aerosol verify that the water-condensational growth maintained the laminar flow, while providing sufficient growth for particle imaging. Particle positions mapped onto particle mobility, in accordance with theoretical expectations. Particles ranging in size from 12 nm to 100 nm were counted with the same efficiency as with a butanol-based ultrafine particlemore » counter, once inlet and line losses were taken into account.« less
Fast and accurate quantum molecular dynamics of dense plasmas across temperature regimes
Sjostrom, Travis; Daligault, Jerome
2014-10-10
Here, we develop and implement a new quantum molecular dynamics approximation that allows fast and accurate simulations of dense plasmas from cold to hot conditions. The method is based on a carefully designed orbital-free implementation of density functional theory. The results for hydrogen and aluminum are in very good agreement with Kohn-Sham (orbital-based) density functional theory and path integral Monte Carlo calculations for microscopic features such as the electron density as well as the equation of state. The present approach does not scale with temperature and hence extends to higher temperatures than is accessible in the Kohn-Sham method and lowermore » temperatures than is accessible by path integral Monte Carlo calculations, while being significantly less computationally expensive than either of those two methods.« less
Preliminary investigation of a water-based method for fast integrating mobility spectrometry
Spielman, Steven R.; Hering, Susanne V.; Kuang, Chongai; ...
2017-06-06
A water-based condensational growth channel was developed for imaging mobility-separated particles within a parallel plate separation channel of the Fast Integrated Mobility Spectrometer (FIMS). Reported are initial tests of that system, in which the alcohol condenser of the FIMS was replaced by a water-based condensational growth channel. Tests with monodispersed sodium chloride aerosol verify that the water-condensational growth maintained the laminar flow, while providing sufficient growth for particle imaging. Particle positions mapped onto particle mobility, in accordance with theoretical expectations. Particles ranging in size from 12 nm to 100 nm were counted with the same efficiency as with a butanol-based ultrafine particlemore » counter, once inlet and line losses were taken into account.« less
Long lifetime fast spectrum reactor for lunar surface power system
NASA Astrophysics Data System (ADS)
Kambe, Mitsuru
1993-01-01
In the framework of innovative reactor research activities, a conceptual design study of fast spectrum reactor and primary system for 800 kWe lunar surface power system to be combined with potassium Rankine cycle power conversion has been conducted to meet the power requirements of the lunar base activities in the next century. The reactor subsystem is characterized by RAPID (Refueling by All Pins Integrated Design) concept to enhance inherent safety and to enable quick and simplifed refueling in every 10 years. RAPID concept affords power plant design lifetime of up to 30 years. Integrity of the reactor structure and replacement of failed primary circuits are also discussed. Substantial reduction in per-kWh cost on considering launch, emplacement, and final disposition can be expected by a long system lifetime.
Hansstein, Francesca Valeria; Hong, Yu; Di, Chen
2017-09-01
In recent decades, China has experienced an exponential growth in the number of internet users, especially among the youngest population, as well as a rapid proliferation of Western-type fast food restaurants. The health consequences of internet availability and fast food consumption among youth have been largely studied in Western countries, but few studies have focused on China. This paper has two goals. The first is to evaluate the differences in new media exposure and preferences for fast foods between rural and urban areas. The second goal is to test the association between new media exposure and fast food consumption. The targets of this analysis are Chinese children and adolescents aged 6-18 attending school at the time of the interview. Research hypotheses were tested using mean-groups comparisons for differences between rural urban sub-samples, and logistic regressions with odds ratios to estimate the relationship between media exposure and preferences towards fast foods. Cross-sectional data from the 2009 China Health and Nutrition Survey were employed. Watching online videos and playing computer games are behaviors associated with higher probabilities of eating at fast food restaurants in both rural and urban young residents, with higher odds in rural areas. Surfing the internet is associated with higher odds of being overweight in both rural and urban settings. Results also show that children living in rural areas spend significantly more time playing computer games, watching TV and videotapes, but less time doing homework than their urban peers. This paper suggests that monitoring the nutritional effects of new media exposure in China is of key importance in order to develop adequate health promotion policies, in both rural and urban areas.
Hallengren, Erik; Almgren, Peter; Rosvall, Maria; Östling, Gerd; Persson, Margaretha; Bergmann, Andreas; Struck, Joachim; Engström, Gunnar; Hedblad, Bo; Melander, Olle
2017-05-16
Growth hormone (GH) has been linked to cardiovascular disease but the exact mechanism of this association is still unclear. We here test if the fasting levels of GH are cross-sectionally associated with carotid intima media thickness (IMT) and whether treatment with fluvastatin affects the fasting level of GH. We examined the association between GH and IMT in 4425 individuals (aged 46-68 years) included in the baseline examination (1991-1994) of the Malmö Diet and Cancer cardiovascular cohort (MDC-CC). From that cohort we then studied 472 individuals (aged 50-70 years) who also participated (1994-1999) in the β-Blocker Cholesterol-Lowering Asymptomatic Plaque Study (BCAPS), a randomized, double blind, placebo-controlled, single-center clinical trial. Using multivariate linear regression models we related the change in GH-levels at 12 months compared with baseline to treatment with 40 mg fluvastatin once daily. In MDC-CC fasting values of GH exhibited a positive cross-sectional relation to the IMT at the carotid bulb independent of traditional cardiovascular risk factors (p = 0.002). In a gender-stratified analysis the correlation were significant for males (p = 0.005), but not for females (p = 0.09). Treatment with fluvastatin was associated with a minor reduction in the fasting levels of hs-GH in males (p = 0.05) and a minor rise in the same levels among females (p = 0.05). We here demonstrate that higher fasting levels of GH are associated with thicker IMT in the carotid bulb in males. Treatment with fluvastatin for 12 months only had a minor, and probably not clinically relevant, effect on the fasting levels of hs-GH.
Simpkins, D.G.; Hubert, W.A.; Del Rio, C.M.; Rule, D.C.
2003-01-01
The physiological traits that allow fish to survive periods of limited food resources are poorly understood. We assessed changes in proximate body composition, relative organ mass, blood metabolites, and relative weight (Wr) of sedentary and actively swimming (15 cm/s) juvenile rainbow trout (154-182 mm total length) over 147 d of fasting. Fasting caused measurable responses that were augmented when fish were swimming. Lipids and plasma triacylglycerides declined over time. Proteins were catabolized simultaneously with lipid reserves, but ammonia concentrations in plasma did not increase. The liver somatic index (LSI) did not change substantially over 105 d, suggesting that gluconeogenesis maintained blood glucose concentrations and hepatic glycogen reserves for a substantial period of fasting. The gut somatic index (GSI) and Wr declined linearly during fasting, but the LSI did not decline until after 105 d of fasting. Consequently, the use of different body condition indices could lead to different conclusions about the condition of juvenile rainbow trout. Swimming activity caused fish to have lower lipid and protein reserves than those of sedentary fish. No mortalities were observed among sedentary fish, but mortalities occurred among actively swimming fish after 97 d of fasting when 3.2% or less lipid remained in their bodies. Body condition indices did not account for differences in proximate body composition between sedentary and actively swimming fish and were relatively poor predictors of lipid content and risk of mortality. The probability of mortality was most accurately predicted by percent lipid content. Therefore, we suggest that fisheries scientists consider using percent lipid content when evaluating the physiological status and risk of mortality due to starvation among juvenile rainbow trout.
Jarrott, L. C.; McGuffey, C.; Beg, F. N.; ...
2017-10-24
Fast electron transport and spatial energy deposition are investigated in integrated cone-guided Fast Ignition experiments by measuring fast electron induced copper K-shell emission using a copper tracer added to deuterated plastic shells with a geometrically reentrant gold cone. Experiments were carried out at the Laboratory for Laser Energetics on the OMEGA/OMEGA-EP Laser where the plastic shells were imploded using 54 of the 60 OMEGA60 beams (3ω, 20 kJ), while the high intensity OMEGA-EP (BL2) beam (1 ω, 10 ps, 500 J, I peak > 10 19 W/cm 2) was focused onto the inner cone tip. Here, a retrograde analysis usingmore » the hybrid-PIC electron transport code, ZUMA, is performed to examine the sensitivity of the copper Kα spatial profile on the laser-produced fast electrons, facilitating the optimization of new target point designs and laser configurations to improve the compressed core areal density by a factor of 4 and the fast electron energy coupling by a factor of 3.5.« less
Bayesian networks for satellite payload testing
NASA Astrophysics Data System (ADS)
Przytula, Krzysztof W.; Hagen, Frank; Yung, Kar
1999-11-01
Satellite payloads are fast increasing in complexity, resulting in commensurate growth in cost of manufacturing and operation. A need exists for a software tool, which would assist engineers in production and operation of satellite systems. We have designed and implemented a software tool, which performs part of this task. The tool aids a test engineer in debugging satellite payloads during system testing. At this stage of satellite integration and testing both the tested payload and the testing equipment represent complicated systems consisting of a very large number of components and devices. When an error is detected during execution of a test procedure, the tool presents to the engineer a ranked list of potential sources of the error and a list of recommended further tests. The engineer decides this on this basis if to perform some of the recommended additional test or replace the suspect component. The tool has been installed in payload testing facility. The tool is based on Bayesian networks, a graphical method of representing uncertainty in terms of probabilistic influences. The Bayesian network was configured using detailed flow diagrams of testing procedures and block diagrams of the payload and testing hardware. The conditional and prior probability values were initially obtained from experts and refined in later stages of design. The Bayesian network provided a very informative model of the payload and testing equipment and inspired many new ideas regarding the future test procedures and testing equipment configurations. The tool is the first step in developing a family of tools for various phases of satellite integration and operation.
Motion streaks in fast motion rivalry cause orientation-selective suppression.
Apthorp, Deborah; Wenderoth, Peter; Alais, David
2009-05-14
We studied binocular rivalry between orthogonally translating arrays of random Gaussian blobs and measured the strength of rivalry suppression for static oriented probes. Suppression depth was quantified by expressing monocular probe thresholds during dominance relative to thresholds during suppression. Rivalry between two fast motions or two slow motions was compared in order to test the suggestion that fast-moving objects leave oriented "motion streaks" due to temporal integration (W. S. Geisler, 1999). If fast motions do produce motion streaks, then fast motion rivalry might also entail rivalry between the orthogonal streak orientations. We tested this using a static oriented probe that was aligned either parallel to the motion trajectory (hence collinear with the "streaks") or was orthogonal to the trajectory, predicting that rivalry suppression would be greater for parallel probes, and only for rivalry between fast motions. Results confirmed that suppression depth did depend on probe orientation for fast motion but not for slow motion. Further experiments showed that threshold elevations for the oriented probe during suppression exhibited clear orientation tuning. However, orientation-tuned elevations were also present during dominance, suggesting within-channel masking as the basis of the extra-deep suppression. In sum, the presence of orientation-dependent suppression in fast motion rivalry is consistent with the "motion streaks" hypothesis.
Singular solution of the Feller diffusion equation via a spectral decomposition.
Gan, Xinjun; Waxman, David
2015-01-01
Feller studied a branching process and found that the distribution for this process approximately obeys a diffusion equation [W. Feller, in Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability (University of California Press, Berkeley and Los Angeles, 1951), pp. 227-246]. This diffusion equation and its generalizations play an important role in many scientific problems, including, physics, biology, finance, and probability theory. We work under the assumption that the fundamental solution represents a probability density and should account for all of the probability in the problem. Thus, under the circumstances where the random process can be irreversibly absorbed at the boundary, this should lead to the presence of a Dirac delta function in the fundamental solution at the boundary. However, such a feature is not present in the standard approach (Laplace transformation). Here we require that the total integrated probability is conserved. This yields a fundamental solution which, when appropriate, contains a term proportional to a Dirac delta function at the boundary. We determine the fundamental solution directly from the diffusion equation via spectral decomposition. We obtain exact expressions for the eigenfunctions, and when the fundamental solution contains a Dirac delta function at the boundary, every eigenfunction of the forward diffusion operator contains a delta function. We show how these combine to produce a weight of the delta function at the boundary which ensures the total integrated probability is conserved. The solution we present covers cases where parameters are time dependent, thereby greatly extending its applicability.
Singular solution of the Feller diffusion equation via a spectral decomposition
NASA Astrophysics Data System (ADS)
Gan, Xinjun; Waxman, David
2015-01-01
Feller studied a branching process and found that the distribution for this process approximately obeys a diffusion equation [W. Feller, in Proceedings of the Second Berkeley Symposium on Mathematical Statistics and Probability (University of California Press, Berkeley and Los Angeles, 1951), pp. 227-246]. This diffusion equation and its generalizations play an important role in many scientific problems, including, physics, biology, finance, and probability theory. We work under the assumption that the fundamental solution represents a probability density and should account for all of the probability in the problem. Thus, under the circumstances where the random process can be irreversibly absorbed at the boundary, this should lead to the presence of a Dirac delta function in the fundamental solution at the boundary. However, such a feature is not present in the standard approach (Laplace transformation). Here we require that the total integrated probability is conserved. This yields a fundamental solution which, when appropriate, contains a term proportional to a Dirac delta function at the boundary. We determine the fundamental solution directly from the diffusion equation via spectral decomposition. We obtain exact expressions for the eigenfunctions, and when the fundamental solution contains a Dirac delta function at the boundary, every eigenfunction of the forward diffusion operator contains a delta function. We show how these combine to produce a weight of the delta function at the boundary which ensures the total integrated probability is conserved. The solution we present covers cases where parameters are time dependent, thereby greatly extending its applicability.
Advanced tokamak research with integrated modeling in JT-60 Upgrade
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hayashi, N.
2010-05-15
Researches on advanced tokamak (AT) have progressed with integrated modeling in JT-60 Upgrade [N. Oyama et al., Nucl. Fusion 49, 104007 (2009)]. Based on JT-60U experimental analyses and first principle simulations, new models were developed and integrated into core, rotation, edge/pedestal, and scrape-off-layer (SOL)/divertor codes. The integrated models clarified complex and autonomous features in AT. An integrated core model was implemented to take account of an anomalous radial transport of alpha particles caused by Alfven eigenmodes. It showed the reduction in the fusion gain by the anomalous radial transport and further escape of alpha particles. Integrated rotation model showed mechanismsmore » of rotation driven by the magnetic-field-ripple loss of fast ions and the charge separation due to fast-ion drift. An inward pinch model of high-Z impurity due to the atomic process was developed and indicated that the pinch velocity increases with the toroidal rotation. Integrated edge/pedestal model clarified causes of collisionality dependence of energy loss due to the edge localized mode and the enhancement of energy loss by steepening a core pressure gradient just inside the pedestal top. An ideal magnetohydrodynamics stability code was developed to take account of toroidal rotation and clarified a destabilizing effect of rotation on the pedestal. Integrated SOL/divertor model clarified a mechanism of X-point multifaceted asymmetric radiation from edge. A model of the SOL flow driven by core particle orbits which partially enter the SOL was developed by introducing the ion-orbit-induced flow to fluid equations.« less
Biochemical Profiles of Submariners: A Longitudinal Health Study
1975-05-15
probability curves . A slight skewing was demonstrated for total calcium, lactic dehydrogenase, alkaline phosphatase, and total protein. Blood urea...values for fasting blood sugar reported by Cutler and as- sociates4 and Craig and Bartholomew2 are essentially identical to our average values of...Longitudinal Health Study examina- tion. Upon reporting for study at 0700, a 7.5-ml vacuum tube without anticoagu- lant was used to obtain blood without
ERIC Educational Resources Information Center
Cameron, Stuart
2010-01-01
Bangladesh's urban population is rising fast. In the capital, Dhaka, some 4 million people live in slums. They are lacking in wealth, power and social connections; probably under-counted in national surveys; and under-served by both government and non-government organisations, many of whom still see poverty as a rural issue or see the urban poor…
On Algorithms for Generating Computationally Simple Piecewise Linear Classifiers
1989-05-01
suffers. - Waveform classification, e.g. speech recognition, seismic analysis (i.e. discrimination between earthquakes and nuclear explosions), target...assuming Gaussian distributions (B-G) d) Bayes classifier with probability densities estimated with the k-N-N method (B- kNN ) e) The -arest neighbour...range of classifiers are chosen including a fast, easy computable and often used classifier (B-G), reliable and complex classifiers (B- kNN and NNR
A fast combination method in DSmT and its application to recommender system
Liu, Yihai
2018-01-01
In many applications involving epistemic uncertainties usually modeled by belief functions, it is often necessary to approximate general (non-Bayesian) basic belief assignments (BBAs) to subjective probabilities (called Bayesian BBAs). This necessity occurs if one needs to embed the fusion result in a system based on the probabilistic framework and Bayesian inference (e.g. tracking systems), or if one needs to make a decision in the decision making problems. In this paper, we present a new fast combination method, called modified rigid coarsening (MRC), to obtain the final Bayesian BBAs based on hierarchical decomposition (coarsening) of the frame of discernment. Regarding this method, focal elements with probabilities are coarsened efficiently to reduce computational complexity in the process of combination by using disagreement vector and a simple dichotomous approach. In order to prove the practicality of our approach, this new approach is applied to combine users’ soft preferences in recommender systems (RSs). Additionally, in order to make a comprehensive performance comparison, the proportional conflict redistribution rule #6 (PCR6) is regarded as a baseline in a range of experiments. According to the results of experiments, MRC is more effective in accuracy of recommendations compared to original Rigid Coarsening (RC) method and comparable in computational time. PMID:29351297
NASA Astrophysics Data System (ADS)
Khryachkov, Vitaly; Goverdovskii, Andrei; Ketlerov, Vladimir; Mitrofanov, Vecheslav; Sergachev, Alexei
2018-03-01
Binary fission of 232Th and 238U induced by fast neutrons were under intent investigation in the IPPE during recent years. These measurements were performed with a twin ionization chamber with Frisch grids. Signals from the detector were digitized for further processing with a specially developed software. It results in information of kinetic energies, masses, directions and Bragg curves of registered fission fragments. Total statistics of a few million fission events were collected during each experiment. It was discovered that for several combinations of fission fragment masses their total kinetic energy was very close to total free energy of the fissioning system. The probability of such fission events for the fast neutron induced fission was found to be much higher than for spontaneous fission of 252Cf and thermal neutron induced fission of 235U. For experiments with 238U target the energy of incident neutrons were 5 MeV and 6.5 MeV. Close analysis of dependence of fission fragment distribution on compound nucleus excitation energy gave us some explanation of the phenomenon. It could be a process in highly excited compound nucleus which leads the fissioning system from the scission point into the fusion valley with high probability.
The Mid-Atlantic Integrated Assessment (MAIA-Estuaries) evaluated ecological conditions in US Mid-Atlantic estuaries during the summers of 1997 and 1998. Over 800 probability-based stations were monitored in four main estuarine systems?Chesapeake Bay, the Delaware Estuary, Maryla...
Simulator platform for fast reactor operation and safety technology demonstration
DOE Office of Scientific and Technical Information (OSTI.GOV)
Vilim, R. B.; Park, Y. S.; Grandy, C.
2012-07-30
A simulator platform for visualization and demonstration of innovative concepts in fast reactor technology is described. The objective is to make more accessible the workings of fast reactor technology innovations and to do so in a human factors environment that uses state-of-the art visualization technologies. In this work the computer codes in use at Argonne National Laboratory (ANL) for the design of fast reactor systems are being integrated to run on this platform. This includes linking reactor systems codes with mechanical structures codes and using advanced graphics to depict the thermo-hydraulic-structure interactions that give rise to an inherently safe responsemore » to upsets. It also includes visualization of mechanical systems operation including advanced concepts that make use of robotics for operations, in-service inspection, and maintenance.« less
Does fast or slow evaluation foster greater certainty?
Tormala, Zakary L; Clarkson, Joshua J; Henderson, Marlone D
2011-03-01
This research investigates the effect of perceived evaluation duration--that is, the perceived time or speed with which one generates an evaluation--on attitude certainty. Integrating diverse findings from past research, the authors propose that perceiving either fast or slow evaluation can augment attitude certainty depending on specifiable factors. Across three studies, it is shown that when people express opinions, evaluate familiar objects, or typically trust their gut reactions, perceiving fast rather than slow evaluation generally promotes greater certainty. In contrast, when people form opinions, evaluate unfamiliar objects, or typically trust more thoughtful responses, perceiving slow rather than fast evaluation generally promotes greater certainty. Mediation analyses reveal that these effects stem from trade-offs between perceived rational thought and the perceived ease of retrieving an attitude. Implications for research on deliberative versus intuitive decision making are discussed.
HI Intensity Mapping with FAST
NASA Astrophysics Data System (ADS)
Bigot-Sazy, M.-A.; Ma, Y.-Z.; Battye, R. A.; Browne, I. W. A.; Chen, T.; Dickinson, C.; Harper, S.; Maffei, B.; Olivari, L. C.; Wilkinsondagger, P. N.
2016-02-01
We discuss the detectability of large-scale HI intensity fluctuations using the FAST telescope. We present forecasts for the accuracy of measuring the Baryonic Acoustic Oscillations and constraining the properties of dark energy. The FAST 19-beam L-band receivers (1.05-1.45 GHz) can provide constraints on the matter power spectrum and dark energy equation of state parameters (w0,wa) that are comparable to the BINGO and CHIME experiments. For one year of integration time we find that the optimal survey area is 6000 deg2. However, observing with larger frequency coverage at higher redshift (0.95-1.35 GHz) improves the projected errorbars on the HI power spectrum by more than 2 σ confidence level. The combined constraints from FAST, CHIME, BINGO and Planck CMB observations can provide reliable, stringent constraints on the dark energy equation of state.
Focused Assessment with Sonography for Trauma in weightlessness: a feasibility study
NASA Technical Reports Server (NTRS)
Kirkpatrick, Andrew W.; Hamilton, Douglas R.; Nicolaou, Savvas; Sargsyan, Ashot E.; Campbell, Mark R.; Feiveson, Alan; Dulchavsky, Scott A.; Melton, Shannon; Beck, George; Dawson, David L.
2003-01-01
BACKGROUND: The Focused Assessment with Sonography for Trauma (FAST) examines for fluid in gravitationally dependent regions. There is no prior experience with this technique in weightlessness, such as on the International Space Station, where sonography is currently the only diagnostic imaging tool. STUDY DESIGN: A ground-based (1 g) porcine model for sonography was developed. We examined both the feasibility and the comparative performance of the FAST examination in parabolic flight. Sonographic detection and fluid behavior were evaluated in four animals during alternating weightlessness (0 g) and hypergravity (1.8 g) periods. During flight, boluses of fluid were incrementally introduced into the peritoneal cavity. Standardized sonographic windows were recorded. Postflight, the video recordings were divided into 169 20-second segments for subsequent interpretation by 12 blinded ultrasonography experts. Reviewers first decided whether a video segment was of sufficient diagnostic quality to analyze (determinate). Determinate segments were then analyzed as containing or not containing fluid. A probit regression model compared the probability of a positive fluid diagnosis to actual fluid levels (0 to 500 mL) under both 0-g and 1.8-g conditions. RESULTS: The in-flight sonographers found real-time scanning and interpretation technically similar to that of terrestrial conditions, as long as restraint was maintained. On blinded review, 80% of the recorded ultrasound segments were considered determinate. The best sensitivity for diagnosis in 0 g was found to be from the subhepatic space, with probability of a positive fluid diagnosis ranging from 9% (no fluid) to 51% (500 mL fluid). CONCLUSIONS: The FAST examination is technically feasible in weightlessness, and merits operational consideration for clinical contingencies in space.
Heavy inertial particles in turbulent flows gain energy slowly but lose it rapidly
NASA Astrophysics Data System (ADS)
Bhatnagar, Akshay; Gupta, Anupam; Mitra, Dhrubaditya; Pandit, Rahul
2018-03-01
We present an extensive numerical study of the time irreversibility of the dynamics of heavy inertial particles in three-dimensional, statistically homogeneous, and isotropic turbulent flows. We show that the probability density function (PDF) of the increment, W (τ ) , of a particle's energy over a time scale τ is non-Gaussian, and skewed toward negative values. This implies that, on average, particles gain energy over a period of time that is longer than the duration over which they lose energy. We call this slow gain and fast loss. We find that the third moment of W (τ ) scales as τ3 for small values of τ . We show that the PDF of power-input p is negatively skewed too; we use this skewness Ir as a measure of the time irreversibility and we demonstrate that it increases sharply with the Stokes number St for small St; this increase slows down at St≃1 . Furthermore, we obtain the PDFs of t+ and t-, the times over which p has, respectively, positive or negative signs, i.e., the particle gains or loses energy. We obtain from these PDFs a direct and natural quantification of the slow gain and fast loss of the energy of the particles, because these PDFs possess exponential tails from which we infer the characteristic loss and gain times tloss and tgain, respectively, and we obtain tloss
Stanley, Elise F
2015-01-01
At fast-transmitting presynaptic terminals Ca2+ enter through voltage gated calcium channels (CaVs) and bind to a synaptic vesicle (SV) -associated calcium sensor (SV-sensor) to gate fusion and discharge. An open CaV generates a high-concentration plume, or nanodomain of Ca2+ that dissipates precipitously with distance from the pore. At most fast synapses, such as the frog neuromuscular junction (NMJ), the SV sensors are located sufficiently close to individual CaVs to be gated by single nanodomains. However, at others, such as the mature rodent calyx of Held (calyx of Held), the physiology is more complex with evidence that CaVs that are both close and distant from the SV sensor and it is argued that release is gated primarily by the overlapping Ca2+ nanodomains from many CaVs. We devised a 'graphic modeling' method to sum Ca2+ from individual CaVs located at varying distances from the SV-sensor to determine the SV release probability and also the fraction of that probability that can be attributed to single domain gating. This method was applied first to simplified, low and high CaV density model release sites and then to published data on the contrasting frog NMJ and the rodent calyx of Held native synapses. We report 3 main predictions: the SV-sensor is positioned very close to the point at which the SV fuses with the membrane; single domain-release gating predominates even at synapses where the SV abuts a large cluster of CaVs, and even relatively remote CaVs can contribute significantly to single domain-based gating. PMID:26457441
Heavy inertial particles in turbulent flows gain energy slowly but lose it rapidly.
Bhatnagar, Akshay; Gupta, Anupam; Mitra, Dhrubaditya; Pandit, Rahul
2018-03-01
We present an extensive numerical study of the time irreversibility of the dynamics of heavy inertial particles in three-dimensional, statistically homogeneous, and isotropic turbulent flows. We show that the probability density function (PDF) of the increment, W(τ), of a particle's energy over a time scale τ is non-Gaussian, and skewed toward negative values. This implies that, on average, particles gain energy over a period of time that is longer than the duration over which they lose energy. We call this slow gain and fast loss. We find that the third moment of W(τ) scales as τ^{3} for small values of τ. We show that the PDF of power-input p is negatively skewed too; we use this skewness Ir as a measure of the time irreversibility and we demonstrate that it increases sharply with the Stokes number St for small St; this increase slows down at St≃1. Furthermore, we obtain the PDFs of t^{+} and t^{-}, the times over which p has, respectively, positive or negative signs, i.e., the particle gains or loses energy. We obtain from these PDFs a direct and natural quantification of the slow gain and fast loss of the energy of the particles, because these PDFs possess exponential tails from which we infer the characteristic loss and gain times t_{loss} and t_{gain}, respectively, and we obtain t_{loss}
NASA Astrophysics Data System (ADS)
Eilert, Tobias; Beckers, Maximilian; Drechsler, Florian; Michaelis, Jens
2017-10-01
The analysis tool and software package Fast-NPS can be used to analyse smFRET data to obtain quantitative structural information about macromolecules in their natural environment. In the algorithm a Bayesian model gives rise to a multivariate probability distribution describing the uncertainty of the structure determination. Since Fast-NPS aims to be an easy-to-use general-purpose analysis tool for a large variety of smFRET networks, we established an MCMC based sampling engine that approximates the target distribution and requires no parameter specification by the user at all. For an efficient local exploration we automatically adapt the multivariate proposal kernel according to the shape of the target distribution. In order to handle multimodality, the sampler is equipped with a parallel tempering scheme that is fully adaptive with respect to temperature spacing and number of chains. Since the molecular surrounding of a dye molecule affects its spatial mobility and thus the smFRET efficiency, we introduce dye models which can be selected for every dye molecule individually. These models allow the user to represent the smFRET network in great detail leading to an increased localisation precision. Finally, a tool to validate the chosen model combination is provided. Programme Files doi:http://dx.doi.org/10.17632/7ztzj63r68.1 Licencing provisions: Apache-2.0 Programming language: GUI in MATLAB (The MathWorks) and the core sampling engine in C++ Nature of problem: Sampling of highly diverse multivariate probability distributions in order to solve for macromolecular structures from smFRET data. Solution method: MCMC algorithm with fully adaptive proposal kernel and parallel tempering scheme.
Telehealth Education in Nursing Curricula.
Ali, Nagia S; Carlton, Kay Hodson; Ali, Omar S
2015-01-01
Telehealth care is a fast-growing avenue of providing health care services at a distance. A descriptive study was conducted to identify trends of telehealth education in 43 schools of nursing. Findings reflected inadequate integration of telehealth in classroom content, simulation, and clinical experiences. Interviews with 4 nursing leaders of telehealth provided some recommendations on how to integrate telehealth education in nursing curricula.
International Space Station Alpha (ISSA) Integrated Traffic Model
NASA Technical Reports Server (NTRS)
Gates, R. E.
1995-01-01
The paper discusses the development process of the International Space Station Alpha (ISSA) Integrated Traffic Model which is a subsystem analyses tool utilized in the ISSA design analysis cycles. Fast-track prototyping of the detailed relationships between daily crew and station consumables, propellant needs, maintenance requirements and crew rotation via spread sheets provide adequate benchmarks to assess cargo vehicle design and performance characteristics.
INTEGRAL gamma-ray upper limits on FRB180309
NASA Astrophysics Data System (ADS)
Savchenko, V.; Ferrigno, C.; Panessa, F.; Bazzano, A.; Ubertini, E. Kuulkers, P.; Keane, E.
2018-03-01
A very high signal-to-noise fast radio burst has been detected at the Parkes Telescope on 2018-03-09 at 02:49:32.99 UTC (ATeL #11385). The INTEGRAL observatory was taking data on a field centered at RA=87.04, Dec=19.32, 130 degrees from the approximate FRB arrival direction (RA=321.2 Dec=-33.8).
INTEGRAL gamma-ray upper limits on FRB180311
NASA Astrophysics Data System (ADS)
Savchenko, V.; Ferrigno, C.; Panessa, F.; Bazzano, A.; Ubertini, E.; Kuulkers, P.; Keane, E.
2018-03-01
A fast radio burst has been detected at the Parkes Telescope on 2018-03-11 at 04:11:54.80 UTC (ATeL #11396). The INTEGRAL observatory was taking data on a field centered at RA=260.177, Dec=-40.105, 43.0 degrees from the approximate FRB arrival direction (RA=21:31:33.42 Dec=-57:44:26.7).
Discrepancy-based error estimates for Quasi-Monte Carlo III. Error distributions and central limits
NASA Astrophysics Data System (ADS)
Hoogland, Jiri; Kleiss, Ronald
1997-04-01
In Quasi-Monte Carlo integration, the integration error is believed to be generally smaller than in classical Monte Carlo with the same number of integration points. Using an appropriate definition of an ensemble of quasi-random point sets, we derive various results on the probability distribution of the integration error, which can be compared to the standard Central Limit Theorem for normal stochastic sampling. In many cases, a Gaussian error distribution is obtained.
Integration of auditory and kinesthetic information in motion: alterations in Parkinson's disease.
Sabaté, Magdalena; Llanos, Catalina; Rodríguez, Manuel
2008-07-01
The main aim in this work was to study the interaction between auditory and kinesthetic stimuli and its influence on motion control. The study was performed on healthy subjects and patients with Parkinson's disease (PD). Thirty-five right-handed volunteers (young, PD, and age-matched healthy participants, and PD-patients) were studied with three different motor tasks (slow cyclic movements, fast cyclic movements, and slow continuous movements) and under the action of kinesthetic stimuli and sounds at different beat rates. The action of kinesthesia was evaluated by comparing real movements with virtual movements (movements imaged but not executed). The fast cyclic task was accelerated by kinesthetic but not by auditory stimuli. The slow cyclic task changed with the beat rate of sounds but not with kinesthetic stimuli. The slow continuous task showed an integrated response to both sensorial modalities. These data show that the influence of the multisensory integration on motion changes with the motor task and that some motor patterns are modulated by the simultaneous action of auditory and kinesthetic information, a cross-modal integration that was different in PD-patients. PsycINFO Database Record (c) 2008 APA, all rights reserved.
A fast non-local means algorithm based on integral image and reconstructed similar kernel
NASA Astrophysics Data System (ADS)
Lin, Zheng; Song, Enmin
2018-03-01
Image denoising is one of the essential methods in digital image processing. The non-local means (NLM) denoising approach is a remarkable denoising technique. However, its time complexity of the computation is high. In this paper, we design a fast NLM algorithm based on integral image and reconstructed similar kernel. First, the integral image is introduced in the traditional NLM algorithm. In doing so, it reduces a great deal of repetitive operations in the parallel processing, which will greatly improves the running speed of the algorithm. Secondly, in order to amend the error of the integral image, we construct a similar window resembling the Gaussian kernel in the pyramidal stacking pattern. Finally, in order to eliminate the influence produced by replacing the Gaussian weighted Euclidean distance with Euclidean distance, we propose a scheme to construct a similar kernel with a size of 3 x 3 in a neighborhood window which will reduce the effect of noise on a single pixel. Experimental results demonstrate that the proposed algorithm is about seventeen times faster than the traditional NLM algorithm, yet produce comparable results in terms of Peak Signal-to- Noise Ratio (the PSNR increased 2.9% in average) and perceptual image quality.
Integration of Fire Control, Flight Control and Propulsion Control Systems.
1983-08-01
process of preparation to a degree imocmpatible with fast moving technology. Undoubtedly something can be done to see that US authors produce papers...PDU provide a true "Up Front" system controller. Fast selection of Air to Air Mode is provided by 2 pushbuttons on the throttle. Pushing one or both...nature of the load snd its effect on the generation system. This calculation will include Load Power, Reactance, Nature - continuous or intermittent
Fast wavelet based algorithms for linear evolution equations
NASA Technical Reports Server (NTRS)
Engquist, Bjorn; Osher, Stanley; Zhong, Sifen
1992-01-01
A class was devised of fast wavelet based algorithms for linear evolution equations whose coefficients are time independent. The method draws on the work of Beylkin, Coifman, and Rokhlin which they applied to general Calderon-Zygmund type integral operators. A modification of their idea is applied to linear hyperbolic and parabolic equations, with spatially varying coefficients. A significant speedup over standard methods is obtained when applied to hyperbolic equations in one space dimension and parabolic equations in multidimensions.
Integrated Knowledge Elicitation and Representation Framework.
1991-04-01
applications. Given the fast paced, time-stressed battlefield environment, expert and decision sup- port systems will play an important role in future...can weaken them a little bit, it will help." I: "Why would they use dummy guns to divert us?" E: "We use imagery a lot from a fast moving airplane. A...found that subjects are often slow to update their assessment of a situation, even when new information conflicts with their old impressions. Leddo and
NASA Technical Reports Server (NTRS)
Keppenne, Christian L.; Rienecker, Michele M.; Kovach, Robin M.; Vernieres, Guillaume; Koster, Randal D. (Editor)
2014-01-01
An attractive property of ensemble data assimilation methods is that they provide flow dependent background error covariance estimates which can be used to update fields of observed variables as well as fields of unobserved model variables. Two methods to estimate background error covariances are introduced which share the above property with ensemble data assimilation methods but do not involve the integration of multiple model trajectories. Instead, all the necessary covariance information is obtained from a single model integration. The Space Adaptive Forecast error Estimation (SAFE) algorithm estimates error covariances from the spatial distribution of model variables within a single state vector. The Flow Adaptive error Statistics from a Time series (FAST) method constructs an ensemble sampled from a moving window along a model trajectory. SAFE and FAST are applied to the assimilation of Argo temperature profiles into version 4.1 of the Modular Ocean Model (MOM4.1) coupled to the GEOS-5 atmospheric model and to the CICE sea ice model. The results are validated against unassimilated Argo salinity data. They show that SAFE and FAST are competitive with the ensemble optimal interpolation (EnOI) used by the Global Modeling and Assimilation Office (GMAO) to produce its ocean analysis. Because of their reduced cost, SAFE and FAST hold promise for high-resolution data assimilation applications.