Design sensitivity analysis of nonlinear structural response
NASA Technical Reports Server (NTRS)
Cardoso, J. B.; Arora, J. S.
1987-01-01
A unified theory is described of design sensitivity analysis of linear and nonlinear structures for shape, nonshape and material selection problems. The concepts of reference volume and adjoint structure are used to develop the unified viewpoint. A general formula for design sensitivity analysis is derived. Simple analytical linear and nonlinear examples are used to interpret various terms of the formula and demonstrate its use.
Design sensitivity analysis using EAL. Part 1: Conventional design parameters
NASA Technical Reports Server (NTRS)
Dopker, B.; Choi, Kyung K.; Lee, J.
1986-01-01
A numerical implementation of design sensitivity analysis of builtup structures is presented, using the versatility and convenience of an existing finite element structural analysis code and its database management system. The finite element code used in the implemenatation presented is the Engineering Analysis Language (EAL), which is based on a hybrid method of analysis. It was shown that design sensitivity computations can be carried out using the database management system of EAL, without writing a separate program and a separate database. Conventional (sizing) design parameters such as cross-sectional area of beams or thickness of plates and plane elastic solid components are considered. Compliance, displacement, and stress functionals are considered as performance criteria. The method presented is being extended to implement shape design sensitivity analysis using a domain method and a design component method.
Sensitivity analysis of Stirling engine design parameters
Naso, V.; Dong, W.; Lucentini, M.; Capata, R.
1998-07-01
In the preliminary Stirling engine design process, the values of some design parameters (temperature ratio, swept volume ratio, phase angle and dead volume ratio) have to be assumed; as a matter of fact it can be difficult to determine the best values of these parameters for a particular engine design. In this paper, a mathematical model is developed to analyze the sensitivity of engine's performance variations corresponding to variations of these parameters.
Design sensitivity analysis of rotorcraft airframe structures for vibration reduction
NASA Technical Reports Server (NTRS)
Murthy, T. Sreekanta
1987-01-01
Optimization of rotorcraft structures for vibration reduction was studied. The objective of this study is to develop practical computational procedures for structural optimization of airframes subject to steady-state vibration response constraints. One of the key elements of any such computational procedure is design sensitivity analysis. A method for design sensitivity analysis of airframes under vibration response constraints is presented. The mathematical formulation of the method and its implementation as a new solution sequence in MSC/NASTRAN are described. The results of the application of the method to a simple finite element stick model of the AH-1G helicopter airframe are presented and discussed. Selection of design variables that are most likely to bring about changes in the response at specified locations in the airframe is based on consideration of forced response strain energy. Sensitivity coefficients are determined for the selected design variable set. Constraints on the natural frequencies are also included in addition to the constraints on the steady-state response. Sensitivity coefficients for these constraints are determined. Results of the analysis and insights gained in applying the method to the airframe model are discussed. The general nature of future work to be conducted is described.
Design sensitivity analysis and optimization tool (DSO) for sizing design applications
NASA Technical Reports Server (NTRS)
Chang, Kuang-Hua; Choi, Kyung K.; Perng, Jyh-Hwa
1992-01-01
The DSO tool, a structural design software system that provides the designer with a graphics-based menu-driven design environment to perform easy design optimization for general applications, is presented. Three design stages, preprocessing, design sensitivity analysis, and postprocessing, are implemented in the DSO to allow the designer to carry out the design process systematically. A framework, including data base, user interface, foundation class, and remote module, has been designed and implemented to facilitate software development for the DSO. A number of dedicated commercial software/packages have been integrated in the DSO to support the design procedures. Instead of parameterizing an FEM, design parameters are defined on a geometric model associated with physical quantities, and the continuum design sensitivity analysis theory is implemented to compute design sensitivity coefficients using postprocessing data from the analysis codes. A tracked vehicle road wheel is given as a sizing design application to demonstrate the DSO's easy and convenient design optimization process.
Aeroacoustic sensitivity analysis and optimal aeroacoustic design of turbomachinery blades
NASA Technical Reports Server (NTRS)
Hall, Kenneth C.
1994-01-01
During the first year of the project, we have developed a theoretical analysis - and wrote a computer code based on this analysis - to compute the sensitivity of unsteady aerodynamic loads acting on airfoils in cascades due to small changes in airfoil geometry. The steady and unsteady flow though a cascade of airfoils is computed using the full potential equation. Once the nominal solutions have been computed, one computes the sensitivity. The analysis takes advantage of the fact that LU decomposition is used to compute the nominal steady and unsteady flow fields. If the LU factors are saved, then the computer time required to compute the sensitivity of both the steady and unsteady flows to changes in airfoil geometry is quite small. The results to date are quite encouraging, and may be summarized as follows: (1) The sensitivity procedure has been validated by comparing the results obtained by 'finite difference' techniques, that is, computing the flow using the nominal flow solver for two slightly different airfoils and differencing the results. The 'analytic' solution computed using the method developed under this grant and the finite difference results are found to be in almost perfect agreement. (2) The present sensitivity analysis is computationally much more efficient than finite difference techniques. We found that using a 129 by 33 node computational grid, the present sensitivity analysis can compute the steady flow sensitivity about ten times more efficiently that the finite difference approach. For the unsteady flow problem, the present sensitivity analysis is about two and one-half times as fast as the finite difference approach. We expect that the relative efficiencies will be even larger for the finer grids which will be used to compute high frequency aeroacoustic solutions. Computational results show that the sensitivity analysis is valid for small to moderate sized design perturbations. (3) We found that the sensitivity analysis provided important
Design Parameters Influencing Reliability of CCGA Assembly: A Sensitivity Analysis
NASA Technical Reports Server (NTRS)
Tasooji, Amaneh; Ghaffarian, Reza; Rinaldi, Antonio
2006-01-01
Area Array microelectronic packages with small pitch and large I/O counts are now widely used in microelectronics packaging. The impact of various package design and materials/process parameters on reliability has been studied through extensive literature review. Reliability of Ceramic Column Grid Array (CCGA) package assemblies has been evaluated using JPL thermal cycle test results (-50(deg)/75(deg)C, -55(deg)/100(deg)C, and -55(deg)/125(deg)C), as well as those reported by other investigators. A sensitivity analysis has been performed using the literature da to study the impact of design parameters and global/local stress conditions on assembly reliability. The applicability of various life-prediction models for CCGA design has been investigated by comparing model's predictions with the experimental thermal cycling data. Finite Element Method (FEM) analysis has been conducted to assess the state of the stress/strain in CCGA assembly under different thermal cycling, and to explain the different failure modes and locations observed in JPL test assemblies.
Treatment of body forces in boundary element design sensitivity analysis
NASA Technical Reports Server (NTRS)
Saigal, Sunil; Kane, James H.; Aithal, R.; Cheng, Jizu
1989-01-01
The inclusion of body forces has received a good deal of attention in boundary element research. The consideration of such forces is essential in the desgin of high performance components such as fan and turbine disks in a gas turbine engine. Due to their critical performance requirements, optimal shapes are often desired for these components. The boundary element method (BEM) offers the possibility of being an efficient method for such iterative analysis as shape optimization. The implicit-differentiation of the boundary integral equations is performed to obtain the sensitivity equations. The body forces are accounted for by either the particular integrals for uniform body forces or by a surface integration for non-uniform body forces. The corresponding sensitivity equations for both these cases are presented. The validity of present formulations is established through a close agreement with exact analytical results.
Observations Regarding Use of Advanced CFD Analysis, Sensitivity Analysis, and Design Codes in MDO
NASA Technical Reports Server (NTRS)
Newman, Perry A.; Hou, Gene J. W.; Taylor, Arthur C., III
1996-01-01
Observations regarding the use of advanced computational fluid dynamics (CFD) analysis, sensitivity analysis (SA), and design codes in gradient-based multidisciplinary design optimization (MDO) reflect our perception of the interactions required of CFD and our experience in recent aerodynamic design optimization studies using CFD. Sample results from these latter studies are summarized for conventional optimization (analysis - SA codes) and simultaneous analysis and design optimization (design code) using both Euler and Navier-Stokes flow approximations. The amount of computational resources required for aerodynamic design using CFD via analysis - SA codes is greater than that required for design codes. Thus, an MDO formulation that utilizes the more efficient design codes where possible is desired. However, in the aerovehicle MDO problem, the various disciplines that are involved have different design points in the flight envelope; therefore, CFD analysis - SA codes are required at the aerodynamic 'off design' points. The suggested MDO formulation is a hybrid multilevel optimization procedure that consists of both multipoint CFD analysis - SA codes and multipoint CFD design codes that perform suboptimizations.
Design component method for sensitivity analysis of built-up structures
NASA Technical Reports Server (NTRS)
Choi, Kyung K.; Seong, Hwai G.
1986-01-01
A 'design component method' that provides a unified and systematic organization of design sensitivity analysis for built-up structures is developed and implemented. Both conventional design variables, such as thickness and cross-sectional area, and shape design variables of components of built-up structures are considered. It is shown that design of components of built-up structures can be characterized and system design sensitivity expressions obtained by simply adding contributions from each component. The method leads to a systematic organization of computations for design sensitivity analysis that is similar to the way in which computations are organized within a finite element code.
Sensitivity analysis for aeroacoustic and aeroelastic design of turbomachinery blades
NASA Technical Reports Server (NTRS)
Lorence, Christopher B.; Hall, Kenneth C.
1995-01-01
A new method for computing the effect that small changes in the airfoil shape and cascade geometry have on the aeroacoustic and aeroelastic behavior of turbomachinery cascades is presented. The nonlinear unsteady flow is assumed to be composed of a nonlinear steady flow plus a small perturbation unsteady flow that is harmonic in time. First, the full potential equation is used to describe the behavior of the nonlinear mean (steady) flow through a two-dimensional cascade. The small disturbance unsteady flow through the cascade is described by the linearized Euler equations. Using rapid distortion theory, the unsteady velocity is split into a rotational part that contains the vorticity and an irrotational part described by a scalar potential. The unsteady vorticity transport is described analytically in terms of the drift and stream functions computed from the steady flow. Hence, the solution of the linearized Euler equations may be reduced to a single inhomogeneous equation for the unsteady potential. The steady flow and small disturbance unsteady flow equations are discretized using bilinear quadrilateral isoparametric finite elements. The nonlinear mean flow solution and streamline computational grid are computed simultaneously using Newton iteration. At each step of the Newton iteration, LU decomposition is used to solve the resulting set of linear equations. The unsteady flow problem is linear, and is also solved using LU decomposition. Next, a sensitivity analysis is performed to determine the effect small changes in cascade and airfoil geometry have on the mean and unsteady flow fields. The sensitivity analysis makes use of the nominal steady and unsteady flow LU decompositions so that no additional matrices need to be factored. Hence, the present method is computationally very efficient. To demonstrate how the sensitivity analysis may be used to redesign cascades, a compressor is redesigned for improved aeroelastic stability and two different fan exit guide
Results of an integrated structure-control law design sensitivity analysis
NASA Technical Reports Server (NTRS)
Gilbert, Michael G.
1988-01-01
Next generation air and space vehicle designs are driven by increased performance requirements, demanding a high level of design integration between traditionally separate design disciplines. Interdisciplinary analysis capabilities have been developed, for aeroservoelastic aircraft and large flexible spacecraft control for instance, but the requisite integrated design methods are only beginning to be developed. One integrated design method which has received attention is based on hierarchal problem decompositions, optimization, and design sensitivity analyses. This paper highlights a design sensitivity analysis method for Linear Quadratic Cost, Gaussian (LQG) optimal control laws, which predicts change in the optimal control law due to changes in fixed problem parameters using analytical sensitivity equations. Numerical results of a design sensitivity analysis for a realistic aeroservoelastic aircraft example are presented. In this example, the sensitivity of the optimally controlled aircraft's response to various problem formulation and physical aircraft parameters is determined. These results are used to predict the aircraft's new optimally controlled response if the parameter was to have some other nominal value during the control law design process. The sensitivity results are validated by recomputing the optimal control law for discrete variations in parameters, computing the new actual aircraft response, and comparing with the predicted response. These results show an improvement in sensitivity accuracy for integrated design purposes over methods which do not include changess in the optimal control law. Use of the analytical LQG sensitivity expressions is also shown to be more efficient that finite difference methods for the computation of the equivalent sensitivity information.
Design tradeoff studies and sensitivity analysis, appendix B
NASA Technical Reports Server (NTRS)
1979-01-01
Further work was performed on the Near Term Hybrid Passenger Vehicle Development Program. Fuel economy on the order of 2 to 3 times that of a conventional vehicle, with a comparable life cycle cost, is possible. The two most significant factors in keeping the life cycle cost down are the retail price increment and the ratio of battery replacement cost to battery life. Both factors can be reduced by reducing the power rating of the electric drive portion of the system relative to the system power requirements. The type of battery most suitable for the hybrid, from the point of view of minimizing life cycle cost, is nickel-iron. The hybrid is much less sensitive than a conventional vehicle is, in terms of the reduction in total fuel consumption and resultant decreases in operating expense, to reductions in vehicle weight, tire rolling resistance, etc., and to propulsion system and drivetrain improvements designed to improve the brake specific fuel consumption of the engine under low road load conditions. It is concluded that modifications to package the propulsion system and battery pack can be easily accommodated within the confines of a modified carryover body such as the Ford Ltd.
Design sensitivity analysis with Applicon IFAD using the adjoint variable method
NASA Technical Reports Server (NTRS)
Frederick, Marjorie C.; Choi, Kyung K.
1984-01-01
A numerical method is presented to implement structural design sensitivity analysis using the versatility and convenience of existing finite element structural analysis program and the theoretical foundation in structural design sensitivity analysis. Conventional design variables, such as thickness and cross-sectional areas, are considered. Structural performance functionals considered include compliance, displacement, and stress. It is shown that calculations can be carried out outside existing finite element codes, using postprocessing data only. That is, design sensitivity analysis software does not have to be imbedded in an existing finite element code. The finite element structural analysis program used in the implementation presented is IFAD. Feasibility of the method is shown through analysis of several problems, including built-up structures. Accurate design sensitivity results are obtained without the uncertainty of numerical accuracy associated with selection of a finite difference perturbation.
On 3-D modeling and automatic regridding in shape design sensitivity analysis
NASA Technical Reports Server (NTRS)
Choi, Kyung K.; Yao, Tse-Min
1987-01-01
The material derivative idea of continuum mechanics and the adjoint variable method of design sensitivity analysis are used to obtain a computable expression for the effect of shape variations on measures of structural performance of three-dimensional elastic solids.
Sensitivity analysis of physiological factors in space habitat design
NASA Technical Reports Server (NTRS)
Billingham, J.
1982-01-01
The costs incurred by design conservatism in space habitat design are discussed from a structural standpoint, and areas of physiological research into less than earth-normal conditions that offer the greatest potential decrease in habitat construction and operating costs are studied. The established range of human tolerance limits is defined for those physiological conditions which directly affect habitat structural design. These entire ranges or portions thereof are set as habitat design constraints as a function of habitat population and degree of ecological closure. Calculations are performed to determine the structural weight and cost associated with each discrete population size and its selected environmental conditions, on the basis of habitable volume equivalence for four basic habitat configurations: sphere, cylinder with hemispherical ends, torus, and crystal palace.
NASA Technical Reports Server (NTRS)
Sobieszczanski-Sobieski, Jaroslaw
1988-01-01
Optimization by decomposition, complex system sensitivity analysis, and a rapid growth of disciplinary sensitivity analysis are some of the recent developments that hold promise of a quantum jump in the support engineers receive from computers in the quantitative aspects of design. Review of the salient points of these techniques is given and illustrated by examples from aircraft design as a process that combines the best of human intellect and computer power to manipulate data.
NASA Technical Reports Server (NTRS)
Sobieszczanski-Sobieski, Jaroslaw
1988-01-01
Optimization by decomposition, complex system sensitivity analysis, and a rapid growth of disciplinary sensitivity analysis are some of the recent developments that hold promise of a quantum jump in the support engineers receive from computers in the quantitative aspects of design. Review of the salient points of these techniques is given and illustrated by examples from aircraft design as a process that combines the best of human intellect and computer power to manipulate data.
NASA Technical Reports Server (NTRS)
Hou, Gene J.-W.; Gumbert, Clyde R.; Newman, Perry A.
2004-01-01
A major step in a most probable point (MPP)-based method for reliability analysis is to determine the MPP. This is usually accomplished by using an optimization search algorithm. The optimal solutions associated with the MPP provide measurements related to safety probability. This study focuses on two commonly used approximate probability integration methods; i.e., the Reliability Index Approach (RIA) and the Performance Measurement Approach (PMA). Their reliability sensitivity equations are first derived in this paper, based on the derivatives of their respective optimal solutions. Examples are then provided to demonstrate the use of these derivatives for better reliability analysis and Reliability-Based Design Optimization (RBDO).
NASA Technical Reports Server (NTRS)
Hou, Gene J.-W; Newman, Perry A. (Technical Monitor)
2004-01-01
A major step in a most probable point (MPP)-based method for reliability analysis is to determine the MPP. This is usually accomplished by using an optimization search algorithm. The minimum distance associated with the MPP provides a measurement of safety probability, which can be obtained by approximate probability integration methods such as FORM or SORM. The reliability sensitivity equations are derived first in this paper, based on the derivatives of the optimal solution. Examples are provided later to demonstrate the use of these derivatives for better reliability analysis and reliability-based design optimization (RBDO).
Parallel-vector design sensitivity analysis in structural dynamics
NASA Technical Reports Server (NTRS)
Zhang, Y.; Nguyen, D. T.
1992-01-01
This paper presents a parallel-vector algorithm for sensitivity calculations in linear structural dynamics. The proposed alternative formulation works efficiently with the reduced system of dynamic equations, since it eliminates the need for expensive and complicated based-vector derivatives, which are required in the conventional reduced system formulation. The relationship between the alternative formulation and the conventional reduced system formulation has been established, and it has been proven analytically that the two approaches are identical when all the mode shapes are included. This paper validates the proposed alternative algorithm through numerical experiments, where only a small number of mode shapes are used. In addition, a modified mode acceleration method is presented, thus not only the displacements but also the velocities and accelerations are shown to be improved.
NASA Technical Reports Server (NTRS)
Nguyen, Duc T.; Storaasli, Olaf O.; Qin, Jiangning; Qamar, Ramzi
1994-01-01
An automatic differentiation tool (ADIFOR) is incorporated into a finite element based structural analysis program for shape and non-shape design sensitivity analysis of structural systems. The entire analysis and sensitivity procedures are parallelized and vectorized for high performance computation. Small scale examples to verify the accuracy of the proposed program and a medium scale example to demonstrate the parallel vector performance on multiple CRAY C90 processors are included.
NASA Technical Reports Server (NTRS)
Ibrahim, A. H.; Tiwari, S. N.; Smith, R. E.
1997-01-01
Variational methods (VM) sensitivity analysis employed to derive the costate (adjoint) equations, the transversality conditions, and the functional sensitivity derivatives. In the derivation of the sensitivity equations, the variational methods use the generalized calculus of variations, in which the variable boundary is considered as the design function. The converged solution of the state equations together with the converged solution of the costate equations are integrated along the domain boundary to uniquely determine the functional sensitivity derivatives with respect to the design function. The application of the variational methods to aerodynamic shape optimization problems is demonstrated for internal flow problems at supersonic Mach number range. The study shows, that while maintaining the accuracy of the functional sensitivity derivatives within the reasonable range for engineering prediction purposes, the variational methods show a substantial gain in computational efficiency, i.e., computer time and memory, when compared with the finite difference sensitivity analysis.
Results of an integrated structure/control law design sensitivity analysis
NASA Technical Reports Server (NTRS)
Gilbert, Michael G.
1989-01-01
A design sensitivity analysis method for Linear Quadratic Cost, Gaussian (LQG) optimal control laws, which predicts change in the optimal control law due to changes in fixed problem parameters using analytical sensitivity equations is discussed. Numerical results of a design sensitivity analysis for a realistic aeroservoelastic aircraft example are presented. In this example, the sensitivity of the optimally controlled aircraft's response to various problem formulation and physical aircraft parameters is determined. These results are used to predict the aircraft's new optimally controlled response if the parameter was to have some other nominal value during the control law design process. The sensitivity results are validated by recomputing the optimal control law for discrete variations in parameters, computing the new actual aircraft response, and comparing with the predicted response. These results show an improvement in sensitivity accuracy for integrated design purposes over methods which do not include changes in the optimal control law. Use of the analytical LQG sensitivity expressions is also shown to be more efficient than finite difference methods for the computation of the equivalent sensitivity information.
NASA Technical Reports Server (NTRS)
Taylor, Arthur C., III; Hou, Gene W.
1996-01-01
An incremental iterative formulation together with the well-known spatially split approximate-factorization algorithm, is presented for solving the large, sparse systems of linear equations that are associated with aerodynamic sensitivity analysis. This formulation is also known as the 'delta' or 'correction' form. For the smaller two dimensional problems, a direct method can be applied to solve these linear equations in either the standard or the incremental form, in which case the two are equivalent. However, iterative methods are needed for larger two-dimensional and three dimensional applications because direct methods require more computer memory than is currently available. Iterative methods for solving these equations in the standard form are generally unsatisfactory due to an ill-conditioned coefficient matrix; this problem is overcome when these equations are cast in the incremental form. The methodology is successfully implemented and tested using an upwind cell-centered finite-volume formulation applied in two dimensions to the thin-layer Navier-Stokes equations for external flow over an airfoil. In three dimensions this methodology is demonstrated with a marching-solution algorithm for the Euler equations to calculate supersonic flow over the High-Speed Civil Transport configuration (HSCT 24E). The sensitivity derivatives obtained with the incremental iterative method from a marching Euler code are used in a design-improvement study of the HSCT configuration that involves thickness. camber, and planform design variables.
NASA Astrophysics Data System (ADS)
Zong, Yali; Hu, Naigang; Duan, Baoyan; Yang, Guigeng; Cao, Hongjun; Xu, Wanye
2016-03-01
Inevitable manufacturing errors and inconsistency between assumed and actual boundary conditions can affect the shape precision and cable tensions of a cable-network antenna, and even result in failure of the structure in service. In this paper, an analytical sensitivity analysis method of the shape precision and cable tensions with respect to the parameters carrying uncertainty was studied. Based on the sensitivity analysis, an optimal design procedure was proposed to alleviate the effects of the parameters that carry uncertainty. The validity of the calculated sensitivities is examined by those computed by a finite difference method. Comparison with a traditional design method shows that the presented design procedure can remarkably reduce the influence of the uncertainties on the antenna performance. Moreover, the results suggest that especially slender front net cables, thick tension ties, relatively slender boundary cables and high tension level can improve the ability of cable-network antenna structures to resist the effects of the uncertainties on the antenna performance.
Geometrically nonlinear design sensitivity analysis on parallel-vector high-performance computers
NASA Technical Reports Server (NTRS)
Baddourah, Majdi A.; Nguyen, Duc T.
1993-01-01
Parallel-vector solution strategies for generation and assembly of element matrices, solution of the resulted system of linear equations, calculations of the unbalanced loads, displacements, stresses, and design sensitivity analysis (DSA) are all incorporated into the Newton Raphson (NR) procedure for nonlinear finite element analysis and DSA. Numerical results are included to show the performance of the proposed method for structural analysis and DSA in a parallel-vector computer environment.
Value-Driven Design and Sensitivity Analysis of Hybrid Energy Systems using Surrogate Modeling
Wenbo Du; Humberto E. Garcia; William R. Binder; Christiaan J. J. Paredis
2001-10-01
A surrogate modeling and analysis methodology is applied to study dynamic hybrid energy systems (HES). The effect of battery size on the smoothing of variability in renewable energy generation is investigated. Global sensitivity indices calculated using surrogate models show the relative sensitivity of system variability to dynamic properties of key components. A value maximization approach is used to consider the tradeoff between system variability and required battery size. Results are found to be highly sensitive to the renewable power profile considered, demonstrating the importance of accurate renewable resource modeling and prediction. The documented computational framework and preliminary results represent an important step towards a comprehensive methodology for HES evaluation, design, and optimization.
Three-Dimensional Simulation And Design Sensitivity Analysis Of The Injection Molding Process
NASA Astrophysics Data System (ADS)
Ilinca, Florin; Hétu, Jean-François
2004-06-01
Getting the proper combination of different process parameters such as injection speed, melt temperature and mold temperature is important in getting a part that minimizes warpage and has the desired mechanical properties. Very often a successful design in injection molding comes at the end of a long trial and error process. Design Sensitivity Analysis (DSA) can help molders improve the design and can produce substantial investment savings in both time and money. This paper investigates the ability of the sensitivity analysis to drive an optimization tool in order to improve the quality of the injected part. The paper presents the solution of the filling stage of the injection molding process by a 3D finite element solution algorithm. The sensitivity of the solution with respect to different process parameters is computed using the continuous sensitivity equation method. Solutions are shown for the non-isothermal filling of a rectangular plate with a polymer melt behaving as a non-Newtonian fluid. The paper presents the equations for the sensitivity of the velocity, pressure and temperature and their solution by finite elements. Sensitivities of the solution with respect to the injection speed, the melt and mold temperatures are shown.
Kim, Min-Geun; Jang, Hong-Lae; Cho, Seonho
2013-05-01
An efficient adjoint design sensitivity analysis method is developed for reduced atomic systems. A reduced atomic system and the adjoint system are constructed in a locally confined region, utilizing generalized Langevin equation (GLE) for periodic lattice structures. Due to the translational symmetry of lattice structures, the size of time history kernel function that accounts for the boundary effects of the reduced atomic systems could be reduced to a single atom’s degrees of freedom. For the problems of highly nonlinear design variables, the finite difference method is impractical for its inefficiency and inaccuracy. However, the adjoint method is very efficient regardless of the number of design variables since one additional time integration is required for the adjoint GLE. Through numerical examples, the derived adjoint sensitivity turns out to be accurate and efficient through the comparison with finite difference sensitivity.
Sensitivity Analysis of Design Variables to Optimize the Performance of the USV
NASA Astrophysics Data System (ADS)
Cao, Xue; Wei, Zifan; Yang, Songlin; Wen, Yiyan
Optimization is an important part of the design on Unmanned Surface Vehicle (USV). In this paper, considering the rapidity, maneuverability, seakeeping and rollover resistance performance of the USV, the design variables of the optimization system of the USV have been determined a mathematical model for comprehensive optimization of USV has been established. Integrated optimization design of multi-target and multi-constrain is achieved by computer programs. However, the influence degree of each design variable are different on the final optimization results, in order to determine the degree of influence of each design variables, find out the key variables for a further optimization analysis and sensitivity studies of the design variables to optimization will be crucial. For solving this problem, a C++ program has been written by genetic algorithm and five discrete variables have been selected which are used to study the sensitivity of optimization. The results showed that different design variables have different effects on the optimization. The length of the ship and the speed of propeller have the greatest effect on the total objective function. The speed of propeller has a greater impact on both rapidity and seakeeping. The length of ship L, the molded breadth of ship B, the draft of ship T and design speed Vs have a greater sensitivity to maneuverability. Also, molded breadth B has the greatest effect on the rollover resistance.
Stratospheric Airship Design Sensitivity
NASA Astrophysics Data System (ADS)
Smith, Ira Steve; Fortenberry, Michael; Noll, . James; Perry, William
2012-07-01
The concept of a stratospheric or high altitude powered platform has been around almost as long as stratospheric free balloons. Airships are defined as Lighter-Than-Air (LTA) vehicles with propulsion and steering systems. Over the past five (5) years there has been an increased interest by the U. S. Department of Defense as well as commercial enterprises in airships at all altitudes. One of these interests is in the area of stratospheric airships. Whereas DoD is primarily interested in things that look down, such platforms offer a platform for science applications, both downward and outward looking. Designing airships to operate in the stratosphere is very challenging due to the extreme high altitude environment. It is significantly different than low altitude airship designs such as observed in the familiar advertising or tourism airships or blimps. The stratospheric airship design is very dependent on the specific application and the particular requirements levied on the vehicle with mass and power limits. The design is a complex iterative process and is sensitive to many factors. In an effort to identify the key factors that have the greatest impacts on the design, a parametric analysis of a simplified airship design has been performed. The results of these studies will be presented.
NASA Astrophysics Data System (ADS)
Augusta Neto, Maria; Yu, Wenbin; Pereira Leal, Rogerio
2008-10-01
This article describes a new approach to design the cross-section layer orientations of composite laminated beam structures. The beams are modelled with realistic cross-sectional geometry and material properties instead of a simplified model. The VABS (the variational asymptotic beam section analysis) methodology is used to compute the cross-sectional model for a generalized Timoshenko model, which was embedded in the finite element solver FEAP. Optimal design is performed with respect to the layers' orientation. The design sensitivity analysis is analytically formulated and implemented. The direct differentiation method is used to evaluate the response sensitivities with respect to the design variables. Thus, the design sensitivities of the Timoshenko stiffness computed by VABS methodology are imbedded into the modified VABS program and linked to the beam finite element solver. The modified method of feasible directions and sequential quadratic programming algorithms are used to seek the optimal continuous solution of a set of numerical examples. The buckling load associated with the twist-bend instability of cantilever composite beams, which may have several cross-section geometries, is improved in the optimization procedure.
System Sensitivity Analysis Applied to the Conceptual Design of a Dual-Fuel Rocket SSTO
NASA Technical Reports Server (NTRS)
Olds, John R.
1994-01-01
This paper reports the results of initial efforts to apply the System Sensitivity Analysis (SSA) optimization method to the conceptual design of a single-stage-to-orbit (SSTO) launch vehicle. SSA is an efficient, calculus-based MDO technique for generating sensitivity derivatives in a highly multidisciplinary design environment. The method has been successfully applied to conceptual aircraft design and has been proven to have advantages over traditional direct optimization methods. The method is applied to the optimization of an advanced, piloted SSTO design similar to vehicles currently being analyzed by NASA as possible replacements for the Space Shuttle. Powered by a derivative of the Russian RD-701 rocket engine, the vehicle employs a combination of hydrocarbon, hydrogen, and oxygen propellants. Three primary disciplines are included in the design - propulsion, performance, and weights & sizing. A complete, converged vehicle analysis depends on the use of three standalone conceptual analysis computer codes. Efforts to minimize vehicle dry (empty) weight are reported in this paper. The problem consists of six system-level design variables and one system-level constraint. Using SSA in a 'manual' fashion to generate gradient information, six system-level iterations were performed from each of two different starting points. The results showed a good pattern of convergence for both starting points. A discussion of the advantages and disadvantages of the method, possible areas of improvement, and future work is included.
Adelman, D.D.; Stansbury, J.
1997-12-31
The Resource Conservation and Recovery Act (RCRA) Subtitle C, Comprehensive Environmental Response, Compensation, And Liability Act (CERCLA), and subsequent amendments have formed a comprehensive framework to deal with hazardous wastes on the national level. Key to this waste management is guidance on design (e.g., cover and bottom leachate control systems) of hazardous waste landfills. The objective of this research was to investigate the sensitivity of leachate volume at hazardous waste disposal sites to climatic, soil cover, and vegetative cover (Leaf Area Index) conditions. The computer model HELP3 which has the capability to simulate double bottom liner systems as called for in hazardous waste disposal sites was used in the analysis. HELP3 was used to model 54 combinations of climatic conditions, disposal site soil surface curve numbers, and leaf area index values to investigate how sensitive disposal site leachate volume was to these three variables. Results showed that leachate volume from the bottom double liner system was not sensitive to these parameters. However, the cover liner system leachate volume was quite sensitive to climatic conditions and less sensitive to Leaf Area Index and curve number values. Since humid locations had considerably more cover liner system leachate volume than and locations, different design standards may be appropriate for humid conditions than for and conditions.
Design of a smart magnetic sensor by sensitivity based covariance analysis
NASA Astrophysics Data System (ADS)
Krishna Kumar, P. T.
2001-08-01
We use the technique of sensitivity based covariance analysis to design a smart magnetic sensor for depth profile studies where a NMR flux meter is used as the sensor in a Van de Graff accelerator (VGA). The minimum detection limit of any sensor tends to the systematic uncertainty, and, using this phenomenology, we estimated the upper and lower bounds for the correlated systematic uncertainties for the proton energy accelerated by the VGA by the technique of determinant inequalities. Knowledge of the bounds would help in the design of a smart magnetic sensor with reduced correlated systematic uncertainty.
Sensitivity Analysis of the Thermal Response of 9975 Packaging Using Factorial Design Methods
Gupta, Narendra K.
2005-10-31
A method is presented for using the statistical design of experiment (2{sup k} Factorial Design) technique in the sensitivity analysis of the thermal response (temperature) of the 9975 radioactive material packaging where multiple thermal properties of the impact absorbing and fire insulating material Celotex and certain boundary conditions are subject to uncertainty. 2{sup k} Factorial Design method is very efficient in the use of available data and is capable of analyzing the impact of main variables (Factors) and their interactions on the component design. The 9975 design is based on detailed finite element (FE) analyses and extensive proof testing to meet the design requirements given in 10CFR71 [1]. However, the FE analyses use Celotex thermal properties that are based on published data and limited experiments. Celotex is an orthotropic material that is used in the home building industry. Its thermal properties are prone to variation due to manufacturing and fabrication processes, and due to long environmental exposure. This paper will evaluate the sensitivity of variations in thermal conductivity of the Celotex, convection coefficient at the drum surface, and drum emissivity (herein called Factors) on the thermal response of 9975 packaging under Normal Conditions of Transport (NCT). Application of this methodology will ascertain the robustness of the 9975 design and it can lead to more specific and useful understanding of the effects of various Factors on 9975 performance.
NASA Technical Reports Server (NTRS)
Taylor, Arthur C., III; Hou, Gene W.
1992-01-01
Fundamental equations of aerodynamic sensitivity analysis and approximate analysis for the two dimensional thin layer Navier-Stokes equations are reviewed, and special boundary condition considerations necessary to apply these equations to isolated lifting airfoils on 'C' and 'O' meshes are discussed in detail. An efficient strategy which is based on the finite element method and an elastic membrane representation of the computational domain is successfully tested, which circumvents the costly 'brute force' method of obtaining grid sensitivity derivatives, and is also useful in mesh regeneration. The issue of turbulence modeling is addressed in a preliminary study. Aerodynamic shape sensitivity derivatives are efficiently calculated, and their accuracy is validated on two viscous test problems, including: (1) internal flow through a double throat nozzle, and (2) external flow over a NACA 4-digit airfoil. An automated aerodynamic design optimization strategy is outlined which includes the use of a design optimization program, an aerodynamic flow analysis code, an aerodynamic sensitivity and approximate analysis code, and a mesh regeneration and grid sensitivity analysis code. Application of the optimization methodology to the two test problems in each case resulted in a new design having a significantly improved performance in the aerodynamic response of interest.
A wideband FMBEM for 2D acoustic design sensitivity analysis based on direct differentiation method
NASA Astrophysics Data System (ADS)
Chen, Leilei; Zheng, Changjun; Chen, Haibo
2013-09-01
This paper presents a wideband fast multipole boundary element method (FMBEM) for two dimensional acoustic design sensitivity analysis based on the direct differentiation method. The wideband fast multipole method (FMM) formed by combining the original FMM and the diagonal form FMM is used to accelerate the matrix-vector products in the boundary element analysis. The Burton-Miller formulation is used to overcome the fictitious frequency problem when using a single Helmholtz boundary integral equation for exterior boundary-value problems. The strongly singular and hypersingular integrals in the sensitivity equations can be evaluated explicitly and directly by using the piecewise constant discretization. The iterative solver GMRES is applied to accelerate the solution of the linear system of equations. A set of optimal parameters for the wideband FMBEM design sensitivity analysis are obtained by observing the performances of the wideband FMM algorithm in terms of computing time and memory usage. Numerical examples are presented to demonstrate the efficiency and validity of the proposed algorithm.
NASA Technical Reports Server (NTRS)
Taylor, Arthur C., III; Newman, James C., III; Barnwell, Richard W.
1997-01-01
A three-dimensional unstructured grid approach to aerodynamic shape sensitivity analysis and design optimization has been developed and is extended to model geometrically complex configurations. The advantage of unstructured grids (when compared with a structured-grid approach) is their inherent ability to discretize irregularly shaped domains with greater efficiency and less effort. Hence, this approach is ideally suited for geometrically complex configurations of practical interest. In this work the nonlinear Euler equations are solved using an upwind, cell-centered, finite-volume scheme. The discrete, linearized systems which result from this scheme are solved iteratively by a preconditioned conjugate-gradient-like algorithm known as GMRES for the two-dimensional geometry and a Gauss-Seidel algorithm for the three-dimensional; similar procedures are used to solve the accompanying linear aerodynamic sensitivity equations in incremental iterative form. As shown, this particular form of the sensitivity equation makes large-scale gradient-based aerodynamic optimization possible by taking advantage of memory efficient methods to construct exact Jacobian matrix-vector products. Simple parameterization techniques are utilized for demonstrative purposes. Once the surface has been deformed, the unstructured grid is adapted by considering the mesh as a system of interconnected springs. Grid sensitivities are obtained by differentiating the surface parameterization and the grid adaptation algorithms with ADIFOR (which is an advanced automatic-differentiation software tool). To demonstrate the ability of this procedure to analyze and design complex configurations of practical interest, the sensitivity analysis and shape optimization has been performed for a two-dimensional high-lift multielement airfoil and for a three-dimensional Boeing 747-200 aircraft.
Design of 3-D Nacelle near Flat-Plate Wing Using Multiblock Sensitivity Analysis (ADOS)
NASA Technical Reports Server (NTRS)
Eleshaky, Mohamed E.; Baysal, Oktay
1994-01-01
One of the major design tasks involved in reducing aircraft drag is the integration of the engine nacelles and airframe. With this impetus, nacelle shapes with and without the presence of a flat-plate wing nearby were optimized. This also served as a demonstration of the 3-D version of the recently developed aerodynamic design optimization methodology using sensitivity analysis, ADOS. The required flow analyses were obtained by solving the three-dimensional, compressible, thin-layer Navier-Stokes equations using an implicit, upwind-biased, finite volume scheme. The sensitivity analyses were performed using the preconditioned version of the SADD scheme (sensitivity analysis on domain decomposition). In addition to demonstrating the present method's capability for automatic optimization, the results offered some insight into two important issues related to optimizing the shapes of multicomponent configurations in close proximity. First, inclusion of the mutual interference between the components resulted in a different shape as opposed to shaping an isolated component. Secondly, exclusion of the viscous effects compromised not only the flow physics but also the optimized shapes even for isolated components.
NASA Astrophysics Data System (ADS)
Bhatia, Manav
Aerospace structures operate under extreme thermal environments. Hot external aerothermal environment at high Mach number flight leads to high structural temperatures. At the same time, cold internal cryogenic-fuel-tanks and thermal management concepts like Thermal Protection System (TPS) and active cooling result in a high temperature gradient through the structure. Multidisciplinary Design Optimization (MDO) of such structures requires a design-oriented approach to this problem. The broad goal of this research effort is to advance the existing state of the art towards MDO of large scale aerospace structures. The components required for this work are the sensitivity analysis formulation encompassing the scope of the physical phenomena being addressed, a set of efficient approximations to cut-down the required CPU cost, and a general purpose design-oriented numerical analysis tool capable of handling problems of this scope. In this work finite element discretization has been used to solve the conduction partial differential equations and the Poljak method has been used to discretize the integral equations for internal cavity radiation. A methodology has been established to couple the conduction finite element analysis to the internal radiation analysis. This formulation is then extended for sensitivity analysis of heat transfer and coupled thermal-structural problems. The most CPU intensive operations in the overall analysis have been identified, and approximation methods have been proposed to reduce the associated CPU cost. Results establish the effectiveness of these approximation methods, which lead to very high savings in CPU cost without any deterioration in the results. The results presented in this dissertation include two cases: a hexahedral cavity with internal and external radiation with conducting walls, and a wing box which is geometrically similar to the orbiter wing.
NASA Astrophysics Data System (ADS)
Hacker, Joshua
2013-04-01
Ensemble sensitivity analysis (ESA) is emerging as a viable alternative to adjoint sensitivity. Several open issues face ESA for forecasts dominated by mesoscale phenomena, including (1) sampling error arising from finite-sized ensembles causing over-estimated sensitivities, and (2) violation of linearity assumptions for strongly nonlinear flows. In an effort to use ESA for predictability studies and observing network design in complex terrain, we present results from experiments designed to address these open issues. Sampling error in ESA arises in two places. First, when hypothetical observations are introduced to test the sensitivity estimates for linearity. Here the same localization that was used in the filter itself can be simply applied. Second and more critical, localization should be considered within the sensitivity calculations. Sensitivity to hypothetical observations, estimated without re-running the ensemble, includes regression of a sample of a final-time (forecast) metric onto a sample of initial states. Derivation to include localization results in two localization coefficients (or factors) applied in separate regression steps. Because the forecast metric is usually a sum, and can also include a sum over a spatial region and multiple physical variables, a spatial localization function is difficult to specify. We present results from experiments to empirically estimate localization factors for ESA to test hypothetical observations for mesoscale data assimilation in complex terrain. Localization factors are first derived for an ensemble filter following the empirical localization methodology. Sensitivities for a fog event over Salt Lake City, and a Colorado downslope wind event, are tested for linearity by approximating assimilation of perfect observations at points of maximum sensitivity, both with and without localization. Observation sensitivity is then estimated, with and without localization, and tested for linearity. The validity of the
Dykes, K.; Ning, A.; King, R.; Graf, P.; Scott, G.; Veers, P.
2014-02-01
This paper introduces the development of a new software framework for research, design, and development of wind energy systems which is meant to 1) represent a full wind plant including all physical and nonphysical assets and associated costs up to the point of grid interconnection, 2) allow use of interchangeable models of varying fidelity for different aspects of the system, and 3) support system level multidisciplinary analyses and optimizations. This paper describes the design of the overall software capability and applies it to a global sensitivity analysis of wind turbine and plant performance and cost. The analysis was performed using three different model configurations involving different levels of fidelity, which illustrate how increasing fidelity can preserve important system interactions that build up to overall system performance and cost. Analyses were performed for a reference wind plant based on the National Renewable Energy Laboratory's 5-MW reference turbine at a mid-Atlantic offshore location within the United States.
NASA Astrophysics Data System (ADS)
Newman, James Charles, III
1997-10-01
The first two steps in the development of an integrated multidisciplinary design optimization procedure capable of analyzing the nonlinear fluid flow about geometrically complex aeroelastic configurations have been accomplished in the present work. For the first step, a three-dimensional unstructured grid approach to aerodynamic shape sensitivity analysis and design optimization has been developed. The advantage of unstructured grids, when compared with a structured-grid approach, is their inherent ability to discretize irregularly shaped domains with greater efficiency and less effort. Hence, this approach is ideally suited for geometrically complex configurations of practical interest. In this work the time-dependent, nonlinear Euler equations are solved using an upwind, cell-centered, finite-volume scheme. The discrete, linearized systems which result from this scheme are solved iteratively by a preconditioned conjugate-gradient-like algorithm known as GMRES for the two-dimensional cases and a Gauss-Seidel algorithm for the three-dimensional; at steady-state, similar procedures are used to solve the accompanying linear aerodynamic sensitivity equations in incremental iterative form. As shown, this particular form of the sensitivity equation makes large-scale gradient-based aerodynamic optimization possible by taking advantage of memory efficient methods to construct exact Jacobian matrix-vector products. Various surface parameterization techniques have been employed in the current study to control the shape of the design surface. Once this surface has been deformed, the interior volume of the unstructured grid is adapted by considering the mesh as a system of interconnected tension springs. Grid sensitivities are obtained by differentiating the surface parameterization and the grid adaptation algorithms with ADIFOR, an advanced automatic-differentiation software tool. To demonstrate the ability of this procedure to analyze and design complex configurations of
Sensitivity analysis of a dry-processed Candu fuel pellet's design parameters
Choi, Hangbok; Ryu, Ho Jin
2007-07-01
Sensitivity analysis was carried out in order to investigate the effect of a fuel pellet's design parameters on the performance of a dry-processed Canada deuterium uranium (CANDU) fuel and to suggest the optimum design modifications. Under a normal operating condition, a dry-processed fuel has a higher internal pressure and plastic strain due to a higher fuel centerline temperature when compared with a standard natural uranium CANDU fuel. Under a condition that the fuel bundle dimensions do not change, sensitivity calculations were performed on a fuel's design parameters such as the axial gap, dish depth, gap clearance and plenum volume. The results showed that the internal pressure and plastic strain of the cladding were most effectively reduced if a fuel's element plenum volume was increased. More specifically, the internal pressure and plastic strain of the dry-processed fuel satisfied the design limits of a standard CANDU fuel when the plenum volume was increased by one half a pellet, 0.5 mm{sup 3}/K. (authors)
Shin, J.; Malusare, P.; Hovland, P. D.; Mathematics and Computer Science
2008-01-01
Automatic differentiation (AD) has been expanding its role in scientific computing. While several AD tools have been actively developed and used, a wide range of problems remain to be solved. Activity analysis allows AD tools to generate derivative code for fewer variables, leading to a faster run time of the output code. This paper describes a new context-sensitive, flow-sensitive (CSFS) activity analysis, which is developed by extending an existing context-sensitive, flow-insensitive (CSFI) activity analysis. Our experiments with eight benchmarks show that the new CSFS activity analysis is more than 27 times slower but reduces 8 overestimations for the MIT General Circulation Model (MITgcm) and 1 for an ODE solver (c2) compared with the existing CSFI activity analysis implementation. Although the number of reduced overestimations looks small, the additionally identified passive variables may significantly reduce tedious human effort in maintaining a large code base such as MITgcm.
NASA Technical Reports Server (NTRS)
Taylor, Arthur C., III; Hou, Gene W.
1994-01-01
The straightforward automatic-differentiation and the hand-differentiated incremental iterative methods are interwoven to produce a hybrid scheme that captures some of the strengths of each strategy. With this compromise, discrete aerodynamic sensitivity derivatives are calculated with the efficient incremental iterative solution algorithm of the original flow code. Moreover, the principal advantage of automatic differentiation is retained (i.e., all complicated source code for the derivative calculations is constructed quickly with accuracy). The basic equations for second-order sensitivity derivatives are presented; four methods are compared. Each scheme requires that large systems are solved first for the first-order derivatives and, in all but one method, for the first-order adjoint variables. Of these latter three schemes, two require no solutions of large systems thereafter. For the other two for which additional systems are solved, the equations and solution procedures are analogous to those for the first order derivatives. From a practical viewpoint, implementation of the second-order methods is feasible only with software tools such as automatic differentiation, because of the extreme complexity and large number of terms. First- and second-order sensitivities are calculated accurately for two airfoil problems, including a turbulent flow example; both geometric-shape and flow-condition design variables are considered. Several methods are tested; results are compared on the basis of accuracy, computational time, and computer memory. For first-order derivatives, the hybrid incremental iterative scheme obtained with automatic differentiation is competitive with the best hand-differentiated method; for six independent variables, it is at least two to four times faster than central finite differences and requires only 60 percent more memory than the original code; the performance is expected to improve further in the future.
NASA Astrophysics Data System (ADS)
Yamazaki, Koetsu; Sakamoto, Jiro; Kitano, Masami
1993-02-01
A design sensitivity calculation technique based on the implicit differentiation method is formulated for isoparametric boundary elements for three-dimensional (3D) shape optimization problems. The practical sensitivity equations for boundary displacements and stresses are derived, and the efficiency and accuracy of the technique are compared with the semi-analytic method by implementing the sensitivity analysis of typical and basic shape design problems numerically. The sensitivity calculation technique is then applied to the minimum weight design problems of 3D bodies under stress constraints, such as the shape optimization of the ellipsoidal cavity in a cube and the connecting rod, where the Taylor series approximation, based on the boundary element sensitivity analysis at current design point, is adopted for the efficient implementation of the optimization.
Application of design sensitivity analysis for greater improvement on machine structural dynamics
NASA Technical Reports Server (NTRS)
Yoshimura, Masataka
1987-01-01
Methodologies are presented for greatly improving machine structural dynamics by using design sensitivity analyses and evaluative parameters. First, design sensitivity coefficients and evaluative parameters of structural dynamics are described. Next, the relations between the design sensitivity coefficients and the evaluative parameters are clarified. Then, design improvement procedures of structural dynamics are proposed for the following three cases: (1) addition of elastic structural members, (2) addition of mass elements, and (3) substantial charges of joint design variables. Cases (1) and (2) correspond to the changes of the initial framework or configuration, and (3) corresponds to the alteration of poor initial design variables. Finally, numerical examples are given for demonstrating the availability of the methods proposed.
NASA Technical Reports Server (NTRS)
Taylor, Arthur C., III; Hou, Gene W.
1993-01-01
In this study involving advanced fluid flow codes, an incremental iterative formulation (also known as the delta or correction form) together with the well-known spatially-split approximate factorization algorithm, is presented for solving the very large sparse systems of linear equations which are associated with aerodynamic sensitivity analysis. For smaller 2D problems, a direct method can be applied to solve these linear equations in either the standard or the incremental form, in which case the two are equivalent. Iterative methods are needed for larger 2D and future 3D applications, however, because direct methods require much more computer memory than is currently available. Iterative methods for solving these equations in the standard form are generally unsatisfactory due to an ill-conditioning of the coefficient matrix; this problem can be overcome when these equations are cast in the incremental form. These and other benefits are discussed. The methodology is successfully implemented and tested in 2D using an upwind, cell-centered, finite volume formulation applied to the thin-layer Navier-Stokes equations. Results are presented for two sample airfoil problems: (1) subsonic low Reynolds number laminar flow; and (2) transonic high Reynolds number turbulent flow.
NASA Technical Reports Server (NTRS)
Rais-Rohani, Masoud
2003-01-01
This report discusses the development and application of two alternative strategies in the form of global and sequential local response surface (RS) techniques for the solution of reliability-based optimization (RBO) problems. The problem of a thin-walled composite circular cylinder under axial buckling instability is used as a demonstrative example. In this case, the global technique uses a single second-order RS model to estimate the axial buckling load over the entire feasible design space (FDS) whereas the local technique uses multiple first-order RS models with each applied to a small subregion of FDS. Alternative methods for the calculation of unknown coefficients in each RS model are explored prior to the solution of the optimization problem. The example RBO problem is formulated as a function of 23 uncorrelated random variables that include material properties, thickness and orientation angle of each ply, cylinder diameter and length, as well as the applied load. The mean values of the 8 ply thicknesses are treated as independent design variables. While the coefficients of variation of all random variables are held fixed, the standard deviations of ply thicknesses can vary during the optimization process as a result of changes in the design variables. The structural reliability analysis is based on the first-order reliability method with reliability index treated as the design constraint. In addition to the probabilistic sensitivity analysis of reliability index, the results of the RBO problem are presented for different combinations of cylinder length and diameter and laminate ply patterns. The two strategies are found to produce similar results in terms of accuracy with the sequential local RS technique having a considerably better computational efficiency.
Radicioni, A.; Corbetta, G.; D`Aloisio, G.; Bjoerset, A.
1995-12-31
The development of a subsea field requires among other things a definition of the methods for pipeline, flow-lines and umbilicals laying and tie-in. The selection of a particular method is the result of a detailed analysis where advantages and drawbacks are highlighted arid weighted versus global costs and reliability. During the conceptual study for the assessment of possible different solutions, engineering tools can be used as a time saving solution for the selection of the best installation method. Different installation methods to perform tie-in have been analyzed and simplified mathematical models have been used to better understand the behavior of the pipeline during its installation. The following methods were considered: first-end pull-in, second-end pull-in, deflect-to-connect. This paper summarizes the results achieved and describes the design tools prepared during the development of the subject sensitivity analysis; such tools can be used as a helpful basis for the selection or design of a tie-in system.
Sensitivity Analysis in Engineering
NASA Technical Reports Server (NTRS)
Adelman, Howard M. (Compiler); Haftka, Raphael T. (Compiler)
1987-01-01
The symposium proceedings presented focused primarily on sensitivity analysis of structural response. However, the first session, entitled, General and Multidisciplinary Sensitivity, focused on areas such as physics, chemistry, controls, and aerodynamics. The other four sessions were concerned with the sensitivity of structural systems modeled by finite elements. Session 2 dealt with Static Sensitivity Analysis and Applications; Session 3 with Eigenproblem Sensitivity Methods; Session 4 with Transient Sensitivity Analysis; and Session 5 with Shape Sensitivity Analysis.
Sensitivity Analysis and Mitigation with Applications to Ballistic and Low-thrust Trajectory Design
NASA Astrophysics Data System (ADS)
Alizadeh, Iman
The ever increasing desire to expand space mission capabilities within the limited budgets of space industries requires new approaches to the old problem of spacecraft trajectory design. For example, recent initiatives for space exploration involve developing new tools to design low-cost, fail-safe trajectories to visit several potential destinations beyond our celestial neighborhood such as Jupiter's moons, asteroids, etc. Designing and navigating spacecraft trajectories to reach these destinations safely are complex and challenging. In particular, fundamental questions of orbital stability imposed by planetary protection requirements are not easily taken into account by standard optimal control schemes. The event of temporary engine loss or an unexpected missed thrust can indeed quickly lead to impact with planetary bodies or other unrecoverable trajectories. While electric propulsion technology provides superior efficiency compared to chemical engines, the very low-control authority and engine performance degradation can impose higher risk to the mission in strongly perturbed orbital environments. The risk is due to the complex gravitational field and its associated chaotic dynamics which causes large navigation dispersions in a short time if left un-controlled. Moreover, in these situations it can be outside the low-thrust propulsion system capability to correct the spacecraft trajectory in a reasonable time frame. These concerns can lead to complete or partial mission failure or even an infeasible mission concept at the early design stage. The goal of this research is to assess and increase orbital stability of ballistic and low-thrust transfer trajectories in multi-body systems. In particular, novel techniques are presented to characterize sensitivity and improve recovery characteristics of ballistic and low-thrust trajectories in unstable orbital environments. The techniques developed are based on perturbation analysis around ballistic trajectories to
Automated divertor target design by adjoint shape sensitivity analysis and a one-shot method
Dekeyser, W.; Reiter, D.; Baelmans, M.
2014-12-01
As magnetic confinement fusion progresses towards the development of first reactor-scale devices, computational tokamak divertor design is a topic of high priority. Presently, edge plasma codes are used in a forward approach, where magnetic field and divertor geometry are manually adjusted to meet design requirements. Due to the complex edge plasma flows and large number of design variables, this method is computationally very demanding. On the other hand, efficient optimization-based design strategies have been developed in computational aerodynamics and fluid mechanics. Such an optimization approach to divertor target shape design is elaborated in the present paper. A general formulation of the design problems is given, and conditions characterizing the optimal designs are formulated. Using a continuous adjoint framework, design sensitivities can be computed at a cost of only two edge plasma simulations, independent of the number of design variables. Furthermore, by using a one-shot method the entire optimization problem can be solved at an equivalent cost of only a few forward simulations. The methodology is applied to target shape design for uniform power load, in simplified edge plasma geometry.
NASA Astrophysics Data System (ADS)
Oloso, Amidu Olawale
A hybrid automatic differentiation/incremental iterative method was implemented in the general purpose advanced computational fluid dynamics code (CFL3D Version 4.1) to yield a new code (CFL3D.ADII) that is capable of computing consistently discrete first order sensitivity derivatives for complex geometries. With the exception of unsteady problems, the new code retains all the useful features and capabilities of the original CFL3D flow analysis code. The superiority of the new code over a carefully applied method of finite-differences is demonstrated. A coarse grain, scalable, distributed-memory, parallel version of CFL3D.ADII was developed based on "derivative stripmining". In this data-parallel approach, an identical copy of CFL3D.ADII is executed on each processor with different derivative input files. The effect of communication overhead on the overall parallel computational efficiency is negligible. However, the fraction of CFL3D.ADII duplicated on all processors has significant impact on the computational efficiency. To reduce the large execution time associated with the sequential 1-D line search in gradient-based aerodynamic optimization, an alternative parallel approach was developed. The execution time of the new approach was reduced effectively to that of one flow analysis, regardless of the number of function evaluations in the 1-D search. The new approach was found to yield design results that are essentially identical to those obtained from the traditional sequential approach but at much smaller execution time. The parallel CFL3D.ADII and the parallel 1-D line search are demonstrated in shape improvement studies of a realistic High Speed Civil Transport (HSCT) wing/body configuration represented by over 100 design variables and 200,000 grid points in inviscid supersonic flow on the 16 node IBM SP2 parallel computer at the Numerical Aerospace Simulation (NAS) facility, NASA Ames Research Center. In addition to making the handling of such a large
Designing novel cellulase systems through agent-based modeling and global sensitivity analysis
Apte, Advait A; Senger, Ryan S; Fong, Stephen S
2014-01-01
Experimental techniques allow engineering of biological systems to modify functionality; however, there still remains a need to develop tools to prioritize targets for modification. In this study, agent-based modeling (ABM) was used to build stochastic models of complexed and non-complexed cellulose hydrolysis, including enzymatic mechanisms for endoglucanase, exoglucanase, and β-glucosidase activity. Modeling results were consistent with experimental observations of higher efficiency in complexed systems than non-complexed systems and established relationships between specific cellulolytic mechanisms and overall efficiency. Global sensitivity analysis (GSA) of model results identified key parameters for improving overall cellulose hydrolysis efficiency including: (1) the cellulase half-life, (2) the exoglucanase activity, and (3) the cellulase composition. Overall, the following parameters were found to significantly influence cellulose consumption in a consolidated bioprocess (CBP): (1) the glucose uptake rate of the culture, (2) the bacterial cell concentration, and (3) the nature of the cellulase enzyme system (complexed or non-complexed). Broadly, these results demonstrate the utility of combining modeling and sensitivity analysis to identify key parameters and/or targets for experimental improvement. PMID:24830736
Eldred, Michael Scott; Vigil, Dena M.; Dalbey, Keith R.; Bohnhoff, William J.; Adams, Brian M.; Swiler, Laura Painton; Lefantzi, Sophia; Hough, Patricia Diane; Eddy, John P.
2011-12-01
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes and iterative analysis methods. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, and stochastic expansion methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a theoretical manual for selected algorithms implemented within the DAKOTA software. It is not intended as a comprehensive theoretical treatment, since a number of existing texts cover general optimization theory, statistical analysis, and other introductory topics. Rather, this manual is intended to summarize a set of DAKOTA-related research publications in the areas of surrogate-based optimization, uncertainty quantification, and optimization under uncertainty that provide the foundation for many of DAKOTA's iterative analysis capabilities.
NASA Astrophysics Data System (ADS)
Unger, André J. A.
2010-02-01
This work is the second installment in a two-part series, and focuses on object-oriented programming methods to implement an augmented-state variable approach to aggregate the PCS index and introduce the Bermudan-style call feature into the proposed CAT bond model. The PCS index is aggregated quarterly using a discrete Asian running-sum formulation. The resulting aggregate PCS index augmented-state variable is used to specify the payoff (principle) on the CAT bond based on reinsurance layers. The purpose of the Bermudan-style call option is to allow the reinsurer to minimize their interest rate risk exposure on making fixed coupon payments under prevailing interest rates. A sensitivity analysis is performed to determine the impact of uncertainty in the frequency and magnitude of hurricanes on the price of the CAT bond. Results indicate that while the CAT bond is highly sensitive to the natural variability in the frequency of landfalling hurricanes between El Ninõ and non-El Ninõ years, it remains relatively insensitive to uncertainty in the magnitude of damages. In addition, results indicate that the maximum price of the CAT bond is insensitive to whether it is engineered to cover low frequency high magnitude events in a 'high' reinsurance layer relative to high frequency low magnitude events in a 'low' reinsurance layer. Also, while it is possible for the reinsurer to minimize their interest rate risk exposure on the fixed coupon payments, the impact of this risk on the price of the CAT bond appears small relative to the natural variability in the CAT bond price, and consequently catastrophic risk, due to uncertainty in the frequency and magnitude of landfalling hurricanes.
Nakata, Satoshi; Nagayama, Masaharu; Kitahata, Hiroyuki; Suematsu, Nobuhiko J; Hasegawa, Takeshi
2015-04-28
The development of self-propelled motors that mimic biological motors is an important challenge for the transport of either themselves or some material in a small space, since biological systems exhibit high autonomy and various types of responses, such as taxis and swarming. In this perspective, we review non-living systems that behave like living matter. We especially focus on nonlinearity to enhance autonomy and the response of the system, since characteristic nonlinear phenomena, such as oscillation, synchronization, pattern formation, bifurcation, and hysteresis, are coupled to self-motion of which driving force is the difference in the interfacial tension. Mathematical modelling based on reaction-diffusion equations and equations of motion as well as physicochemical analysis from the point of view of the molecular structure are also important for the design of non-living motors that mimic living motors. PMID:25826144
Adams, Brian M.; Ebeida, Mohamed Salah; Eldred, Michael S.; Jakeman, John Davis; Swiler, Laura Painton; Stephens, John Adam; Vigil, Dena M.; Wildey, Timothy Michael; Bohnhoff, William J.; Eddy, John P.; Hu, Kenneth T.; Dalbey, Keith R.; Bauman, Lara E; Hough, Patricia Diane
2014-05-01
The Dakota (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a exible and extensible interface between simulation codes and iterative analysis methods. Dakota contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quanti cation with sampling, reliability, and stochastic expansion methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the Dakota toolkit provides a exible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a user's manual for the Dakota software and provides capability overviews and procedures for software execution, as well as a variety of example studies.
1992-02-20
SENSIT,MUSIG,COMSEN is a set of three related programs for sensitivity test analysis. SENSIT conducts sensitivity tests. These tests are also known as threshold tests, LD50 tests, gap tests, drop weight tests, etc. SENSIT interactively instructs the experimenter on the proper level at which to stress the next specimen, based on the results of previous responses. MUSIG analyzes the results of a sensitivity test to determine the mean and standard deviation of the underlying population bymore » computing maximum likelihood estimates of these parameters. MUSIG also computes likelihood ratio joint confidence regions and individual confidence intervals. COMSEN compares the results of two sensitivity tests to see if the underlying populations are significantly different. COMSEN provides an unbiased method of distinguishing between statistical variation of the estimates of the parameters of the population and true population difference.« less
LISA Telescope Sensitivity Analysis
NASA Technical Reports Server (NTRS)
Waluschka, Eugene; Krebs, Carolyn (Technical Monitor)
2001-01-01
The results of a LISA telescope sensitivity analysis will be presented, The emphasis will be on the outgoing beam of the Dall-Kirkham' telescope and its far field phase patterns. The computed sensitivity analysis will include motions of the secondary with respect to the primary, changes in shape of the primary and secondary, effect of aberrations of the input laser beam and the effect the telescope thin film coatings on polarization. An end-to-end optical model will also be discussed.
Rearden, B.T.; Anderson, W.J.; Harms, G.A.
2005-08-15
Framatome ANP, Sandia National Laboratories (SNL), Oak Ridge National Laboratory (ORNL), and the University of Florida are cooperating on the U.S. Department of Energy Nuclear Energy Research Initiative (NERI) project 2001-0124 to design, assemble, execute, analyze, and document a series of critical experiments to validate reactor physics and criticality safety codes for the analysis of commercial power reactor fuels consisting of UO{sub 2} with {sup 235}U enrichments {>=}5 wt%. The experiments will be conducted at the SNL Pulsed Reactor Facility.Framatome ANP and SNL produced two series of conceptual experiment designs based on typical parameters, such as fuel-to-moderator ratios, that meet the programmatic requirements of this project within the given restraints on available materials and facilities. ORNL used the Tools for Sensitivity and Uncertainty Analysis Methodology Implementation (TSUNAMI) to assess, from a detailed physics-based perspective, the similarity of the experiment designs to the commercial systems they are intended to validate. Based on the results of the TSUNAMI analysis, one series of experiments was found to be preferable to the other and will provide significant new data for the validation of reactor physics and criticality safety codes.
Hahn, K D; Cooper, G W; Ruiz, C L; Fehl, D L; Chandler, G A; Knapp, P F; Leeper, R J; Nelson, A J; Smelser, R M; Torres, J A
2014-04-01
We present a general methodology to determine the diagnostic sensitivity that is directly applicable to neutron-activation diagnostics fielded on a wide variety of neutron-producing experiments, which include inertial-confinement fusion (ICF), dense plasma focus, and ion beam-driven concepts. This approach includes a combination of several effects: (1) non-isotropic neutron emission; (2) the 1/r(2) decrease in neutron fluence in the activation material; (3) the spatially distributed neutron scattering, attenuation, and energy losses due to the fielding environment and activation material itself; and (4) temporally varying neutron emission. As an example, we describe the copper-activation diagnostic used to measure secondary deuterium-tritium fusion-neutron yields on ICF experiments conducted on the pulsed-power Z Accelerator at Sandia National Laboratories. Using this methodology along with results from absolute calibrations and Monte Carlo simulations, we find that for the diagnostic configuration on Z, the diagnostic sensitivity is 0.037% ± 17% counts/neutron per cm(2) and is ∼ 40% less sensitive than it would be in an ideal geometry due to neutron attenuation, scattering, and energy-loss effects. PMID:24784607
Hahn, K. D. Ruiz, C. L.; Fehl, D. L.; Chandler, G. A.; Knapp, P. F.; Smelser, R. M.; Torres, J. A.; Cooper, G. W.; Nelson, A. J.; Leeper, R. J.
2014-04-15
We present a general methodology to determine the diagnostic sensitivity that is directly applicable to neutron-activation diagnostics fielded on a wide variety of neutron-producing experiments, which include inertial-confinement fusion (ICF), dense plasma focus, and ion beam-driven concepts. This approach includes a combination of several effects: (1) non-isotropic neutron emission; (2) the 1/r{sup 2} decrease in neutron fluence in the activation material; (3) the spatially distributed neutron scattering, attenuation, and energy losses due to the fielding environment and activation material itself; and (4) temporally varying neutron emission. As an example, we describe the copper-activation diagnostic used to measure secondary deuterium-tritium fusion-neutron yields on ICF experiments conducted on the pulsed-power Z Accelerator at Sandia National Laboratories. Using this methodology along with results from absolute calibrations and Monte Carlo simulations, we find that for the diagnostic configuration on Z, the diagnostic sensitivity is 0.037% ± 17% counts/neutron per cm{sup 2} and is ∼ 40% less sensitive than it would be in an ideal geometry due to neutron attenuation, scattering, and energy-loss effects.
NASA Astrophysics Data System (ADS)
Hahn, K. D.; Cooper, G. W.; Ruiz, C. L.; Fehl, D. L.; Chandler, G. A.; Knapp, P. F.; Leeper, R. J.; Nelson, A. J.; Smelser, R. M.; Torres, J. A.
2014-04-01
We present a general methodology to determine the diagnostic sensitivity that is directly applicable to neutron-activation diagnostics fielded on a wide variety of neutron-producing experiments, which include inertial-confinement fusion (ICF), dense plasma focus, and ion beam-driven concepts. This approach includes a combination of several effects: (1) non-isotropic neutron emission; (2) the 1/r2 decrease in neutron fluence in the activation material; (3) the spatially distributed neutron scattering, attenuation, and energy losses due to the fielding environment and activation material itself; and (4) temporally varying neutron emission. As an example, we describe the copper-activation diagnostic used to measure secondary deuterium-tritium fusion-neutron yields on ICF experiments conducted on the pulsed-power Z Accelerator at Sandia National Laboratories. Using this methodology along with results from absolute calibrations and Monte Carlo simulations, we find that for the diagnostic configuration on Z, the diagnostic sensitivity is 0.037% ± 17% counts/neutron per cm2 and is ˜ 40% less sensitive than it would be in an ideal geometry due to neutron attenuation, scattering, and energy-loss effects.
Sensitivity Analysis Without Assumptions
VanderWeele, Tyler J.
2016-01-01
Unmeasured confounding may undermine the validity of causal inference with observational studies. Sensitivity analysis provides an attractive way to partially circumvent this issue by assessing the potential influence of unmeasured confounding on causal conclusions. However, previous sensitivity analysis approaches often make strong and untestable assumptions such as having an unmeasured confounder that is binary, or having no interaction between the effects of the exposure and the confounder on the outcome, or having only one unmeasured confounder. Without imposing any assumptions on the unmeasured confounder or confounders, we derive a bounding factor and a sharp inequality such that the sensitivity analysis parameters must satisfy the inequality if an unmeasured confounder is to explain away the observed effect estimate or reduce it to a particular level. Our approach is easy to implement and involves only two sensitivity parameters. Surprisingly, our bounding factor, which makes no simplifying assumptions, is no more conservative than a number of previous sensitivity analysis techniques that do make assumptions. Our new bounding factor implies not only the traditional Cornfield conditions that both the relative risk of the exposure on the confounder and that of the confounder on the outcome must satisfy but also a high threshold that the maximum of these relative risks must satisfy. Furthermore, this new bounding factor can be viewed as a measure of the strength of confounding between the exposure and the outcome induced by a confounder. PMID:26841057
Witholder, R.E.
1980-04-01
The Solar Energy Research Institute has conducted a limited sensitivity analysis on a System for Projecting the Utilization of Renewable Resources (SPURR). The study utilized the Domestic Policy Review scenario for SPURR agricultural and industrial process heat and utility market sectors. This sensitivity analysis determines whether variations in solar system capital cost, operation and maintenance cost, and fuel cost (biomass only) correlate with intuitive expectations. The results of this effort contribute to a much larger issue: validation of SPURR. Such a study has practical applications for engineering improvements in solar technologies and is useful as a planning tool in the R and D allocation process.
Design tradeoff studies and sensitivity analysis, appendices B1 - B4. [hybrid electric vehicles
NASA Technical Reports Server (NTRS)
1979-01-01
Documentation is presented for a program which separately computes fuel and energy consumption for the two modes of operation of a hybrid electric vehicle. The distribution of daily travel is specified as input data as well as the weights which the component driving cycles are given in each of the composite cycles. The possibility of weight reduction through the substitution of various materials is considered as well as the market potential for hybrid vehicles. Data relating to battery compartment weight distribution and vehicle handling analysis is tabulated.
NASA Technical Reports Server (NTRS)
Contreras, Michael T.; Trease, Brian P.; Bojanowski, Cezary; Kulakx, Ronald F.
2013-01-01
A wheel experiencing sinkage and slippage events poses a high risk to planetary rover missions as evidenced by the mobility challenges endured by the Mars Exploration Rover (MER) project. Current wheel design practice utilizes loads derived from a series of events in the life cycle of the rover which do not include (1) failure metrics related to wheel sinkage and slippage and (2) performance trade-offs based on grouser placement/orientation. Wheel designs are rigorously tested experimentally through a variety of drive scenarios and simulated soil environments; however, a robust simulation capability is still in development due to myriad of complex interaction phenomena that contribute to wheel sinkage and slippage conditions such as soil composition, large deformation soil behavior, wheel geometry, nonlinear contact forces, terrain irregularity, etc. For the purposes of modeling wheel sinkage and slippage at an engineering scale, meshfree nite element approaches enable simulations that capture su cient detail of wheel-soil interaction while remaining computationally feasible. This study implements the JPL wheel-soil benchmark problem in the commercial code environment utilizing the large deformation modeling capability of Smooth Particle Hydrodynamics (SPH) meshfree methods. The nominal, benchmark wheel-soil interaction model that produces numerically stable and physically realistic results is presented and simulations are shown for both wheel traverse and wheel sinkage cases. A sensitivity analysis developing the capability and framework for future ight applications is conducted to illustrate the importance of perturbations to critical material properties and parameters. Implementation of the proposed soil-wheel interaction simulation capability and associated sensitivity framework has the potential to reduce experimentation cost and improve the early stage wheel design proce
Pyle, Matt Christopher
2012-01-01
In this thesis, we illustrate how the complex E- field geometry produced by interdigitated electrodes at alternating voltage biases naturally encodes 3D fiducial volume information into the charge and phonon signals and thus is a natural geometry for our next generation dark matter detectors. Secondly, we will study in depth the physics of import to our devices including transition edge sensor dynamics, quasi- particle dynamics in our Al collection fins, and phonon physics in the crystal itself so that we can both understand the performance of our previous CDMS II device as well as optimize the design of our future devices. Of interest to the broader physics community is the derivation of the ideal athermal phonon detector resolution and it's T^{3} _{c} scaling behavior which suggests that the athermal phonon detector technology developed by CDMS could also be used to discover coherent neutrino scattering and search for non-standard neutrino interaction and sterile neutrinos. These proposed resolution optimized devices can also be used in searches for exotic MeV-GeV dark matter as well as novel background free searches for 8GeV light WIMPs.
NASA Astrophysics Data System (ADS)
Muneepeerakul, Chitsomanus; Huffaker, Ray; Munoz-Carpena, Rafael
2016-04-01
The weather index insurance promises financial resilience to farmers struck by harsh weather conditions with swift compensation at affordable premium thanks to its minimal adverse selection and moral hazard. Despite these advantages, the very nature of indexing causes the presence of "production basis risk" that the selected weather indexes and their thresholds do not correspond to actual damages. To reduce basis risk without additional data collection cost, we propose the use of rain intensity and frequency as indexes as it could offer better protection at the lower premium by avoiding basis risk-strike trade-off inherent in the total rainfall index. We present empirical evidences and modeling results that even under the similar cumulative rainfall and temperature environment, yield can significantly differ especially for drought sensitive crops. We further show that deriving the trigger level and payoff function from regression between historical yield and total rainfall data may pose significant basis risk owing to their non-unique relationship in the insured range of rainfall. Lastly, we discuss the design of index insurance in terms of contract specifications based on the results from global sensitivity analysis.
RESRAD parameter sensitivity analysis
Cheng, J.J.; Yu, C.; Zielen, A.J.
1991-08-01
Three methods were used to perform a sensitivity analysis of RESRAD code input parameters -- enhancement of RESRAD by the Gradient Enhanced Software System (GRESS) package, direct parameter perturbation, and graphic comparison. Evaluation of these methods indicated that (1) the enhancement of RESRAD by GRESS has limitations and should be used cautiously, (2) direct parameter perturbation is tedious to implement, and (3) the graphics capability of RESRAD 4.0 is the most direct and convenient method for performing sensitivity analyses. This report describes procedures for implementing these methods and presents a comparison of results. 3 refs., 9 figs., 8 tabs.
Naval Waste Package Design Sensitivity
T. Schmitt
2006-12-13
The purpose of this calculation is to determine the sensitivity of the structural response of the Naval waste packages to varying inner cavity dimensions when subjected to a comer drop and tip-over from elevated surface. This calculation will also determine the sensitivity of the structural response of the Naval waste packages to the upper bound of the naval canister masses. The scope of this document is limited to reporting the calculation results in terms of through-wall stress intensities in the outer corrosion barrier. This calculation is intended for use in support of the preliminary design activities for the license application design of the Naval waste package. It examines the effects of small changes between the naval canister and the inner vessel, and in these dimensions, the Naval Long waste package and Naval Short waste package are similar. Therefore, only the Naval Long waste package is used in this calculation and is based on the proposed potential designs presented by the drawings and sketches in References 2.1.10 to 2.1.17 and 2.1.20. All conclusions are valid for both the Naval Long and Naval Short waste packages.
NASA Technical Reports Server (NTRS)
Leininger, G. G.; Borysiak, M. L.
1978-01-01
In the feedback control design of multivariable systems, closed loop performance evaluations must include the dynamic behavior of variables unavailable to the feedback controller. For the multivariable Nyquist array method, a set of sensitivity functions are proposed to simplify the adjustment of compensator parameters when the dynamic response of the unmeasurable output variables is unacceptable. A sensitivity study to improve thrust and turbine temperature responses for the Pratt-Whitney F100 turbofan engine demonstrates the utility of the proposed method.
Scaling in sensitivity analysis
Link, W.A.; Doherty, P.F., Jr.
2002-01-01
Population matrix models allow sets of demographic parameters to be summarized by a single value 8, the finite rate of population increase. The consequences of change in individual demographic parameters are naturally measured by the corresponding changes in 8; sensitivity analyses compare demographic parameters on the basis of these changes. These comparisons are complicated by issues of scale. Elasticity analysis attempts to deal with issues of scale by comparing the effects of proportional changes in demographic parameters, but leads to inconsistencies in evaluating demographic rates. We discuss this and other problems of scaling in sensitivity analysis, and suggest a simple criterion for choosing appropriate scales. We apply our suggestions to data for the killer whale, Orcinus orca.
LISA Telescope Sensitivity Analysis
NASA Technical Reports Server (NTRS)
Waluschka, Eugene; Krebs, Carolyn (Technical Monitor)
2002-01-01
The Laser Interferometer Space Antenna (LISA) for the detection of Gravitational Waves is a very long baseline interferometer which will measure the changes in the distance of a five million kilometer arm to picometer accuracies. As with any optical system, even one with such very large separations between the transmitting and receiving, telescopes, a sensitivity analysis should be performed to see how, in this case, the far field phase varies when the telescope parameters change as a result of small temperature changes.
Sensitivity testing and analysis
Neyer, B.T.
1991-01-01
New methods of sensitivity testing and analysis are proposed. The new test method utilizes Maximum Likelihood Estimates to pick the next test level in order to maximize knowledge of both the mean, {mu}, and the standard deviation, {sigma} of the population. Simulation results demonstrate that this new test provides better estimators (less bias and smaller variance) of both {mu} and {sigma} than the other commonly used tests (Probit, Bruceton, Robbins-Monro, Langlie). A new method of analyzing sensitivity tests is also proposed. It uses the Likelihood Ratio Test to compute regions of arbitrary confidence. It can calculate confidence regions, for {mu}, {sigma}, and arbitrary percentiles. Unlike presently used methods, such as the program ASENT which is based on the Cramer-Rao theorem, it can analyze the results of all sensitivity tests, and it does not significantly underestimate the size of the confidence regions. The new test and analysis methods will be explained and compared to the presently used methods. 19 refs., 12 figs.
WASTE PACKAGE DESIGN SENSITIVITY REPORT
P. Mecharet
2001-03-09
The purpose of this technical report is to present the current designs for waste packages and determine which designs will be evaluated for the Site Recommendation (SR) or Licence Application (LA), to demonstrate how the design will be shown to comply with the applicable design criteria. The evaluations to support SR or LA are based on system description document criteria. The objective is to determine those system description document criteria for which compliance is to be demonstrated for SR; and, having identified the criteria, to refer to the documents that show compliance. In addition, those system description document criteria for which compliance will be addressed for LA are identified, with a distinction made between two steps of the LA process: the LA-Construction Authorization (LA-CA) phase on one hand, and the LA-Receive and Possess (LA-R&P) phase on the other hand. The scope of this work encompasses the Waste Package Project disciplines for criticality, shielding, structural, and thermal analysis.
Eldred, Michael Scott; Dalbey, Keith R.; Bohnhoff, William J.; Adams, Brian M.; Swiler, Laura Painton; Hough, Patricia Diane; Gay, David M.; Eddy, John P.; Haskell, Karen H.
2010-05-01
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes and iterative analysis methods. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, and stochastic finite element methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a reference manual for the commands specification for the DAKOTA software, providing input overviews, option descriptions, and example specifications.
Griffin, Joshua D. (Sandai National Labs, Livermore, CA); Eldred, Michael Scott; Martinez-Canales, Monica L.; Watson, Jean-Paul; Kolda, Tamara Gibson; Adams, Brian M.; Swiler, Laura Painton; Williams, Pamela J.; Hough, Patricia Diane; Gay, David M.; Dunlavy, Daniel M.; Eddy, John P.; Hart, William Eugene; Guinta, Anthony A.; Brown, Shannon L.
2006-10-01
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes and iterative analysis methods. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, and stochastic finite element methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a reference manual for the commands specification for the DAKOTA software, providing input overviews, option descriptions, and example specifications.
Sensitivity analysis in computational aerodynamics
NASA Technical Reports Server (NTRS)
Bristow, D. R.
1984-01-01
Information on sensitivity analysis in computational aerodynamics is given in outline, graphical, and chart form. The prediction accuracy if the MCAERO program, a perturbation analysis method, is discussed. A procedure for calculating perturbation matrix, baseline wing paneling for perturbation analysis test cases and applications of an inviscid sensitivity matrix are among the topics covered.
Solid rocket motor nozzle flexseal design sensitivity
NASA Astrophysics Data System (ADS)
Donat, James R.
1993-02-01
On solid rocket motors, direction is controlled by controlling the thrust vector. To achieve this, the nozzle usually incorporates a flexseal that allows the nozzle to vector (or rotate) in any direction. The flexseal has a core of alternating layers of elastomer pads and metal or composite shims. Flexseal core design is an iterative process. An estimate of the flexseal core geometry is made. The core is then analyzed for performance characteristics such as stress, weight, and the torque required to vector the core. Based on a comparison between the requirements/constraints and analysis results, another estimate of the geometry is then made. Understanding the effects changes in the core geometry have on the performance characteristics greatly decreases the number of iterations and time required to optimize the design. This paper documents a study undertaken to better understand these effects and how sensitive performance characteristics are to core geometry changes.
Sensitivity and Uncertainty Analysis Shell
1999-04-20
SUNS (Sensitivity and Uncertainty Analysis Shell) is a 32-bit application that runs under Windows 95/98 and Windows NT. It is designed to aid in statistical analyses for a broad range of applications. The class of problems for which SUNS is suitable is generally defined by two requirements: 1. A computer code is developed or acquired that models some processes for which input is uncertain and the user is interested in statistical analysis of the outputmore » of that code. 2. The statistical analysis of interest can be accomplished using the Monte Carlo analysis. The implementation then requires that the user identify which input to the process model is to be manipulated for statistical analysis. With this information, the changes required to loosely couple SUNS with the process model can be completed. SUNS is then used to generate the required statistical sample and the user-supplied process model analyses the sample. The SUNS post processor displays statistical results from any existing file that contains sampled input and output values.« less
Accurate adjoint design sensitivities for nano metal optics.
Hansen, Paul; Hesselink, Lambertus
2015-09-01
We present a method for obtaining accurate numerical design sensitivities for metal-optical nanostructures. Adjoint design sensitivity analysis, long used in fluid mechanics and mechanical engineering for both optimization and structural analysis, is beginning to be used for nano-optics design, but it fails for sharp-cornered metal structures because the numerical error in electromagnetic simulations of metal structures is highest at sharp corners. These locations feature strong field enhancement and contribute strongly to design sensitivities. By using high-accuracy FEM calculations and rounding sharp features to a finite radius of curvature we obtain highly-accurate design sensitivities for 3D metal devices. To provide a bridge to the existing literature on adjoint methods in other fields, we derive the sensitivity equations for Maxwell's equations in the PDE framework widely used in fluid mechanics. PMID:26368483
1981-02-02
Version: 00 SENSIT computes the sensitivity and uncertainty of a calculated integral response (such as a dose rate) due to input cross sections and their uncertainties. Sensitivity profiles are computed for neutron and gamma-ray reaction cross sections (of standard multigroup cross-section sets) and for secondary energy distributions (SED's) of multigroup scattering matrices.
NASA Astrophysics Data System (ADS)
Tucek, Kamil; Carlsson, Johan; Wider, Hartmut
2006-04-01
We evaluated the sensitivity of several design and safety parameters with regard to five different nuclear data libraries, JEF2.2, JEFF3.0, ENDF/B-VI.8, JENDL3.2, and JENDL3.3. More specifically, the effective multiplication factor, burn-up reactivity swing and decay heat generation in available LFR and SFR designs were estimated. Monte Carlo codes MCNP and MCB were used in the analyses of the neutronic and burn-up performance of the systems. Thermo-hydraulic safety calculations were performed by the STAR-CD CFD code. For the LFR, ENDF/B-VI.8 and JEF2.2 showed to give a harder neutron spectrum than JEFF3.0, JENDL3.2, and JENDL3.3 data due to the lower inelastic scattering cross-section of lead in these libraries. Hence, the neutron economy of the system becomes more favourable and keff is higher when calculated with ENDF/B-VI.8 and JEF2.2 data. As for actinide cross-section data, the uncertainties in the keff values appeared to be mainly due to 239Pu, 240Pu and 241Am. Differences in the estimated burn-up reactivity swings proved to be significant, for an SFR as large as a factor of three (when comparing ENDF/B-VI.8 results to those of JENDL3.2). Uncertainties in the evaluation of short-term decay heat generation showed to be of the order of several per cent. Significant differences were, understandably, observed between decay heat generation data quoted in literature for LWR-UOX and those calculated for an LFR (U,TRU)O2 spent fuel. A corresponding difference in calculated core parameters (outlet coolant temperature) during protected total Loss-of-Power was evaluated.
Knopman, D.S.; Voss, C.I.
1988-01-01
Sensitivities of solute concentration to parameters associated with first-order chemical decay, boundary conditions, initial conditions, and multilayer transport are examined. A sensitivity is a change in solute concentration resulting from a change in a model parameter. Minimum information required in regression on chemical data for the estimation of model parameters by regression is expressed in terms of sensitivities. Nonlinear regression models were tested on sets of noiseless observations from known models that exceeded the minimum sensitivity information requirements. Results demonstrate that the regression models consistently converged to the correct parameters when the initial sets of parameter values substantially deviated from the correct parameters. -from Authors
Griffin, Joshua D. (Sandia National lababoratory, Livermore, CA); Eldred, Michael Scott; Martinez-Canales, Monica L.; Watson, Jean-Paul; Kolda, Tamara Gibson (Sandia National lababoratory, Livermore, CA); Giunta, Anthony Andrew; Adams, Brian M.; Swiler, Laura Painton; Williams, Pamela J.; Hough, Patricia Diane (Sandia National lababoratory, Livermore, CA); Gay, David M.; Dunlavy, Daniel M.; Eddy, John P.; Hart, William Eugene; Brown, Shannon L.
2006-10-01
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes and iterative analysis methods. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, and stochastic finite element methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a developers manual for the DAKOTA software and describes the DAKOTA class hierarchies and their interrelationships. It derives directly from annotation of the actual source code and provides detailed class documentation, including all member functions and attributes.
Griffin, Joshua D. (Sandai National Labs, Livermore, CA); Eldred, Michael Scott; Martinez-Canales, Monica L.; Watson, Jean-Paul; Kolda, Tamara Gibson; Giunta, Anthony Andrew; Adams, Brian M.; Swiler, Laura Painton; Williams, Pamela J. (Sandai National Labs, Livermore, CA); Hough, Patricia Diane (Sandai National Labs, Livermore, CA); Gay, David M.; Dunlavy, Daniel M.; Eddy, John P.; Hart, William Eugene; Brown, Shannon L.
2006-10-01
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes and iterative analysis methods. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, and stochastic finite element methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a user's manual for the DAKOTA software and provides capability overviews and procedures for software execution, as well as a variety of example studies.
Eldred, Michael Scott; Dalbey, Keith R.; Bohnhoff, William J.; Adams, Brian M.; Swiler, Laura Painton; Hough, Patricia Diane; Gay, David M.; Eddy, John P.; Haskell, Karen H.
2010-05-01
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes and iterative analysis methods. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, and stochastic finite element methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a user's manual for the DAKOTA software and provides capability overviews and procedures for software execution, as well as a variety of example studies.
Eldred, Michael Scott; Dalbey, Keith R.; Bohnhoff, William J.; Adams, Brian M.; Swiler, Laura Painton; Hough, Patricia Diane; Gay, David M.; Eddy, John P.; Haskell, Karen H.
2010-05-01
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes and iterative analysis methods. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, and stochastic finite element methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and performance analysis of computational models on high performance computers. This report serves as a developers manual for the DAKOTA software and describes the DAKOTA class hierarchies and their interrelationships. It derives directly from annotation of the actual source code and provides detailed class documentation, including all member functions and attributes.
Michaelis, Marc; Leopold, Claudia S
2015-12-30
The tack of a pressure sensitive adhesive (PSA) is not an inherent material property and strongly depends on the measurement conditions. Following the concept of a measurement system analysis (MSA), influencing factors of the probe tack test were investigated by a design of experiments (DoE) approach. A response surface design with 38 runs was built to evaluate the influence of detachment speed, dwell time, contact force, adhesive film thickness and API content on tack, determined as the maximum of the stress strain curve (σmax). It could be shown that all investigated factors have a significant effect on the response and that the DoE approach allowed to detect two-factorial interactions between the dwell time, the contact force, the adhesive film thickness and the API content. Surprisingly, it was found that tack increases with decreasing and not with increasing adhesive film thickness. PMID:26428630
Adjoint sensitivity analysis of an ultrawideband antenna
Stephanson, M B; White, D A
2011-07-28
The frequency domain finite element method using H(curl)-conforming finite elements is a robust technique for full-wave analysis of antennas. As computers become more powerful, it is becoming feasible to not only predict antenna performance, but also to compute sensitivity of antenna performance with respect to multiple parameters. This sensitivity information can then be used for optimization of the design or specification of manufacturing tolerances. In this paper we review the Adjoint Method for sensitivity calculation, and apply it to the problem of optimizing a Ultrawideband antenna.
Guan, Yanwei; Gao, Shiqiao; Liu, Haipeng; Jin, Lei; Niu, Shaohua
2016-01-01
In this paper, a new micromachined tuning fork gyroscope (TFG) with an anchored diamond coupling mechanism is proposed while the mode ordering and the vibration sensitivity are also investigated. The sense-mode of the proposed TFG was optimized through use of an anchored diamond coupling spring, which enables the in-phase mode frequency to be 108.3% higher than the anti-phase one. The frequencies of the in- and anti-phase modes in the sense direction are 9799.6 Hz and 4705.3 Hz, respectively. The analytical solutions illustrate that the stiffness difference ratio of the in- and anti-phase modes is inversely proportional to the output induced by the vibration from the sense direction. Additionally, FEM simulations demonstrate that the stiffness difference ratio of the anchored diamond coupling TFG is 16.08 times larger than the direct coupling one while the vibration output is reduced by 94.1%. Consequently, the proposed new anchored diamond coupling TFG can structurally increase the stiffness difference ratio to improve the mode ordering and considerably reduce the vibration sensitivity without sacrificing the scale factor. PMID:27049385
Guan, Yanwei; Gao, Shiqiao; Liu, Haipeng; Jin, Lei; Niu, Shaohua
2016-01-01
In this paper, a new micromachined tuning fork gyroscope (TFG) with an anchored diamond coupling mechanism is proposed while the mode ordering and the vibration sensitivity are also investigated. The sense-mode of the proposed TFG was optimized through use of an anchored diamond coupling spring, which enables the in-phase mode frequency to be 108.3% higher than the anti-phase one. The frequencies of the in- and anti-phase modes in the sense direction are 9799.6 Hz and 4705.3 Hz, respectively. The analytical solutions illustrate that the stiffness difference ratio of the in- and anti-phase modes is inversely proportional to the output induced by the vibration from the sense direction. Additionally, FEM simulations demonstrate that the stiffness difference ratio of the anchored diamond coupling TFG is 16.08 times larger than the direct coupling one while the vibration output is reduced by 94.1%. Consequently, the proposed new anchored diamond coupling TFG can structurally increase the stiffness difference ratio to improve the mode ordering and considerably reduce the vibration sensitivity without sacrificing the scale factor. PMID:27049385
An analysis of sensitivity tests
Neyer, B.T.
1992-03-06
A new method of analyzing sensitivity tests is proposed. It uses the Likelihood Ratio Test to compute regions of arbitrary confidence. It can calculate confidence regions for the parameters of the distribution (e.g., the mean, {mu}, and the standard deviation, {sigma}) as well as various percentiles. Unlike presently used methods, such as those based on asymptotic analysis, it can analyze the results of all sensitivity tests, and it does not significantly underestimate the size of the confidence regions. The main disadvantage of this method is that it requires much more computation to calculate the confidence regions. However, these calculations can be easily and quickly performed on most computers.
Design oriented structural analysis
NASA Technical Reports Server (NTRS)
Giles, Gary L.
1994-01-01
Desirable characteristics and benefits of design oriented analysis methods are described and illustrated by presenting a synoptic description of the development and uses of the Equivalent Laminated Plate Solution (ELAPS) computer code. ELAPS is a design oriented structural analysis method which is intended for use in the early design of aircraft wing structures. Model preparation is minimized by using a few large plate segments to model the wing box structure. Computational efficiency is achieved by using a limited number of global displacement functions that encompass all segments over the wing planform. Coupling with other codes is facilitated since the output quantities such as deflections and stresses are calculated as continuous functions over the plate segments. Various aspects of the ELAPS development are discussed including the analytical formulation, verification of results by comparison with finite element analysis results, coupling with other codes, and calculation of sensitivity derivatives. The effectiveness of ELAPS for multidisciplinary design application is illustrated by describing its use in design studies of high speed civil transport wing structures.
Sensitivity analysis and application in exploration geophysics
NASA Astrophysics Data System (ADS)
Tang, R.
2013-12-01
In exploration geophysics, the usual way of dealing with geophysical data is to form an Earth model describing underground structure in the area of investigation. The resolved model, however, is based on the inversion of survey data which is unavoidable contaminated by various noises and is sampled in a limited number of observation sites. Furthermore, due to the inherent non-unique weakness of inverse geophysical problem, the result is ambiguous. And it is not clear that which part of model features is well-resolved by the data. Therefore the interpretation of the result is intractable. We applied a sensitivity analysis to address this problem in magnetotelluric(MT). The sensitivity, also named Jacobian matrix or the sensitivity matrix, is comprised of the partial derivatives of the data with respect to the model parameters. In practical inversion, the matrix can be calculated by direct modeling of the theoretical response for the given model perturbation, or by the application of perturbation approach and reciprocity theory. We now acquired visualized sensitivity plot by calculating the sensitivity matrix and the solution is therefore under investigation that the less-resolved part is indicated and should not be considered in interpretation, while the well-resolved parameters can relatively be convincing. The sensitivity analysis is hereby a necessary and helpful tool for increasing the reliability of inverse models. Another main problem of exploration geophysics is about the design strategies of joint geophysical survey, i.e. gravity, magnetic & electromagnetic method. Since geophysical methods are based on the linear or nonlinear relationship between observed data and subsurface parameters, an appropriate design scheme which provides maximum information content within a restricted budget is quite difficult. Here we firstly studied sensitivity of different geophysical methods by mapping the spatial distribution of different survey sensitivity with respect to the
Precision of Sensitivity in the Design Optimization of Indeterminate Structures
NASA Technical Reports Server (NTRS)
Patnaik, Surya N.; Pai, Shantaram S.; Hopkins, Dale A.
2006-01-01
Design sensitivity is central to most optimization methods. The analytical sensitivity expression for an indeterminate structural design optimization problem can be factored into a simple determinate term and a complicated indeterminate component. Sensitivity can be approximated by retaining only the determinate term and setting the indeterminate factor to zero. The optimum solution is reached with the approximate sensitivity. The central processing unit (CPU) time to solution is substantially reduced. The benefit that accrues from using the approximate sensitivity is quantified by solving a set of problems in a controlled environment. Each problem is solved twice: first using the closed-form sensitivity expression, then using the approximation. The problem solutions use the CometBoards testbed as the optimization tool with the integrated force method as the analyzer. The modification that may be required, to use the stiffener method as the analysis tool in optimization, is discussed. The design optimization problem of an indeterminate structure contains many dependent constraints because of the implicit relationship between stresses, as well as the relationship between the stresses and displacements. The design optimization process can become problematic because the implicit relationship reduces the rank of the sensitivity matrix. The proposed approximation restores the full rank and enhances the robustness of the design optimization method.
Involute composite design evaluation using global design sensitivity derivatives
NASA Technical Reports Server (NTRS)
Hart, J. K.; Stanton, E. L.
1989-01-01
An optimization capability for involute structures has been developed. Its key feature is the use of global material geometry variables which are so chosen that all combinations of design variables within a set of lower and upper bounds correspond to manufacturable designs. A further advantage of global variables is that their number does not increase with increasing mesh density. The accuracy of the sensitivity derivatives has been verified both through finite difference tests and through the successful use of the derivatives by an optimizer. The state of the art in composite design today is still marked by point design algorithms linked together using ad hoc methods not directly related to a manufacturing procedure. The global design sensitivity approach presented here for involutes can be applied to filament wound shells and other composite constructions using material form features peculiar to each construction. The present involute optimization technology is being applied to the Space Shuttle SRM nozzle boot ring redesigns by PDA Engineering.
Stellarator Coil Design and Plasma Sensitivity
Long-Poe Ku and Allen H. Boozer
2010-11-03
The rich information contained in the plasma response to external magnetic perturbations can be used to help design stellarator coils more effectively. We demonstrate the feasibility by first devel- oping a simple, direct method to study perturbations in stellarators that do not break stellarator symmetry and periodicity. The method applies a small perturbation to the plasma boundary and evaluates the resulting perturbed free-boundary equilibrium to build up a sensitivity matrix for the important physics attributes of the underlying configuration. Using this sensitivity information, design methods for better stellarator coils are then developed. The procedure and a proof-of-principle application are given that (1) determine the spatial distributions of external normal magnetic field at the location of the unperturbed plasma boundary to which the plasma properties are most sen- sitive, (2) determine the distributions of external normal magnetic field that can be produced most efficiently by distant coils, (3) choose the ratios of the magnitudes of the the efficiently produced magnetic distributions so the sensitive plasma properties can be controlled. Using these methods, sets of modular coils are found for the National Compact Stellarator Experiment (NCSX) that are either smoother or can be located much farther from the plasma boundary than those of the present design.
Sensitivity Analysis in the Model Web
NASA Astrophysics Data System (ADS)
Jones, R.; Cornford, D.; Boukouvalas, A.
2012-04-01
The Model Web, and in particular the Uncertainty enabled Model Web being developed in the UncertWeb project aims to allow model developers and model users to deploy and discover models exposed as services on the Web. In particular model users will be able to compose model and data resources to construct and evaluate complex workflows. When discovering such workflows and models on the Web it is likely that the users might not have prior experience of the model behaviour in detail. It would be particularly beneficial if users could undertake a sensitivity analysis of the models and workflows they have discovered and constructed to allow them to assess the sensitivity to their assumptions and parameters. This work presents a Web-based sensitivity analysis tool which provides computationally efficient sensitivity analysis methods for models exposed on the Web. In particular the tool is tailored to the UncertWeb profiles for both information models (NetCDF and Observations and Measurements) and service specifications (WPS and SOAP/WSDL). The tool employs emulation technology where this is found to be possible, constructing statistical surrogate models for the models or workflows, to allow very fast variance based sensitivity analysis. Where models are too complex for emulation to be possible, or evaluate too fast for this to be necessary the original models are used with a carefully designed sampling strategy. A particular benefit of constructing emulators of the models or workflow components is that within the framework these can be communicated and evaluated at any physical location. The Web-based tool and backend API provide several functions to facilitate the process of creating an emulator and performing sensitivity analysis. A user can select a model exposed on the Web and specify the input ranges. Once this process is complete, they are able to perform screening to discover important inputs, train an emulator, and validate the accuracy of the trained emulator. In
A numerical comparison of sensitivity analysis techniques
Hamby, D.M.
1993-12-31
Engineering and scientific phenomena are often studied with the aid of mathematical models designed to simulate complex physical processes. In the nuclear industry, modeling the movement and consequence of radioactive pollutants is extremely important for environmental protection and facility control. One of the steps in model development is the determination of the parameters most influential on model results. A {open_quotes}sensitivity analysis{close_quotes} of these parameters is not only critical to model validation but also serves to guide future research. A previous manuscript (Hamby) detailed many of the available methods for conducting sensitivity analyses. The current paper is a comparative assessment of several methods for estimating relative parameter sensitivity. Method practicality is based on calculational ease and usefulness of the results. It is the intent of this report to demonstrate calculational rigor and to compare parameter sensitivity rankings resulting from various sensitivity analysis techniques. An atmospheric tritium dosimetry model (Hamby) is used here as an example, but the techniques described can be applied to many different modeling problems. Other investigators (Rose; Dalrymple and Broyd) present comparisons of sensitivity analyses methodologies, but none as comprehensive as the current work.
Zwickl, Titus; Carleer, Bart; Kubli, Waldemar
2005-08-05
In the past decade, sheet metal forming simulation became a well established tool to predict the formability of parts. In the automotive industry, this has enabled significant reduction in the cost and time for vehicle design and development, and has helped to improve the quality and performance of vehicle parts. However, production stoppages for troubleshooting and unplanned die maintenance, as well as production quality fluctuations continue to plague manufacturing cost and time. The focus therefore has shifted in recent times beyond mere feasibility to robustness of the product and process being engineered. Ensuring robustness is the next big challenge for the virtual tryout / simulation technology.We introduce new methods, based on systematic stochastic simulations, to visualize the behavior of the part during the whole forming process -- in simulation as well as in production. Sensitivity analysis explains the response of the part to changes in influencing parameters. Virtual tryout allows quick exploration of changed designs and conditions. Robust design and manufacturing guarantees quality and process capability for the production process. While conventional simulations helped to reduce development time and cost by ensuring feasible processes, robustness engineering tools have the potential for far greater cost and time savings.Through examples we illustrate how expected and unexpected behavior of deep drawing parts may be tracked down, identified and assigned to the influential parameters. With this knowledge, defects can be eliminated or springback can be compensated e.g.; the response of the part to uncontrollable noise can be predicted and minimized. The newly introduced methods enable more reliable and predictable stamping processes in general.
Rotary absorption heat pump sensitivity analysis
NASA Astrophysics Data System (ADS)
Bamberger, J. A.; Zalondek, F. R.
1990-03-01
Conserve Resources, Incorporated is currently developing an innovative, patented absorption heat pump. The heat pump uses rotation and thin film technology to enhance the absorption process and to provide a more efficient, compact system. The results are presented of a sensitivity analysis of the rotary absorption heat pump (RAHP) performance conducted to further the development of a 1-ton RAHP. The objective of the uncertainty analysis was to determine the sensitivity of RAHP steady state performance to uncertainties in design parameters. Prior to conducting the uncertainty analysis, a computer model was developed to describe the performance of the RAHP thermodynamic cycle. The RAHP performance is based on many interrelating factors, not all of which could be investigated during the sensitivity analysis. Confirmatory measurements of LiBr/H2O properties during absorber/generator operation will provide experimental verification that the system is operating as it was designed to operate. Quantities to be measured include: flow rate in the absorber and generator, film thickness, recirculation rate, and the effects of rotational speed on these parameters.
NASA Technical Reports Server (NTRS)
Gordon, Steven C.
1993-01-01
Spacecraft in orbit near libration point L1 in the Sun-Earth system are excellent platforms for research concerning solar effects on the terrestrial environment. One spacecraft mission launched in 1978 used an L1 orbit for nearly 4 years, and future L1 orbital missions are also being planned. Orbit determination and station-keeping are, however, required for these orbits. In particular, orbit determination error analysis may be used to compute the state uncertainty after a predetermined tracking period; the predicted state uncertainty levels then will impact the control costs computed in station-keeping simulations. Error sources, such as solar radiation pressure and planetary mass uncertainties, are also incorporated. For future missions, there may be some flexibility in the type and size of the spacecraft's nominal trajectory, but different orbits may produce varying error analysis and station-keeping results. The nominal path, for instance, can be (nearly) periodic or distinctly quasi-periodic. A periodic 'halo' orbit may be constructed to be significantly larger than a quasi-periodic 'Lissajous' path; both may meet mission requirements, but perhaps the required control costs for these orbits are probably different. Also for this spacecraft tracking and control simulation problem, experimental design methods can be used to determine the most significant uncertainties. That is, these methods can determine the error sources in the tracking and control problem that most impact the control cost (output); it also produces an equation that gives the approximate functional relationship between the error inputs and the output.
Wirtz, Tom; Fleming, Yves; Gerard, Mathieu; Gysin, Urs; Glatzel, Thilo; Meyer, Ernst; Wegmann, Urs; Maier, Urs; Odriozola, Aitziber Herrero; Uehli, Daniel
2012-06-15
State-of-the-art secondary ion mass spectrometry (SIMS) instruments allow producing 3D chemical mappings with excellent sensitivity and spatial resolution. Several important artifacts however arise from the fact that SIMS 3D mapping does not take into account the surface topography of the sample. In order to correct these artifacts, we have integrated a specially developed scanning probe microscopy (SPM) system into a commercial Cameca NanoSIMS 50 instrument. This new SPM module, which was designed as a DN200CF flange-mounted bolt-on accessory, includes a new high-precision sample stage, a scanner with a range of 100 {mu}m in x and y direction, and a dedicated SPM head which can be operated in the atomic force microscopy (AFM) and Kelvin probe force microscopy modes. Topographical information gained from AFM measurements taken before, during, and after SIMS analysis as well as the SIMS data are automatically compiled into an accurate 3D reconstruction using the software program 'SARINA,' which was developed for this first combined SIMS-SPM instrument. The achievable lateral resolutions are 6 nm in the SPM mode and 45 nm in the SIMS mode. Elemental 3D images obtained with our integrated SIMS-SPM instrument on Al/Cu and polystyrene/poly(methyl methacrylate) samples demonstrate the advantages of the combined SIMS-SPM approach.
Using Dynamic Sensitivity Analysis to Assess Testability
NASA Technical Reports Server (NTRS)
Voas, Jeffrey; Morell, Larry; Miller, Keith
1990-01-01
This paper discusses sensitivity analysis and its relationship to random black box testing. Sensitivity analysis estimates the impact that a programming fault at a particular location would have on the program's input/output behavior. Locations that are relatively \\"insensitive" to faults can render random black box testing unlikely to uncover programming faults. Therefore, sensitivity analysis gives new insight when interpreting random black box testing results. Although sensitivity analysis is computationally intensive, it requires no oracle and no human intervention.
Stiff DAE integrator with sensitivity analysis capabilities
2007-11-26
IDAS is a general purpose (serial and parallel) solver for differential equation (ODE) systems with senstivity analysis capabilities. It provides both forward and adjoint sensitivity analysis options.
Point Source Location Sensitivity Analysis
NASA Astrophysics Data System (ADS)
Cox, J. Allen
1986-11-01
This paper presents the results of an analysis of point source location accuracy and sensitivity as a function of focal plane geometry, optical blur spot, and location algorithm. Five specific blur spots are treated: gaussian, diffraction-limited circular aperture with and without central obscuration (obscured and clear bessinc, respectively), diffraction-limited rectangular aperture, and a pill box distribution. For each blur spot, location accuracies are calculated for square, rectangular, and hexagonal detector shapes of equal area. The rectangular detectors are arranged on a hexagonal lattice. The two location algorithms consist of standard and generalized centroid techniques. Hexagonal detector arrays are shown to give the best performance under a wide range of conditions.
Grid sensitivity for aerodynamic optimization and flow analysis
NASA Technical Reports Server (NTRS)
Sadrehaghighi, I.; Tiwari, S. N.
1993-01-01
After reviewing relevant literature, it is apparent that one aspect of aerodynamic sensitivity analysis, namely grid sensitivity, has not been investigated extensively. The grid sensitivity algorithms in most of these studies are based on structural design models. Such models, although sufficient for preliminary or conceptional design, are not acceptable for detailed design analysis. Careless grid sensitivity evaluations, would introduce gradient errors within the sensitivity module, therefore, infecting the overall optimization process. Development of an efficient and reliable grid sensitivity module with special emphasis on aerodynamic applications appear essential. The organization of this study is as follows. The physical and geometric representations of a typical model are derived in chapter 2. The grid generation algorithm and boundary grid distribution are developed in chapter 3. Chapter 4 discusses the theoretical formulation and aerodynamic sensitivity equation. The method of solution is provided in chapter 5. The results are presented and discussed in chapter 6. Finally, some concluding remarks are provided in chapter 7.
Nursing-sensitive indicators: a concept analysis
Heslop, Liza; Lu, Sai
2014-01-01
Aim To report a concept analysis of nursing-sensitive indicators within the applied context of the acute care setting. Background The concept of ‘nursing sensitive indicators’ is valuable to elaborate nursing care performance. The conceptual foundation, theoretical role, meaning, use and interpretation of the concept tend to differ. The elusiveness of the concept and the ambiguity of its attributes may have hindered research efforts to advance its application in practice. Design Concept analysis. Data sources Using ‘clinical indicators’ or ‘quality of nursing care’ as subject headings and incorporating keyword combinations of ‘acute care’ and ‘nurs*’, CINAHL and MEDLINE with full text in EBSCOhost databases were searched for English language journal articles published between 2000–2012. Only primary research articles were selected. Methods A hybrid approach was undertaken, incorporating traditional strategies as per Walker and Avant and a conceptual matrix based on Holzemer's Outcomes Model for Health Care Research. Results The analysis revealed two main attributes of nursing-sensitive indicators. Structural attributes related to health service operation included: hours of nursing care per patient day, nurse staffing. Outcome attributes related to patient care included: the prevalence of pressure ulcer, falls and falls with injury, nosocomial selective infection and patient/family satisfaction with nursing care. Conclusion This concept analysis may be used as a basis to advance understandings of the theoretical structures that underpin both research and practical application of quality dimensions of nursing care performance. PMID:25113388
Longitudinal Genetic Analysis of Anxiety Sensitivity
ERIC Educational Resources Information Center
Zavos, Helena M. S.; Gregory, Alice M.; Eley, Thalia C.
2012-01-01
Anxiety sensitivity is associated with both anxiety and depression and has been shown to be heritable. Little, however, is known about the role of genetic influence on continuity and change of symptoms over time. The authors' aim was to examine the stability of anxiety sensitivity during adolescence. By using a genetically sensitive design, the…
NASA Astrophysics Data System (ADS)
Wong, Li Wah; Watanabe, Norihiro; Fuchs, Sven; Bauer, Klaus; Cacace, Mauro; Blöcher, Guido; Kastner, Oliver; Zimmermann, Günter
2013-04-01
In order to show the impacts of natural internal fault zones and well design on geothermal energy production, two main deep geothermal reservoir sites in Germany, Groß Schönebeck (GrSk) and Berlin Tempelhof which are part of the North German Basin (NGB) are investigated. Groß Schönebeck is located at about 40km away from the Berlin centre whereas Berlin Tempelhof is situated in the south-central Berlin. Hydrothermal power plant shows complex coupling between four major components, the deep geothermal reservoir, the boreholes, the heat exchangers of the primary thermal water cycle and the power plant unit. In order to study the lifetime behavior of the overall Enhanced Geothermal System (EGS), it is mandatory to develop a combined transient model representing all relevant components as whole and their inter-relations. In this regards, the framework of Groß Schönebeck (GrSk) project is posed as the first scenario. The hydrothermal power plant is subdivided logically into components modeled separately and subsequently a standalone 3D transient hydro-thermal FEM (finite element method) reservoir model which consists of reservoir, fractures, wells and fault zones is weighted in the first place and its hydro-thermal processes are simulated for a period of 35 years. Using COMSOL Multiphysics, two significant objectives are achieved. Deviated geometries such as the production well and dipping geometries such as natural internal fault zones are successfully implemented into the 3D transient reservoir model which is constructed with the integration of hydraulically induced fractures and reservoir rock layers which are conducive to geothermal power production. Using OpenGeoSys (OGS), sensitivity analysis of varied conductivity of natural internal fault zones due to different permeability and apertures, on current fluid flow and heat transfer pattern is carried out. Study shows that natural internal fault zones play a significant role in the generation of production
A review of sensitivity analysis techniques
Hamby, D.M.
1993-12-31
Mathematical models are utilized to approximate various highly complex engineering, physical, environmental, social, and economic phenomena. Model parameters exerting the most influence on model results are identified through a {open_quotes}sensitivity analysis.{close_quotes} A comprehensive review is presented of more than a dozen sensitivity analysis methods. The most fundamental of sensitivity techniques utilizes partial differentiation whereas the simplest approach requires varying parameter values one-at-a-time. Correlation analysis is used to determine relationships between independent and dependent variables. Regression analysis provides the most comprehensive sensitivity measure and is commonly utilized to build response surfaces that approximate complex models.
NASA Astrophysics Data System (ADS)
Richard, Christopher L.
At the core of the geothermal industry is a need to identify how policy incentives can better be applied for optimal return. Literature from Bloomquist (1999), Doris et al. (2009), and McIlveen (2011) suggest that a more tailored approach to crafting geothermal policy is warranted. In this research the guiding theory is based on those suggestions and is structured to represent a policy analysis approach using analytical methods. The methods being used are focus on qualitative and quantitative results. To address the qualitative sections of this research an extensive review of contemporary literature is used to identify the frequency of use for specific barriers, and is followed upon with an industry survey to determine existing gaps. As a result there is support for certain barriers and justification for expanding those barriers found within the literature. This method of inquiry is an initial point for structuring modeling tools to further quantify the research results as part of the theoretical framework. Analytical modeling utilizes the levelized cost of energy as a foundation for comparative assessment of policy incentives. Model parameters use assumptions to draw conclusions from literature and survey results to reflect unique attributes held by geothermal power technologies. Further testing by policy option provides an opportunity to assess the sensitivity of each variable with respect to applied policy. Master limited partnerships, feed in tariffs, RD&D, and categorical exclusions all result as viable options for mitigating specific barriers associated to developing geothermal power. The results show reductions of levelized cost based upon the model's exclusive parameters. These results are also compared to contemporary policy options highlighting the need for tailored policy, as discussed by Bloomquist (1999), Doris et al. (2009), and McIlveen (2011). It is the intent of this research to provide the reader with a descriptive understanding of the role of
Sensitivity Analysis of Wing Aeroelastic Responses
NASA Technical Reports Server (NTRS)
Issac, Jason Cherian
1995-01-01
Design for prevention of aeroelastic instability (that is, the critical speeds leading to aeroelastic instability lie outside the operating range) is an integral part of the wing design process. Availability of the sensitivity derivatives of the various critical speeds with respect to shape parameters of the wing could be very useful to a designer in the initial design phase, when several design changes are made and the shape of the final configuration is not yet frozen. These derivatives are also indispensable for a gradient-based optimization with aeroelastic constraints. In this study, flutter characteristic of a typical section in subsonic compressible flow is examined using a state-space unsteady aerodynamic representation. The sensitivity of the flutter speed of the typical section with respect to its mass and stiffness parameters, namely, mass ratio, static unbalance, radius of gyration, bending frequency, and torsional frequency is calculated analytically. A strip theory formulation is newly developed to represent the unsteady aerodynamic forces on a wing. This is coupled with an equivalent plate structural model and solved as an eigenvalue problem to determine the critical speed of the wing. Flutter analysis of the wing is also carried out using a lifting-surface subsonic kernel function aerodynamic theory (FAST) and an equivalent plate structural model. Finite element modeling of the wing is done using NASTRAN so that wing structures made of spars and ribs and top and bottom wing skins could be analyzed. The free vibration modes of the wing obtained from NASTRAN are input into FAST to compute the flutter speed. An equivalent plate model which incorporates first-order shear deformation theory is then examined so it can be used to model thick wings, where shear deformations are important. The sensitivity of natural frequencies to changes in shape parameters is obtained using ADIFOR. A simple optimization effort is made towards obtaining a minimum weight
Sensitivity analysis of a wing aeroelastic response
NASA Technical Reports Server (NTRS)
Kapania, Rakesh K.; Eldred, Lloyd B.; Barthelemy, Jean-Francois M.
1991-01-01
A variation of Sobieski's Global Sensitivity Equations (GSE) approach is implemented to obtain the sensitivity of the static aeroelastic response of a three-dimensional wing model. The formulation is quite general and accepts any aerodynamics and structural analysis capability. An interface code is written to convert one analysis's output to the other's input, and visa versa. Local sensitivity derivatives are calculated by either analytic methods or finite difference techniques. A program to combine the local sensitivities, such as the sensitivity of the stiffness matrix or the aerodynamic kernel matrix, into global sensitivity derivatives is developed. The aerodynamic analysis package FAST, using a lifting surface theory, and a structural package, ELAPS, implementing Giles' equivalent plate model are used.
Recent developments in structural sensitivity analysis
NASA Technical Reports Server (NTRS)
Haftka, Raphael T.; Adelman, Howard M.
1988-01-01
Recent developments are reviewed in two major areas of structural sensitivity analysis: sensitivity of static and transient response; and sensitivity of vibration and buckling eigenproblems. Recent developments from the standpoint of computational cost, accuracy, and ease of implementation are presented. In the area of static response, current interest is focused on sensitivity to shape variation and sensitivity of nonlinear response. Two general approaches are used for computing sensitivities: differentiation of the continuum equations followed by discretization, and the reverse approach of discretization followed by differentiation. It is shown that the choice of methods has important accuracy and implementation implications. In the area of eigenproblem sensitivity, there is a great deal of interest and significant progress in sensitivity of problems with repeated eigenvalues. In addition to reviewing recent contributions in this area, the paper raises the issue of differentiability and continuity associated with the occurrence of repeated eigenvalues.
Boundary formulations for sensitivity analysis without matrix derivatives
NASA Technical Reports Server (NTRS)
Kane, J. H.; Guru Prasad, K.
1993-01-01
A new hybrid approach to continuum structural shape sensitivity analysis employing boundary element analysis (BEA) is presented. The approach uses iterative reanalysis to obviate the need to factor perturbed matrices in the determination of surface displacement and traction sensitivities via a univariate perturbation/finite difference (UPFD) step. The UPFD approach makes it possible to immediately reuse existing subroutines for computation of BEA matrix coefficients in the design sensitivity analysis process. The reanalysis technique computes economical response of univariately perturbed models without factoring perturbed matrices. The approach provides substantial computational economy without the burden of a large-scale reprogramming effort.
Sensitivity Analysis of Differential-Algebraic Equations and Partial Differential Equations
Petzold, L; Cao, Y; Li, S; Serban, R
2005-08-09
Sensitivity analysis generates essential information for model development, design optimization, parameter estimation, optimal control, model reduction and experimental design. In this paper we describe the forward and adjoint methods for sensitivity analysis, and outline some of our recent work on theory, algorithms and software for sensitivity analysis of differential-algebraic equation (DAE) and time-dependent partial differential equation (PDE) systems.
Sensitivity Analysis for some Water Pollution Problem
NASA Astrophysics Data System (ADS)
Le Dimet, François-Xavier; Tran Thu, Ha; Hussaini, Yousuff
2014-05-01
Sensitivity Analysis for Some Water Pollution Problems Francois-Xavier Le Dimet1 & Tran Thu Ha2 & M. Yousuff Hussaini3 1Université de Grenoble, France, 2Vietnamese Academy of Sciences, 3 Florida State University Sensitivity analysis employs some response function and the variable with respect to which its sensitivity is evaluated. If the state of the system is retrieved through a variational data assimilation process, then the observation appears only in the Optimality System (OS). In many cases, observations have errors and it is important to estimate their impact. Therefore, sensitivity analysis has to be carried out on the OS, and in that sense sensitivity analysis is a second order property. The OS can be considered as a generalized model because it contains all the available information. This presentation proposes a method to carry out sensitivity analysis in general. The method is demonstrated with an application to water pollution problem. The model involves shallow waters equations and an equation for the pollutant concentration. These equations are discretized using a finite volume method. The response function depends on the pollutant source, and its sensitivity with respect to the source term of the pollutant is studied. Specifically, we consider: • Identification of unknown parameters, and • Identification of sources of pollution and sensitivity with respect to the sources. We also use a Singular Evolutive Interpolated Kalman Filter to study this problem. The presentation includes a comparison of the results from these two methods. .
Extended Forward Sensitivity Analysis for Uncertainty Quantification
Haihua Zhao; Vincent A. Mousseau
2008-09-01
This report presents the forward sensitivity analysis method as a means for quantification of uncertainty in system analysis. The traditional approach to uncertainty quantification is based on a “black box” approach. The simulation tool is treated as an unknown signal generator, a distribution of inputs according to assumed probability density functions is sent in and the distribution of the outputs is measured and correlated back to the original input distribution. This approach requires large number of simulation runs and therefore has high computational cost. Contrary to the “black box” method, a more efficient sensitivity approach can take advantage of intimate knowledge of the simulation code. In this approach equations for the propagation of uncertainty are constructed and the sensitivity is solved for as variables in the same simulation. This “glass box” method can generate similar sensitivity information as the above “black box” approach with couples of runs to cover a large uncertainty region. Because only small numbers of runs are required, those runs can be done with a high accuracy in space and time ensuring that the uncertainty of the physical model is being measured and not simply the numerical error caused by the coarse discretization. In the forward sensitivity method, the model is differentiated with respect to each parameter to yield an additional system of the same size as the original one, the result of which is the solution sensitivity. The sensitivity of any output variable can then be directly obtained from these sensitivities by applying the chain rule of differentiation. We extend the forward sensitivity method to include time and spatial steps as special parameters so that the numerical errors can be quantified against other physical parameters. This extension makes the forward sensitivity method a much more powerful tool to help uncertainty analysis. By knowing the relative sensitivity of time and space steps with other
Extended Forward Sensitivity Analysis for Uncertainty Quantification
Haihua Zhao; Vincent A. Mousseau
2013-01-01
This paper presents the extended forward sensitivity analysis as a method to help uncertainty qualification. By including time step and potentially spatial step as special sensitivity parameters, the forward sensitivity method is extended as one method to quantify numerical errors. Note that by integrating local truncation errors over the whole system through the forward sensitivity analysis process, the generated time step and spatial step sensitivity information reflect global numerical errors. The discretization errors can be systematically compared against uncertainties due to other physical parameters. This extension makes the forward sensitivity method a much more powerful tool to help uncertainty qualification. By knowing the relative sensitivity of time and space steps with other interested physical parameters, the simulation is allowed to run at optimized time and space steps without affecting the confidence of the physical parameter sensitivity results. The time and space steps forward sensitivity analysis method can also replace the traditional time step and grid convergence study with much less computational cost. Two well-defined benchmark problems with manufactured solutions are utilized to demonstrate the method.
A strategy to design novel structure photochromic sensitizers for dye-sensitized solar cells
NASA Astrophysics Data System (ADS)
Wu, Wenjun; Wang, Jiaxing; Zheng, Zhiwei; Hu, Yue; Jin, Jiayu; Zhang, Qiong; Hua, Jianli
2015-02-01
Two sensitizers with novel structure were designed and synthetized by introducing photochromic bisthienylethene (BTE) group into the conjugated system. Thanks to the photochromic effect the sensitizers have under ultraviolet and visible light, the conjugated bridge can be restructured and the resulting two photoisomers showed different behaviors in photovoltaic devices. This opens up a new research way for the dye-sensitized solar cells (DSSCs).
Updated Chemical Kinetics and Sensitivity Analysis Code
NASA Technical Reports Server (NTRS)
Radhakrishnan, Krishnan
2005-01-01
An updated version of the General Chemical Kinetics and Sensitivity Analysis (LSENS) computer code has become available. A prior version of LSENS was described in "Program Helps to Determine Chemical-Reaction Mechanisms" (LEW-15758), NASA Tech Briefs, Vol. 19, No. 5 (May 1995), page 66. To recapitulate: LSENS solves complex, homogeneous, gas-phase, chemical-kinetics problems (e.g., combustion of fuels) that are represented by sets of many coupled, nonlinear, first-order ordinary differential equations. LSENS has been designed for flexibility, convenience, and computational efficiency. The present version of LSENS incorporates mathematical models for (1) a static system; (2) steady, one-dimensional inviscid flow; (3) reaction behind an incident shock wave, including boundary layer correction; (4) a perfectly stirred reactor; and (5) a perfectly stirred reactor followed by a plug-flow reactor. In addition, LSENS can compute equilibrium properties for the following assigned states: enthalpy and pressure, temperature and pressure, internal energy and volume, and temperature and volume. For static and one-dimensional-flow problems, including those behind an incident shock wave and following a perfectly stirred reactor calculation, LSENS can compute sensitivity coefficients of dependent variables and their derivatives, with respect to the initial values of dependent variables and/or the rate-coefficient parameters of the chemical reactions.
Extended Forward Sensitivity Analysis for Uncertainty Quantification
Haihua Zhao; Vincent A. Mousseau
2011-09-01
Verification and validation (V&V) are playing more important roles to quantify uncertainties and realize high fidelity simulations in engineering system analyses, such as transients happened in a complex nuclear reactor system. Traditional V&V in the reactor system analysis focused more on the validation part or did not differentiate verification and validation. The traditional approach to uncertainty quantification is based on a 'black box' approach. The simulation tool is treated as an unknown signal generator, a distribution of inputs according to assumed probability density functions is sent in and the distribution of the outputs is measured and correlated back to the original input distribution. The 'black box' method mixes numerical errors with all other uncertainties. It is also not efficient to perform sensitivity analysis. Contrary to the 'black box' method, a more efficient sensitivity approach can take advantage of intimate knowledge of the simulation code. In these types of approaches equations for the propagation of uncertainty are constructed and the sensitivities are directly solved for as variables in the simulation. This paper presents the forward sensitivity analysis as a method to help uncertainty qualification. By including time step and potentially spatial step as special sensitivity parameters, the forward sensitivity method is extended as one method to quantify numerical errors. Note that by integrating local truncation errors over the whole system through the forward sensitivity analysis process, the generated time step and spatial step sensitivity information reflect global numerical errors. The discretization errors can be systematically compared against uncertainties due to other physical parameters. This extension makes the forward sensitivity method a much more powerful tool to help uncertainty qualification. By knowing the relative sensitivity of time and space steps with other interested physical parameters, the simulation is allowed
Imaging system sensitivity analysis with NV-IPM
NASA Astrophysics Data System (ADS)
Fanning, Jonathan; Teaney, Brian
2014-05-01
This paper describes the sensitivity analysis capabilities to be added to version 1.2 of the NVESD imaging sensor model NV-IPM. Imaging system design always involves tradeoffs to design the best system possible within size, weight, and cost constraints. In general, the performance of a well designed system will be limited by the largest, heaviest, and most expensive components. Modeling is used to analyze system designs before the system is built. Traditionally, NVESD models were only used to determine the performance of a given system design. NV-IPM has the added ability to automatically determine the sensitivity of any system output to changes in the system parameters. The component-based structure of NV-IPM tracks the dependence between outputs and inputs such that only the relevant parameters are varied in the sensitivity analysis. This allows sensitivity analysis of an output such as probability of identification to determine the limiting parameters of the system. Individual components can be optimized by doing sensitivity analysis of outputs such as NETD or SNR. This capability will be demonstrated by analyzing example imaging systems.
Aero-Structural Interaction, Analysis, and Shape Sensitivity
NASA Technical Reports Server (NTRS)
Newman, James C., III
1999-01-01
A multidisciplinary sensitivity analysis technique that has been shown to be independent of step-size selection is examined further. The accuracy of this step-size independent technique, which uses complex variables for determining sensitivity derivatives, has been previously established. The primary focus of this work is to validate the aero-structural analysis procedure currently being used. This validation consists of comparing computed and experimental data obtained for an Aeroelastic Research Wing (ARW-2). Since the aero-structural analysis procedure has the complex variable modifications already included into the software, sensitivity derivatives can automatically be computed. Other than for design purposes, sensitivity derivatives can be used for predicting the solution at nearby conditions. The use of sensitivity derivatives for predicting the aero-structural characteristics of this configuration is demonstrated.
Coal Transportation Rate Sensitivity Analysis
2005-01-01
On December 21, 2004, the Surface Transportation Board (STB) requested that the Energy Information Administration (EIA) analyze the impact of changes in coal transportation rates on projected levels of electric power sector energy use and emissions. Specifically, the STB requested an analysis of changes in national and regional coal consumption and emissions resulting from adjustments in railroad transportation rates for Wyoming's Powder River Basin (PRB) coal using the National Energy Modeling System (NEMS). However, because NEMS operates at a relatively aggregate regional level and does not represent the costs of transporting coal over specific rail lines, this analysis reports on the impacts of interregional changes in transportation rates from those used in the Annual Energy Outlook 2005 (AEO2005) reference case.
Simultaneous analysis and design
NASA Technical Reports Server (NTRS)
Haftka, R. T.
1984-01-01
Optimization techniques are increasingly being used for performing nonlinear structural analysis. The development of element by element (EBE) preconditioned conjugate gradient (CG) techniques is expected to extend this trend to linear analysis. Under these circumstances the structural design problem can be viewed as a nested optimization problem. There are computational benefits to treating this nested problem as a large single optimization problem. The response variables (such as displacements) and the structural parameters are all treated as design variables in a unified formulation which performs simultaneously the design and analysis. Two examples are used for demonstration. A seventy-two bar truss is optimized subject to linear stress constraints and a wing box structure is optimized subject to nonlinear collapse constraints. Both examples show substantial computational savings with the unified approach as compared to the traditional nested approach.
Sensitivity analysis of distributed volcanic source inversion
NASA Astrophysics Data System (ADS)
Cannavo', Flavio; Camacho, Antonio G.; González, Pablo J.; Puglisi, Giuseppe; Fernández, José
2016-04-01
A recently proposed algorithm (Camacho et al., 2011) claims to rapidly estimate magmatic sources from surface geodetic data without any a priori assumption about source geometry. The algorithm takes the advantages of fast calculation from the analytical models and adds the capability to model free-shape distributed sources. Assuming homogenous elastic conditions, the approach can determine general geometrical configurations of pressured and/or density source and/or sliding structures corresponding to prescribed values of anomalous density, pressure and slip. These source bodies are described as aggregation of elemental point sources for pressure, density and slip, and they fit the whole data (keeping some 3D regularity conditions). Although some examples and applications have been already presented to demonstrate the ability of the algorithm in reconstructing a magma pressure source (e.g. Camacho et al., 2011,Cannavò et al., 2015), a systematic analysis of sensitivity and reliability of the algorithm is still lacking. In this explorative work we present results from a large statistical test designed to evaluate the advantages and limitations of the methodology by assessing its sensitivity to the free and constrained parameters involved in inversions. In particular, besides the source parameters, we focused on the ground deformation network topology, and noise in measurements. The proposed analysis can be used for a better interpretation of the algorithm results in real-case applications. Camacho, A. G., González, P. J., Fernández, J. & Berrino, G. (2011) Simultaneous inversion of surface deformation and gravity changes by means of extended bodies with a free geometry: Application to deforming calderas. J. Geophys. Res. 116. Cannavò F., Camacho A.G., González P.J., Mattia M., Puglisi G., Fernández J. (2015) Real Time Tracking of Magmatic Intrusions by means of Ground Deformation Modeling during Volcanic Crises, Scientific Reports, 5 (10970) doi:10.1038/srep
Tilt-Sensitivity Analysis for Space Telescopes
NASA Technical Reports Server (NTRS)
Papalexandris, Miltiadis; Waluschka, Eugene
2003-01-01
A report discusses a computational-simulation study of phase-front propagation in the Laser Interferometer Space Antenna (LISA), in which space telescopes would transmit and receive metrological laser beams along 5-Gm interferometer arms. The main objective of the study was to determine the sensitivity of the average phase of a beam with respect to fluctuations in pointing of the beam. The simulations account for the effects of obscurations by a secondary mirror and its supporting struts in a telescope, and for the effects of optical imperfections (especially tilt) of a telescope. A significant innovation introduced in this study is a methodology, applicable to space telescopes in general, for predicting the effects of optical imperfections. This methodology involves a Monte Carlo simulation in which one generates many random wavefront distortions and studies their effects through computational simulations of propagation. Then one performs a statistical analysis of the results of the simulations and computes the functional relations among such important design parameters as the sizes of distortions and the mean value and the variance of the loss of performance. These functional relations provide information regarding position and orientation tolerances relevant to design and operation.
Sensitivity analysis for solar plates
NASA Technical Reports Server (NTRS)
Aster, R. W.
1986-01-01
Economic evaluation methods and analyses of emerging photovoltaic (PV) technology since 1976 was prepared. This type of analysis was applied to the silicon research portion of the PV Program in order to determine the importance of this research effort in relationship to the successful development of commercial PV systems. All four generic types of PV that use silicon were addressed: crystal ingots grown either by the Czochralski method or an ingot casting method; ribbons pulled directly from molten silicon; an amorphous silicon thin film; and use of high concentration lenses. Three technologies were analyzed: the Union Carbide fluidized bed reactor process, the Hemlock process, and the Union Carbide Komatsu process. The major components of each process were assessed in terms of the costs of capital equipment, labor, materials, and utilities. These assessments were encoded as the probabilities assigned by experts for achieving various cost values or production rates.
Sensitivity analysis for solar plates
NASA Astrophysics Data System (ADS)
Aster, R. W.
1986-02-01
Economic evaluation methods and analyses of emerging photovoltaic (PV) technology since 1976 was prepared. This type of analysis was applied to the silicon research portion of the PV Program in order to determine the importance of this research effort in relationship to the successful development of commercial PV systems. All four generic types of PV that use silicon were addressed: crystal ingots grown either by the Czochralski method or an ingot casting method; ribbons pulled directly from molten silicon; an amorphous silicon thin film; and use of high concentration lenses. Three technologies were analyzed: the Union Carbide fluidized bed reactor process, the Hemlock process, and the Union Carbide Komatsu process. The major components of each process were assessed in terms of the costs of capital equipment, labor, materials, and utilities. These assessments were encoded as the probabilities assigned by experts for achieving various cost values or production rates.
Evolution of Geometric Sensitivity Derivatives from Computer Aided Design Models
NASA Technical Reports Server (NTRS)
Jones, William T.; Lazzara, David; Haimes, Robert
2010-01-01
The generation of design parameter sensitivity derivatives is required for gradient-based optimization. Such sensitivity derivatives are elusive at best when working with geometry defined within the solid modeling context of Computer-Aided Design (CAD) systems. Solid modeling CAD systems are often proprietary and always complex, thereby necessitating ad hoc procedures to infer parameter sensitivity. A new perspective is presented that makes direct use of the hierarchical associativity of CAD features to trace their evolution and thereby track design parameter sensitivity. In contrast to ad hoc methods, this method provides a more concise procedure following the model design intent and determining the sensitivity of CAD geometry directly to its respective defining parameters.
Multiple predictor smoothing methods for sensitivity analysis.
Helton, Jon Craig; Storlie, Curtis B.
2006-08-01
The use of multiple predictor smoothing methods in sampling-based sensitivity analyses of complex models is investigated. Specifically, sensitivity analysis procedures based on smoothing methods employing the stepwise application of the following nonparametric regression techniques are described: (1) locally weighted regression (LOESS), (2) additive models, (3) projection pursuit regression, and (4) recursive partitioning regression. The indicated procedures are illustrated with both simple test problems and results from a performance assessment for a radioactive waste disposal facility (i.e., the Waste Isolation Pilot Plant). As shown by the example illustrations, the use of smoothing procedures based on nonparametric regression techniques can yield more informative sensitivity analysis results than can be obtained with more traditional sensitivity analysis procedures based on linear regression, rank regression or quadratic regression when nonlinear relationships between model inputs and model predictions are present.
On the sensitivity analysis of porous material models
NASA Astrophysics Data System (ADS)
Ouisse, Morvan; Ichchou, Mohamed; Chedly, Slaheddine; Collet, Manuel
2012-11-01
Porous materials are used in many vibroacoustic applications. Different available models describe their behaviors according to materials' intrinsic characteristics. For instance, in the case of porous material with rigid frame, and according to the Champoux-Allard model, five parameters are employed. In this paper, an investigation about this model sensitivity to parameters according to frequency is conducted. Sobol and FAST algorithms are used for sensitivity analysis. A strong parametric frequency dependent hierarchy is shown. Sensitivity investigations confirm that resistivity is the most influent parameter when acoustic absorption and surface impedance of porous materials with rigid frame are considered. The analysis is first performed on a wide category of porous materials, and then restricted to a polyurethane foam analysis in order to illustrate the impact of the reduction of the design space. In a second part, a sensitivity analysis is performed using the Biot-Allard model with nine parameters including mechanical effects of the frame and conclusions are drawn through numerical simulations.
Ceramic tubesheet design analysis
Mallett, R.H.; Swindeman, R.W.
1996-06-01
A transport combustor is being commissioned at the Southern Services facility in Wilsonville, Alabama to provide a gaseous product for the assessment of hot-gas filtering systems. One of the barrier filters incorporates a ceramic tubesheet to support candle filters. The ceramic tubesheet, designed and manufactured by Industrial Filter and Pump Manufacturing Company (EF&PM), is unique and offers distinct advantages over metallic systems in terms of density, resistance to corrosion, and resistance to creep at operating temperatures above 815{degrees}C (1500{degrees}F). Nevertheless, the operational requirements of the ceramic tubesheet are severe. The tubesheet is almost 1.5 m in (55 in.) in diameter, has many penetrations, and must support the weight of the ceramic filters, coal ash accumulation, and a pressure drop (one atmosphere). Further, thermal stresses related to steady state and transient conditions will occur. To gain a better understanding of the structural performance limitations, a contract was placed with Mallett Technology, Inc. to perform a thermal and structural analysis of the tubesheet design. The design analysis specification and a preliminary design analysis were completed in the early part of 1995. The analyses indicated that modifications to the design were necessary to reduce thermal stress, and it was necessary to complete the redesign before the final thermal/mechanical analysis could be undertaken. The preliminary analysis identified the need to confirm that the physical and mechanical properties data used in the design were representative of the material in the tubesheet. Subsequently, few exploratory tests were performed at ORNL to evaluate the ceramic structural material.
An analytical sensitivity method for use in integrated aeroservoelastic aircraft design
NASA Technical Reports Server (NTRS)
Gilbert, Michael G.
1989-01-01
Interdisciplinary analysis capabilities have been developed for aeroservoelastic aircraft and large flexible spacecraft, but the requisite integrated design methods are only beginning to be developed. One integrated design method which has received attention is based on hierarchal problem decompositions, optimization, and design sensitivity analyses. This paper highlights a design sensitivity analysis method for Linear Quadratic Gaussian (LQG) optimal control laws, enabling the use of LQG techniques in the hierarchal design methodology. The LQG sensitivity analysis method calculates the change in the optimal control law and resulting controlled system responses due to changes in fixed design integration parameters using analytical sensitivity equations. Numerical results of an LQG design sensitivity analysis for a realistic aeroservoelastic aircraft example are presented. In this example, the sensitivity of the optimal control law and aircraft response for various parameters such as wing bending natural frequency is determined. The sensitivity results computed from the analytical expressions are used to estimate changes in response resulting from changes in the parameters. Comparisons of the estimates with exact calculated responses show they are reasonably accurate for + or - 15 percent changes in the parameters. Evaluation of the analytical expressions is computationally faster than equivalent finite difference calculations.
An analytical sensitivity method for use in integrated aeroservoelastic aircraft design
NASA Technical Reports Server (NTRS)
Gilbert, Michael G.
1989-01-01
Interdisciplinary analysis capabilities have been developed for aeroservoelastic aircraft and large flexible spacecraft, but the requisite integrated design methods are only beginning to be developed. One integrated design method which has received attention is based on hierarchal problem decompositions, optimization, and design sensitivity analyses. This paper highlights a design sensitivity analysis method for Linear Quadratic Gaussian (LQG) optimal control laws, enabling the use of LQG techniques in the hierarchal design methodology. The LQG sensitivity analysis method calculates the change in the optimal control law and resulting controlled system responses due to changes in fixed design integration parameters using analytical sensitivity equations. Numerical results of a LQG design sensitivity analysis for a realistic aeroservoelastic aircraft example are presented. In this example, the sensitivity of the optimal control law and aircraft response for various parameters such as wing bending natural frequency is determined. The sensitivity results computed from the analytical expressions are used to estimate changes in response resulting from changes in the parameters. Comparisons of the estimates with exact calculated responses show they are reasonably accurate for + or - 15 percent changes in the parameters. Evaluation of the analytical expressions is computationally faster than equivalent finite difference calculations.
Methodology for a stormwater sensitive urban watershed design
NASA Astrophysics Data System (ADS)
Romnée, Ambroise; Evrard, Arnaud; Trachte, Sophie
2015-11-01
In urban stormwater management, decentralized systems are nowadays worldwide experimented, including stormwater best management practices. However, a watershed-scale approach, relevant for urban hydrology, is almost always neglected when designing a stormwater management plan with best management practices. As a consequence, urban designers fail to convince public authorities of the actual hydrologic effectiveness of such an approach to urban watershed stormwater management. In this paper, we develop a design oriented methodology for studying the morphology of an urban watershed in terms of sustainable stormwater management. The methodology is a five-step method, firstly based on the cartographic analysis of many stormwater relevant indicators regarding the landscape, the urban fabric and the governance. The second step focuses on the identification of many territorial stakes and their corresponding strategies of a decentralized stormwater management. Based on the indicators, the stakes and the strategies, the third step defines many spatial typologies regarding the roadway system and the urban fabric system. The fourth step determines many stormwater management scenarios to be applied to both spatial typologies systems. The fifth step is the design of decentralized stormwater management projects integrating BMPs into each spatial typology. The methodology aims to advise urban designers and engineering offices in the right location and selection of BMPs without given them a hypothetical unique solution. Since every location and every watershed is different due to local guidelines and stakeholders, this paper provide a methodology for a stormwater sensitive urban watershed design that could be reproduced everywhere. As an example, the methodology is applied as a case study to an urban watershed in Belgium, confirming that the method is applicable to any urban watershed. This paper should be helpful for engineering and design offices in urban hydrology to define a
Liquid Acquisition Device Design Sensitivity Study
NASA Technical Reports Server (NTRS)
VanDyke, M. K.; Hastings, L. J.
2012-01-01
In-space propulsion often necessitates the use of a capillary liquid acquisition device (LAD) to assure that gas-free liquid propellant is available to support engine restarts in microgravity. If a capillary screen-channel device is chosen, then the designer must determine the appropriate combination screen mesh and channel geometry. A screen mesh selection which results in the smallest LAD width when compared to any other screen candidate (for a constant length) is desirable; however, no best screen exists for all LAD design requirements. Flow rate, percent fill, and acceleration are the most influential drivers for determining screen widths. Increased flow rates and reduced percent fills increase the through-the-screen flow pressure losses, which drive the LAD to increased widths regardless of screen choice. Similarly, increased acceleration levels and corresponding liquid head pressures drive the screen mesh selection toward a higher bubble point (liquid retention capability). After ruling out some screens on the basis of acceleration requirements alone, candidates can be identified by examining screens with small flow-loss-to-bubble point ratios for a given condition (i.e., comparing screens at certain flow rates and fill levels). Within the same flow rate and fill level, the screen constants inertia resistance coefficient, void fraction, screen pore or opening diameter, and bubble point can become the driving forces in identifying the smaller flow-loss-to-bubble point ratios.
A strategy to design novel structure photochromic sensitizers for dye-sensitized solar cells
Wu, Wenjun; Wang, Jiaxing; Zheng, Zhiwei; Hu, Yue; Jin, Jiayu; Zhang, Qiong; Hua, Jianli
2015-01-01
Two sensitizers with novel structure were designed and synthetized by introducing photochromic bisthienylethene (BTE) group into the conjugated system. Thanks to the photochromic effect the sensitizers have under ultraviolet and visible light, the conjugated bridge can be restructured and the resulting two photoisomers showed different behaviors in photovoltaic devices. This opens up a new research way for the dye-sensitized solar cells (DSSCs). PMID:25716204
A strategy to design novel structure photochromic sensitizers for dye-sensitized solar cells.
Wu, Wenjun; Wang, Jiaxing; Zheng, Zhiwei; Hu, Yue; Jin, Jiayu; Zhang, Qiong; Hua, Jianli
2015-01-01
Two sensitizers with novel structure were designed and synthetized by introducing photochromic bisthienylethene (BTE) group into the conjugated system. Thanks to the photochromic effect the sensitizers have under ultraviolet and visible light, the conjugated bridge can be restructured and the resulting two photoisomers showed different behaviors in photovoltaic devices. This opens up a new research way for the dye-sensitized solar cells (DSSCs). PMID:25716204
SEP thrust subsystem performance sensitivity analysis
NASA Technical Reports Server (NTRS)
Atkins, K. L.; Sauer, C. G., Jr.; Kerrisk, D. J.
1973-01-01
This is a two-part report on solar electric propulsion (SEP) performance sensitivity analysis. The first part describes the preliminary analysis of the SEP thrust system performance for an Encke rendezvous mission. A detailed description of thrust subsystem hardware tolerances on mission performance is included together with nominal spacecraft parameters based on these tolerances. The second part describes the method of analysis and graphical techniques used in generating the data for Part 1. Included is a description of both the trajectory program used and the additional software developed for this analysis. Part 2 also includes a comprehensive description of the use of the graphical techniques employed in this performance analysis.
Probabilistic sensitivity analysis in health economics.
Baio, Gianluca; Dawid, A Philip
2015-12-01
Health economic evaluations have recently become an important part of the clinical and medical research process and have built upon more advanced statistical decision-theoretic foundations. In some contexts, it is officially required that uncertainty about both parameters and observable variables be properly taken into account, increasingly often by means of Bayesian methods. Among these, probabilistic sensitivity analysis has assumed a predominant role. The objective of this article is to review the problem of health economic assessment from the standpoint of Bayesian statistical decision theory with particular attention to the philosophy underlying the procedures for sensitivity analysis. PMID:21930515
A PDE Sensitivity Equation Method for Optimal Aerodynamic Design
NASA Technical Reports Server (NTRS)
Borggaard, Jeff; Burns, John
1996-01-01
The use of gradient based optimization algorithms in inverse design is well established as a practical approach to aerodynamic design. A typical procedure uses a simulation scheme to evaluate the objective function (from the approximate states) and its gradient, then passes this information to an optimization algorithm. Once the simulation scheme (CFD flow solver) has been selected and used to provide approximate function evaluations, there are several possible approaches to the problem of computing gradients. One popular method is to differentiate the simulation scheme and compute design sensitivities that are then used to obtain gradients. Although this black-box approach has many advantages in shape optimization problems, one must compute mesh sensitivities in order to compute the design sensitivity. In this paper, we present an alternative approach using the PDE sensitivity equation to develop algorithms for computing gradients. This approach has the advantage that mesh sensitivities need not be computed. Moreover, when it is possible to use the CFD scheme for both the forward problem and the sensitivity equation, then there are computational advantages. An apparent disadvantage of this approach is that it does not always produce consistent derivatives. However, for a proper combination of discretization schemes, one can show asymptotic consistency under mesh refinement, which is often sufficient to guarantee convergence of the optimal design algorithm. In particular, we show that when asymptotically consistent schemes are combined with a trust-region optimization algorithm, the resulting optimal design method converges. We denote this approach as the sensitivity equation method. The sensitivity equation method is presented, convergence results are given and the approach is illustrated on two optimal design problems involving shocks.
Comparative Sensitivity Analysis of Muscle Activation Dynamics
Rockenfeller, Robert; Günther, Michael; Schmitt, Syn; Götz, Thomas
2015-01-01
We mathematically compared two models of mammalian striated muscle activation dynamics proposed by Hatze and Zajac. Both models are representative for a broad variety of biomechanical models formulated as ordinary differential equations (ODEs). These models incorporate parameters that directly represent known physiological properties. Other parameters have been introduced to reproduce empirical observations. We used sensitivity analysis to investigate the influence of model parameters on the ODE solutions. In addition, we expanded an existing approach to treating initial conditions as parameters and to calculating second-order sensitivities. Furthermore, we used a global sensitivity analysis approach to include finite ranges of parameter values. Hence, a theoretician striving for model reduction could use the method for identifying particularly low sensitivities to detect superfluous parameters. An experimenter could use it for identifying particularly high sensitivities to improve parameter estimation. Hatze's nonlinear model incorporates some parameters to which activation dynamics is clearly more sensitive than to any parameter in Zajac's linear model. Other than Zajac's model, Hatze's model can, however, reproduce measured shifts in optimal muscle length with varied muscle activity. Accordingly we extracted a specific parameter set for Hatze's model that combines best with a particular muscle force-length relation. PMID:26417379
Comparative Sensitivity Analysis of Muscle Activation Dynamics.
Rockenfeller, Robert; Günther, Michael; Schmitt, Syn; Götz, Thomas
2015-01-01
We mathematically compared two models of mammalian striated muscle activation dynamics proposed by Hatze and Zajac. Both models are representative for a broad variety of biomechanical models formulated as ordinary differential equations (ODEs). These models incorporate parameters that directly represent known physiological properties. Other parameters have been introduced to reproduce empirical observations. We used sensitivity analysis to investigate the influence of model parameters on the ODE solutions. In addition, we expanded an existing approach to treating initial conditions as parameters and to calculating second-order sensitivities. Furthermore, we used a global sensitivity analysis approach to include finite ranges of parameter values. Hence, a theoretician striving for model reduction could use the method for identifying particularly low sensitivities to detect superfluous parameters. An experimenter could use it for identifying particularly high sensitivities to improve parameter estimation. Hatze's nonlinear model incorporates some parameters to which activation dynamics is clearly more sensitive than to any parameter in Zajac's linear model. Other than Zajac's model, Hatze's model can, however, reproduce measured shifts in optimal muscle length with varied muscle activity. Accordingly we extracted a specific parameter set for Hatze's model that combines best with a particular muscle force-length relation. PMID:26417379
Variational Methods in Sensitivity Analysis and Optimization for Aerodynamic Applications
NASA Technical Reports Server (NTRS)
Ibrahim, A. H.; Hou, G. J.-W.; Tiwari, S. N. (Principal Investigator)
1996-01-01
Variational methods (VM) sensitivity analysis, which is the continuous alternative to the discrete sensitivity analysis, is employed to derive the costate (adjoint) equations, the transversality conditions, and the functional sensitivity derivatives. In the derivation of the sensitivity equations, the variational methods use the generalized calculus of variations, in which the variable boundary is considered as the design function. The converged solution of the state equations together with the converged solution of the costate equations are integrated along the domain boundary to uniquely determine the functional sensitivity derivatives with respect to the design function. The determination of the sensitivity derivatives of the performance index or functional entails the coupled solutions of the state and costate equations. As the stable and converged numerical solution of the costate equations with their boundary conditions are a priori unknown, numerical stability analysis is performed on both the state and costate equations. Thereafter, based on the amplification factors obtained by solving the generalized eigenvalue equations, the stability behavior of the costate equations is discussed and compared with the state (Euler) equations. The stability analysis of the costate equations suggests that the converged and stable solution of the costate equation is possible only if the computational domain of the costate equations is transformed to take into account the reverse flow nature of the costate equations. The application of the variational methods to aerodynamic shape optimization problems is demonstrated for internal flow problems at supersonic Mach number range. The study shows, that while maintaining the accuracy of the functional sensitivity derivatives within the reasonable range for engineering prediction purposes, the variational methods show a substantial gain in computational efficiency, i.e., computer time and memory, when compared with the finite
Pediatric Pain, Predictive Inference, and Sensitivity Analysis.
ERIC Educational Resources Information Center
Weiss, Robert
1994-01-01
Coping style and effects of counseling intervention on pain tolerance was studied for 61 elementary school students through immersion of hands in cold water. Bayesian predictive inference tools are able to distinguish between subject characteristics and manipulable treatments. Sensitivity analysis strengthens the certainty of conclusions about…
NASA Technical Reports Server (NTRS)
Zumwalt, Kenneth W.; El-Sayed, Mohamed E. M.
1990-01-01
This paper presents an analytical approach for incorporating design sensitivity calculations directly into the finite element analysis. The formulation depends on the implicit differentiation approach and requires few additional calculations to obtain the design sensitivity derivatives. In order to evaluate this approach, it is compared with the semi-analytical approach which is based on commonly used finite difference formulations. Both approaches are implemented to calculate the design sensitivities for continuum and structural isoparametric elements. To demonstrate the accuracy and robustness of the developed analytical approach compared to the semi-analytical approach, some test cases using different structural and continuum element types are presented.
NIR sensitivity analysis with the VANE
NASA Astrophysics Data System (ADS)
Carrillo, Justin T.; Goodin, Christopher T.; Baylot, Alex E.
2016-05-01
Near infrared (NIR) cameras, with peak sensitivity around 905-nm wavelengths, are increasingly used in object detection applications such as pedestrian detection, occupant detection in vehicles, and vehicle detection. In this work, we present the results of simulated sensitivity analysis for object detection with NIR cameras. The analysis was conducted using high performance computing (HPC) to determine the environmental effects on object detection in different terrains and environmental conditions. The Virtual Autonomous Navigation Environment (VANE) was used to simulate highresolution models for environment, terrain, vehicles, and sensors. In the experiment, an active fiducial marker was attached to the rear bumper of a vehicle. The camera was mounted on a following vehicle that trailed at varying standoff distances. Three different terrain conditions (rural, urban, and forest), two environmental conditions (clear and hazy), three different times of day (morning, noon, and evening), and six different standoff distances were used to perform the sensor sensitivity analysis. The NIR camera that was used for the simulation is the DMK firewire monochrome on a pan-tilt motor. Standoff distance was varied along with environment and environmental conditions to determine the critical failure points for the sensor. Feature matching was used to detect the markers in each frame of the simulation, and the percentage of frames in which one of the markers was detected was recorded. The standoff distance produced the biggest impact on the performance of the camera system, while the camera system was not sensitive to environment conditions.
Geothermal well cost sensitivity analysis: current status
Carson, C.C.; Lin, Y.T.
1980-01-01
The geothermal well-cost model developed by Sandia National Laboratories is being used to analyze the sensitivity of well costs to improvements in geothermal drilling technology. Three interim results from this modeling effort are discussed. The sensitivity of well costs to bit parameters, rig parameters, and material costs; an analysis of the cost reduction potential of an advanced bit; and a consideration of breakeven costs for new cementing technology. All three results illustrate that the well-cost savings arising from any new technology will be highly site-dependent but that in specific wells the advances considered can result in significant cost reductions.
Sensitivity analysis for magnetic induction tomography.
Soleimani, Manuchehr; Jersey-Willuhn, Karen
2004-01-01
This work focuses on sensitivity analysis of magnetic induction tomography in terms of theoretical modelling and numerical implementation. We will explain a new and efficient method to determine the Jacobian matrix, directly from the results of the forward solution. The results presented are for the eddy current approximation, and are given in terms of magnetic vector potential, which is computationally convenient, and which may be extracted directly from the FE solution of the forward problem. Examples of sensitivity maps for an opposite sensor geometry are also shown. PMID:17271947
Improving Discrete-Sensitivity-Based Approach for Practical Design Optimization
NASA Technical Reports Server (NTRS)
Baysal, Oktay; Cordero, Yvette; Pandya, Mohagna J.
1997-01-01
In developing the automated methodologies for simulation-based optimal shape designs, their accuracy, efficiency and practicality are the defining factors to their success. To that end, four recent improvements to the building blocks of such a methodology, intended for more practical design optimization, have been reported. First, in addition to a polynomial-based parameterization, a partial differential equation (PDE) based parameterization was shown to be a practical tool for a number of reasons. Second, an alternative has been incorporated to one of the tedious phases of developing such a methodology, namely, the automatic differentiation of the computer code for the flow analysis in order to generate the sensitivities. Third, by extending the methodology for the thin-layer Navier-Stokes (TLNS) based flow simulations, the more accurate flow physics was made available. However, the computer storage requirement for a shape optimization of a practical configuration with the -fidelity simulations (TLNS and dense-grid based simulations), required substantial computational resources. Therefore, the final improvement reported herein responded to this point by including the alternating-direct-implicit (ADI) based system solver as an alternative to the preconditioned biconjugate (PbCG) and other direct solvers.
Fault Tree Reliability Analysis and Design-for-reliability
1998-05-05
WinR provides a fault tree analysis capability for performing systems reliability and design-for-reliability analyses. The package includes capabilities for sensitivity and uncertainity analysis, field failure data analysis, and optimization.
Software Performs Complex Design Analysis
NASA Technical Reports Server (NTRS)
2008-01-01
Designers use computational fluid dynamics (CFD) to gain greater understanding of the fluid flow phenomena involved in components being designed. They also use finite element analysis (FEA) as a tool to help gain greater understanding of the structural response of components to loads, stresses and strains, and the prediction of failure modes. Automated CFD and FEA engineering design has centered on shape optimization, which has been hindered by two major problems: 1) inadequate shape parameterization algorithms, and 2) inadequate algorithms for CFD and FEA grid modification. Working with software engineers at Stennis Space Center, a NASA commercial partner, Optimal Solutions Software LLC, was able to utilize its revolutionary, one-of-a-kind arbitrary shape deformation (ASD) capability-a major advancement in solving these two aforementioned problems-to optimize the shapes of complex pipe components that transport highly sensitive fluids. The ASD technology solves the problem of inadequate shape parameterization algorithms by allowing the CFD designers to freely create their own shape parameters, therefore eliminating the restriction of only being able to use the computer-aided design (CAD) parameters. The problem of inadequate algorithms for CFD grid modification is solved by the fact that the new software performs a smooth volumetric deformation. This eliminates the extremely costly process of having to remesh the grid for every shape change desired. The program can perform a design change in a markedly reduced amount of time, a process that would traditionally involve the designer returning to the CAD model to reshape and then remesh the shapes, something that has been known to take hours, days-even weeks or months-depending upon the size of the model.
Sensitivity Analysis for Coupled Aero-structural Systems
NASA Technical Reports Server (NTRS)
Giunta, Anthony A.
1999-01-01
A novel method has been developed for calculating gradients of aerodynamic force and moment coefficients for an aeroelastic aircraft model. This method uses the Global Sensitivity Equations (GSE) to account for the aero-structural coupling, and a reduced-order modal analysis approach to condense the coupling bandwidth between the aerodynamic and structural models. Parallel computing is applied to reduce the computational expense of the numerous high fidelity aerodynamic analyses needed for the coupled aero-structural system. Good agreement is obtained between aerodynamic force and moment gradients computed with the GSE/modal analysis approach and the same quantities computed using brute-force, computationally expensive, finite difference approximations. A comparison between the computational expense of the GSE/modal analysis method and a pure finite difference approach is presented. These results show that the GSE/modal analysis approach is the more computationally efficient technique if sensitivity analysis is to be performed for two or more aircraft design parameters.
Microelectromechanical Resonant Accelerometer Designed with a High Sensitivity.
Zhang, Jing; Su, Yan; Shi, Qin; Qiu, An-Ping
2015-01-01
This paper describes the design and experimental evaluation of a silicon micro-machined resonant accelerometer (SMRA). This type of accelerometer works on the principle that a proof mass under acceleration applies force to two double-ended tuning fork (DETF) resonators, and the frequency output of two DETFs exhibits a differential shift. The dies of an SMRA are fabricated using silicon-on-insulator (SOI) processing and wafer-level vacuum packaging. This research aims to design a high-sensitivity SMRA because a high sensitivity allows for the acceleration signal to be easily demodulated by frequency counting techniques and decreases the noise level. This study applies the energy-consumed concept and the Nelder-Mead algorithm in the SMRA to address the design issues and further increase its sensitivity. Using this novel method, the sensitivity of the SMRA has been increased by 66.1%, which attributes to both the re-designed DETF and the reduced energy loss on the micro-lever. The results of both the closed-form and finite-element analyses are described and are in agreement with one another. A resonant frequency of approximately 22 kHz, a frequency sensitivity of over 250 Hz per g, a one-hour bias stability of 55 μg, a bias repeatability (1σ) of 48 μg and the bias-instability of 4.8 μg have been achieved. PMID:26633425
Microelectromechanical Resonant Accelerometer Designed with a High Sensitivity
Zhang, Jing; Su, Yan; Shi, Qin; Qiu, An-Ping
2015-01-01
This paper describes the design and experimental evaluation of a silicon micro-machined resonant accelerometer (SMRA). This type of accelerometer works on the principle that a proof mass under acceleration applies force to two double-ended tuning fork (DETF) resonators, and the frequency output of two DETFs exhibits a differential shift. The dies of an SMRA are fabricated using silicon-on-insulator (SOI) processing and wafer-level vacuum packaging. This research aims to design a high-sensitivity SMRA because a high sensitivity allows for the acceleration signal to be easily demodulated by frequency counting techniques and decreases the noise level. This study applies the energy-consumed concept and the Nelder-Mead algorithm in the SMRA to address the design issues and further increase its sensitivity. Using this novel method, the sensitivity of the SMRA has been increased by 66.1%, which attributes to both the re-designed DETF and the reduced energy loss on the micro-lever. The results of both the closed-form and finite-element analyses are described and are in agreement with one another. A resonant frequency of approximately 22 kHz, a frequency sensitivity of over 250 Hz per g, a one-hour bias stability of 55 μg, a bias repeatability (1σ) of 48 μg and the bias-instability of 4.8 μg have been achieved. PMID:26633425
Passive solar design handbook. Volume 3: Passive solar design analysis
NASA Astrophysics Data System (ADS)
Jones, R. W.; Bascomb, J. D.; Kosiewicz, C. E.; Lazarus, G. S.; McFarland, R. D.; Wray, W. O.
1982-07-01
Simple analytical methods concerning the design of passive solar heating systems are presented with an emphasis on the average annual heating energy consumption. Key terminology and methods are reviewed. The solar load ratio (SLR) is defined, and its relationship to analysis methods is reviewed. The annual calculation, or Load Collector Ratio (LCR) method, is outlined. Sensitivity data are discussed. Information is presented on balancing conservation and passive solar strategies in building design. Detailed analysis data are presented for direct gain and sunspace systems, and details of the systems are described. Key design parameters are discussed in terms of their impact on annual heating performance of the building. These are the sensitivity data. The SLR correlations for the respective system types are described. The monthly calculation, or SLR method, based on the SLR correlations, is reviewed. Performance data are given for 9 direct gain systems and 15 water wall and 42 Trombe wall systems.
Moridis, G.; Reagan, M.T.
2011-01-15
In the second paper of this series, we evaluate two additional well designs for production from permafrost-associated (PA) hydrate deposits. Both designs are within the capabilities of conventional technology. We determine that large volumes of gas can be produced at high rates (several MMSCFD) for long times using either well design. The production approach involves initial fluid withdrawal from the water zone underneath the hydrate-bearing layer (HBL). The production process follows a cyclical pattern, with each cycle composed of two stages: a long stage (months to years) of increasing gas production and decreasing water production, and a short stage (days to weeks) that involves destruction of the secondary hydrate (mainly through warm water injection) that evolves during the first stage, and is followed by a reduction in the fluid withdrawal rate. A well configuration with completion throughout the HBL leads to high production rates, but also the creation of a secondary hydrate barrier around the well that needs to be destroyed regularly by water injection. However, a configuration that initially involves heating of the outer surface of the wellbore and later continuous injection of warm water at low rates (Case C) appears to deliver optimum performance over the period it takes for the exhaustion of the hydrate deposit. Using Case C as the standard, we determine that gas production from PA hydrate deposits increases with the fluid withdrawal rate, the initial hydrate saturation and temperature, and with the formation permeability.
New Methods for Sensitivity Analysis in Chaotic, Turbulent Fluid Flows
NASA Astrophysics Data System (ADS)
Blonigan, Patrick; Wang, Qiqi
2012-11-01
Computational methods for sensitivity analysis are invaluable tools for fluid mechanics research and engineering design. These methods are used in many applications, including aerodynamic shape optimization and adaptive grid refinement. However, traditional sensitivity analysis methods break down when applied to long-time averaged quantities in chaotic fluid flowfields, such as those obtained using high-fidelity turbulence simulations. Also, a number of dynamical properties of chaotic fluid flows, most notably the ``Butterfly Effect,'' make the formulation of new sensitivity analysis methods difficult. This talk will outline two chaotic sensitivity analysis methods. The first method, the Fokker-Planck adjoint method, forms a probability density function on the strange attractor associated with the system and uses its adjoint to find gradients. The second method, the Least Squares Sensitivity method, finds some ``shadow trajectory'' in phase space for which perturbations do not grow exponentially. This method is formulated as a quadratic programing problem with linear constraints. This talk is concluded with demonstrations of these new methods on some example problems, including the Lorenz attractor and flow around an airfoil at a high angle of attack.
Spacecraft design sensitivity for a disaster warning satellite system
NASA Technical Reports Server (NTRS)
Maloy, J. E.; Provencher, C. E.; Leroy, B. E.; Braley, R. C.; Shumaker, H. A.
1977-01-01
A disaster warning satellite (DWS) is described for warning the general public of impending natural catastrophes. The concept is responsive to NOAA requirements and maximizes the use of ATS-6 technology. Upon completion of concept development, the study was extended to establishing the sensitivity of the DWSS spacecraft power, weight, and cost to variations in both warning and conventional communications functions. The results of this sensitivity analysis are presented.
Computational methods for efficient structural reliability and reliability sensitivity analysis
NASA Technical Reports Server (NTRS)
Wu, Y.-T.
1993-01-01
This paper presents recent developments in efficient structural reliability analysis methods. The paper proposes an efficient, adaptive importance sampling (AIS) method that can be used to compute reliability and reliability sensitivities. The AIS approach uses a sampling density that is proportional to the joint PDF of the random variables. Starting from an initial approximate failure domain, sampling proceeds adaptively and incrementally with the goal of reaching a sampling domain that is slightly greater than the failure domain to minimize over-sampling in the safe region. Several reliability sensitivity coefficients are proposed that can be computed directly and easily from the above AIS-based failure points. These probability sensitivities can be used for identifying key random variables and for adjusting design to achieve reliability-based objectives. The proposed AIS methodology is demonstrated using a turbine blade reliability analysis problem.
Sensitivity analysis of the critical speed in railway vehicle dynamics
NASA Astrophysics Data System (ADS)
Bigoni, D.; True, H.; Engsig-Karup, A. P.
2014-05-01
We present an approach to global sensitivity analysis aiming at the reduction of its computational cost without compromising the results. The method is based on sampling methods, cubature rules, high-dimensional model representation and total sensitivity indices. It is applied to a half car with a two-axle Cooperrider bogie, in order to study the sensitivity of the critical speed with respect to the suspension parameters. The importance of a certain suspension component is expressed by the variance in critical speed that is ascribable to it. This proves to be useful in the identification of parameters for which the accuracy of their values is critically important. The approach has a general applicability in many engineering fields and does not require the knowledge of the particular solver of the dynamical system. This analysis can be used as part of the virtual homologation procedure and to help engineers during the design phase of complex systems.
Li, Peiyue; Qian, Hui; Wu, Jianhua; Chen, Jie
2013-03-01
Sensitivity analysis is becoming increasingly widespread in many fields of engineering and sciences and has become a necessary step to verify the feasibility and reliability of a model or a method. The sensitivity of the Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) method in water quality assessment mainly includes sensitivity to the parameter weights and sensitivity to the index input data. In the present study, the sensitivity of TOPSIS to the parameter weights was discussed in detail. The present study assumed the original parameter weights to be equal to each other, and then each weight was changed separately to see how the assessment results would be affected. Fourteen schemes were designed to investigate the sensitivity to the variation of each weight. The variation ranges that keep the assessment results unchangeable were also derived theoretically. The results show that the final assessment results will change when the weights increase or decrease by ±20 to ±50 %. The feedback of different samples to the variation of a given weight is different, and the feedback of a given sample to the variation of different weights is also different. The final assessment results can keep relatively stable when a given weight is disturbed as long as the initial variation ratios meet one of the eight derived requirements. PMID:22752962
Tang, Qiaohong; Mo, Zhongjun; Yao, Jie; Li, Qi; Du, Chenfei; Wang, Lizhen; Fan, Yubo
2014-12-01
This study was aimed to estimate the effect of different ProDisc-C arthroplasty designs after it was implanted to C5-C6 cervicalspine. Finite element (FE) model of intact C5-C6 segments including the vertebrae and disc was developed and validated. Ball-and-socket artificial disc prosthesis model (ProDisc-C, Synthes) was implanted into the validated FE model and the curvature of the ProDisc-C prosthesis was varied. All models were loaded with compressed force 74 N and the pure moment of 1.8 Nm along flexion-extension and bilateral bending and axial torsion separately. The results indicated that the variation in the curvature of ball and socket configuration would influence the range of motion in flexion/extension, while there were not apparently differences under other conditions of loads. The method increasing the curvature will solve the stress concentration of the polyethylene, but it will also bring adverse outcomes, such as facet joint force increasing and ligament tension increasing. Therefore, the design of artificial discs should be considered comprehensively to reserve the range of motion as well as to avoid the adverse problems, so as not to affect the long-term clinical results. PMID:25868242
Diagnostic Analysis of Middle Atmosphere Climate Sensitivity
NASA Astrophysics Data System (ADS)
Zhu, X.; Cai, M.; Swartz, W. H.; Coy, L.; Yee, J.; Talaat, E. R.
2013-12-01
Both the middle atmosphere climate sensitivity associated with the cooling trend and its uncertainty due to a complex system of drivers increase with altitude. Furthermore, the combined effect of middle atmosphere cooling due to long-lived greenhouse gases and ozone is also associated with natural climate variations due to solar activity. To understand and predict climate change from a global perspective, we use the recently developed climate feedback-response analysis method (CFRAM) to identify and isolate the signals from the external forcing and from different feedback processes in the middle atmosphere climate system. By use of the JHU/APL middle atmosphere radiation algorithm, the CFRAM is applied to the model output fields of the high-altitude GEOS-5 climate model in the middle atmosphere to delineate the individual contributions of radiative forcing to middle atmosphere climate sensitivity.
Sensitivity analysis and approximation methods for general eigenvalue problems
NASA Technical Reports Server (NTRS)
Murthy, D. V.; Haftka, R. T.
1986-01-01
Optimization of dynamic systems involving complex non-hermitian matrices is often computationally expensive. Major contributors to the computational expense are the sensitivity analysis and reanalysis of a modified design. The present work seeks to alleviate this computational burden by identifying efficient sensitivity analysis and approximate reanalysis methods. For the algebraic eigenvalue problem involving non-hermitian matrices, algorithms for sensitivity analysis and approximate reanalysis are classified, compared and evaluated for efficiency and accuracy. Proper eigenvector normalization is discussed. An improved method for calculating derivatives of eigenvectors is proposed based on a more rational normalization condition and taking advantage of matrix sparsity. Important numerical aspects of this method are also discussed. To alleviate the problem of reanalysis, various approximation methods for eigenvalues are proposed and evaluated. Linear and quadratic approximations are based directly on the Taylor series. Several approximation methods are developed based on the generalized Rayleigh quotient for the eigenvalue problem. Approximation methods based on trace theorem give high accuracy without needing any derivatives. Operation counts for the computation of the approximations are given. General recommendations are made for the selection of appropriate approximation technique as a function of the matrix size, number of design variables, number of eigenvalues of interest and the number of design points at which approximation is sought.
Sensitivity method for integrated structure/active control law design
NASA Technical Reports Server (NTRS)
Gilbert, Michael G.
1987-01-01
The development is described of an integrated structure/active control law design methodology for aeroelastic aircraft applications. A short motivating introduction to aeroservoelasticity is given along with the need for integrated structures/controls design algorithms. Three alternative approaches to development of an integrated design method are briefly discussed with regards to complexity, coordination and tradeoff strategies, and the nature of the resulting solutions. This leads to the formulation of the proposed approach which is based on the concepts of sensitivity of optimum solutions and multi-level decompositions. The concept of sensitivity of optimum is explained in more detail and compared with traditional sensitivity concepts of classical control theory. The analytical sensitivity expressions for the solution of the linear, quadratic cost, Gaussian (LQG) control problem are summarized in terms of the linear regulator solution and the Kalman Filter solution. Numerical results for a state space aeroelastic model of the DAST ARW-II vehicle are given, showing the changes in aircraft responses to variations of a structural parameter, in this case first wing bending natural frequency.
The Theoretical Foundation of Sensitivity Analysis for GPS
NASA Astrophysics Data System (ADS)
Shikoska, U.; Davchev, D.; Shikoski, J.
2008-10-01
In this paper the equations of sensitivity analysis are derived and established theoretical underpinnings for the analyses. Paper propounds a land-vehicle navigation concepts and definition for sensitivity analysis. Equations of sensitivity analysis are presented for a linear Kalman filter and case study is given to illustrate the use of sensitivity analysis to the reader. At the end of the paper, extensions that are required for this research are made to the basic equations of sensitivity analysis specifically; the equations of sensitivity analysis are re-derived for a linearized Kalman filter.
Efficient sensitivity analysis and optimization of a helicopter rotor
NASA Technical Reports Server (NTRS)
Lim, Joon W.; Chopra, Inderjit
1989-01-01
Aeroelastic optimization of a system essentially consists of the determination of the optimum values of design variables which minimize the objective function and satisfy certain aeroelastic and geometric constraints. The process of aeroelastic optimization analysis is illustrated. To carry out aeroelastic optimization effectively, one needs a reliable analysis procedure to determine steady response and stability of a rotor system in forward flight. The rotor dynamic analysis used in the present study developed inhouse at the University of Maryland is based on finite elements in space and time. The analysis consists of two major phases: vehicle trim and rotor steady response (coupled trim analysis), and aeroelastic stability of the blade. For a reduction of helicopter vibration, the optimization process requires the sensitivity derivatives of the objective function and aeroelastic stability constraints. For this, the derivatives of steady response, hub loads and blade stability roots are calculated using a direct analytical approach. An automated optimization procedure is developed by coupling the rotor dynamic analysis, design sensitivity analysis and constrained optimization code CONMIN.
LCA data quality: sensitivity and uncertainty analysis.
Guo, M; Murphy, R J
2012-10-01
Life cycle assessment (LCA) data quality issues were investigated by using case studies on products from starch-polyvinyl alcohol based biopolymers and petrochemical alternatives. The time horizon chosen for the characterization models was shown to be an important sensitive parameter for the environmental profiles of all the polymers. In the global warming potential and the toxicity potential categories the comparison between biopolymers and petrochemical counterparts altered as the time horizon extended from 20 years to infinite time. These case studies demonstrated that the use of a single time horizon provide only one perspective on the LCA outcomes which could introduce an inadvertent bias into LCA outcomes especially in toxicity impact categories and thus dynamic LCA characterization models with varying time horizons are recommended as a measure of the robustness for LCAs especially comparative assessments. This study also presents an approach to integrate statistical methods into LCA models for analyzing uncertainty in industrial and computer-simulated datasets. We calibrated probabilities for the LCA outcomes for biopolymer products arising from uncertainty in the inventory and from data variation characteristics this has enabled assigning confidence to the LCIA outcomes in specific impact categories for the biopolymer vs. petrochemical polymer comparisons undertaken. Uncertainty combined with the sensitivity analysis carried out in this study has led to a transparent increase in confidence in the LCA findings. We conclude that LCAs lacking explicit interpretation of the degree of uncertainty and sensitivities are of limited value as robust evidence for decision making or comparative assertions. PMID:22854094
High Sensitivity MEMS Strain Sensor: Design and Simulation
Mohammed, Ahmed A. S.; Moussa, Walied A.; Lou, Edmond
2008-01-01
In this article, we report on the new design of a miniaturized strain microsensor. The proposed sensor utilizes the piezoresistive properties of doped single crystal silicon. Employing the Micro Electro Mechanical Systems (MEMS) technology, high sensor sensitivities and resolutions have been achieved. The current sensor design employs different levels of signal amplifications. These amplifications include geometric, material and electronic levels. The sensor and the electronic circuits can be integrated on a single chip, and packaged as a small functional unit. The sensor converts input strain to resistance change, which can be transformed to bridge imbalance voltage. An analog output that demonstrates high sensitivity (0.03mV/με), high absolute resolution (1με) and low power consumption (100μA) with a maximum range of ±4000με has been reported. These performance characteristics have been achieved with high signal stability over a wide temperature range (±50°C), which introduces the proposed MEMS strain sensor as a strong candidate for wireless strain sensing applications under harsh environmental conditions. Moreover, this sensor has been designed, verified and can be easily modified to measure other values such as force, torque…etc. In this work, the sensor design is achieved using Finite Element Method (FEM) with the application of the piezoresistivity theory. This design process and the microfabrication process flow to prototype the design have been presented.
A comparison of two sampling methods for global sensitivity analysis
NASA Astrophysics Data System (ADS)
Tarantola, Stefano; Becker, William; Zeitz, Dirk
2012-05-01
We compare the convergence properties of two different quasi-random sampling designs - Sobol's quasi-Monte Carlo, and Latin supercube sampling in variance-based global sensitivity analysis. We use the non-monotonic V-function of Sobol' as base case-study, and compare the performance of both sampling strategies at increasing sample size and dimensionality against analytical values. The results indicate that in almost all cases investigated here, the Sobol' design performs better. This, coupled with the fact that effective Latin supercube sampling requires a priori knowledge of the interaction properties of the function, leads us to recommend Sobol' sampling in most practical cases.
Design and Synthesis of a Calcium-Sensitive Photocage.
Heckman, Laurel M; Grimm, Jonathan B; Schreiter, Eric R; Kim, Charles; Verdecia, Mark A; Shields, Brenda C; Lavis, Luke D
2016-07-11
Photolabile protecting groups (or "photocages") enable precise spatiotemporal control of chemical functionality and facilitate advanced biological experiments. Extant photocages exhibit a simple input-output relationship, however, where application of light elicits a photochemical reaction irrespective of the environment. Herein, we refine and extend the concept of photolabile groups, synthesizing the first Ca(2+) -sensitive photocage. This system functions as a chemical coincidence detector, releasing small molecules only in the presence of both light and elevated [Ca(2+) ]. Caging a fluorophore with this ion-sensitive moiety yields an "ion integrator" that permanently marks cells undergoing high Ca(2+) flux during an illumination-defined time period. Our general design concept demonstrates a new class of light-sensitive material for cellular imaging, sensing, and targeted molecular delivery. PMID:27218487
Simple Sensitivity Analysis for Orion GNC
NASA Technical Reports Server (NTRS)
Pressburger, Tom; Hoelscher, Brian; Martin, Rodney; Sricharan, Kumar
2013-01-01
The performance of Orion flight software, especially its GNC software, is being analyzed by running Monte Carlo simulations of Orion spacecraft flights. The simulated performance is analyzed for conformance with flight requirements, expressed as performance constraints. Flight requirements include guidance (e.g. touchdown distance from target) and control (e.g., control saturation) as well as performance (e.g., heat load constraints). The Monte Carlo simulations disperse hundreds of simulation input variables, for everything from mass properties to date of launch.We describe in this paper a sensitivity analysis tool (Critical Factors Tool or CFT) developed to find the input variables or pairs of variables which by themselves significantly influence satisfaction of requirements or significantly affect key performance metrics (e.g., touchdown distance from target). Knowing these factors can inform robustness analysis, can inform where engineering resources are most needed, and could even affect operations. The contributions of this paper include the introduction of novel sensitivity measures, such as estimating success probability, and a technique for determining whether pairs of factors are interacting dependently or independently. The tool found that input variables such as moments, mass, thrust dispersions, and date of launch were found to be significant factors for success of various requirements. Examples are shown in this paper as well as a summary and physics discussion of EFT-1 driving factors that the tool found.
Bayesian sensitivity analysis of bifurcating nonlinear models
NASA Astrophysics Data System (ADS)
Becker, W.; Worden, K.; Rowson, J.
2013-01-01
Sensitivity analysis allows one to investigate how changes in input parameters to a system affect the output. When computational expense is a concern, metamodels such as Gaussian processes can offer considerable computational savings over Monte Carlo methods, albeit at the expense of introducing a data modelling problem. In particular, Gaussian processes assume a smooth, non-bifurcating response surface. This work highlights a recent extension to Gaussian processes which uses a decision tree to partition the input space into homogeneous regions, and then fits separate Gaussian processes to each region. In this way, bifurcations can be modelled at region boundaries and different regions can have different covariance properties. To test this method, both the treed and standard methods were applied to the bifurcating response of a Duffing oscillator and a bifurcating FE model of a heart valve. It was found that the treed Gaussian process provides a practical way of performing uncertainty and sensitivity analysis on large, potentially-bifurcating models, which cannot be dealt with by using a single GP, although an open problem remains how to manage bifurcation boundaries that are not parallel to coordinate axes.
A Post-Monte-Carlo Sensitivity Analysis Code
2000-04-04
SATOOL (Sensitivity Analysis TOOL) is a code for sensitivity analysis, following an uncertainity analysis with Monte Carlo simulations. Sensitivity analysis identifies those input variables, whose variance contributes dominatly to the variance in the output. This analysis can be used to reduce the variance in the output variables by redefining the "sensitive" variables with greater precision, i.e. with lower variance. The code identifies a group of sensitive variables, ranks them in the order of importance andmore » also quantifies the relative importance among the sensitive variables.« less
Multiplexed analysis of chromosome conformation at vastly improved sensitivity
Davies, James O.J.; Telenius, Jelena M.; McGowan, Simon; Roberts, Nigel A.; Taylor, Stephen; Higgs, Douglas R.; Hughes, Jim R.
2015-01-01
Since methods for analysing chromosome conformation in mammalian cells are either low resolution or low throughput and are technically challenging they are not widely used outside of specialised laboratories. We have re-designed the Capture-C method producing a new approach, called next generation (NG) Capture-C. This produces unprecedented levels of sensitivity and reproducibility and can be used to analyse many genetic loci and samples simultaneously. Importantly, high-resolution data can be produced on as few as 100,000 cells and SNPs can be used to generate allele specific tracks. The method is straightforward to perform and should therefore greatly facilitate the task of linking SNPs identified by genome wide association studies with the genes they influence. The complete and detailed protocol presented here, with new publicly available tools for library design and data analysis, will allow most laboratories to analyse chromatin conformation at levels of sensitivity and throughput that were previously impossible. PMID:26595209
Sensitivity Analysis of Chaotic Flow around Two-Dimensional Airfoil
NASA Astrophysics Data System (ADS)
Blonigan, Patrick; Wang, Qiqi; Nielsen, Eric; Diskin, Boris
2015-11-01
Computational methods for sensitivity analysis are invaluable tools for fluid dynamics research and engineering design. These methods are used in many applications, including aerodynamic shape optimization and adaptive grid refinement. However, traditional sensitivity analysis methods, including the adjoint method, break down when applied to long-time averaged quantities in chaotic fluid flow fields, such as high-fidelity turbulence simulations. This break down is due to the ``Butterfly Effect'' the high sensitivity of chaotic dynamical systems to the initial condition. A new sensitivity analysis method developed by the authors, Least Squares Shadowing (LSS), can compute useful and accurate gradients for quantities of interest in chaotic dynamical systems. LSS computes gradients using the ``shadow trajectory'', a phase space trajectory (or solution) for which perturbations to the flow field do not grow exponentially in time. To efficiently compute many gradients for one objective function, we use an adjoint version of LSS. This talk will briefly outline Least Squares Shadowing and demonstrate it on chaotic flow around a Two-Dimensional airfoil.
Sensitivity Analysis Of Technological And Material Parameters In Roll Forming
NASA Astrophysics Data System (ADS)
Gehring, Albrecht; Saal, Helmut
2007-05-01
Roll forming is applied for several decades to manufacture thin gauged profiles. However, the knowledge about this technology is still based on empirical approaches. Due to the complexity of the forming process, the main effects on profile properties are difficult to identify. This is especially true for the interaction of technological parameters and material parameters. General considerations for building a finite-element model of the roll forming process are given in this paper. A sensitivity analysis is performed on base of a statistical design approach in order to identify the effects and interactions of different parameters on profile properties. The parameters included in the analysis are the roll diameter, the rolling speed, the sheet thickness, friction between the tools and the sheet and the strain hardening behavior of the sheet material. The analysis includes an isotropic hardening model and a nonlinear kinematic hardening model. All jobs are executed parallel to reduce the overall time as the sensitivity analysis requires much CPU-time. The results of the sensitivity analysis demonstrate the opportunities to improve the properties of roll formed profiles by adjusting technological and material parameters to their optimum interacting performance.
Scalable analysis tools for sensitivity analysis and UQ (3160) results.
Karelitz, David B.; Ice, Lisa G.; Thompson, David C.; Bennett, Janine C.; Fabian, Nathan; Scott, W. Alan; Moreland, Kenneth D.
2009-09-01
The 9/30/2009 ASC Level 2 Scalable Analysis Tools for Sensitivity Analysis and UQ (Milestone 3160) contains feature recognition capability required by the user community for certain verification and validation tasks focused around sensitivity analysis and uncertainty quantification (UQ). These feature recognition capabilities include crater detection, characterization, and analysis from CTH simulation data; the ability to call fragment and crater identification code from within a CTH simulation; and the ability to output fragments in a geometric format that includes data values over the fragments. The feature recognition capabilities were tested extensively on sample and actual simulations. In addition, a number of stretch criteria were met including the ability to visualize CTH tracer particles and the ability to visualize output from within an S3D simulation.
NASA Technical Reports Server (NTRS)
1982-01-01
Design of military and civil helicopters, produced by Bell Helicopter Textron, and aided by the use of COSMIC'S computer program VASP enables performance of more accurate analyses to insure product safety and improved production efficiency.
Stochastic Simulations and Sensitivity Analysis of Plasma Flow
Lin, Guang; Karniadakis, George E.
2008-08-01
For complex physical systems with large number of random inputs, it will be very expensive to perform stochastic simulations for all of the random inputs. Stochastic sensitivity analysis is introduced in this paper to rank the significance of random inputs, provide information on which random input has more influence on the system outputs and the coupling or interaction effect among different random inputs. There are two types of numerical methods in stochastic sensitivity analysis: local and global methods. The local approach, which relies on a partial derivative of output with respect to parameters, is used to measure the sensitivity around a local operating point. When the system has strong nonlinearities and parameters fluctuate within a wide range from their nominal values, the local sensitivity does not provide full information to the system operators. On the other side, the global approach examines the sensitivity from the entire range of the parameter variations. The global screening methods, based on One-At-a-Time (OAT) perturbation of parameters, rank the significant parameters and identify their interaction among a large number of parameters. Several screening methods have been proposed in literature, i.e., the Morris method, Cotter's method, factorial experimentation, and iterated fractional factorial design. In this paper, the Morris method, Monte Carlo sampling method, Quasi-Monte Carlo method and collocation method based on sparse grids are studied. Additionally, two MHD examples are presented to demonstrate the capability and efficiency of the stochastic sensitivity analysis, which can be used as a pre-screening technique for reducing the dimensionality and hence the cost in stochastic simulations.
Global sensitivity analysis of groundwater transport
NASA Astrophysics Data System (ADS)
Cvetkovic, V.; Soltani, S.; Vigouroux, G.
2015-12-01
In this work we address the model and parametric sensitivity of groundwater transport using the Lagrangian-Stochastic Advection-Reaction (LaSAR) methodology. The 'attenuation index' is used as a relevant and convenient measure of the coupled transport mechanisms. The coefficients of variation (CV) for seven uncertain parameters are assumed to be between 0.25 and 3.5, the highest value being for the lower bound of the mass transfer coefficient k0 . In almost all cases, the uncertainties in the macro-dispersion (CV = 0.35) and in the mass transfer rate k0 (CV = 3.5) are most significant. The global sensitivity analysis using Sobol and derivative-based indices yield consistent rankings on the significance of different models and/or parameter ranges. The results presented here are generic however the proposed methodology can be easily adapted to specific conditions where uncertainty ranges in models and/or parameters can be estimated from field and/or laboratory measurements.
Multicomponent dynamical nucleation theory and sensitivity analysis.
Kathmann, Shawn M; Schenter, Gregory K; Garrett, Bruce C
2004-05-15
Vapor to liquid multicomponent nucleation is a dynamical process governed by a delicate interplay between condensation and evaporation. Since the population of the vapor phase is dominated by monomers at reasonable supersaturations, the formation of clusters is governed by monomer association and dissociation reactions. Although there is no intrinsic barrier in the interaction potential along the minimum energy path for the association process, the formation of a cluster is impeded by a free energy barrier. Dynamical nucleation theory provides a framework in which equilibrium evaporation rate constants can be calculated and the corresponding condensation rate constants determined from detailed balance. The nucleation rate can then be obtained by solving the kinetic equations. The rate constants governing the multistep kinetics of multicomponent nucleation including sensitivity analysis and the potential influence of contaminants will be presented and discussed. PMID:15267849
Sensitivity analysis of periodic matrix population models.
Caswell, Hal; Shyu, Esther
2012-12-01
Periodic matrix models are frequently used to describe cyclic temporal variation (seasonal or interannual) and to account for the operation of multiple processes (e.g., demography and dispersal) within a single projection interval. In either case, the models take the form of periodic matrix products. The perturbation analysis of periodic models must trace the effects of parameter changes, at each phase of the cycle, on output variables that are calculated over the entire cycle. Here, we apply matrix calculus to obtain the sensitivity and elasticity of scalar-, vector-, or matrix-valued output variables. We apply the method to linear models for periodic environments (including seasonal harvest models), to vec-permutation models in which individuals are classified by multiple criteria, and to nonlinear models including both immediate and delayed density dependence. The results can be used to evaluate management strategies and to study selection gradients in periodic environments. PMID:23316494
Sensitivity Analysis of the Static Aeroelastic Response of a Wing
NASA Technical Reports Server (NTRS)
Eldred, Lloyd B.
1993-01-01
A technique to obtain the sensitivity of the static aeroelastic response of a three dimensional wing model is designed and implemented. The formulation is quite general and accepts any aerodynamic and structural analysis capability. A program to combine the discipline level, or local, sensitivities into global sensitivity derivatives is developed. A variety of representations of the wing pressure field are developed and tested to determine the most accurate and efficient scheme for representing the field outside of the aerodynamic code. Chebyshev polynomials are used to globally fit the pressure field. This approach had some difficulties in representing local variations in the field, so a variety of local interpolation polynomial pressure representations are also implemented. These panel based representations use a constant pressure value, a bilinearly interpolated value. or a biquadraticallv interpolated value. The interpolation polynomial approaches do an excellent job of reducing the numerical problems of the global approach for comparable computational effort. Regardless of the pressure representation used. sensitivity and response results with excellent accuracy have been produced for large integrated quantities such as wing tip deflection and trim angle of attack. The sensitivities of such things as individual generalized displacements have been found with fair accuracy. In general, accuracy is found to be proportional to the relative size of the derivatives to the quantity itself.
Design and performance of a positron-sensitive surgical probe
NASA Astrophysics Data System (ADS)
Liu, Fang
We report the design and performance of a portable positron-sensitive surgical imaging probe. The probe is designed to be sensitive to positrons and capable of rejecting background gammas including 511 keV. The probe consists of a multi-anode PMT and an 8 x 8 array of thin 2 mm x 2 mm plastic scintillators coupled 1:1 to GSO crystals. The probe uses three selection criteria to identify positrons. An energy threshold on the plastic signals reduces the false positron signals in the plastic due to background gammas; a second energy threshold on the PMT sum signal greatly reduces background gammas in the GSO. Finally, a timing window accepts only 511 keV gammas from the GSO that arrive within 15 ns of the plastic signals, reducing accidental coincidences to a negligible level. The first application being investigated is sentinel lymph node (SLN) surgery, to identify in real-time the location of SLNs in the axilla with high 18F-FDG uptake, which may indicate metastasis. Our simulations and measurements show that the probe's pixel separation ability in terms of peak-to-valley ratio is ˜3.5. The performance measurements also show that the 64-pixel probe has a sensitivity of 4.7 kcps/muCi using optimal signal selection criteria. For example, it is able to detect in 10 seconds a ˜4 mm lesion with a true-to-background ratio of ˜3 at a tumor uptake ratio of ˜8:1. The signal selection criteria can be fine-tuned, either for higher sensitivity, or for a higher image contrast.
Global sensitivity analysis in wind energy assessment
NASA Astrophysics Data System (ADS)
Tsvetkova, O.; Ouarda, T. B.
2012-12-01
Wind energy is one of the most promising renewable energy sources. Nevertheless, it is not yet a common source of energy, although there is enough wind potential to supply world's energy demand. One of the most prominent obstacles on the way of employing wind energy is the uncertainty associated with wind energy assessment. Global sensitivity analysis (SA) studies how the variation of input parameters in an abstract model effects the variation of the variable of interest or the output variable. It also provides ways to calculate explicit measures of importance of input variables (first order and total effect sensitivity indices) in regard to influence on the variation of the output variable. Two methods of determining the above mentioned indices were applied and compared: the brute force method and the best practice estimation procedure In this study a methodology for conducting global SA of wind energy assessment at a planning stage is proposed. Three sampling strategies which are a part of SA procedure were compared: sampling based on Sobol' sequences (SBSS), Latin hypercube sampling (LHS) and pseudo-random sampling (PRS). A case study of Masdar City, a showcase of sustainable living in the UAE, is used to exemplify application of the proposed methodology. Sources of uncertainty in wind energy assessment are very diverse. In the case study the following were identified as uncertain input parameters: the Weibull shape parameter, the Weibull scale parameter, availability of a wind turbine, lifetime of a turbine, air density, electrical losses, blade losses, ineffective time losses. Ineffective time losses are defined as losses during the time when the actual wind speed is lower than the cut-in speed or higher than the cut-out speed. The output variable in the case study is the lifetime energy production. Most influential factors for lifetime energy production are identified with the ranking of the total effect sensitivity indices. The results of the present
Multitarget global sensitivity analysis of n-butanol combustion.
Zhou, Dingyu D Y; Davis, Michael J; Skodje, Rex T
2013-05-01
A model for the combustion of butanol is studied using a recently developed theoretical method for the systematic improvement of the kinetic mechanism. The butanol mechanism includes 1446 reactions, and we demonstrate that it is straightforward and computationally feasible to implement a full global sensitivity analysis incorporating all the reactions. In addition, we extend our previous analysis of ignition-delay targets to include species targets. The combination of species and ignition targets leads to multitarget global sensitivity analysis, which allows for a more complete mechanism validation procedure than we previously implemented. The inclusion of species sensitivity analysis allows for a direct comparison between reaction pathway analysis and global sensitivity analysis. PMID:23530815
Rheological Models of Blood: Sensitivity Analysis and Benchmark Simulations
NASA Astrophysics Data System (ADS)
Szeliga, Danuta; Macioł, Piotr; Banas, Krzysztof; Kopernik, Magdalena; Pietrzyk, Maciej
2010-06-01
Modeling of blood flow with respect to rheological parameters of the blood is the objective of this paper. Casson type equation was selected as a blood model and the blood flow was analyzed based on Backward Facing Step benchmark. The simulations were performed using ADINA-CFD finite element code. Three output parameters were selected, which characterize the accuracy of flow simulation. Sensitivity analysis of the results with Morris Design method was performed to identify rheological parameters and the model output, which control the blood flow to significant extent. The paper is the part of the work on identification of parameters controlling process of clotting.
Sensitivity analysis of discrete structural systems: A survey
NASA Technical Reports Server (NTRS)
Adelman, H. M.; Haftka, R. T.
1984-01-01
Methods for calculating sensitivity derivatives for discrete structural systems are surveyed, primarily covering literature published during the past two decades. Methods are described for calculating derivatives of static displacements and stresses, eigenvalues and eigenvectors, transient structural response, and derivatives of optimum structural designs with respect to problem parameters. The survey is focused on publications addressed to structural analysis, but also includes a number of methods developed in nonstructural fields such as electronics, controls, and physical chemistry which are directly applicable to structural problems. Most notable among the nonstructural-based methods are the adjoint variable technique from control theory, and the Green's function and FAST methods from physical chemistry.
An analytical approach to grid sensitivity analysis for NACA four-digit wing sections
NASA Technical Reports Server (NTRS)
Sadrehaghighi, I.; Tiwari, S. N.
1992-01-01
Sensitivity analysis in computational fluid dynamics with emphasis on grids and surface parameterization is described. An interactive algebraic grid-generation technique is employed to generate C-type grids around NACA four-digit wing sections. An analytical procedure is developed for calculating grid sensitivity with respect to design parameters of a wing section. A comparison of the sensitivity with that obtained using a finite difference approach is made. Grid sensitivity with respect to grid parameters, such as grid-stretching coefficients, are also investigated. Using the resultant grid sensitivity, aerodynamic sensitivity is obtained using the compressible two-dimensional thin-layer Navier-Stokes equations.
Wear-Out Sensitivity Analysis Project Abstract
NASA Technical Reports Server (NTRS)
Harris, Adam
2015-01-01
During the course of the Summer 2015 internship session, I worked in the Reliability and Maintainability group of the ISS Safety and Mission Assurance department. My project was a statistical analysis of how sensitive ORU's (Orbital Replacement Units) are to a reliability parameter called the wear-out characteristic. The intended goal of this was to determine a worst case scenario of how many spares would be needed if multiple systems started exhibiting wear-out characteristics simultaneously. The goal was also to determine which parts would be most likely to do so. In order to do this, my duties were to take historical data of operational times and failure times of these ORU's and use them to build predictive models of failure using probability distribution functions, mainly the Weibull distribution. Then, I ran Monte Carlo Simulations to see how an entire population of these components would perform. From here, my final duty was to vary the wear-out characteristic from the intrinsic value, to extremely high wear-out values and determine how much the probability of sufficiency of the population would shift. This was done for around 30 different ORU populations on board the ISS.
Sensitivity analysis of volume scattering phase functions.
Tuchow, Noah; Broughton, Jennifer; Kudela, Raphael
2016-08-01
To solve the radiative transfer equation and relate inherent optical properties (IOPs) to apparent optical properties (AOPs), knowledge of the volume scattering phase function is required. Due to the difficulty of measuring the phase function, it is frequently approximated. We explore the sensitivity of derived AOPs to the phase function parameterization, and compare measured and modeled values of both the AOPs and estimated phase functions using data from Monterey Bay, California during an extreme "red tide" bloom event. Using in situ measurements of absorption and attenuation coefficients, as well as two sets of measurements of the volume scattering function (VSF), we compared output from the Hydrolight radiative transfer model to direct measurements. We found that several common assumptions used in parameterizing the radiative transfer model consistently introduced overestimates of modeled versus measured remote-sensing reflectance values. Phase functions from VSF data derived from measurements at multiple wavelengths and a single scattering single angle significantly overestimated reflectances when using the manufacturer-supplied corrections, but were substantially improved using newly published corrections; phase functions calculated from VSF measurements using three angles and three wavelengths and processed using manufacture-supplied corrections were comparable, demonstrating that reasonable predictions can be made using two commercially available instruments. While other studies have reached similar conclusions, our work extends the analysis to coastal waters dominated by an extreme algal bloom with surface chlorophyll concentrations in excess of 100 mg m^{-3}. PMID:27505819
Designing and Building to ``Impossible'' Tolerances for Vibration Sensitive Equipment
NASA Astrophysics Data System (ADS)
Hertlein, Bernard H.
2003-03-01
As the precision and production capabilities of modern machines and factories increase, our expectations of them rise commensurately. Facility designers and engineers find themselves increasingly involved with measurement needs and design tolerances that were almost unthinkable a few years ago. An area of expertise that demonstrates this very clearly is the field of vibration measurement and control. Magnetic Resonance Imaging, Semiconductor manufacturing, micro-machining, surgical microscopes — These are just a few examples of equipment or techniques that need an extremely stable vibration environment. The challenge to architects, engineers and contractors is to provide that level of stability without undue cost or sacrificing the aesthetics and practicality of a structure. In addition, many facilities have run out of expansion room, so the design is often hampered by the need to reuse all or part of an existing structure, or to site vibration-sensitive equipment close to an existing vibration source. High resolution measurements and nondestructive testing techniques have proven to be invaluable additions to the engineer's toolbox in meeting these challenges. The author summarizes developments in this field over the last fifteen years or so, and lists some common errors of design and construction that can cost a lot of money in retrofit if missed, but can easily be avoided with a little foresight, an appropriate testing program and a carefully thought out checklist.
Sensitivity analysis of hydrodynamic stability operators
NASA Technical Reports Server (NTRS)
Schmid, Peter J.; Henningson, Dan S.; Khorrami, Mehdi R.; Malik, Mujeeb R.
1992-01-01
The eigenvalue sensitivity for hydrodynamic stability operators is investigated. Classical matrix perturbation techniques as well as the concept of epsilon-pseudoeigenvalues are applied to show that parts of the spectrum are highly sensitive to small perturbations. Applications are drawn from incompressible plane Couette, trailing line vortex flow and compressible Blasius boundary layer flow. Parametric studies indicate a monotonically increasing effect of the Reynolds number on the sensitivity. The phenomenon of eigenvalue sensitivity is due to the non-normality of the operators and their discrete matrix analogs and may be associated with large transient growth of the corresponding initial value problem.
Overview of Sensitivity Analysis and Shape Optimization for Complex Aerodynamic Configurations
NASA Technical Reports Server (NTRS)
Newman, Perry A.; Newman, James C., III; Barnwell, Richard W.; Taylor, Arthur C., III; Hou, Gene J.-W.
1998-01-01
This paper presents a brief overview of some of the more recent advances in steady aerodynamic shape-design sensitivity analysis and optimization, based on advanced computational fluid dynamics. The focus here is on those methods particularly well- suited to the study of geometrically complex configurations and their potentially complex associated flow physics. When nonlinear state equations are considered in the optimization process, difficulties are found in the application of sensitivity analysis. Some techniques for circumventing such difficulties are currently being explored and are included here. Attention is directed to methods that utilize automatic differentiation to obtain aerodynamic sensitivity derivatives for both complex configurations and complex flow physics. Various examples of shape-design sensitivity analysis for unstructured-grid computational fluid dynamics algorithms are demonstrated for different formulations of the sensitivity equations. Finally, the use of advanced, unstructured-grid computational fluid dynamics in multidisciplinary analyses and multidisciplinary sensitivity analyses within future optimization processes is recommended and encouraged.
Overview of Sensitivity Analysis and Shape Optimization for Complex Aerodynamic Configurations
NASA Technical Reports Server (NTRS)
Newman, James C., III; Taylor, Arthur C., III; Barnwell, Richard W.; Newman, Perry A.; Hou, Gene J.-W.
1999-01-01
This paper presents a brief overview of some of the more recent advances in steady aerodynamic shape-design sensitivity analysis and optimization, based on advanced computational fluid dynamics (CFD). The focus here is on those methods particularly well-suited to the study of geometrically complex configurations and their potentially complex associated flow physics. When nonlinear state equations are considered in the optimization process, difficulties are found in the application of sensitivity analysis. Some techniques for circumventing such difficulties are currently being explored and are included here. Attention is directed to methods that utilize automatic differentiation to obtain aerodynamic sensitivity derivatives for both complex configurations and complex flow physics. Various examples of shape-design sensitivity analysis for unstructured-grid CFD algorithms are demonstrated for different formulations of the sensitivity equations. Finally, the use of advanced, unstructured-grid CFDs in multidisciplinary analyses and multidisciplinary sensitivity analyses within future optimization processes is recommended and encouraged.
Integrating ethics in design through the value-sensitive design approach.
Cummings, Mary L
2006-10-01
The Accreditation Board of Engineering and Technology (ABET) has declared that to achieve accredited status, 'engineering programs must demonstrate that their graduates have an understanding of professional and ethical responsibility.' Many engineering professors struggle to integrate this required ethics instruction in technical classes and projects because of the lack of a formalized ethics-in-design approach. However, one methodology developed in human-computer interaction research, the Value-Sensitive Design approach, can serve as an engineering education tool which bridges the gap between design and ethics for many engineering disciplines. The three major components of Value-Sensitive Design, conceptual, technical, and empirical, exemplified through a case study which focuses on the development of a command and control supervisory interface for a military cruise missile. PMID:17199145
Optimizing human activity patterns using global sensitivity analysis
Hickmann, Kyle S.; Mniszewski, Susan M.; Del Valle, Sara Y.; Hyman, James M.
2014-01-01
Implementing realistic activity patterns for a population is crucial for modeling, for example, disease spread, supply and demand, and disaster response. Using the dynamic activity simulation engine, DASim, we generate schedules for a population that capture regular (e.g., working, eating, and sleeping) and irregular activities (e.g., shopping or going to the doctor). We use the sample entropy (SampEn) statistic to quantify a schedule’s regularity for a population. We show how to tune an activity’s regularity by adjusting SampEn, thereby making it possible to realistically design activities when creating a schedule. The tuning process sets up a computationally intractable high-dimensional optimization problem. To reduce the computational demand, we use Bayesian Gaussian process regression to compute global sensitivity indices and identify the parameters that have the greatest effect on the variance of SampEn. We use the harmony search (HS) global optimization algorithm to locate global optima. Our results show that HS combined with global sensitivity analysis can efficiently tune the SampEn statistic with few search iterations. We demonstrate how global sensitivity analysis can guide statistical emulation and global optimization algorithms to efficiently tune activities and generate realistic activity patterns. Though our tuning methods are applied to dynamic activity schedule generation, they are general and represent a significant step in the direction of automated tuning and optimization of high-dimensional computer simulations. PMID:25580080
Optimizing human activity patterns using global sensitivity analysis
Fairchild, Geoffrey; Hickmann, Kyle S.; Mniszewski, Susan M.; Del Valle, Sara Y.; Hyman, James M.
2013-12-10
Implementing realistic activity patterns for a population is crucial for modeling, for example, disease spread, supply and demand, and disaster response. Using the dynamic activity simulation engine, DASim, we generate schedules for a population that capture regular (e.g., working, eating, and sleeping) and irregular activities (e.g., shopping or going to the doctor). We use the sample entropy (SampEn) statistic to quantify a schedule’s regularity for a population. We show how to tune an activity’s regularity by adjusting SampEn, thereby making it possible to realistically design activities when creating a schedule. The tuning process sets up a computationally intractable high-dimensional optimization problem. To reduce the computational demand, we use Bayesian Gaussian process regression to compute global sensitivity indices and identify the parameters that have the greatest effect on the variance of SampEn. Here we use the harmony search (HS) global optimization algorithm to locate global optima. Our results show that HS combined with global sensitivity analysis can efficiently tune the SampEn statistic with few search iterations. We demonstrate how global sensitivity analysis can guide statistical emulation and global optimization algorithms to efficiently tune activities and generate realistic activity patterns. Finally, though our tuning methods are applied to dynamic activity schedule generation, they are general and represent a significant step in the direction of automated tuning and optimization of high-dimensional computer simulations.
Optimizing human activity patterns using global sensitivity analysis
Fairchild, Geoffrey; Hickmann, Kyle S.; Mniszewski, Susan M.; Del Valle, Sara Y.; Hyman, James M.
2013-12-10
Implementing realistic activity patterns for a population is crucial for modeling, for example, disease spread, supply and demand, and disaster response. Using the dynamic activity simulation engine, DASim, we generate schedules for a population that capture regular (e.g., working, eating, and sleeping) and irregular activities (e.g., shopping or going to the doctor). We use the sample entropy (SampEn) statistic to quantify a schedule’s regularity for a population. We show how to tune an activity’s regularity by adjusting SampEn, thereby making it possible to realistically design activities when creating a schedule. The tuning process sets up a computationally intractable high-dimensional optimizationmore » problem. To reduce the computational demand, we use Bayesian Gaussian process regression to compute global sensitivity indices and identify the parameters that have the greatest effect on the variance of SampEn. Here we use the harmony search (HS) global optimization algorithm to locate global optima. Our results show that HS combined with global sensitivity analysis can efficiently tune the SampEn statistic with few search iterations. We demonstrate how global sensitivity analysis can guide statistical emulation and global optimization algorithms to efficiently tune activities and generate realistic activity patterns. Finally, though our tuning methods are applied to dynamic activity schedule generation, they are general and represent a significant step in the direction of automated tuning and optimization of high-dimensional computer simulations.« less
Sensitive detection of human insulin using a designed combined pore approach.
Lei, Chang; Noonan, Owen; Jambhrunkar, Siddharth; Qian, Kun; Xu, Chun; Zhang, Jun; Nouwens, Amanda; Yu, Chengzhong
2014-06-25
A unique combined pore approach to the sensitive detection of human insulin is developed. Through a systematic study to understand the impact of pore size and surface chemistry of nanoporous materials on their enrichment and purification performance, the advantages of selected porous materials are integrated to enhance detection sensitivity in a unified two-step process. In the first purification step, a rationally designed large pore material (ca. 100 nm in diameter) is chosen to repel the interferences from nontarget molecules. In the second enrichment step, a hydrophobically modified mesoporous material with a pore size of 5 nm is selected to enrich insulin molecules. A low detection limit of 0.05 ng mL(-1) in artificial urine is achieved by this advanced approach, similar to most antibody-based analysis protocols. This designer approach is efficient and low cost, and thus has great potential in the sensitive detection of biomolecules in complex biological systems. PMID:24599559
NASA Astrophysics Data System (ADS)
Penn, Jay P.
1996-03-01
It is generally believed by those skilled in launch system design that Single-Stage-To-Orbit (SSTO) designs are more technically challenging, more performance sensitive, and yield larger lift-off weights than do Two-Stage-To-Orbit designs (TSTO's) offering similar payload delivery capability. Without additional insight into the other considerations which drive the development, recurring costs, operability, and reliability of a launch fleet, an analyst may easily conclude that the higher performing, less sensitive TSTO designs, thus yield a better solution to achieving low cost payload delivery. This limited insight could justify an argument to eliminate the X-33 SSTO technology/demonstration development effort, and thus proceed directly to less risky TSTO designs. Insight into real world design considerations of launch vehicles makes the choice of SSTO vs TSTO much less clear. The presentation addresses a more comprehensive evaluation of the general class of SSTO and TSTO concepts. These include pure SSTO's, augmented SSTO's, Siamese Twin, and Pure TSTO designs. The assessment considers vehicle performance and scaling relationships which characterize real vehicle designs. The assessment also addresses technology requirements, operations and supportability, cost implications, and sensitivities. Results of the assessment indicate that the trade space between various SSTO and TSTO design approaches is complex and not yet fully understood. The results of the X-33 technology demonstrators, as well as additional parametric analysis is required to better define the relative performance and costs of the various design approaches. The results also indicate that with modern technologies and today's better understanding of vehicle design considerations, the perception that SSTO's are dramatically heavier and more sensitive than TSTO designs is more of a myth, than reality.
Sensitivity analysis of textural parameters for vertebroplasty
NASA Astrophysics Data System (ADS)
Tack, Gye Rae; Lee, Seung Y.; Shin, Kyu-Chul; Lee, Sung J.
2002-05-01
Vertebroplasty is one of the newest surgical approaches for the treatment of the osteoporotic spine. Recent studies have shown that it is a minimally invasive, safe, promising procedure for patients with osteoporotic fractures while providing structural reinforcement of the osteoporotic vertebrae as well as immediate pain relief. However, treatment failures due to excessive bone cement injection have been reported as one of complications. It is believed that control of bone cement volume seems to be one of the most critical factors in preventing complications. We believed that an optimal bone cement volume could be assessed based on CT data of a patient. Gray-level run length analysis was used to extract textural information of the trabecular. At initial stage of the project, four indices were used to represent the textural information: mean width of intertrabecular space, mean width of trabecular, area of intertrabecular space, and area of trabecular. Finally, the area of intertrabecular space was selected as a parameter to estimate an optimal bone cement volume and it was found that there was a strong linear relationship between these 2 variables (correlation coefficient = 0.9433, standard deviation = 0.0246). In this study, we examined several factors affecting overall procedures. The threshold level, the radius of rolling ball and the size of region of interest were selected for the sensitivity analysis. As the level of threshold varied with 9, 10, and 11, the correlation coefficient varied from 0.9123 to 0.9534. As the radius of rolling ball varied with 45, 50, and 55, the correlation coefficient varied from 0.9265 to 0.9730. As the size of region of interest varied with 58 x 58, 64 x 64, and 70 x 70, the correlation coefficient varied from 0.9685 to 0.9468. Finally, we found that strong correlation between actual bone cement volume (Y) and the area (X) of the intertrabecular space calculated from the binary image and the linear equation Y = 0.001722 X - 2
Topographic Avalanche Risk: DEM Sensitivity Analysis
NASA Astrophysics Data System (ADS)
Nazarkulova, Ainura; Strobl, Josef
2015-04-01
GIS-based models are frequently used to assess the risk and trigger probabilities of (snow) avalanche releases, based on parameters and geomorphometric derivatives like elevation, exposure, slope, proximity to ridges and local relief energy. Numerous models, and model-based specific applications and project results have been published based on a variety of approaches and parametrizations as well as calibrations. Digital Elevation Models (DEM) come with many different resolution (scale) and quality (accuracy) properties, some of these resulting from sensor characteristics and DEM generation algorithms, others from different DEM processing workflows and analysis strategies. This paper explores the impact of using different types and characteristics of DEMs for avalanche risk modeling approaches, and aims at establishing a framework for assessing the uncertainty of results. The research question is derived from simply demonstrating the differences in release risk areas and intensities by applying identical models to DEMs with different properties, and then extending this into a broader sensitivity analysis. For the quantification and calibration of uncertainty parameters different metrics are established, based on simple value ranges, probabilities, as well as fuzzy expressions and fractal metrics. As a specific approach the work on DEM resolution-dependent 'slope spectra' is being considered and linked with the specific application of geomorphometry-base risk assessment. For the purpose of this study focusing on DEM characteristics, factors like land cover, meteorological recordings and snowpack structure and transformation are kept constant, i.e. not considered explicitly. Key aims of the research presented here are the development of a multi-resolution and multi-scale framework supporting the consistent combination of large area basic risk assessment with local mitigation-oriented studies, and the transferability of the latter into areas without availability of
[Ecological sensitivity of Shanghai City based on GIS spatial analysis].
Cao, Jian-jun; Liu, Yong-juan
2010-07-01
In this paper, five sensitivity factors affecting the eco-environment of Shanghai City, i.e., rivers and lakes, historical relics and forest parks, geological disasters, soil pollution, and land use, were selected, and their weights were determined by analytic hierarchy process. Combining with GIS spatial analysis technique, the sensitivities of these factors were classified into four grades, i.e., highly sensitive, moderately sensitive, low sensitive, and insensitive, and the spatial distribution of the ecological sensitivity of Shanghai City was figured out. There existed a significant spatial differentiation in the ecological sensitivity of the City, and the insensitive, low sensitive, moderately sensitive, and highly sensitive areas occupied 37.07%, 5.94%, 38.16%, and 18.83%, respectively. Some suggestions on the City's zoning protection and construction were proposed. This study could provide scientific references for the City's environmental protection and economic development. PMID:20879541
5 CFR 732.201 - Sensitivity level designations and investigative requirements.
Code of Federal Regulations, 2012 CFR
2012-01-01
... 5 Administrative Personnel 2 2012-01-01 2012-01-01 false Sensitivity level designations and... Requirements § 732.201 Sensitivity level designations and investigative requirements. (a) For purposes of this... material adverse effect on the national security as a sensitive position at one of three sensitivity...
5 CFR 732.201 - Sensitivity level designations and investigative requirements.
Code of Federal Regulations, 2011 CFR
2011-01-01
... 5 Administrative Personnel 2 2011-01-01 2011-01-01 false Sensitivity level designations and... Requirements § 732.201 Sensitivity level designations and investigative requirements. (a) For purposes of this... material adverse effect on the national security as a sensitive position at one of three sensitivity...
5 CFR 732.201 - Sensitivity level designations and investigative requirements.
Code of Federal Regulations, 2010 CFR
2010-01-01
... 5 Administrative Personnel 2 2010-01-01 2010-01-01 false Sensitivity level designations and... Requirements § 732.201 Sensitivity level designations and investigative requirements. (a) For purposes of this... material adverse effect on the national security as a sensitive position at one of three sensitivity...
5 CFR 732.201 - Sensitivity level designations and investigative requirements.
Code of Federal Regulations, 2014 CFR
2014-01-01
... 5 Administrative Personnel 2 2014-01-01 2014-01-01 false Sensitivity level designations and... Requirements § 732.201 Sensitivity level designations and investigative requirements. (a) For purposes of this... material adverse effect on the national security as a sensitive position at one of three sensitivity...
5 CFR 732.201 - Sensitivity level designations and investigative requirements.
Code of Federal Regulations, 2013 CFR
2013-01-01
... 5 Administrative Personnel 2 2013-01-01 2013-01-01 false Sensitivity level designations and... Requirements § 732.201 Sensitivity level designations and investigative requirements. (a) For purposes of this... material adverse effect on the national security as a sensitive position at one of three sensitivity...
NASA Astrophysics Data System (ADS)
Zhan, Tony Liang-tong; Mu, Qing-yi; Chen, Yun-min; Ke, Han
2015-04-01
The time domain reflectometry (TDR) penetrometer, which can measure both the apparent dielectric permittivity and the bulk electrical conductivity of soils, is an important tool for the site investigation of contaminated land. This paper presents a theoretical method for evaluating the measurement sensitivity and an improved design of the TDR penetrometer. The sensitivity evaluation method is based on a spatial weighting analysis of the electromagnetic field using a seepage analysis software. This method is used to quantify the measurement sensitivity for the three types of TDR penetrometers reported in literature as well as guide the design improvement of the TDR penetrometer. The improved design includes the use of semicircle-shaped conductors and the optimization of the conductor diameter. The measurement sensitivity to the targeted medium for the improved TDR penetrometer is evaluated to be greater than those of the three types of TDR penetrometers reported in literature. The performance of the improved TDR penetrometer was demonstrated by conducting an experimental calibration of the probe and penetration tests in a chamber containing a silty soil column. The experimental results demonstrate that the measurements from the improved TDR penetrometer are able to capture the variation in the water content profiles as well as the leachate contaminated soil.
Design and characterisation of bodipy sensitizers for dye-sensitized NiO solar cells.
Summers, Gareth H; Lefebvre, Jean-François; Black, Fiona A; Davies, E Stephen; Gibson, Elizabeth A; Pullerits, Tönu; Wood, Christopher J; Zidek, Karel
2016-01-14
A series of photosensitizers for NiO-based dye-sensitized solar cells is presented. Three model compounds containing a triphenylamine donor appended to a boron dipyrromethene (bodipy) chromophore have been successfully prepared and characterised using emission spectroscopy, electrochemistry and spectroelectrochemistry, to ultimately direct the design of dyes with more complex structures. Carboxylic acid anchoring groups and thiophene spacers were appended to the model compounds to provide five dyes which were adsorbed onto NiO and integrated into dye-sensitized solar cells. Solar cells incorporating the simple Bodipy-CO₂H dye were surprisingly promising relative to the more complex dye 4. Cell performances were improved with dyes which had increased electronic communication between the donor and acceptor, achieved by incorporating a less hindered bodipy moiety. Further increases in performances were obtained from dyes which contained a thiophene spacer. Thus, the best performance was obtained for 7 which generated a very promising photocurrent density of 5.87 mA cm(-2) and an IPCE of 53%. Spectroelectrochemistry combined with time-resolved transient absorption spectroscopy were used to determine the identity and lifetime of excited state species. Short-lived (ps) transients were recorded for 4, 5 and 7 which are consistent with previous studies. Despite a longer lived (25 ns) charge-separated state for 6/NiO, there was no increase in the photocurrent generated by the corresponding solar cell. PMID:26660278
1991-03-12
Version 00 SUSD calculates sensitivity coefficients for one- and two-dimensional transport problems. Variance and standard deviation of detector responses or design parameters can be obtained using cross-section covariance matrices. In neutron transport problems, this code can perform sensitivity-uncertainty analysis for secondary angular distribution (SAD) or secondary energy distribution (SED).
Extended forward sensitivity analysis of one-dimensional isothermal flow
Johnson, M.; Zhao, H.
2013-07-01
Sensitivity analysis and uncertainty quantification is an important part of nuclear safety analysis. In this work, forward sensitivity analysis is used to compute solution sensitivities on 1-D fluid flow equations typical of those found in system level codes. Time step sensitivity analysis is included as a method for determining the accumulated error from time discretization. The ability to quantify numerical error arising from the time discretization is a unique and important feature of this method. By knowing the relative sensitivity of time step with other physical parameters, the simulation is allowed to run at optimized time steps without affecting the confidence of the physical parameter sensitivity results. The time step forward sensitivity analysis method can also replace the traditional time step convergence studies that are a key part of code verification with much less computational cost. One well-defined benchmark problem with manufactured solutions is utilized to verify the method; another test isothermal flow problem is used to demonstrate the extended forward sensitivity analysis process. Through these sample problems, the paper shows the feasibility and potential of using the forward sensitivity analysis method to quantify uncertainty in input parameters and time step size for a 1-D system-level thermal-hydraulic safety code. (authors)
Support systems design and analysis
NASA Technical Reports Server (NTRS)
Ferguson, R. M.
1985-01-01
The integration of Kennedy Space Center (KSC) ground support systems with the new launch processing system and new launch vehicle provided KSC with a unique challenge in system design and analysis for the Space Transportation System. Approximately 70 support systems are controlled and monitored by the launch processing system. Typical systems are main propulsion oxygen and hydrogen loading systems, environmental control life support system, hydraulics, etc. An End-to-End concept of documentation and analysis was chosen and applied to these systems. Unique problems were resolved in the areas of software analysis, safing under emergency conditions, sampling rates, and control loop analysis. New methods of performing End-to-End reliability analyses were implemented. The systems design approach selected and the resolution of major problem areas are discussed.
Ju, Jonghyun; Han, Yun-ah; Kim, Seok-min
2013-01-01
The effects of structural design parameters on the performance of nano-replicated photonic crystal (PC) label-free biosensors were examined by the analysis of simulated reflection spectra of PC structures. The grating pitch, duty, scaled grating height and scaled TiO2 layer thickness were selected as the design factors to optimize the PC structure. The peak wavelength value (PWV), full width at half maximum of the peak, figure of merit for the bulk and surface sensitivities, and surface/bulk sensitivity ratio were also selected as the responses to optimize the PC label-free biosensor performance. A parametric study showed that the grating pitch was the dominant factor for PWV, and that it had low interaction effects with other scaled design factors. Therefore, we can isolate the effect of grating pitch using scaled design factors. For the design of PC-label free biosensor, one should consider that: (1) the PWV can be measured by the reflection peak measurement instruments, (2) the grating pitch and duty can be manufactured using conventional lithography systems, and (3) the optimum design is less sensitive to the grating height and TiO2 layer thickness variations in the fabrication process. In this paper, we suggested a design guide for highly sensitive PC biosensor in which one select the grating pitch and duty based on the limitations of the lithography and measurement system, and conduct a multi objective optimization of the grating height and TiO2 layer thickness for maximizing performance and minimizing the influence of parameter variation. Through multi-objective optimization of a PC structure with a fixed grating height of 550 nm and a duty of 50%, we obtained a surface FOM of 66.18 RIU-1 and an S/B ratio of 34.8%, with a grating height of 117 nm and TiO2 height of 210 nm. PMID:23470487
Probabilistic Finite Element Analysis & Design Optimization for Structural Designs
NASA Astrophysics Data System (ADS)
Deivanayagam, Arumugam
This study focuses on implementing probabilistic nature of material properties (Kevlar® 49) to the existing deterministic finite element analysis (FEA) of fabric based engine containment system through Monte Carlo simulations (MCS) and implementation of probabilistic analysis in engineering designs through Reliability Based Design Optimization (RBDO). First, the emphasis is on experimental data analysis focusing on probabilistic distribution models which characterize the randomness associated with the experimental data. The material properties of Kevlar® 49 are modeled using experimental data analysis and implemented along with an existing spiral modeling scheme (SMS) and user defined constitutive model (UMAT) for fabric based engine containment simulations in LS-DYNA. MCS of the model are performed to observe the failure pattern and exit velocities of the models. Then the solutions are compared with NASA experimental tests and deterministic results. MCS with probabilistic material data give a good prospective on results rather than a single deterministic simulation results. The next part of research is to implement the probabilistic material properties in engineering designs. The main aim of structural design is to obtain optimal solutions. In any case, in a deterministic optimization problem even though the structures are cost effective, it becomes highly unreliable if the uncertainty that may be associated with the system (material properties, loading etc.) is not represented or considered in the solution process. Reliable and optimal solution can be obtained by performing reliability optimization along with the deterministic optimization, which is RBDO. In RBDO problem formulation, in addition to structural performance constraints, reliability constraints are also considered. This part of research starts with introduction to reliability analysis such as first order reliability analysis, second order reliability analysis followed by simulation technique that
Multivariate Sensitivity Analysis of Time-of-Flight Sensor Fusion
NASA Astrophysics Data System (ADS)
Schwarz, Sebastian; Sjöström, Mårten; Olsson, Roger
2014-09-01
Obtaining three-dimensional scenery data is an essential task in computer vision, with diverse applications in various areas such as manufacturing and quality control, security and surveillance, or user interaction and entertainment. Dedicated Time-of-Flight sensors can provide detailed scenery depth in real-time and overcome short-comings of traditional stereo analysis. Nonetheless, they do not provide texture information and have limited spatial resolution. Therefore such sensors are typically combined with high resolution video sensors. Time-of-Flight Sensor Fusion is a highly active field of research. Over the recent years, there have been multiple proposals addressing important topics such as texture-guided depth upsampling and depth data denoising. In this article we take a step back and look at the underlying principles of ToF sensor fusion. We derive the ToF sensor fusion error model and evaluate its sensitivity to inaccuracies in camera calibration and depth measurements. In accordance with our findings, we propose certain courses of action to ensure high quality fusion results. With this multivariate sensitivity analysis of the ToF sensor fusion model, we provide an important guideline for designing, calibrating and running a sophisticated Time-of-Flight sensor fusion capture systems.
Probability density adjoint for sensitivity analysis of the Mean of Chaos
Blonigan, Patrick J. Wang, Qiqi
2014-08-01
Sensitivity analysis, especially adjoint based sensitivity analysis, is a powerful tool for engineering design which allows for the efficient computation of sensitivities with respect to many parameters. However, these methods break down when used to compute sensitivities of long-time averaged quantities in chaotic dynamical systems. This paper presents a new method for sensitivity analysis of ergodic chaotic dynamical systems, the density adjoint method. The method involves solving the governing equations for the system's invariant measure and its adjoint on the system's attractor manifold rather than in phase-space. This new approach is derived for and demonstrated on one-dimensional chaotic maps and the three-dimensional Lorenz system. It is found that the density adjoint computes very finely detailed adjoint distributions and accurate sensitivities, but suffers from large computational costs.
Design of a High Sensitivity GNSS receiver for Lunar missions
NASA Astrophysics Data System (ADS)
Musumeci, Luciano; Dovis, Fabio; Silva, João S.; da Silva, Pedro F.; Lopes, Hugo D.
2016-06-01
This paper presents the design of a satellite navigation receiver architecture tailored for future Lunar exploration missions, demonstrating the feasibility of using Global Navigation Satellite Systems signals integrated with an orbital filter to achieve such a scope. It analyzes the performance of a navigation solution based on pseudorange and pseudorange rate measurements, generated through the processing of very weak signals of the Global Positioning System (GPS) L1/L5 and Galileo E1/E5 frequency bands. In critical scenarios (e.g. during manoeuvres) acceleration and attitude measurements from additional sensors complementing the GNSS measurements are integrated with the GNSS measurement to match the positioning requirement. A review of environment characteristics (dynamics, geometry and signal power) for the different phases of a reference Lunar mission is provided, focusing on the stringent requirements of the Descent, Approach and Hazard Detection and Avoidance phase. The design of High Sensitivity acquisition and tracking schemes is supported by an extensive simulation test campaign using a software receiver implementation and navigation results are validated by means of an end-to-end software simulator. Acquisition and tracking of GPS and Galileo signals of the L1/E1 and L5/E5a bands was successfully demonstrated for Carrier-to-Noise density ratios as low as 5-8 dB-Hz. The proposed navigation architecture provides acceptable performances during the considered critical phases, granting position and velocity errors below 61.4 m and 3.2 m/s, respectively, for the 99.7% of the mission time.
Automated sensitivity analysis using the GRESS language
Pin, F.G.; Oblow, E.M.; Wright, R.Q.
1986-04-01
An automated procedure for performing large-scale sensitivity studies based on the use of computer calculus is presented. The procedure is embodied in a FORTRAN precompiler called GRESS, which automatically processes computer models and adds derivative-taking capabilities to the normal calculated results. In this report, the GRESS code is described, tested against analytic and numerical test problems, and then applied to a major geohydrological modeling problem. The SWENT nuclear waste repository modeling code is used as the basis for these studies. Results for all problems are discussed in detail. Conclusions are drawn as to the applicability of GRESS in the problems at hand and for more general large-scale modeling sensitivity studies.
Discrete analysis of spatial-sensitivity models
NASA Technical Reports Server (NTRS)
Nielsen, Kenneth R. K.; Wandell, Brian A.
1988-01-01
Procedures for reducing the computational burden of current models of spatial vision are described, the simplifications being consistent with the prediction of the complete model. A method for using pattern-sensitivity measurements to estimate the initial linear transformation is also proposed which is based on the assumption that detection performance is monotonic with the vector length of the sensor responses. It is shown how contrast-threshold data can be used to estimate the linear transformation needed to characterize threshold performance.
Structural Analysis and Design Software
NASA Technical Reports Server (NTRS)
1997-01-01
Collier Research and Development Corporation received a one-of-a-kind computer code for designing exotic hypersonic aircraft called ST-SIZE in the first ever Langley Research Center software copyright license agreement. Collier transformed the NASA computer code into a commercial software package called HyperSizer, which integrates with other Finite Element Modeling and Finite Analysis private-sector structural analysis program. ST-SIZE was chiefly conceived as a means to improve and speed the structural design of a future aerospace plane for Langley Hypersonic Vehicles Office. Including the NASA computer code into HyperSizer has enabled the company to also apply the software to applications other than aerospace, including improved design and construction for offices, marine structures, cargo containers, commercial and military aircraft, rail cars, and a host of everyday consumer products.
Towards More Efficient and Effective Global Sensitivity Analysis
NASA Astrophysics Data System (ADS)
Razavi, Saman; Gupta, Hoshin
2014-05-01
Sensitivity analysis (SA) is an important paradigm in the context of model development and application. There are a variety of approaches towards sensitivity analysis that formally describe different "intuitive" understandings of the sensitivity of a single or multiple model responses to different factors such as model parameters or forcings. These approaches are based on different philosophies and theoretical definitions of sensitivity and range from simple local derivatives to rigorous Sobol-type analysis-of-variance approaches. In general, different SA methods focus and identify different properties of the model response and may lead to different, sometimes even conflicting conclusions about the underlying sensitivities. This presentation revisits the theoretical basis for sensitivity analysis, critically evaluates the existing approaches in the literature, and demonstrates their shortcomings through simple examples. Important properties of response surfaces that are associated with the understanding and interpretation of sensitivities are outlined. A new approach towards global sensitivity analysis is developed that attempts to encompass the important, sensitivity-related properties of response surfaces. Preliminary results show that the new approach is superior to the standard approaches in the literature in terms of effectiveness and efficiency.
Fuzzy sensitivity analysis for reliability assessment of building structures
NASA Astrophysics Data System (ADS)
Kala, Zdeněk
2016-06-01
The mathematical concept of fuzzy sensitivity analysis, which studies the effects of the fuzziness of input fuzzy numbers on the fuzziness of the output fuzzy number, is described in the article. The output fuzzy number is evaluated using Zadeh's general extension principle. The contribution of stochastic and fuzzy uncertainty in reliability analysis tasks of building structures is discussed. The algorithm of fuzzy sensitivity analysis is an alternative to stochastic sensitivity analysis in tasks in which input and output variables are considered as fuzzy numbers.
Global sensitivity analysis of the Indian monsoon during the Pleistocene
NASA Astrophysics Data System (ADS)
Araya-Melo, P. A.; Crucifix, M.; Bounceur, N.
2015-01-01
The sensitivity of the Indian monsoon to the full spectrum of climatic conditions experienced during the Pleistocene is estimated using the climate model HadCM3. The methodology follows a global sensitivity analysis based on the emulator approach of Oakley and O'Hagan (2004) implemented following a three-step strategy: (1) development of an experiment plan, designed to efficiently sample a five-dimensional input space spanning Pleistocene astronomical configurations (three parameters), CO2 concentration and a Northern Hemisphere glaciation index; (2) development, calibration and validation of an emulator of HadCM3 in order to estimate the response of the Indian monsoon over the full input space spanned by the experiment design; and (3) estimation and interpreting of sensitivity diagnostics, including sensitivity measures, in order to synthesise the relative importance of input factors on monsoon dynamics, estimate the phase of the monsoon intensity response with respect to that of insolation, and detect potential non-linear phenomena. By focusing on surface temperature, precipitation, mixed-layer depth and sea-surface temperature over the monsoon region during the summer season (June-July-August-September), we show that precession controls the response of four variables: continental temperature in phase with June to July insolation, high glaciation favouring a late-phase response, sea-surface temperature in phase with May insolation, continental precipitation in phase with July insolation, and mixed-layer depth in antiphase with the latter. CO2 variations control temperature variance with an amplitude similar to that of precession. The effect of glaciation is dominated by the albedo forcing, and its effect on precipitation competes with that of precession. Obliquity is a secondary effect, negligible on most variables except sea-surface temperature. It is also shown that orography forcing reduces the glacial cooling, and even has a positive effect on precipitation
NASA Technical Reports Server (NTRS)
Winters, J. M.; Stark, L.
1984-01-01
Original results for a newly developed eight-order nonlinear limb antagonistic muscle model of elbow flexion and extension are presented. A wider variety of sensitivity analysis techniques are used and a systematic protocol is established that shows how the different methods can be used efficiently to complement one another for maximum insight into model sensitivity. It is explicitly shown how the sensitivity of output behaviors to model parameters is a function of the controller input sequence, i.e., of the movement task. When the task is changed (for instance, from an input sequence that results in the usual fast movement task to a slower movement that may also involve external loading, etc.) the set of parameters with high sensitivity will in general also change. Such task-specific use of sensitivity analysis techniques identifies the set of parameters most important for a given task, and even suggests task-specific model reduction possibilities.
Habitat Design Optimization and Analysis
NASA Technical Reports Server (NTRS)
SanSoucie, Michael P.; Hull, Patrick V.; Tinker, Michael L.
2006-01-01
Long-duration surface missions to the Moon and Mars will require habitats for the astronauts. The materials chosen for the habitat walls play a direct role in the protection against the harsh environments found on the surface. Choosing the best materials, their configuration, and the amount required is extremely difficult due to the immense size of the design region. Advanced optimization techniques are necessary for habitat wall design. Standard optimization techniques are not suitable for problems with such large search spaces; therefore, a habitat design optimization tool utilizing genetic algorithms has been developed. Genetic algorithms use a "survival of the fittest" philosophy, where the most fit individuals are more likely to survive and reproduce. This habitat design optimization tool is a multi-objective formulation of structural analysis, heat loss, radiation protection, and meteoroid protection. This paper presents the research and development of this tool.
Global Sensitivity Analysis of Environmental Models: Convergence, Robustness and Validation
NASA Astrophysics Data System (ADS)
Sarrazin, Fanny; Pianosi, Francesca; Khorashadi Zadeh, Farkhondeh; Van Griensven, Ann; Wagener, Thorsten
2015-04-01
Global Sensitivity Analysis aims to characterize the impact that variations in model input factors (e.g. the parameters) have on the model output (e.g. simulated streamflow). In sampling-based Global Sensitivity Analysis, the sample size has to be chosen carefully in order to obtain reliable sensitivity estimates while spending computational resources efficiently. Furthermore, insensitive parameters are typically identified through the definition of a screening threshold: the theoretical value of their sensitivity index is zero but in a sampling-base framework they regularly take non-zero values. There is little guidance available for these two steps in environmental modelling though. The objective of the present study is to support modellers in making appropriate choices, regarding both sample size and screening threshold, so that a robust sensitivity analysis can be implemented. We performed sensitivity analysis for the parameters of three hydrological models with increasing level of complexity (Hymod, HBV and SWAT), and tested three widely used sensitivity analysis methods (Elementary Effect Test or method of Morris, Regional Sensitivity Analysis, and Variance-Based Sensitivity Analysis). We defined criteria based on a bootstrap approach to assess three different types of convergence: the convergence of the value of the sensitivity indices, of the ranking (the ordering among the parameters) and of the screening (the identification of the insensitive parameters). We investigated the screening threshold through the definition of a validation procedure. The results showed that full convergence of the value of the sensitivity indices is not necessarily needed to rank or to screen the model input factors. Furthermore, typical values of the sample sizes that are reported in the literature can be well below the sample sizes that actually ensure convergence of ranking and screening.
Sensitivity Analysis of Offshore Wind Cost of Energy (Poster)
Dykes, K.; Ning, A.; Graf, P.; Scott, G.; Damiami, R.; Hand, M.; Meadows, R.; Musial, W.; Moriarty, P.; Veers, P.
2012-10-01
No matter the source, offshore wind energy plant cost estimates are significantly higher than for land-based projects. For instance, a National Renewable Energy Laboratory (NREL) review on the 2010 cost of wind energy found baseline cost estimates for onshore wind energy systems to be 71 dollars per megawatt-hour ($/MWh), versus 225 $/MWh for offshore systems. There are many ways that innovation can be used to reduce the high costs of offshore wind energy. However, the use of such innovation impacts the cost of energy because of the highly coupled nature of the system. For example, the deployment of multimegawatt turbines can reduce the number of turbines, thereby reducing the operation and maintenance (O&M) costs associated with vessel acquisition and use. On the other hand, larger turbines may require more specialized vessels and infrastructure to perform the same operations, which could result in higher costs. To better understand the full impact of a design decision on offshore wind energy system performance and cost, a system analysis approach is needed. In 2011-2012, NREL began development of a wind energy systems engineering software tool to support offshore wind energy system analysis. The tool combines engineering and cost models to represent an entire offshore wind energy plant and to perform system cost sensitivity analysis and optimization. Initial results were collected by applying the tool to conduct a sensitivity analysis on a baseline offshore wind energy system using 5-MW and 6-MW NREL reference turbines. Results included information on rotor diameter, hub height, power rating, and maximum allowable tip speeds.
Context-sensitive design and human interaction principles for usable, useful, and adoptable radars
NASA Astrophysics Data System (ADS)
McNamara, Laura A.; Klein, Laura M.
2016-05-01
The evolution of exquisitely sensitive Synthetic Aperture Radar (SAR) systems is positioning this technology for use in time-critical environments, such as search-and-rescue missions and improvised explosive device (IED) detection. SAR systems should be playing a keystone role in the United States' Intelligence, Surveillance, and Reconnaissance activities. Yet many in the SAR community see missed opportunities for incorporating SAR into existing remote sensing data collection and analysis challenges. Drawing on several years' of field research with SAR engineering and operational teams, this paper examines the human and organizational factors that mitigate against the adoption and use of SAR for tactical ISR and operational support. We suggest that SAR has a design problem, and that context-sensitive, human and organizational design frameworks are required if the community is to realize SAR's tactical potential.
Accelerated Sensitivity Analysis in High-Dimensional Stochastic Reaction Networks
Arampatzis, Georgios; Katsoulakis, Markos A.; Pantazis, Yannis
2015-01-01
Existing sensitivity analysis approaches are not able to handle efficiently stochastic reaction networks with a large number of parameters and species, which are typical in the modeling and simulation of complex biochemical phenomena. In this paper, a two-step strategy for parametric sensitivity analysis for such systems is proposed, exploiting advantages and synergies between two recently proposed sensitivity analysis methodologies for stochastic dynamics. The first method performs sensitivity analysis of the stochastic dynamics by means of the Fisher Information Matrix on the underlying distribution of the trajectories; the second method is a reduced-variance, finite-difference, gradient-type sensitivity approach relying on stochastic coupling techniques for variance reduction. Here we demonstrate that these two methods can be combined and deployed together by means of a new sensitivity bound which incorporates the variance of the quantity of interest as well as the Fisher Information Matrix estimated from the first method. The first step of the proposed strategy labels sensitivities using the bound and screens out the insensitive parameters in a controlled manner. In the second step of the proposed strategy, a finite-difference method is applied only for the sensitivity estimation of the (potentially) sensitive parameters that have not been screened out in the first step. Results on an epidermal growth factor network with fifty parameters and on a protein homeostasis with eighty parameters demonstrate that the proposed strategy is able to quickly discover and discard the insensitive parameters and in the remaining potentially sensitive parameters it accurately estimates the sensitivities. The new sensitivity strategy can be several times faster than current state-of-the-art approaches that test all parameters, especially in “sloppy” systems. In particular, the computational acceleration is quantified by the ratio between the total number of parameters over
Global sensitivity analysis of analytical vibroacoustic transmission models
NASA Astrophysics Data System (ADS)
Christen, Jean-Loup; Ichchou, Mohamed; Troclet, Bernard; Bareille, Olivier; Ouisse, Morvan
2016-04-01
Noise reduction issues arise in many engineering problems. One typical vibroacoustic problem is the transmission loss (TL) optimisation and control. The TL depends mainly on the mechanical parameters of the considered media. At early stages of the design, such parameters are not well known. Decision making tools are therefore needed to tackle this issue. In this paper, we consider the use of the Fourier Amplitude Sensitivity Test (FAST) for the analysis of the impact of mechanical parameters on features of interest. FAST is implemented with several structural configurations. FAST method is used to estimate the relative influence of the model parameters while assuming some uncertainty or variability on their values. The method offers a way to synthesize the results of a multiparametric analysis with large variability. Results are presented for transmission loss of isotropic, orthotropic and sandwich plates excited by a diffuse field on one side. Qualitative trends found to agree with the physical expectation. Design rules can then be set up for vibroacoustic indicators. The case of a sandwich plate is taken as an example of the use of this method inside an optimisation process and for uncertainty quantification.
Estimate design sensitivity to process variation for the 14nm node
NASA Astrophysics Data System (ADS)
Landié, Guillaume; Farys, Vincent
2016-03-01
Looking for the highest density and best performance, the 14nm technological node saw the development of aggressive designs, with design rules as close as possible to the limit of the process. Edge placement error (EPE) budget is now tighter and Reticle Enhancement Techniques (RET) must take into account the highest number of parameters to be able to get the best printability and guaranty yield requirements. Overlay is a parameter that must be taken into account earlier during the design library development to avoid design structures presenting a high risk of performance failure. This paper presents a method taking into account the overlay variation and the Resist Image simulation across the process window variation to estimate the design sensitivity to overlay. Areas in the design are classified with specific metrics, from the highest to the lowest overlay sensitivity. This classification can be used to evaluate the robustness of a full chip product to process variability or to work with designers during the design library development. The ultimate goal is to evaluate critical structures in different contexts and report the most critical ones. In this paper, we study layers interacting together, such as Contact/Poly area overlap or Contact/Active distance. ASML-Brion tooling allowed simulating the different resist contours and applying the overlay value to one of the layers. Lithography Manufacturability Check (LMC) detectors are then set to extract the desired values for analysis. Two different approaches have been investigated. The first one is a systematic overlay where we apply the same overlay everywhere on the design. The second one is using a real overlay map which has been measured and applied to the LMC tools. The data are then post-processed and compared to the design target to create a classification and show the error distribution. Figure:
Design sensitivity and Hessian matrix of generalized eigenproblems
NASA Astrophysics Data System (ADS)
Li, Li; Hu, Yujin; Wang, Xuelin
2014-02-01
A generalized eigenproblem is formed and its normalizations are presented and discussed. Then a unified consideration of the computation of the sensitivity and Hessian matrix is studied for both the self-adjoint and non-self-adjoint cases. In the self-adjoint case, a direct algebraic method is presented to determine the eigensolution derivatives simultaneously by solving a linear system with a symmetric coefficient matrix. In the non-self-adjoint case, an algebraic method is presented to determine the eigensolution derivatives directly and simultaneously without having to use the left eigenvectors. In this sense, the method has advantages in computational cost and storage capacity. It is shown that the second order derivatives of eigensolutions can also be obtained by solving a linear system and the computational effort of obtaining Hessian matrix is reduced remarkably since only the recalculation of the right-hand vector of the linear system is required. The presented methods are accurate, compact, numerically stable and easy to implement. Finally, two transcendental eigenproblem examples are used to demonstrate the validity of the presented methods. The first example is considered as an example of the case of non-self-adjoint systems, which can result from feedback control systems. The other example is used to illustrate the case of self-adjoint systems by considering the three bar truss structure which is a viscoelastic composite structure and consists of two aluminum truss components and one viscoelastic truss. In addition, the capacity of predicting the changes of eigenvalues and eigenvectors with respect to the changes of design parameters is studied.
Knopman, Debra S.; Voss, Clifford I.
1987-01-01
The spatial and temporal variability of sensitivities has a significant impact on parameter estimation and sampling design for studies of solute transport in porous media. Physical insight into the behavior of sensitivities is offered through an analysis of analytically derived sensitivities for the one-dimensional form of the advection-dispersion equation. When parameters are estimated in regression models of one-dimensional transport, the spatial and temporal variability in sensitivities influences variance and covariance of parameter estimates. Several principles account for the observed influence of sensitivities on parameter uncertainty. (1) Information about a physical parameter may be most accurately gained at points in space and time. (2) As the distance of observation points from the upstream boundary increases, maximum sensitivity to velocity during passage of the solute front increases. (3) The frequency of sampling must be 'in phase' with the S shape of the dispersion sensitivity curve to yield the most information on dispersion. (4) The sensitivity to the dispersion coefficient is usually at least an order of magnitude less than the sensitivity to velocity. (5) The assumed probability distribution of random error in observations of solute concentration determines the form of the sensitivities. (6) If variance in random error in observations is large, trends in sensitivities of observation points may be obscured by noise. (7) Designs that minimize the variance of one parameter may not necessarily minimize the variance of other parameters.
Partial Differential Algebraic Sensitivity Analysis Code
1995-05-15
PDASAC solves stiff, nonlinear initial-boundary-value in a timelike dimension t and a space dimension x. Plane, circular cylindrical or spherical boundaries can be handled. Mixed-order systems of partial differential and algebraic equations can be analyzed with members of order or 0 or 1 in t, 0,1 or 2 in x. Parametric sensitivities of the calculated states are compted simultaneously on request, via the Jacobian of the state equations. Initial and boundary conditions are efficiently reconciled.more » Local error control (in the max-norm or the 2-norm) is provided for the state vector and can include the parametric sensitivites if desired.« less
Automating sensitivity analysis of computer models using computer calculus
Oblow, E.M.; Pin, F.G.
1985-01-01
An automated procedure for performing sensitivity analyses has been developed. The procedure uses a new FORTRAN compiler with computer calculus capabilities to generate the derivatives needed to set up sensitivity equations. The new compiler is called GRESS - Gradient Enhanced Software System. Application of the automated procedure with ''direct'' and ''adjoint'' sensitivity theory for the analysis of non-linear, iterative systems of equations is discussed. Calculational efficiency consideration and techniques for adjoint sensitivity analysis are emphasized. The new approach is found to preserve the traditional advantages of adjoint theory while removing the tedious human effort previously needed to apply this theoretical methodology. Conclusions are drawn about the applicability of the automated procedure in numerical analysis and large-scale modelling sensitivity studies. 24 refs., 2 figs.
Automated procedure for sensitivity analysis using computer calculus
Oblow, E.M.
1983-05-01
An automated procedure for performing sensitivity analyses has been developed. The procedure uses a new FORTRAN compiler with computer calculus capabilities to generate the derivatives needed to set up sensitivity equations. The new compiler is called GRESS - Gradient Enhanced Software System. Application of the automated procedure with direct and adjoint sensitivity theory for the analysis of non-linear, iterative systems of equations is discussed. Calculational efficiency consideration and techniques for adjoint sensitivity analysis are emphasized. The new approach was found to preserve the traditional advantages of adjoint theory while removing the tedious human effort previously needed to apply this theoretical methodology. Conclusions are drawn about the applicability of the automated procedure in numerical analysis and large-scale modelling sensitivity studies.
Magnetic Coil Design and Analysis
NASA Astrophysics Data System (ADS)
Bulatowicz, Michael
2012-06-01
Modified magnetic field coil geometries as described in U.S. Patent Applications US20100194506 and US20110247414 can produce substantially greater magnetic field homogeneity as compared to the traditional realized versions of idealized magnetic coil geometries such as spherical or Helmholtz. The new coil geometries will be described in detail and will be compared and contrasted to realized versions of idealized geometries, including discussion of errors not typically accounted for in traditional coil design and analysis.
NASA Astrophysics Data System (ADS)
Jayant, Krishna; Auluck, Kshitij; Rodriguez, Sergio; Cao, Yingqiu; Kan, Edwin C.
2014-05-01
We report on factors that affect DNA hybridization detection using ion-sensitive field-effect transistors (ISFETs). Signal generation at the interface between the transistor and immobilized biomolecules is widely ascribed to unscreened molecular charges causing a shift in surface potential and hence the transistor output current. Traditionally, the interaction between DNA and the dielectric or metal sensing interface is modeled by treating the molecular layer as a sheet charge and the ionic profile with a Poisson-Boltzmann distribution. The surface potential under this scenario is described by the Graham equation. This approximation, however, often fails to explain large hybridization signals on the order of tens of mV. More realistic descriptions of the DNA-transistor interface which include factors such as ion permeation, exclusion, and packing constraints have been proposed with little or no corroboration against experimental findings. In this study, we examine such physical models by their assumptions, range of validity, and limitations. We compare simulations against experiments performed on electrolyte-oxide-semiconductor capacitors and foundry-ready floating-gate ISFETs. We find that with weakly charged interfaces (i.e., low intrinsic interface charge), pertinent to the surfaces used in this study, the best agreement between theory and experiment exists when ions are completely excluded from the DNA layer. The influence of various factors such as bulk pH, background salinity, chemical reactivity of surface groups, target molecule concentration, and surface coatings on signal generation is studied. Furthermore, in order to overcome Debye screening limited detection, we suggest two signal enhancement strategies. We first describe frequency domain biosensing, highlighting the ability to sort short DNA strands based on molecular length, and then describe DNA biosensing in multielectrolytes comprising trace amounts of higher-valency salt in a background of
Jayant, Krishna; Auluck, Kshitij; Rodriguez, Sergio; Cao, Yingqiu; Kan, Edwin C
2014-05-01
We report on factors that affect DNA hybridization detection using ion-sensitive field-effect transistors (ISFETs). Signal generation at the interface between the transistor and immobilized biomolecules is widely ascribed to unscreened molecular charges causing a shift in surface potential and hence the transistor output current. Traditionally, the interaction between DNA and the dielectric or metal sensing interface is modeled by treating the molecular layer as a sheet charge and the ionic profile with a Poisson-Boltzmann distribution. The surface potential under this scenario is described by the Graham equation. This approximation, however, often fails to explain large hybridization signals on the order of tens of mV. More realistic descriptions of the DNA-transistor interface which include factors such as ion permeation, exclusion, and packing constraints have been proposed with little or no corroboration against experimental findings. In this study, we examine such physical models by their assumptions, range of validity, and limitations. We compare simulations against experiments performed on electrolyte-oxide-semiconductor capacitors and foundry-ready floating-gate ISFETs. We find that with weakly charged interfaces (i.e., low intrinsic interface charge), pertinent to the surfaces used in this study, the best agreement between theory and experiment exists when ions are completely excluded from the DNA layer. The influence of various factors such as bulk pH, background salinity, chemical reactivity of surface groups, target molecule concentration, and surface coatings on signal generation is studied. Furthermore, in order to overcome Debye screening limited detection, we suggest two signal enhancement strategies. We first describe frequency domain biosensing, highlighting the ability to sort short DNA strands based on molecular length, and then describe DNA biosensing in multielectrolytes comprising trace amounts of higher-valency salt in a background of
A topological approach to computer-aided sensitivity analysis
NASA Technical Reports Server (NTRS)
Chan, S. P.; Munoz, R. M.
1971-01-01
Sensitivities of any arbitrary system are calculated using general purpose digital computer with available software packages for transfer function analysis. Sensitivity shows how element variation within system affects system performance. Signal flow graph illustrates topological system behavior and relationship among parameters in system.
Global and Local Sensitivity Analysis Methods for a Physical System
ERIC Educational Resources Information Center
Morio, Jerome
2011-01-01
Sensitivity analysis is the study of how the different input variations of a mathematical model influence the variability of its output. In this paper, we review the principle of global and local sensitivity analyses of a complex black-box system. A simulated case of application is given at the end of this paper to compare both approaches.…
Software Design for Smile Analysis
Sodagar, A.; Rafatjoo, R.; Gholami Borujeni, D.; Noroozi, H.; Sarkhosh, A.
2010-01-01
Introduction: Esthetics and attractiveness of the smile is one of the major demands in contemporary orthodontic treatment. In order to improve a smile design, it is necessary to record “posed smile” as an intentional, non-pressure, static, natural and reproducible smile. The record then should be analyzed to determine its characteristics. In this study, we intended to design and introduce a software to analyze the smile rapidly and precisely in order to produce an attractive smile for the patients. Materials and Methods: For this purpose, a practical study was performed to design multimedia software “Smile Analysis” which can receive patients’ photographs and videographs. After giving records to the software, the operator should mark the points and lines which are displayed on the system’s guide and also define the correct scale for each image. Thirty-three variables are measured by the software and displayed on the report page. Reliability of measurements in both image and video was significantly high (α=0.7–1). Results: In order to evaluate intra- operator and inter-operator reliability, five cases were selected randomly. Statistical analysis showed that calculations performed in smile analysis software were both valid and highly reliable (for both video and photo). Conclusion: The results obtained from smile analysis could be used in diagnosis, treatment planning and evaluation of the treatment progress. PMID:21998792
Advanced Fuel Cycle Economic Sensitivity Analysis
David Shropshire; Kent Williams; J.D. Smith; Brent Boore
2006-12-01
A fuel cycle economic analysis was performed on four fuel cycles to provide a baseline for initial cost comparison using the Gen IV Economic Modeling Work Group G4 ECON spreadsheet model, Decision Programming Language software, the 2006 Advanced Fuel Cycle Cost Basis report, industry cost data, international papers, the nuclear power related cost study from MIT, Harvard, and the University of Chicago. The analysis developed and compared the fuel cycle cost component of the total cost of energy for a wide range of fuel cycles including: once through, thermal with fast recycle, continuous fast recycle, and thermal recycle.
Sensitivity Analysis in Complex Plasma Chemistry Models
NASA Astrophysics Data System (ADS)
Turner, Miles
2015-09-01
The purpose of a plasma chemistry model is prediction of chemical species densities, including understanding the mechanisms by which such species are formed. These aims are compromised by an uncertain knowledge of the rate constants included in the model, which directly causes uncertainty in the model predictions. We recently showed that this predictive uncertainty can be large--a factor of ten or more in some cases. There is probably no context in which a plasma chemistry model might be used where the existence of uncertainty on this scale could not be a matter of concern. A question that at once follows is: Which rate constants cause such uncertainty? In the present paper we show how this question can be answered by applying a systematic screening procedure--the so-called Morris method--to identify sensitive rate constants. We investigate the topical example of the helium-oxygen chemistry. Beginning with a model with almost four hundred reactions, we show that only about fifty rate constants materially affect the model results, and as few as ten cause most of the uncertainty. This means that the model can be improved, and the uncertainty substantially reduced, by focussing attention on this tractably small set of rate constants. Work supported by Science Foundation Ireland under grant08/SRC/I1411, and by COST Action MP1101 ``Biomedical Applications of Atmospheric Pressure Plasmas.''
Selecting step sizes in sensitivity analysis by finite differences
NASA Technical Reports Server (NTRS)
Iott, J.; Haftka, R. T.; Adelman, H. M.
1985-01-01
This paper deals with methods for obtaining near-optimum step sizes for finite difference approximations to first derivatives with particular application to sensitivity analysis. A technique denoted the finite difference (FD) algorithm, previously described in the literature and applicable to one derivative at a time, is extended to the calculation of several simultaneously. Both the original and extended FD algorithms are applied to sensitivity analysis for a data-fitting problem in which derivatives of the coefficients of an interpolation polynomial are calculated with respect to uncertainties in the data. The methods are also applied to sensitivity analysis of the structural response of a finite-element-modeled swept wing. In a previous study, this sensitivity analysis of the swept wing required a time-consuming trial-and-error effort to obtain a suitable step size, but it proved to be a routine application for the extended FD algorithm herein.
Parameter sensitivity analysis for pesticide impacts on honeybee colonies
We employ Monte Carlo simulation and linear sensitivity analysis techniques to describe the dynamics of a bee exposure model, VarroaPop. Daily simulations are performed that simulate hive population trajectories, taking into account queen strength, foraging success, weather, colo...
SYSTEMATIC SENSITIVITY ANALYSIS OF AIR QUALITY SIMULATION MODELS
This report reviews and assesses systematic sensitivity and uncertainty analysis methods for applications to air quality simulation models. The discussion of the candidate methods presents their basic variables, mathematical foundations, user motivations and preferences, computer...
Validation of FSP Reactor Design with Sensitivity Studies of Beryllium-Reflected Critical Assemblies
John D. Bess; Margaret A. Marshall
2013-02-01
The baseline design for space nuclear power is a fission surface power (FSP) system: sodium-potassium (NaK) cooled, fast spectrum reactor with highly-enriched-uranium (HEU)-O2 fuel, stainless steel (SS) cladding, and beryllium reflectors with B4C control drums. Previous studies were performed to evaluate modeling capabilities and quantify uncertainties and biases associated with analysis methods and nuclear data. Comparison of Zero Power Plutonium Reactor (ZPPR)-20 benchmark experiments with the FSP design indicated that further reduction of the total design model uncertainty requires the reduction in uncertainties pertaining to beryllium and uranium cross-section data. Further comparison with three beryllium-reflected HEU-metal benchmark experiments performed at the Oak Ridge Critical Experiments Facility (ORCEF) concluded the requirement that experimental validation data have similar cross section sensitivities to those found in the FSP design. A series of critical experiments was performed at ORCEF in the 1960s to support the Medium Power Reactor Experiment (MPRE) space reactor design. The small, compact critical assembly (SCCA) experiments were graphite- or beryllium-reflected assemblies of SS-clad, HEU-O2 fuel on a vertical lift machine. All five configurations were evaluated as benchmarks. Two of the five configurations were beryllium reflected, and further evaluated using the sensitivity and uncertainty analysis capabilities of SCALE 6.1. Validation of the example FSP design model was successful in reducing the primary uncertainty constituent, the Be(n,n) reaction, from 0.28 %dk/k to 0.0004 %dk/k. Further assessment of additional reactor physics measurements performed on the SCCA experiments may serve to further validate FSP design and operation.
Analysis-Based Message Design: Rethinking Screen Design Guidelines.
ERIC Educational Resources Information Center
Beriswill, Joanne E.
This article describes the evolution of computer interface research issues from text-based interface design guidelines to more complex issues, including media selection, interface design, and visual design. This research is then integrated into the Analysis-based Message Design (AMD) process. The AMD process divides the interface design process…
Sensitivity Analysis of the Gap Heat Transfer Model in BISON.
Swiler, Laura Painton; Schmidt, Rodney C.; Williamson, Richard; Perez, Danielle
2014-10-01
This report summarizes the result of a NEAMS project focused on sensitivity analysis of the heat transfer model in the gap between the fuel rod and the cladding used in the BISON fuel performance code of Idaho National Laboratory. Using the gap heat transfer models in BISON, the sensitivity of the modeling parameters and the associated responses is investigated. The study results in a quantitative assessment of the role of various parameters in the analysis of gap heat transfer in nuclear fuel.
Fixed point sensitivity analysis of interacting structured populations.
Barabás, György; Meszéna, Géza; Ostling, Annette
2014-03-01
Sensitivity analysis of structured populations is a useful tool in population ecology. Historically, methodological development of sensitivity analysis has focused on the sensitivity of eigenvalues in linear matrix models, and on single populations. More recently there have been extensions to the sensitivity of nonlinear models, and to communities of interacting populations. Here we derive a fully general mathematical expression for the sensitivity of equilibrium abundances in communities of interacting structured populations. Our method yields the response of an arbitrary function of the stage class abundances to perturbations of any model parameters. As a demonstration, we apply this sensitivity analysis to a two-species model of ontogenetic niche shift where each species has two stage classes, juveniles and adults. In the context of this model, we demonstrate that our theory is quite robust to violating two of its technical assumptions: the assumption that the community is at a point equilibrium and the assumption of infinitesimally small parameter perturbations. Our results on the sensitivity of a community are also interpreted in a niche theoretical context: we determine how the niche of a structured population is composed of the niches of the individual states, and how the sensitivity of the community depends on niche segregation. PMID:24368160
Advancing sensitivity analysis to precisely characterize temporal parameter dominance
NASA Astrophysics Data System (ADS)
Guse, Björn; Pfannerstill, Matthias; Strauch, Michael; Reusser, Dominik; Lüdtke, Stefan; Volk, Martin; Gupta, Hoshin; Fohrer, Nicola
2016-04-01
Parameter sensitivity analysis is a strategy for detecting dominant model parameters. A temporal sensitivity analysis calculates daily sensitivities of model parameters. This allows a precise characterization of temporal patterns of parameter dominance and an identification of the related discharge conditions. To achieve this goal, the diagnostic information as derived from the temporal parameter sensitivity is advanced by including discharge information in three steps. In a first step, the temporal dynamics are analyzed by means of daily time series of parameter sensitivities. As sensitivity analysis method, we used the Fourier Amplitude Sensitivity Test (FAST) applied directly onto the modelled discharge. Next, the daily sensitivities are analyzed in combination with the flow duration curve (FDC). Through this step, we determine whether high sensitivities of model parameters are related to specific discharges. Finally, parameter sensitivities are separately analyzed for five segments of the FDC and presented as monthly averaged sensitivities. In this way, seasonal patterns of dominant model parameter are provided for each FDC segment. For this methodical approach, we used two contrasting catchments (upland and lowland catchment) to illustrate how parameter dominances change seasonally in different catchments. For all of the FDC segments, the groundwater parameters are dominant in the lowland catchment, while in the upland catchment the controlling parameters change seasonally between parameters from different runoff components. The three methodical steps lead to clear temporal patterns, which represent the typical characteristics of the study catchments. Our methodical approach thus provides a clear idea of how the hydrological dynamics are controlled by model parameters for certain discharge magnitudes during the year. Overall, these three methodical steps precisely characterize model parameters and improve the understanding of process dynamics in hydrological
Design Spectrum Analysis in NASTRAN
NASA Technical Reports Server (NTRS)
Butler, T. G.
1984-01-01
The utility of Design Spectrum Analysis is to give a mode by mode characterization of the behavior of a design under a given loading. The theory of design spectrum is discussed after operations are explained. User instructions are taken up here in three parts: Transient Preface, Maximum Envelope Spectrum, and RMS Average Spectrum followed by a Summary Table. A single DMAP ALTER packet will provide for all parts of the design spectrum operations. The starting point for getting a modal break-down of the response to acceleration loading is the Modal Transient rigid format. After eigenvalue extraction, modal vectors need to be isolated in the full set of physical coordinates (P-sized as opposed to the D-sized vectors in RF 12). After integration for transient response the results are scanned over the solution time interval for the peak values and for the times that they occur. A module called SCAN was written to do this job, that organizes these maxima into a diagonal output matrix. The maximum amplifier in each mode is applied to the eigenvector of each mode which then reveals the maximum displacements, stresses, forces and boundary reactions that the structure will experience for a load history, mode by mode. The standard NASTRAN output processors have been modified for this task. It is required that modes be normalized to mass.
MAP stability, design, and analysis
NASA Technical Reports Server (NTRS)
Ericsson-Jackson, A. J.; Andrews, S. F.; O'Donnell, J. R., Jr.; Markley, F. L.
1998-01-01
The Microwave Anisotropy Probe (MAP) is a follow-on to the Differential Microwave Radiometer (DMR) instrument on the Cosmic Background Explorer (COBE) spacecraft. The design and analysis of the MAP attitude control system (ACS) have been refined since work previously reported. The full spacecraft and instrument flexible model was developed in NASTRAN, and the resulting flexible modes were plotted and reduced with the Modal Significance Analysis Package (MSAP). The reduced-order model was used to perform the linear stability analysis for each control mode, the results of which are presented in this paper. Although MAP is going to a relatively disturbance-free Lissajous orbit around the Earth-Sun L(2) Lagrange point, a detailed disturbance-torque analysis is required because there are only a small number of opportunities for momentum unloading each year. Environmental torques, including solar pressure at L(2), aerodynamic and gravity gradient during phasing-loop orbits, were calculated and simulated. Thruster plume impingement torques that could affect the performance of the thruster modes were estimated and simulated, and a simple model of fuel slosh was derived to model its effect on the motion of the spacecraft. In addition, a thruster mode linear impulse controller was developed to meet the accuracy requirements of the phasing loop burns. A dynamic attitude error limiter was added to improve the performance of the ACS during large attitude slews. The result of this analysis is a stable ACS subsystem that meets all of the mission's requirements.
Lock Acquisition and Sensitivity Analysis of Advanced LIGO Interferometers
NASA Astrophysics Data System (ADS)
Martynov, Denis
Laser interferometer gravitational wave observatory (LIGO) consists of two complex large-scale laser interferometers designed for direct detection of gravitational waves from distant astrophysical sources in the frequency range 10Hz - 5kHz. Direct detection of space-time ripples will support Einstein's general theory of relativity and provide invaluable information and new insight into physics of the Universe. The initial phase of LIGO started in 2002, and since then data was collected during the six science runs. Instrument sensitivity improved from run to run due to the effort of commissioning team. Initial LIGO has reached designed sensitivity during the last science run, which ended in October 2010. In parallel with commissioning and data analysis with the initial detector, LIGO group worked on research and development of the next generation of detectors. Major instrument upgrade from initial to advanced LIGO started in 2010 and lasted until 2014. This thesis describes results of commissioning work done at the LIGO Livingston site from 2013 until 2015 in parallel with and after the installation of the instrument. This thesis also discusses new techniques and tools developed at the 40m prototype including adaptive filtering, estimation of quantization noise in digital filters and design of isolation kits for ground seismometers. The first part of this thesis is devoted to the description of methods for bringing the interferometer into linear regime when collection of data becomes possible. States of longitudinal and angular controls of interferometer degrees of freedom during lock acquisition process and in low noise configuration are discussed in details. Once interferometer is locked and transitioned to low noise regime, instrument produces astrophysics data that should be calibrated to units of meters or strain. The second part of this thesis describes online calibration technique set up in both observatories to monitor the quality of the collected data in
Sensitivity analysis of geometric errors in additive manufacturing medical models.
Pinto, Jose Miguel; Arrieta, Cristobal; Andia, Marcelo E; Uribe, Sergio; Ramos-Grez, Jorge; Vargas, Alex; Irarrazaval, Pablo; Tejos, Cristian
2015-03-01
Additive manufacturing (AM) models are used in medical applications for surgical planning, prosthesis design and teaching. For these applications, the accuracy of the AM models is essential. Unfortunately, this accuracy is compromised due to errors introduced by each of the building steps: image acquisition, segmentation, triangulation, printing and infiltration. However, the contribution of each step to the final error remains unclear. We performed a sensitivity analysis comparing errors obtained from a reference with those obtained modifying parameters of each building step. Our analysis considered global indexes to evaluate the overall error, and local indexes to show how this error is distributed along the surface of the AM models. Our results show that the standard building process tends to overestimate the AM models, i.e. models are larger than the original structures. They also show that the triangulation resolution and the segmentation threshold are critical factors, and that the errors are concentrated at regions with high curvatures. Errors could be reduced choosing better triangulation and printing resolutions, but there is an important need for modifying some of the standard building processes, particularly the segmentation algorithms. PMID:25649961
DESIGN ANALYSIS FOR THE NAVAL SNF WASTE PACKAGE
T.L. Mitchell
2000-05-31
The purpose of this analysis is to demonstrate the design of the naval spent nuclear fuel (SNF) waste package (WP) using the Waste Package Department's (WPD) design methodologies and processes described in the ''Waste Package Design Methodology Report'' (CRWMS M&O [Civilian Radioactive Waste Management System Management and Operating Contractor] 2000b). The calculations that support the design of the naval SNF WP will be discussed; however, only a sub-set of such analyses will be presented and shall be limited to those identified in the ''Waste Package Design Sensitivity Report'' (CRWMS M&O 2000c). The objective of this analysis is to describe the naval SNF WP design method and to show that the design of the naval SNF WP complies with the ''Naval Spent Nuclear Fuel Disposal Container System Description Document'' (CRWMS M&O 1999a) and Interface Control Document (ICD) criteria for Site Recommendation. Additional criteria for the design of the naval SNF WP have been outlined in Section 6.2 of the ''Waste Package Design Sensitivity Report'' (CRWMS M&O 2000c). The scope of this analysis is restricted to the design of the naval long WP containing one naval long SNF canister. This WP is representative of the WPs that will contain both naval short SNF and naval long SNF canisters. The following items are included in the scope of this analysis: (1) Providing a general description of the applicable design criteria; (2) Describing the design methodology to be used; (3) Presenting the design of the naval SNF waste package; and (4) Showing compliance with all applicable design criteria. The intended use of this analysis is to support Site Recommendation reports and assist in the development of WPD drawings. Activities described in this analysis were conducted in accordance with the technical product development plan (TPDP) ''Design Analysis for the Naval SNF Waste Package (CRWMS M&O 2000a).
Designing novel nano-immunoassays: antibody orientation versus sensitivity
NASA Astrophysics Data System (ADS)
Puertas, S.; Moros, M.; Fernández-Pacheco, R.; Ibarra, M. R.; Grazú, V.; de la Fuente, J. M.
2010-12-01
There is a growing interest in the use of magnetic nanoparticles (MNPs) for their application in quantitative and highly sensitive biosensors. Their use as labels of biological recognition events and their detection by means of some magnetic method constitute a very promising strategy for quantitative high-sensitive lateral-flow assays. In this paper, we report the importance of nanoparticle functionalization for the improvement of sensitivity for a lateral-flow immunoassay. More precisely, we have found that immobilization of IgG anti-hCG through its polysaccharide moieties on MNPs allows more successful recognition of the hCG hormone. Although we have used the detection of hCG as a model in this work, the strategy of binding antibodies to MNPs through its sugar chains reported here is applicable to other antibodies. It has huge potential as it will be very useful for the development of quantitative and high-sensitive lateral-flow assays for its use on human and veterinary, medicine, food and beverage manufacturing, pharmaceutical, medical biologics and personal care product production, environmental remediation, etc.
NASA Astrophysics Data System (ADS)
Edouard, C.; Petit, M.; Forgez, C.; Bernard, J.; Revel, R.
2016-09-01
In this work, a simplified electrochemical and thermal model that can predict both physicochemical and aging behavior of Li-ion batteries is studied. A sensitivity analysis of all its physical parameters is performed in order to find out their influence on the model output based on simulations under various conditions. The results gave hints on whether a parameter needs particular attention when measured or identified and on the conditions (e.g. temperature, discharge rate) under which it is the most sensitive. A specific simulation profile is designed for parameters involved in aging equations in order to determine their sensitivity. Finally, a step-wise method is followed to limit the influence of parameter values when identifying some of them, according to their relative sensitivity from the study. This sensitivity analysis and the subsequent step-wise identification method show very good results, such as a better fitting of the simulated cell voltage with experimental data.
Loganathan, Muthukumaran; Bristow, Douglas A.
2014-04-15
This paper presents a method and cantilever design for improving the mechanical measurement sensitivity in the atomic force microscopy (AFM) tapping mode. The method uses two harmonics in the drive signal to generate a bi-harmonic tapping trajectory. Mathematical analysis demonstrates that the wide-valley bi-harmonic tapping trajectory is as much as 70% more sensitive to changes in the sample topography than the standard single-harmonic trajectory typically used. Although standard AFM cantilevers can be driven in the bi-harmonic tapping trajectory, they require large forcing at the second harmonic. A design is presented for a bi-harmonic cantilever that has a second resonant mode at twice its first resonant mode, thereby capable of generating bi-harmonic trajectories with small forcing signals. Bi-harmonic cantilevers are fabricated by milling a small cantilever on the interior of a standard cantilever probe using a focused ion beam. Bi-harmonic drive signals are derived for standard cantilevers and bi-harmonic cantilevers. Experimental results demonstrate better than 30% improvement in measurement sensitivity using the bi-harmonic cantilever. Images obtained through bi-harmonic tapping exhibit improved sharpness and surface tracking, especially at high scan speeds and low force fields.
Loganathan, Muthukumaran; Bristow, Douglas A
2014-04-01
This paper presents a method and cantilever design for improving the mechanical measurement sensitivity in the atomic force microscopy (AFM) tapping mode. The method uses two harmonics in the drive signal to generate a bi-harmonic tapping trajectory. Mathematical analysis demonstrates that the wide-valley bi-harmonic tapping trajectory is as much as 70% more sensitive to changes in the sample topography than the standard single-harmonic trajectory typically used. Although standard AFM cantilevers can be driven in the bi-harmonic tapping trajectory, they require large forcing at the second harmonic. A design is presented for a bi-harmonic cantilever that has a second resonant mode at twice its first resonant mode, thereby capable of generating bi-harmonic trajectories with small forcing signals. Bi-harmonic cantilevers are fabricated by milling a small cantilever on the interior of a standard cantilever probe using a focused ion beam. Bi-harmonic drive signals are derived for standard cantilevers and bi-harmonic cantilevers. Experimental results demonstrate better than 30% improvement in measurement sensitivity using the bi-harmonic cantilever. Images obtained through bi-harmonic tapping exhibit improved sharpness and surface tracking, especially at high scan speeds and low force fields. PMID:24784614
Design and analysis of a micromachined gyroscope
NASA Astrophysics Data System (ADS)
Zarei, Nilgoon; Leung, Albert; Jones, John D.
2012-03-01
This paper describes the simulation and design of a MEMS thermal gyroscope and optimizing the design for increased sensitivity through the use of the Comsol Multiphysics software package. Two different designs are described, and the effects of working fluid properties are explored. A prototype of this device has been fabricated using techniques for rapid prototyping of MEMS transducers.
Mun, J S; Han, M Y
2012-01-01
The appropriate design and evaluation of a rainwater harvesting (RWH) system is necessary to improve system performance and the stability of the water supply. The main design parameters (DPs) of an RWH system are rainfall, catchment area, collection efficiency, tank volume and water demand. Its operational parameters (OPs) include rainwater use efficiency (RUE), water saving efficiency (WSE) and cycle number (CN). The sensitivity analysis of a rooftop RWH system's DPs to its OPs reveals that the ratio of tank volume to catchment area (V/A) for an RWH system in Seoul, South Korea is recommended between 0.03 and 0.08 in terms of rate of change in RUE. The appropriate design value of V/A is varied with D/A. The extra tank volume up to V/A of 0.15∼0.2 is also available, if necessary to secure more water. Accordingly, we should figure out suitable value or range of DPs based on the sensitivity analysis to optimize design of an RWH system or improve operation efficiency. The operational data employed in this study, which was carried out to validate the design and evaluation method of an RWH system, were obtained from the system in use at a dormitory complex at Seoul National University (SNU) in Korea. The results of these operational data are in good agreement with those used in the initial simulation. The proposed method and the results of this research will be useful in evaluating and comparing the performance of RWH systems. It is found that RUE can be increased by expanding the variety of rainwater uses, particularly in the high rainfall season. PMID:22054581
Space Propulsion Design and Analysis
NASA Technical Reports Server (NTRS)
Sinha, Neeraj; Brinckman, Kevin; Ayyalasomayajula, Hanitha; Dash, Sanford
2007-01-01
This software provides an improved methodology for predicting launcher base pressure and heat loads for RSRM (Reusable Solid Rocket Motor) launchers by accounting for complex anisotropic stress/strains and variable turbulent Prandtl and Schmidt numbers. A "building block" approach to turbulence model development, and validation has been applied for improved missile/launcher base region analysis. Modifications to existing kappa - epsilon turbulence models and application of scalar variance models are incorporated into a RANS-based method for aeropropulsive flow modeling, directly related to base flow methodology. (RANS stands for Reynolds-averaged Navier-Stokes.) The models are applied in a RANS solver framework and can improve analysis of other complex flow fields. The enhanced models provide a more accurate predictive capability for improving the design and analysis of RSRM launcher configuration. The kappa - epsilon model enhancements have been shown to improve the capability for predicting turbulence effects in base blow environments. The scalar variance models have been assessed over a wide range of flow configurations to improve prediction of turbulent scalar mixing.
Sensitivity Analysis of the Integrated Medical Model for ISS Programs
NASA Technical Reports Server (NTRS)
Goodenow, D. A.; Myers, J. G.; Arellano, J.; Boley, L.; Garcia, Y.; Saile, L.; Walton, M.; Kerstman, E.; Reyes, D.; Young, M.
2016-01-01
Sensitivity analysis estimates the relative contribution of the uncertainty in input values to the uncertainty of model outputs. Partial Rank Correlation Coefficient (PRCC) and Standardized Rank Regression Coefficient (SRRC) are methods of conducting sensitivity analysis on nonlinear simulation models like the Integrated Medical Model (IMM). The PRCC method estimates the sensitivity using partial correlation of the ranks of the generated input values to each generated output value. The partial part is so named because adjustments are made for the linear effects of all the other input values in the calculation of correlation between a particular input and each output. In SRRC, standardized regression-based coefficients measure the sensitivity of each input, adjusted for all the other inputs, on each output. Because the relative ranking of each of the inputs and outputs is used, as opposed to the values themselves, both methods accommodate the nonlinear relationship of the underlying model. As part of the IMM v4.0 validation study, simulations are available that predict 33 person-missions on ISS and 111 person-missions on STS. These simulated data predictions feed the sensitivity analysis procedures. The inputs to the sensitivity procedures include the number occurrences of each of the one hundred IMM medical conditions generated over the simulations and the associated IMM outputs: total quality time lost (QTL), number of evacuations (EVAC), and number of loss of crew lives (LOCL). The IMM team will report the results of using PRCC and SRRC on IMM v4.0 predictions of the ISS and STS missions created as part of the external validation study. Tornado plots will assist in the visualization of the condition-related input sensitivities to each of the main outcomes. The outcomes of this sensitivity analysis will drive review focus by identifying conditions where changes in uncertainty could drive changes in overall model output uncertainty. These efforts are an integral
Cacuci, Dan G.; Ionescu-Bujor, Mihaela
2004-07-15
Part II of this review paper highlights the salient features of the most popular statistical methods currently used for local and global sensitivity and uncertainty analysis of both large-scale computational models and indirect experimental measurements. These statistical procedures represent sampling-based methods (random sampling, stratified importance sampling, and Latin Hypercube sampling), first- and second-order reliability algorithms (FORM and SORM, respectively), variance-based methods (correlation ratio-based methods, the Fourier Amplitude Sensitivity Test, and the Sobol Method), and screening design methods (classical one-at-a-time experiments, global one-at-a-time design methods, systematic fractional replicate designs, and sequential bifurcation designs). It is emphasized that all statistical uncertainty and sensitivity analysis procedures first commence with the 'uncertainty analysis' stage and only subsequently proceed to the 'sensitivity analysis' stage; this path is the exact reverse of the conceptual path underlying the methods of deterministic sensitivity and uncertainty analysis where the sensitivities are determined prior to using them for uncertainty analysis. By comparison to deterministic methods, statistical methods for uncertainty and sensitivity analysis are relatively easier to develop and use but cannot yield exact values of the local sensitivities. Furthermore, current statistical methods have two major inherent drawbacks as follows: 1. Since many thousands of simulations are needed to obtain reliable results, statistical methods are at best expensive (for small systems) or, at worst, impracticable (e.g., for large time-dependent systems).2. Since the response sensitivities and parameter uncertainties are inherently and inseparably amalgamated in the results produced by these methods, improvements in parameter uncertainties cannot be directly propagated to improve response uncertainties; rather, the entire set of simulations and
Least Squares Shadowing Sensitivity Analysis of Chaotic Flow Around a Two-Dimensional Airfoil
NASA Technical Reports Server (NTRS)
Blonigan, Patrick J.; Wang, Qiqi; Nielsen, Eric J.; Diskin, Boris
2016-01-01
Gradient-based sensitivity analysis has proven to be an enabling technology for many applications, including design of aerospace vehicles. However, conventional sensitivity analysis methods break down when applied to long-time averages of chaotic systems. This breakdown is a serious limitation because many aerospace applications involve physical phenomena that exhibit chaotic dynamics, most notably high-resolution large-eddy and direct numerical simulations of turbulent aerodynamic flows. A recently proposed methodology, Least Squares Shadowing (LSS), avoids this breakdown and advances the state of the art in sensitivity analysis for chaotic flows. The first application of LSS to a chaotic flow simulated with a large-scale computational fluid dynamics solver is presented. The LSS sensitivity computed for this chaotic flow is verified and shown to be accurate, but the computational cost of the current LSS implementation is high.
Computational aspects of sensitivity calculations in linear transient structural analysis
NASA Technical Reports Server (NTRS)
Greene, W. H.; Haftka, R. T.
1991-01-01
The calculation of sensitivities of displacements, velocities, accelerations, and stresses in linear, structural, and transient response problems is studied. Several existing sensitivity calculation methods and two new methods are compared for three example problems. Approximation vectors such as vibration mode shapes are used to reduce the dimensionality of the finite model. To be computationally practical in large-order problems, the overall finite difference methods must use the approximation vectors from the original design in the analyses of the perturbed models. This was found to result in poor convergence of stress sensitivities in several cases. Two semianalytical techniques are developed to overcome this poor convergence. Both new methods result in very good convergence of the stress sensitivities; the computational cost is much less than would result if the vibration modes were recalculated and then used in an overall finite difference method.
Sensitivity analysis technique for application to deterministic models
Ishigami, T.; Cazzoli, E.; Khatib-Rahbar, M.; Unwin, S.D.
1987-01-01
The characterization of sever accident source terms for light water reactors should include consideration of uncertainties. An important element of any uncertainty analysis is an evaluation of the sensitivity of the output probability distributions reflecting source term uncertainties to assumptions regarding the input probability distributions. Historically, response surface methods (RSMs) were developed to replace physical models using, for example, regression techniques, with simplified models for example, regression techniques, with simplified models for extensive calculations. The purpose of this paper is to present a new method for sensitivity analysis that does not utilize RSM, but instead relies directly on the results obtained from the original computer code calculations. The merits of this approach are demonstrated by application of the proposed method to the suppression pool aerosol removal code (SPARC), and the results are compared with those obtained by sensitivity analysis with (a) the code itself, (b) a regression model, and (c) Iman's method.
Sensitivity analysis for missing data in regulatory submissions.
Permutt, Thomas
2016-07-30
The National Research Council Panel on Handling Missing Data in Clinical Trials recommended that sensitivity analyses have to be part of the primary reporting of findings from clinical trials. Their specific recommendations, however, seem not to have been taken up rapidly by sponsors of regulatory submissions. The NRC report's detailed suggestions are along rather different lines than what has been called sensitivity analysis in the regulatory setting up to now. Furthermore, the role of sensitivity analysis in regulatory decision-making, although discussed briefly in the NRC report, remains unclear. This paper will examine previous ideas of sensitivity analysis with a view to explaining how the NRC panel's recommendations are different and possibly better suited to coping with present problems of missing data in the regulatory setting. It will also discuss, in more detail than the NRC report, the relevance of sensitivity analysis to decision-making, both for applicants and for regulators. Published 2015. This article is a U.S. Government work and is in the public domain in the USA. PMID:26567763
Multiobjective sensitivity analysis and optimization of distributed hydrologic model MOBIDIC
NASA Astrophysics Data System (ADS)
Yang, J.; Castelli, F.; Chen, Y.
2014-10-01
Calibration of distributed hydrologic models usually involves how to deal with the large number of distributed parameters and optimization problems with multiple but often conflicting objectives that arise in a natural fashion. This study presents a multiobjective sensitivity and optimization approach to handle these problems for the MOBIDIC (MOdello di Bilancio Idrologico DIstribuito e Continuo) distributed hydrologic model, which combines two sensitivity analysis techniques (the Morris method and the state-dependent parameter (SDP) method) with multiobjective optimization (MOO) approach ɛ-NSGAII (Non-dominated Sorting Genetic Algorithm-II). This approach was implemented to calibrate MOBIDIC with its application to the Davidson watershed, North Carolina, with three objective functions, i.e., the standardized root mean square error (SRMSE) of logarithmic transformed discharge, the water balance index, and the mean absolute error of the logarithmic transformed flow duration curve, and its results were compared with those of a single objective optimization (SOO) with the traditional Nelder-Mead simplex algorithm used in MOBIDIC by taking the objective function as the Euclidean norm of these three objectives. Results show that (1) the two sensitivity analysis techniques are effective and efficient for determining the sensitive processes and insensitive parameters: surface runoff and evaporation are very sensitive processes to all three objective functions, while groundwater recession and soil hydraulic conductivity are not sensitive and were excluded in the optimization. (2) Both MOO and SOO lead to acceptable simulations; e.g., for MOO, the average Nash-Sutcliffe value is 0.75 in the calibration period and 0.70 in the validation period. (3) Evaporation and surface runoff show similar importance for watershed water balance, while the contribution of baseflow can be ignored. (4) Compared to SOO, which was dependent on the initial starting location, MOO provides more
Sensitivity analysis approach to multibody systems described by natural coordinates
NASA Astrophysics Data System (ADS)
Li, Xiufeng; Wang, Yabin
2014-03-01
The classical natural coordinate modeling method which removes the Euler angles and Euler parameters from the governing equations is particularly suitable for the sensitivity analysis and optimization of multibody systems. However, the formulation has so many principles in choosing the generalized coordinates that it hinders the implementation of modeling automation. A first order direct sensitivity analysis approach to multibody systems formulated with novel natural coordinates is presented. Firstly, a new selection method for natural coordinate is developed. The method introduces 12 coordinates to describe the position and orientation of a spatial object. On the basis of the proposed natural coordinates, rigid constraint conditions, the basic constraint elements as well as the initial conditions for the governing equations are derived. Considering the characteristics of the governing equations, the newly proposed generalized-α integration method is used and the corresponding algorithm flowchart is discussed. The objective function, the detailed analysis process of first order direct sensitivity analysis and related solving strategy are provided based on the previous modeling system. Finally, in order to verify the validity and accuracy of the method presented, the sensitivity analysis of a planar spinner-slider mechanism and a spatial crank-slider mechanism are conducted. The test results agree well with that of the finite difference method, and the maximum absolute deviation of the results is less than 3%. The proposed approach is not only convenient for automatic modeling, but also helpful for the reduction of the complexity of sensitivity analysis, which provides a practical and effective way to obtain sensitivity for the optimization problems of multibody systems.
Shape design sensitivities using fully automatic 3-D mesh generation
NASA Technical Reports Server (NTRS)
Botkin, M. E.
1990-01-01
Previous work in three dimensional shape optimization involved specifying design variables by associating parameters directly with mesh points. More recent work has shown the use of fully-automatic mesh generation based upon a parameterized geometric representation. Design variables have been associated with a mathematical model of the part rather than the discretized representation. The mesh generation procedure uses a nonuniform grid intersection technique to place nodal points directly on the surface geometry. Although there exists an associativity between the mesh and the geometrical/topological entities, there is no mathematical functional relationship. This poses a problem during certain steps in the optimization process in which geometry modification is required. For the large geometrical changes which occur at the beginning of each optimization step, a completely new mesh is created. However, for gradient calculations many small changes must be made and it would be too costly to regenerate the mesh for each design variable perturbation. For that reason, a local remeshing procedure has been implemented which operates only on the specific edges and faces associated with the design variable being perturbed. Two realistic design problems are presented which show the efficiency of this process and test the accuracy of the gradient computations.
NASA Technical Reports Server (NTRS)
Hornberger, G. M.; Rastetter, E. B.
1982-01-01
A literature review of the use of sensitivity analyses in modelling nonlinear, ill-defined systems, such as ecological interactions is presented. Discussions of previous work, and a proposed scheme for generalized sensitivity analysis applicable to ill-defined systems are included. This scheme considers classes of mathematical models, problem-defining behavior, analysis procedures (especially the use of Monte-Carlo methods), sensitivity ranking of parameters, and extension to control system design.
Sensitivity analysis of the fission gas behavior model in BISON.
Swiler, Laura Painton; Pastore, Giovanni; Perez, Danielle; Williamson, Richard
2013-05-01
This report summarizes the result of a NEAMS project focused on sensitivity analysis of a new model for the fission gas behavior (release and swelling) in the BISON fuel performance code of Idaho National Laboratory. Using the new model in BISON, the sensitivity of the calculated fission gas release and swelling to the involved parameters and the associated uncertainties is investigated. The study results in a quantitative assessment of the role of intrinsic uncertainties in the analysis of fission gas behavior in nuclear fuel.
Sensitivity analysis for handling uncertainty in an economic evaluation.
Limwattananon, Supon
2014-05-01
To meet updated international standards, this paper revises the previous Thai guidelines for conducting sensitivity analyses as part of the decision analysis model for health technology assessment. It recommends both deterministic and probabilistic sensitivity analyses to handle uncertainty of the model parameters, which are best represented graphically. Two new methodological issues are introduced-a threshold analysis of medicines' unit prices for fulfilling the National Lists of Essential Medicines' requirements and the expected value of information for delaying decision-making in contexts where there are high levels of uncertainty. Further research is recommended where parameter uncertainty is significant and where the cost of conducting the research is not prohibitive. PMID:24964700
Self-validated Variance-based Methods for Sensitivity Analysis of Model Outputs
Tong, C
2009-04-20
Global sensitivity analysis (GSA) has the advantage over local sensitivity analysis in that GSA does not require strong model assumptions such as linearity or monotonicity. As a result, GSA methods such as those based on variance decomposition are well-suited to multi-physics models, which are often plagued by large nonlinearities. However, as with many other sampling-based methods, inadequate sample size can badly pollute the result accuracies. A natural remedy is to adaptively increase the sample size until sufficient accuracy is obtained. This paper proposes an iterative methodology comprising mechanisms for guiding sample size selection and self-assessing result accuracy. The elegant features in the the proposed methodology are the adaptive refinement strategies for stratified designs. We first apply this iterative methodology to the design of a self-validated first-order sensitivity analysis algorithm. We also extend this methodology to design a self-validated second-order sensitivity analysis algorithm based on refining replicated orthogonal array designs. Several numerical experiments are given to demonstrate the effectiveness of these methods.
NASA Astrophysics Data System (ADS)
Razavi, Saman; Gupta, Hoshin; Haghnegahdar, Amin
2016-04-01
Global sensitivity analysis (GSA) is a systems theoretic approach to characterizing the overall (average) sensitivity of one or more model responses across the factor space, by attributing the variability of those responses to different controlling (but uncertain) factors (e.g., model parameters, forcings, and boundary and initial conditions). GSA can be very helpful to improve the credibility and utility of Earth and Environmental System Models (EESMs), as these models are continually growing in complexity and dimensionality with continuous advances in understanding and computing power. However, conventional approaches to GSA suffer from (1) an ambiguous characterization of sensitivity, and (2) poor computational efficiency, particularly as the problem dimension grows. Here, we identify several important sensitivity-related characteristics of response surfaces that must be considered when investigating and interpreting the ''global sensitivity'' of a model response (e.g., a metric of model performance) to its parameters/factors. Accordingly, we present a new and general sensitivity and uncertainty analysis framework, Variogram Analysis of Response Surfaces (VARS), based on an analogy to 'variogram analysis', that characterizes a comprehensive spectrum of information on sensitivity. We prove, theoretically, that Morris (derivative-based) and Sobol (variance-based) methods and their extensions are special cases of VARS, and that their SA indices are contained within the VARS framework. We also present a practical strategy for the application of VARS to real-world problems, called STAR-VARS, including a new sampling strategy, called "star-based sampling". Our results across several case studies show the STAR-VARS approach to provide reliable and stable assessments of "global" sensitivity, while being at least 1-2 orders of magnitude more efficient than the benchmark Morris and Sobol approaches.
Sensitivity Analysis and Optimization of Aerodynamic Configurations with Blend Surfaces
NASA Technical Reports Server (NTRS)
Thomas, A. M.; Tiwari, S. N.
1997-01-01
A novel (geometrical) parametrization procedure using solutions to a suitably chosen fourth order partial differential equation is used to define a class of airplane configurations. Inclusive in this definition are surface grids, volume grids, and grid sensitivity. The general airplane configuration has wing, fuselage, vertical tail and horizontal tail. The design variables are incorporated into the boundary conditions, and the solution is expressed as a Fourier series. The fuselage has circular cross section, and the radius is an algebraic function of four design parameters and an independent computational variable. Volume grids are obtained through an application of the Control Point Form method. A graphic interface software is developed which dynamically changes the surface of the airplane configuration with the change in input design variable. The software is made user friendly and is targeted towards the initial conceptual development of any aerodynamic configurations. Grid sensitivity with respect to surface design parameters and aerodynamic sensitivity coefficients based on potential flow is obtained using an Automatic Differentiation precompiler software tool ADIFOR. Aerodynamic shape optimization of the complete aircraft with twenty four design variables is performed. Unstructured and structured volume grids and Euler solutions are obtained with standard software to demonstrate the feasibility of the new surface definition.
Efficient sensitivity analysis method for chaotic dynamical systems
NASA Astrophysics Data System (ADS)
Liao, Haitao
2016-05-01
The direct differentiation and improved least squares shadowing methods are both developed for accurately and efficiently calculating the sensitivity coefficients of time averaged quantities for chaotic dynamical systems. The key idea is to recast the time averaged integration term in the form of differential equation before applying the sensitivity analysis method. An additional constraint-based equation which forms the augmented equations of motion is proposed to calculate the time averaged integration variable and the sensitivity coefficients are obtained as a result of solving the augmented differential equations. The application of the least squares shadowing formulation to the augmented equations results in an explicit expression for the sensitivity coefficient which is dependent on the final state of the Lagrange multipliers. The LU factorization technique to calculate the Lagrange multipliers leads to a better performance for the convergence problem and the computational expense. Numerical experiments on a set of problems selected from the literature are presented to illustrate the developed methods. The numerical results demonstrate the correctness and effectiveness of the present approaches and some short impulsive sensitivity coefficients are observed by using the direct differentiation sensitivity analysis method.
Adjoint-based sensitivity analysis for reactor-safety applications
Parks, C.V.
1985-01-01
The application and usefulness of an adjoint-based methodology for performing sensitivity analysis on reactor safety computer codes is investigated. The adjoint-based methodology, referred to as differential sensitivity theory (DST), provides first-order derivatives of the calculated quantities of interest (responses) with respect to the input parameters. The basic theoretical development of DST is presented along with the needed general extensions for consideration of model discontinuities and a variety of useful response definitions. A simple analytic problem is used to highlight the general DST procedures. Finally, DST procedures presented in this work are applied to two highly nonlinear reactor accident analysis codes: (1) FASTGAS, a relatively small code for analysis of loss-of-decay-heat-removal accident in a gas-cooled fast reactor, and (2) an existing code called VENUS-II which is typically employed for analyzing the core disassembly phase of a hypothetical fast reactor accident. The two codes are different both in terms of complexity and in terms of the facets of DST which can be illustrated. Sensitivity results from the adjoint codes ADJGAS and VENUS-ADJ are verified with direct recalculations using perturbed input parameters. The effectiveness of the DST results for parameter ranking, prediction of response changes, and uncertainty analysis are illustrated. The conclusion drawn from this study is that DST is a viable, cost-effective methodology for accurate sensitivity analysis.
Bayesian sensitivity analysis of a nonlinear finite element model
NASA Astrophysics Data System (ADS)
Becker, W.; Oakley, J. E.; Surace, C.; Gili, P.; Rowson, J.; Worden, K.
2012-10-01
A major problem in uncertainty and sensitivity analysis is that the computational cost of propagating probabilistic uncertainty through large nonlinear models can be prohibitive when using conventional methods (such as Monte Carlo methods). A powerful solution to this problem is to use an emulator, which is a mathematical representation of the model built from a small set of model runs at specified points in input space. Such emulators are massively cheaper to run and can be used to mimic the "true" model, with the result that uncertainty analysis and sensitivity analysis can be performed for a greatly reduced computational cost. The work here investigates the use of an emulator known as a Gaussian process (GP), which is an advanced probabilistic form of regression. The GP is particularly suited to uncertainty analysis since it is able to emulate a wide class of models, and accounts for its own emulation uncertainty. Additionally, uncertainty and sensitivity measures can be estimated analytically, given certain assumptions. The GP approach is explained in detail here, and a case study of a finite element model of an airship is used to demonstrate the method. It is concluded that the GP is a very attractive way of performing uncertainty and sensitivity analysis on large models, provided that the dimensionality is not too high.
A Global Sensitivity Analysis Methodology for Multi-physics Applications
Tong, C H; Graziani, F R
2007-02-02
Experiments are conducted to draw inferences about an entire ensemble based on a selected number of observations. This applies to both physical experiments as well as computer experiments, the latter of which are performed by running the simulation models at different input configurations and analyzing the output responses. Computer experiments are instrumental in enabling model analyses such as uncertainty quantification and sensitivity analysis. This report focuses on a global sensitivity analysis methodology that relies on a divide-and-conquer strategy and uses intelligent computer experiments. The objective is to assess qualitatively and/or quantitatively how the variabilities of simulation output responses can be accounted for by input variabilities. We address global sensitivity analysis in three aspects: methodology, sampling/analysis strategies, and an implementation framework. The methodology consists of three major steps: (1) construct credible input ranges; (2) perform a parameter screening study; and (3) perform a quantitative sensitivity analysis on a reduced set of parameters. Once identified, research effort should be directed to the most sensitive parameters to reduce their uncertainty bounds. This process is repeated with tightened uncertainty bounds for the sensitive parameters until the output uncertainties become acceptable. To accommodate the needs of multi-physics application, this methodology should be recursively applied to individual physics modules. The methodology is also distinguished by an efficient technique for computing parameter interactions. Details for each step will be given using simple examples. Numerical results on large scale multi-physics applications will be available in another report. Computational techniques targeted for this methodology have been implemented in a software package called PSUADE.
NASA Astrophysics Data System (ADS)
Zhang, Xianxi; Chen, Qianqian; Sun, Huafei; Pan, Tingting; Hu, Guiqi; Ma, Ruimin; Dou, Jianmin; Li, Dacheng; Pan, Xu
2014-01-01
Alkyne bridged porphyrins have been proved very promising sensitizers for dye-sensitized solar cells (DSSCs) with the highest photo-to-electric conversion efficiencies of 11.9% solely and 12.3% co-sensitized with other sensitizers achieved. Developing better porphyrin sensitizers with wider electronic absorption spectra to further improve the efficiencies of corresponding solar cells is still of great significance for the application of DSSCs. A series of triphenyl zinc porphyrins (ZnTriPP) differing in the nature of a pendant acceptor group and the conjugated bridge between the porphyrin nucleus and the acceptor unit were modeled and their electronic and spectral properties calculated using density functional theory. As compared with each other and the experimental results of the compounds used in DSSCs previously, the molecules with a relatively longer conjugative linker and a strong electron-withdrawing group such as cyanide adjacent to the carboxyl acid group seem to provide wider electronic absorption spectra and higher photo-to-electric conversion efficiencies. The dye candidates ZnTriPPE, ZnTriPPM, ZnTriPPQ, ZnTriPPR and ZnTriPPS designed in the current work were found promising to provide comparable photo-to-electric conversion efficiencies to the record 11.9% of the alkyne bridged porphyrin sensitizer YD2-o-C8 reported previously.
Adjoint sensitivity analysis of plasmonic structures using the FDTD method.
Zhang, Yu; Ahmed, Osman S; Bakr, Mohamed H
2014-05-15
We present an adjoint variable method for estimating the sensitivities of arbitrary responses with respect to the parameters of dispersive discontinuities in nanoplasmonic devices. Our theory is formulated in terms of the electric field components at the vicinity of perturbed discontinuities. The adjoint sensitivities are computed using at most one extra finite-difference time-domain (FDTD) simulation regardless of the number of parameters. Our approach is illustrated through the sensitivity analysis of an add-drop coupler consisting of a square ring resonator between two parallel waveguides. The computed adjoint sensitivities of the scattering parameters are compared with those obtained using the accurate but computationally expensive central finite difference approach. PMID:24978258
Variability-based global sensitivity analysis of circuit response
NASA Astrophysics Data System (ADS)
Opalski, Leszek J.
2014-11-01
The research problem of interest to this paper is: how to determine efficiently and objectively the most and the least influential parameters of a multimodule electronic system - given the system model f and the module parameter variation ranges. The author investigates if existing generic global sensitivity methods are applicable for electronic circuit design, even if they were developed (and successfully applied) in quite distant engineering areas. A photodiode detector analog front-end system response time is used to reveal capability of the selected global sensitivity approaches under study.
Sensitivity analysis in a Lassa fever deterministic mathematical model
NASA Astrophysics Data System (ADS)
Abdullahi, Mohammed Baba; Doko, Umar Chado; Mamuda, Mamman
2015-05-01
Lassa virus that causes the Lassa fever is on the list of potential bio-weapons agents. It was recently imported into Germany, the Netherlands, the United Kingdom and the United States as a consequence of the rapid growth of international traffic. A model with five mutually exclusive compartments related to Lassa fever is presented and the basic reproduction number analyzed. A sensitivity analysis of the deterministic model is performed. This is done in order to determine the relative importance of the model parameters to the disease transmission. The result of the sensitivity analysis shows that the most sensitive parameter is the human immigration, followed by human recovery rate, then person to person contact. This suggests that control strategies should target human immigration, effective drugs for treatment and education to reduced person to person contact.
The Volatility of Data Space: Topology Oriented Sensitivity Analysis
Du, Jing; Ligmann-Zielinska, Arika
2015-01-01
Despite the difference among specific methods, existing Sensitivity Analysis (SA) technologies are all value-based, that is, the uncertainties in the model input and output are quantified as changes of values. This paradigm provides only limited insight into the nature of models and the modeled systems. In addition to the value of data, a potentially richer information about the model lies in the topological difference between pre-model data space and post-model data space. This paper introduces an innovative SA method called Topology Oriented Sensitivity Analysis, which defines sensitivity as the volatility of data space. It extends SA into a deeper level that lies in the topology of data. PMID:26368929
Uncertainty and sensitivity analysis and its applications in OCD measurements
NASA Astrophysics Data System (ADS)
Vagos, Pedro; Hu, Jiangtao; Liu, Zhuan; Rabello, Silvio
2009-03-01
This article describes an Uncertainty & Sensitivity Analysis package, a mathematical tool that can be an effective time-shortcut for optimizing OCD models. By including real system noises in the model, an accurate method for predicting measurements uncertainties is shown. The assessment, in an early stage, of the uncertainties, sensitivities and correlations of the parameters to be measured drives the user in the optimization of the OCD measurement strategy. Real examples are discussed revealing common pitfalls like hidden correlations and simulation results are compared with real measurements. Special emphasis is given to 2 different cases: 1) the optimization of the data set of multi-head metrology tools (NI-OCD, SE-OCD), 2) the optimization of the azimuth measurement angle in SE-OCD. With the uncertainty and sensitivity analysis result, the right data set and measurement mode (NI-OCD, SE-OCD or NI+SE OCD) can be easily selected to achieve the best OCD model performance.
Blurring the Inputs: A Natural Language Approach to Sensitivity Analysis
NASA Technical Reports Server (NTRS)
Kleb, William L.; Thompson, Richard A.; Johnston, Christopher O.
2007-01-01
To document model parameter uncertainties and to automate sensitivity analyses for numerical simulation codes, a natural-language-based method to specify tolerances has been developed. With this new method, uncertainties are expressed in a natural manner, i.e., as one would on an engineering drawing, namely, 5.25 +/- 0.01. This approach is robust and readily adapted to various application domains because it does not rely on parsing the particular structure of input file formats. Instead, tolerances of a standard format are added to existing fields within an input file. As a demonstration of the power of this simple, natural language approach, a Monte Carlo sensitivity analysis is performed for three disparate simulation codes: fluid dynamics (LAURA), radiation (HARA), and ablation (FIAT). Effort required to harness each code for sensitivity analysis was recorded to demonstrate the generality and flexibility of this new approach.
NASA Astrophysics Data System (ADS)
Luo, Jiannan; Lu, Wenxi
2014-06-01
Sobol‧ sensitivity analyses based on different surrogates were performed on a trichloroethylene (TCE)-contaminated aquifer to assess the sensitivity of the design variables of remediation duration, surfactant concentration and injection rates at four wells to remediation efficiency First, the surrogate models of a multi-phase flow simulation model were constructed by applying radial basis function artificial neural network (RBFANN) and Kriging methods, and the two models were then compared. Based on the developed surrogate models, the Sobol‧ method was used to calculate the sensitivity indices of the design variables which affect the remediation efficiency. The coefficient of determination (R2) and the mean square error (MSE) of these two surrogate models demonstrated that both models had acceptable approximation accuracy, furthermore, the approximation accuracy of the Kriging model was slightly better than that of the RBFANN model. Sobol‧ sensitivity analysis results demonstrated that the remediation duration was the most important variable influencing remediation efficiency, followed by rates of injection at wells 1 and 3, while rates of injection at wells 2 and 4 and the surfactant concentration had negligible influence on remediation efficiency. In addition, high-order sensitivity indices were all smaller than 0.01, which indicates that interaction effects of these six factors were practically insignificant. The proposed Sobol‧ sensitivity analysis based on surrogate is an effective tool for calculating sensitivity indices, because it shows the relative contribution of the design variables (individuals and interactions) to the output performance variability with a limited number of runs of a computationally expensive simulation model. The sensitivity analysis results lay a foundation for the optimal groundwater remediation process optimization.
Parameter sensitivity analysis of IL-6 signalling pathways.
Chu, Y; Jayaraman, A; Hahn, J
2007-11-01
Signal transduction pathways generally consist of a large number of individual components and have an even greater number of parameters describing their reaction kinetics. Although the structure of some signalling pathways can be found in the literature, many of the parameters are not well known and they would need to be re-estimated from experimental data for each specific case. However it is not feasible to estimate hundreds of parameters because of the cost of the experiments associated with generating data. Parameter sensitivity analysis can address this situation as it investigates how the system behaviour is changed by variations of parameters and the analysis identifies which parameters play a key role in signal transduction. Only these important parameters need then be re-estimated using data from further experiments. This article presents a detailed parameter sensitivity analysis of the JAK/STAT and MAPK signal transduction pathway that is used for signalling by the cytokine IL-6. As no parameter sensitivity analysis technique is known to work best for all situations, a comparison of the results returned by four techniques is presented: differential analysis, the Morris method, a sampling-based approach and the Fourier amplitude sensitivity test. The recruitment of the transcription factor STAT3 to the dimer of the phosphorylated receptor complex is determined as the most important step by the sensitivity analysis. Additionally, the desphosphorylation of the nuclear STAT3 dimer by PP2 as well as feedback inhibition by SOCS3 are found to play an important role for signal transduction. PMID:18203580
Multicriteria Evaluation and Sensitivity Analysis on Information Security
NASA Astrophysics Data System (ADS)
Syamsuddin, Irfan
2013-05-01
Information security plays a significant role in recent information society. Increasing number and impact of cyber attacks on information assets have resulted the increasing awareness among managers that attack on information is actually attack on organization itself. Unfortunately, particular model for information security evaluation for management levels is still not well defined. In this study, decision analysis based on Ternary Analytic Hierarchy Process (T-AHP) is proposed as a novel model to aid managers who responsible in making strategic evaluation related to information security issues. In addition, sensitivity analysis is applied to extend our analysis by using several "what-if" scenarios in order to measure the consistency of the final evaluation. Finally, we conclude that the final evaluation made by managers has a significant consistency shown by sensitivity analysis results.
The Design and Operation of Ultra-Sensitive and Tunable Radio-Frequency Interferometers
Cui, Yan; Wang, Pingshan
2015-01-01
Dielectric spectroscopy (DS) is an important technique for scientific and technological investigations in various areas. DS sensitivity and operating frequency ranges are critical for many applications, including lab-on-chip development where sample volumes are small with a wide range of dynamic processes to probe. In this work, we present the design and operation considerations of radio-frequency (RF) interferometers that are based on power-dividers (PDs) and quadrature-hybrids (QHs). Such interferometers are proposed to address the sensitivity and frequency tuning challenges of current DS techniques. Verified algorithms together with mathematical models are presented to quantify material properties from scattering parameters for three common transmission line sensing structures, i.e., coplanar waveguides (CPWs), conductor-backed CPWs, and microstrip lines. A high-sensitivity and stable QH-based interferometer is demonstrated by measuring glucose–water solution at a concentration level that is ten times lower than some recent RF sensors while our sample volume is ~1 nL. Composition analysis of ternary mixture solutions are also demonstrated with a PD-based interferometer. Further work is needed to address issues like system automation, model improvement at high frequencies, and interferometer scaling. PMID:26549891
Beyond the GUM: variance-based sensitivity analysis in metrology
NASA Astrophysics Data System (ADS)
Lira, I.
2016-07-01
Variance-based sensitivity analysis is a well established tool for evaluating the contribution of the uncertainties in the inputs to the uncertainty in the output of a general mathematical model. While the literature on this subject is quite extensive, it has not found widespread use in metrological applications. In this article we present a succinct review of the fundamentals of sensitivity analysis, in a form that should be useful to most people familiarized with the Guide to the Expression of Uncertainty in Measurement (GUM). Through two examples, it is shown that in linear measurement models, no new knowledge is gained by using sensitivity analysis that is not already available after the terms in the so-called ‘law of propagation of uncertainties’ have been computed. However, if the model behaves non-linearly in the neighbourhood of the best estimates of the input quantities—and if these quantities are assumed to be statistically independent—sensitivity analysis is definitely advantageous for gaining insight into how they can be ranked according to their importance in establishing the uncertainty of the measurand.
Sensitivity analysis of the Ohio phosphorus risk index
Technology Transfer Automated Retrieval System (TEKTRAN)
The Phosphorus (P) Index is a widely used tool for assessing the vulnerability of agricultural fields to P loss; yet, few of the P Indices developed in the U.S. have been evaluated for their accuracy. Sensitivity analysis is one approach that can be used prior to calibration and field-scale testing ...
Omitted Variable Sensitivity Analysis with the Annotated Love Plot
ERIC Educational Resources Information Center
Hansen, Ben B.; Fredrickson, Mark M.
2014-01-01
The goal of this research is to make sensitivity analysis accessible not only to empirical researchers but also to the various stakeholders for whom educational evaluations are conducted. To do this it derives anchors for the omitted variable (OV)-program participation association intrinsically, using the Love plot to present a wide range of…
Cost effective, weight sensitive design for military airborne electronic systems
Peck, W.M.
1996-12-31
Thermal management of Military airborne electronic systems is governed by many trade-offs. While the trade-offs may change depending on the customer and system requirements, minimizing weight is usually the primary engineering concern because it saves aircraft fuel. Fuel savings provides increased range and time aloft for the aircraft. The most common approach to achieving meaningful reductions in equipment weight is to reduce system volume. Reduced volume is achieved by increasing electronic packaging density which is accomplished by incorporating new materials, processes, and technologies into the system design. The following four considerations are currently under study in the development of an 8 KW high altitude Military electronic system in order to reduce system volume: (1) identifying design parameters and performing trade-off studies between the use of liquid vs. forced air for system cooling; (2) modeling the total system thermal resistance path to identify possible areas for reducing component temperature rise in order to provide enhanced system reliability; (3) substituting commercial plastic integrated circuits (ICs) for Mil ceramic components to reduce material cost while still meeting system requirements; and (4) using TC1050 material technology in: Militarizing Commercial Off The Shelf (COTS) Circuit Card Assemblies (CCAs), developing high conductivity ceramic packaging for Multi Chip Modules (MCMs), and for developing low coefficient of thermal expansion (CTE) composites for use at both the chip and chassis level. Current results from these studies have yielded a design that has a 2:1 reduction in system volume and a weight reduction of 480 lbs from a currently fielded system.
Global sensitivity analysis in control-augmented structural synthesis
NASA Technical Reports Server (NTRS)
Bloebaum, Christina L.
1989-01-01
In this paper, an integrated approach to structural/control design is proposed in which variables in both the passive (structural) and active (control) disciplines of an optimization process are changed simultaneously. The global sensitivity equation (GSE) method of Sobieszczanski-Sobieski (1988) is used to obtain the behavior sensitivity derivatives necessary for the linear approximations used in the parallel multidisciplinary synthesis problem. The GSE allows for the decoupling of large systems into smaller subsystems and thus makes it possible to determine the local sensitivities of each subsystem's outputs to its inputs and parameters. The advantages in using the GSE method are demonstrated using a finite-element representation of a truss structure equipped with active lateral displacement controllers, which is undergoing forced vibration.
Adjoint-based sensitivity analysis for reactor safety applications
Parks, C.V.
1986-08-01
The application and usefulness of an adjoint-based methodology for performing sensitivity analysis on reactor safety computer codes is investigated. The adjoint-based methodology, referred to as differential sensitivity theory (DST), provides first-order derivatives of the calculated quantities of interest (responses) with respect to the input parameters. The basic theoretical development of DST is presented along with the needed general extensions for consideration of model discontinuities and a variety of useful response definitions. A simple analytic problem is used to highlight the general DST procedures. finally, DST procedures presented in this work are applied to two highly nonlinear reactor accident analysis codes: (1) FASTGAS, a relatively small code for analysis of a loss-of-decay-heat-removal accident in a gas-cooled fast reactor, and (2) an existing code called VENUS-II which has been employed for analyzing the core disassembly phase of a hypothetical fast reactor accident. The two codes are different both in terms of complexity and in terms of the facets of DST which can be illustrated. Sensitivity results from the adjoint codes ADJGAS and VENUS-ADJ are verified with direct recalcualtions using perturbed input parameters. The effectiveness of the DST results for parameter ranking, prediction of response changes, and uncertainty analysis are illustrated. The conclusion drawn from this study is that DST is a viable, cost-effective methodology for accurate sensitivity analysis. In addition, a useful sensitivity tool for use in the fast reactor safety area has been developed in VENUS-ADJ. Future work needs to concentrate on combining the accurate first-order derivatives/results from DST with existing methods (based solely on direct recalculations) for higher-order response surfaces.
Integrative "omic" analysis for tamoxifen sensitivity through cell based models.
Weng, Liming; Ziliak, Dana; Lacroix, Bonnie; Geeleher, Paul; Huang, R Stephanie
2014-01-01
It has long been observed that tamoxifen sensitivity varies among breast cancer patients. Further, ethnic differences of tamoxifen therapy between Caucasian and African American have also been reported. Since most studies have been focused on Caucasian people, we sought to comprehensively evaluate genetic variants related to tamoxifen therapy in African-derived samples. An integrative "omic" approach developed by our group was used to investigate relationships among endoxifen (an active metabolite of tamoxifen) sensitivity, SNP genotype, mRNA and microRNA expressions in 58 HapMap YRI lymphoblastoid cell lines. We identified 50 SNPs that associate with cellular sensitivity to endoxifen through their effects on 34 genes and 30 microRNA expression. Some of these findings are shared in both Caucasian and African samples, while others are unique in the African samples. Among gene/microRNA that were identified in both ethnic groups, the expression of TRAF1 is also correlated with tamoxifen sensitivity in a collection of 44 breast cancer cell lines. Further, knock-down TRAF1 and over-expression of hsa-let-7i confirmed the roles of hsa-let-7i and TRAF1 in increasing tamoxifen sensitivity in the ZR-75-1 breast cancer cell line. Our integrative omic analysis facilitated the discovery of pharmacogenomic biomarkers that potentially affect tamoxifen sensitivity. PMID:24699530
Design, analysis and test verification of advanced encapsulation systems
NASA Technical Reports Server (NTRS)
Garcia, A., III
1982-01-01
An analytical methodology for advanced encapsulation designs was developed. From these methods design sensitivities are established for the development of photovoltaic module criteria and the definition of needed research tasks. Analytical models were developed to perform optical, thermal, electrical and analyses on candidate encapsulation systems. From these analyses several candidate systems were selected for qualification testing. Additionally, test specimens of various types are constructed and tested to determine the validity of the analysis methodology developed. Identified deficiencies and/or discrepancies between analytical models and relevant test data are corrected. Prediction capability of analytical models is improved. Encapsulation engineering generalities, principles, and design aids for photovoltaic module designers is generated.
LSENS - GENERAL CHEMICAL KINETICS AND SENSITIVITY ANALYSIS CODE
NASA Technical Reports Server (NTRS)
Bittker, D. A.
1994-01-01
which provides the relationships between the predictions of a kinetics model and the input parameters of the problem. LSENS provides for efficient and accurate chemical kinetics computations and includes sensitivity analysis for a variety of problems, including nonisothermal conditions. LSENS replaces the previous NASA general chemical kinetics codes GCKP and GCKP84. LSENS is designed for flexibility, convenience and computational efficiency. A variety of chemical reaction models can be considered. The models include static system, steady one-dimensional inviscid flow, reaction behind an incident shock wave including boundary layer correction, and the perfectly stirred (highly backmixed) reactor. In addition, computations of equilibrium properties can be performed for the following assigned states, enthalpy and pressure, temperature and pressure, internal energy and volume, and temperature and volume. For static problems LSENS computes sensitivity coefficients with respect to the initial values of the dependent variables and/or the three rates coefficient parameters of each chemical reaction. To integrate the ODEs describing chemical kinetics problems, LSENS uses the packaged code LSODE, the Livermore Solver for Ordinary Differential Equations, because it has been shown to be the most efficient and accurate code for solving such problems. The sensitivity analysis computations use the decoupled direct method, as implemented by Dunker and modified by Radhakrishnan. This method has shown greater efficiency and stability with equal or better accuracy than other methods of sensitivity analysis. LSENS is written in FORTRAN 77 with the exception of the NAMELIST extensions used for input. While this makes the code fairly machine independent, execution times on IBM PC compatibles would be unacceptable to most users. LSENS has been successfully implemented on a Sun4 running SunOS and a DEC VAX running VMS. With minor modifications, it should also be easily implemented on other
Li, Peiyue; Wu, Jianhua; Qian, Hui; Chen, Jie
2013-03-01
This is the second part of the study on sensitivity analysis of the technique for order preference by similarity to ideal solution (TOPSIS) method in water quality assessment. In the present study, the sensitivity of the TOPSIS method to the index input data was investigated. The sensitivity was first theoretically analyzed under two major assumptions. One assumption was that one index or more of the samples were perturbed with the same ratio while other indices kept unchanged. The other one was that all indices of a given sample were changed simultaneously with the same ratio, while the indices of other samples were unchanged. Furthermore, a case study under assumption 2 was also carried out in this paper. When the same indices of different water samples are changed simultaneously with the same variation ratio, the final water quality assessment results will not be influenced at all. When the input data of all indices of a given sample are perturbed with the same variation ratio, the assessment values of all samples will be influenced theoretically. However, the case study shows that only the perturbed sample is sensitive to the variation, and a simple linear equation representing the relation between the closeness coefficient (CC) values of the perturbed sample and variation ratios can be derived under the assumption 2. This linear equation can be used for determining the sample orders under various variation ratios. PMID:22832843
Automated Simulation For Analysis And Design
NASA Technical Reports Server (NTRS)
Cantwell, E.; Shenk, Tim; Robinson, Peter; Upadhye, R.
1992-01-01
Design Assistant Workstation (DAWN) software being developed to facilitate simulation of qualitative and quantitative aspects of behavior of life-support system in spacecraft, chemical-processing plant, heating and cooling system of large building, or any of variety of systems including interacting process streams and processes. Used to analyze alternative design scenarios or specific designs of such systems. Expert system will automate part of design analysis: reason independently by simulating design scenarios and return to designer with overall evaluations and recommendations.
Nonlinear mathematical modeling and sensitivity analysis of hydraulic drive unit
NASA Astrophysics Data System (ADS)
Kong, Xiangdong; Yu, Bin; Quan, Lingxiao; Ba, Kaixian; Wu, Liujie
2015-09-01
The previous sensitivity analysis researches are not accurate enough and also have the limited reference value, because those mathematical models are relatively simple and the change of the load and the initial displacement changes of the piston are ignored, even experiment verification is not conducted. Therefore, in view of deficiencies above, a nonlinear mathematical model is established in this paper, including dynamic characteristics of servo valve, nonlinear characteristics of pressure-flow, initial displacement of servo cylinder piston and friction nonlinearity. The transfer function block diagram is built for the hydraulic drive unit closed loop position control, as well as the state equations. Through deriving the time-varying coefficient items matrix and time-varying free items matrix of sensitivity equations respectively, the expression of sensitivity equations based on the nonlinear mathematical model are obtained. According to structure parameters of hydraulic drive unit, working parameters, fluid transmission characteristics and measured friction-velocity curves, the simulation analysis of hydraulic drive unit is completed on the MATLAB/Simulink simulation platform with the displacement step 2 mm, 5 mm and 10 mm, respectively. The simulation results indicate that the developed nonlinear mathematical model is sufficient by comparing the characteristic curves of experimental step response and simulation step response under different constant load. Then, the sensitivity function time-history curves of seventeen parameters are obtained, basing on each state vector time-history curve of step response characteristic. The maximum value of displacement variation percentage and the sum of displacement variation absolute values in the sampling time are both taken as sensitivity indexes. The sensitivity indexes values above are calculated and shown visually in histograms under different working conditions, and change rules are analyzed. Then the sensitivity
Probabilistic approach to identify sensitive parameter distributions in multimedia pathway analysis.
Kamboj, S.; Gnanapragasam, E.; LePoire, D.; Biwer, B. M.; Cheng, J.; Arnish, J.; Yu, C.; Chen, S. Y.; Mo, T.; Abu-Eid, R.; Thaggard, M.; Environmental Assessment; NRC
2002-01-01
Sensitive parameter distributions were identified with the use of probabilistic analysis in the RESRAD computer code. RESRAD is a multimedia pathway analysis code designed to evaluate radiological exposures resulting from radiological contamination in soil. The dose distribution was obtained by using a set of default parameter distribution/values. Most of the variations in the output dose distribution could be attributed to uncertainty in a small set of input parameters that could be considered as sensitive parameter distributions. The identification of the sensitive parameters is a first step in the prioritization of future research and information gathering. When site-specific parameter distribution/values are available for an actual site, the same process should be used with these site-specific data. Regression analysis used to identify sensitive parameters indicated that the dominant pathways depended on the radionuclide and source configurations. However, two parameter distributions were sensitive for many radionuclides: the external shielding factor when external exposure was the dominant pathway and the plant transfer factor when plant ingestion was the dominant pathway. No single correlation or regression coefficient can be used alone to identify sensitive parameters in all the cases. The coefficients are useful guides, but they have to be used in conjunction with other aids, such as scatter plots, and should undergo further analysis.
Double Precision Differential/Algebraic Sensitivity Analysis Code
1995-06-02
DDASAC solves nonlinear initial-value problems involving stiff implicit systems of ordinary differential and algebraic equations. Purely algebraic nonlinear systems can also be solved, given an initial guess within the region of attraction of a solution. Options include automatic reconciliation of inconsistent initial states and derivatives, automatic initial step selection, direct concurrent parametric sensitivity analysis, and stopping at a prescribed value of any user-defined functional of the current solution vector. Local error control (in the max-normmore » or the 2-norm) is provided for the state vector and can include the sensitivities on request.« less
Superconducting Accelerating Cavity Pressure Sensitivity Analysis and Stiffening
Rodnizki, J; Ben Aliz, Y; Grin, A; Horvitz, Z; Perry, A; Weissman, L; Davis, G Kirk; Delayen, Jean R.
2014-12-01
The Soreq Applied Research Accelerator Facility (SARAF) design is based on a 40 MeV 5 mA light ions superconducting RF linac. Phase-I of SARAF delivers up to 2 mA CW proton beams in an energy range of 1.5 - 4.0 MeV. The maximum beam power that we have reached is 5.7 kW. Today, the main limiting factor to reach higher ion energy and beam power is related to the HWR sensitivity to the liquid helium coolant pressure fluctuations. The HWR sensitivity to helium pressure is about 60 Hz/mbar. The cavities had been designed, a decade ago, to be soft in order to enable tuning of their novel shape. However, the cavities turned out to be too soft. In this work we found that increasing the rigidity of the cavities in the vicinity of the external drift tubes may reduce the cavity sensitivity by a factor of three. A preliminary design to increase the cavity rigidity is presented.
Design Sensitivities of the Superconducting Parallel-Bar Cavity
De Silva, Subashini U.; Delayen, Jean D.
2010-09-01
The superconducting parallel-bar cavity has properties that makes it attractive as a deflecting or crabbing rf structure. For example it is under consideration as an rf separator for the Jefferson Lab 12 GeV upgrade and as a crabbing structure for a possible LHC luminosity upgrade. In order to maintain the purity of the deflecting mode and avoid mixing with the near accelerating mode caused by geometrical imperfection, a minimum frequency separation is needed which depends on the expected deviations from perfect symmetry. We have done an extensive analysis of the impact of several geometrical imperfections on the properties of the parallel-bar cavities and the effects on the beam, and present the results in this paper.
Pressure-Sensitive Paints Advance Rotorcraft Design Testing
NASA Technical Reports Server (NTRS)
2013-01-01
The rotors of certain helicopters can spin at speeds as high as 500 revolutions per minute. As the blades slice through the air, they flex, moving into the wind and back out, experiencing pressure changes on the order of thousands of times a second and even higher. All of this makes acquiring a true understanding of rotorcraft aerodynamics a difficult task. A traditional means of acquiring aerodynamic data is to conduct wind tunnel tests using a vehicle model outfitted with pressure taps and other sensors. These sensors add significant costs to wind tunnel testing while only providing measurements at discrete locations on the model's surface. In addition, standard sensor solutions do not work for pulling data from a rotor in motion. "Typical static pressure instrumentation can't handle that," explains Neal Watkins, electronics engineer in Langley Research Center s Advanced Sensing and Optical Measurement Branch. "There are dynamic pressure taps, but your costs go up by a factor of five to ten if you use those. In addition, recovery of the pressure tap readings is accomplished through slip rings, which allow only a limited amount of sensors and can require significant maintenance throughout a typical rotor test." One alternative to sensor-based wind tunnel testing is pressure sensitive paint (PSP). A coating of a specialized paint containing luminescent material is applied to the model. When exposed to an LED or laser light source, the material glows. The glowing material tends to be reactive to oxygen, explains Watkins, which causes the glow to diminish. The more oxygen that is present (or the more air present, since oxygen exists in a fixed proportion in air), the less the painted surface glows. Imaged with a camera, the areas experiencing greater air pressure show up darker than areas of less pressure. "The paint allows for a global pressure map as opposed to specific points," says Watkins. With PSP, each pixel recorded by the camera becomes an optical pressure
System analysis in rotorcraft design: The past decade
NASA Technical Reports Server (NTRS)
Galloway, Thomas L.
1988-01-01
Rapid advances in the technology of electronic digital computers and the need for an integrated synthesis approach in developing future rotorcraft programs has led to increased emphasis on system analysis techniques in rotorcraft design. The task in systems analysis is to deal with complex, interdependent, and conflicting requirements in a structured manner so rational and objective decisions can be made. Whether the results are wisdom or rubbish depends upon the validity and sometimes more importantly, the consistency of the inputs, the correctness of the analysis, and a sensible choice of measures of effectiveness to draw conclusions. In rotorcraft design this means combining design requirements, technology assessment, sensitivity analysis and reviews techniques currently in use by NASA and Army organizations in developing research programs and vehicle specifications for rotorcraft. These procedures span simple graphical approaches to comprehensive analysis on large mainframe computers. Examples of recent applications to military and civil missions are highlighted.
Sin, Gürkan; Gernaey, Krist V; Neumann, Marc B; van Loosdrecht, Mark C M; Gujer, Willi
2011-01-01
This study demonstrates the usefulness of global sensitivity analysis in wastewater treatment plant (WWTP) design to prioritize sources of uncertainty and quantify their impact on performance criteria. The study, which is performed with the Benchmark Simulation Model no. 1 plant design, complements a previous paper on input uncertainty characterisation and propagation (Sin et al., 2009). A sampling-based sensitivity analysis is conducted to compute standardized regression coefficients. It was found that this method is able to decompose satisfactorily the variance of plant performance criteria (with R(2) > 0.9) for effluent concentrations, sludge production and energy demand. This high extent of linearity means that the plant performance criteria can be described as linear functions of the model inputs under the defined plant conditions. In effect, the system of coupled ordinary differential equations can be replaced by multivariate linear models, which can be used as surrogate models. The importance ranking based on the sensitivity measures demonstrates that the most influential factors involve ash content and influent inert particulate COD among others, largely responsible for the uncertainty in predicting sludge production and effluent ammonium concentration. While these results were in agreement with process knowledge, the added value is that the global sensitivity methods can quantify the contribution of the variance of significant parameters, e.g., ash content explains 70% of the variance in sludge production. Further the importance of formulating appropriate sensitivity analysis scenarios that match the purpose of the model application needs to be highlighted. Overall, the global sensitivity analysis proved a powerful tool for explaining and quantifying uncertainties as well as providing insight into devising useful ways for reducing uncertainties in the plant performance. This information can help engineers design robust WWTP plants. PMID:20828785
Shape sensitivity analysis of flutter response of a laminated wing
NASA Technical Reports Server (NTRS)
Bergen, Fred D.; Kapania, Rakesh K.
1988-01-01
A method is presented for calculating the shape sensitivity of a wing aeroelastic response with respect to changes in geometric shape. Yates' modified strip method is used in conjunction with Giles' equivalent plate analysis to predict the flutter speed, frequency, and reduced frequency of the wing. Three methods are used to calculate the sensitivity of the eigenvalue. The first method is purely a finite difference calculation of the eigenvalue derivative directly from the solution of the flutter problem corresponding to the two different values of the shape parameters. The second method uses an analytic expression for the eigenvalue sensitivities of a general complex matrix, where the derivatives of the aerodynamic, mass, and stiffness matrices are computed using a finite difference approximation. The third method also uses an analytic expression for the eigenvalue sensitivities, but the aerodynamic matrix is computed analytically. All three methods are found to be in good agreement with each other. The sensitivities of the eigenvalues were used to predict the flutter speed, frequency, and reduced frequency. These approximations were found to be in good agreement with those obtained using a complete reanalysis.
Graphical methods for the sensitivity analysis in discriminant analysis
Kim, Youngil; Anderson-Cook, Christine M.; Dae-Heung, Jang
2015-09-30
Similar to regression, many measures to detect influential data points in discriminant analysis have been developed. Many follow similar principles as the diagnostic measures used in linear regression in the context of discriminant analysis. Here we focus on the impact on the predicted classification posterior probability when a data point is omitted. The new method is intuitive and easily interpretative compared to existing methods. We also propose a graphical display to show the individual movement of the posterior probability of other data points when a specific data point is omitted. This enables the summaries to capture the overall pattern of the change.
Graphical methods for the sensitivity analysis in discriminant analysis
Kim, Youngil; Anderson-Cook, Christine M.; Dae-Heung, Jang
2015-09-30
Similar to regression, many measures to detect influential data points in discriminant analysis have been developed. Many follow similar principles as the diagnostic measures used in linear regression in the context of discriminant analysis. Here we focus on the impact on the predicted classification posterior probability when a data point is omitted. The new method is intuitive and easily interpretative compared to existing methods. We also propose a graphical display to show the individual movement of the posterior probability of other data points when a specific data point is omitted. This enables the summaries to capture the overall pattern ofmore » the change.« less
Planar Inlet Design and Analysis Process (PINDAP)
NASA Technical Reports Server (NTRS)
Slater, John W.; Gruber, Christopher R.
2005-01-01
The Planar Inlet Design and Analysis Process (PINDAP) is a collection of software tools that allow the efficient aerodynamic design and analysis of planar (two-dimensional and axisymmetric) inlets. The aerodynamic analysis is performed using the Wind-US computational fluid dynamics (CFD) program. A major element in PINDAP is a Fortran 90 code named PINDAP that can establish the parametric design of the inlet and efficiently model the geometry and generate the grid for CFD analysis with design changes to those parameters. The use of PINDAP is demonstrated for subsonic, supersonic, and hypersonic inlets.
Sensitivity Analysis and Optimal Control of Anthroponotic Cutaneous Leishmania
Zamir, Muhammad; Zaman, Gul; Alshomrani, Ali Saleh
2016-01-01
This paper is focused on the transmission dynamics and optimal control of Anthroponotic Cutaneous Leishmania. The threshold condition R0 for initial transmission of infection is obtained by next generation method. Biological sense of the threshold condition is investigated and discussed in detail. The sensitivity analysis of the reproduction number is presented and the most sensitive parameters are high lighted. On the basis of sensitivity analysis, some control strategies are introduced in the model. These strategies positively reduce the effect of the parameters with high sensitivity indices, on the initial transmission. Finally, an optimal control strategy is presented by taking into account the cost associated with control strategies. It is also shown that an optimal control exists for the proposed control problem. The goal of optimal control problem is to minimize, the cost associated with control strategies and the chances of infectious humans, exposed humans and vector population to become infected. Numerical simulations are carried out with the help of Runge-Kutta fourth order procedure. PMID:27505634
Post-Optimality Analysis In Aerospace Vehicle Design
NASA Technical Reports Server (NTRS)
Braun, Robert D.; Kroo, Ilan M.; Gage, Peter J.
1993-01-01
This analysis pertains to the applicability of optimal sensitivity information to aerospace vehicle design. An optimal sensitivity (or post-optimality) analysis refers to computations performed once the initial optimization problem is solved. These computations may be used to characterize the design space about the present solution and infer changes in this solution as a result of constraint or parameter variations, without reoptimizing the entire system. The present analysis demonstrates that post-optimality information generated through first-order computations can be used to accurately predict the effect of constraint and parameter perturbations on the optimal solution. This assessment is based on the solution of an aircraft design problem in which the post-optimality estimates are shown to be within a few percent of the true solution over the practical range of constraint and parameter variations. Through solution of a reusable, single-stage-to-orbit, launch vehicle design problem, this optimal sensitivity information is also shown to improve the efficiency of the design process, For a hierarchically decomposed problem, this computational efficiency is realized by estimating the main-problem objective gradient through optimal sep&ivity calculations, By reducing the need for finite differentiation of a re-optimized subproblem, a significant decrease in the number of objective function evaluations required to reach the optimal solution is obtained.
NASA Astrophysics Data System (ADS)
Li, J.; Duan, Q. Y.; Gong, W.; Ye, A.; Dai, Y.; Miao, C.; Di, Z.; Tong, C.; Sun, Y.
2013-08-01
Proper specification of model parameters is critical to the performance of land surface models (LSMs). Due to high dimensionality and parameter interaction, estimating parameters of an LSM is a challenging task. Sensitivity analysis (SA) is a tool that can screen out the most influential parameters on model outputs. In this study, we conducted parameter screening for six output fluxes for the Common Land Model: sensible heat, latent heat, upward longwave radiation, net radiation, soil temperature and soil moisture. A total of 40 adjustable parameters were considered. Five qualitative SA methods, including local, sum-of-trees, multivariate adaptive regression splines, delta test and Morris methods, were compared. The proper sampling design and sufficient sample size necessary to effectively screen out the sensitive parameters were examined. We found that there are 2-8 sensitive parameters, depending on the output type, and about 400 samples are adequate to reliably identify the most sensitive parameters. We also employed a revised Sobol' sensitivity method to quantify the importance of all parameters. The total effects of the parameters were used to assess the contribution of each parameter to the total variances of the model outputs. The results confirmed that global SA methods can generally identify the most sensitive parameters effectively, while local SA methods result in type I errors (i.e., sensitive parameters labeled as insensitive) or type II errors (i.e., insensitive parameters labeled as sensitive). Finally, we evaluated and confirmed the screening results for their consistency with the physical interpretation of the model parameters.
NASA Astrophysics Data System (ADS)
Li, J. D.; Duan, Q. Y.; Gong, W.; Ye, A. Z.; Dai, Y. J.; Miao, C. Y.; Di, Z. H.; Tong, C.; Sun, Y. W.
2013-02-01
Proper specification of model parameters is critical to the performance of land surface models (LSMs). Due to high dimensionality and parameter interaction, estimating parameters of a LSM is a challenging task. Sensitivity analysis (SA) is a tool that can screen out the most influential parameters on model outputs. In this study, we conducted parameter screening for six output fluxes for the Common Land Model: sensible heat, latent heat, upward longwave radiation, net radiation, soil temperature and soil moisture. A total of 40 adjustable parameters were considered. Five qualitative SA methods, including local, sum-of-trees, multivariate adaptive regression splines, delta test and Morris methods, were compared. The proper sampling design and sufficient sample size necessary to effectively screen out the sensitive parameters were examined. We found that there are 2-8 sensitive parameters, depending on the output type, and about 400 samples are adequate to reliably identify the most sensitive parameters. We also employed a revised Sobol' sensitivity method to quantify the importance of all parameters. The total effects of the parameters were used to assess the contribution of each parameter to the total variances of the model outputs. The results confirmed that global SA methods can generally identify the most sensitive parameters effectively, while local SA methods result in type I errors (i.e. sensitive parameters labeled as insensitive) or type II errors (i.e. insensitive parameters labeled as sensitive). Finally, we evaluated and confirmed the screening results for their consistence with the physical interpretation of the model parameters.
Objective analysis of the ARM IOP data: method and sensitivity
Cedarwall, R; Lin, J L; Xie, S C; Yio, J J; Zhang, M H
1999-04-01
Motivated by the need of to obtain accurate objective analysis of field experimental data to force physical parameterizations in numerical models, this paper -first reviews the existing objective analysis methods and interpolation schemes that are used to derive atmospheric wind divergence, vertical velocity, and advective tendencies. Advantages and disadvantages of each method are discussed. It is shown that considerable uncertainties in the analyzed products can result from the use of different analysis schemes and even more from different implementations of a particular scheme. The paper then describes a hybrid approach to combine the strengths of the regular grid method and the line-integral method, together with a variational constraining procedure for the analysis of field experimental data. In addition to the use of upper air data, measurements at the surface and at the top-of-the-atmosphere are used to constrain the upper air analysis to conserve column-integrated mass, water, energy, and momentum. Analyses are shown for measurements taken in the Atmospheric Radiation Measurement Programs (ARM) July 1995 Intensive Observational Period (IOP). Sensitivity experiments are carried out to test the robustness of the analyzed data and to reveal the uncertainties in the analysis. It is shown that the variational constraining process significantly reduces the sensitivity of the final data products.
Sensitivity analysis of transport modeling in a fractured gneiss aquifer
NASA Astrophysics Data System (ADS)
Abdelaziz, Ramadan; Merkel, Broder J.
2015-03-01
Modeling solute transport in fractured aquifers is still challenging for scientists and engineers. Tracer tests are a powerful tool to investigate fractured aquifers with complex geometry and variable heterogeneity. This research focuses on obtaining hydraulic and transport parameters from an experimental site with several wells. At the site, a tracer test with NaCl was performed under natural gradient conditions. Observed concentrations of tracer test were used to calibrate a conservative solute transport model by inverse modeling based on UCODE2013, MODFLOW, and MT3DMS. In addition, several statistics are employed for sensitivity analysis. Sensitivity analysis results indicate that hydraulic conductivity and immobile porosity play important role in the late arrive for breakthrough curve. The results proved that the calibrated model fits well with the observed data set.
Least Squares Shadowing sensitivity analysis of chaotic limit cycle oscillations
NASA Astrophysics Data System (ADS)
Wang, Qiqi; Hu, Rui; Blonigan, Patrick
2014-06-01
The adjoint method, among other sensitivity analysis methods, can fail in chaotic dynamical systems. The result from these methods can be too large, often by orders of magnitude, when the result is the derivative of a long time averaged quantity. This failure is known to be caused by ill-conditioned initial value problems. This paper overcomes this failure by replacing the initial value problem with the well-conditioned "least squares shadowing (LSS) problem". The LSS problem is then linearized in our sensitivity analysis algorithm, which computes a derivative that converges to the derivative of the infinitely long time average. We demonstrate our algorithm in several dynamical systems exhibiting both periodic and chaotic oscillations.
Control of a mechanical aeration process via topological sensitivity analysis
NASA Astrophysics Data System (ADS)
Abdelwahed, M.; Hassine, M.; Masmoudi, M.
2009-06-01
The topological sensitivity analysis method gives the variation of a criterion with respect to the creation of a small hole in the domain. In this paper, we use this method to control the mechanical aeration process in eutrophic lakes. A simplified model based on incompressible Navier-Stokes equations is used, only considering the liquid phase, which is the dominant one. The injected air is taken into account through local boundary conditions for the velocity, on the injector holes. A 3D numerical simulation of the aeration effects is proposed using a mixed finite element method. In order to generate the best motion in the fluid for aeration purposes, the optimization of the injector location is considered. The main idea is to carry out topological sensitivity analysis with respect to the insertion of an injector. Finally, a topological optimization algorithm is proposed and some numerical results, showing the efficiency of our approach, are presented.
Sensitivity analysis techniques for models of human behavior.
Bier, Asmeret Brooke
2010-09-01
Human and social modeling has emerged as an important research area at Sandia National Laboratories due to its potential to improve national defense-related decision-making in the presence of uncertainty. To learn about which sensitivity analysis techniques are most suitable for models of human behavior, different promising methods were applied to an example model, tested, and compared. The example model simulates cognitive, behavioral, and social processes and interactions, and involves substantial nonlinearity, uncertainty, and variability. Results showed that some sensitivity analysis methods create similar results, and can thus be considered redundant. However, other methods, such as global methods that consider interactions between inputs, can generate insight not gained from traditional methods.
On the design of wave digital filters with low sensitivity properties.
NASA Technical Reports Server (NTRS)
Renner, K.; Gupta, S. C.
1973-01-01
The wave digital filter patterned after doubly terminated maximum available power (MAP) networks by means of the Richard's transformation has been shown to have low-coefficient-sensitivity properties. This paper examines the exact nature of the relationship between the wave-digital-filter structure and the MAP networks and how the sensitivity property arises, which permits implementation of the digital structure with a lower coefficient word length than that possible with the conventional structures. The proper design procedure is specified and the nature of the unique complementary outputs is discussed. Finally, an example is considered which illustrates the design, the conversion techniques, and the low sensitivity properties.
Initial Multidisciplinary Design and Analysis Framework
NASA Technical Reports Server (NTRS)
Ozoroski, L. P.; Geiselhart, K. A.; Padula, S. L.; Li, W.; Olson, E. D.; Campbell, R. L.; Shields, E. W.; Berton, J. J.; Gray, J. S.; Jones, S. M.; Naiman, C. G.; Seidel, J. A.; Moore, K. T.; Naylor, B. A.; Townsend, S.
2010-01-01
Within the Supersonics (SUP) Project of the Fundamental Aeronautics Program (FAP), an initial multidisciplinary design & analysis framework has been developed. A set of low- and intermediate-fidelity discipline design and analysis codes were integrated within a multidisciplinary design and analysis framework and demonstrated on two challenging test cases. The first test case demonstrates an initial capability to design for low boom and performance. The second test case demonstrates rapid assessment of a well-characterized design. The current system has been shown to greatly increase the design and analysis speed and capability, and many future areas for development were identified. This work has established a state-of-the-art capability for immediate use by supersonic concept designers and systems analysts at NASA, while also providing a strong base to build upon for future releases as more multifidelity capabilities are developed and integrated.
Ion thruster design and analysis
NASA Technical Reports Server (NTRS)
Kami, S.; Schnelker, D. E.
1976-01-01
Questions concerning the mechanical design of a thruster are considered, taking into account differences in the design of an 8-cm and a 30-cm model. The components of a thruster include the thruster shell assembly, the ion extraction electrode assembly, the cathode isolator vaporizer assembly, the neutralizer isolator vaporizer assembly, ground screen and mask, and the main isolator vaporizer assembly. Attention is given to the materials used in thruster fabrication, the advanced manufacturing methods used, details of thruster performance, an evaluation of thruster life, structural and thermal design considerations, and questions of reliability and quality assurance.
Global sensitivity analysis for DSMC simulations of hypersonic shocks
NASA Astrophysics Data System (ADS)
Strand, James S.; Goldstein, David B.
2013-08-01
Two global, Monte Carlo based sensitivity analyses were performed to determine which reaction rates most affect the results of Direct Simulation Monte Carlo (DSMC) simulations for a hypersonic shock in five-species air. The DSMC code was written and optimized with shock tube simulations in mind, and includes modifications to allow for the efficient simulation of a 1D hypersonic shock. The TCE model is used to convert Arrhenius-form reaction rate constants into reaction cross-sections, after modification to allow accurate modeling of reactions with arbitrarily large rates relative to the VHS collision rate. The square of the Pearson correlation coefficient was used as the measure for sensitivity in the first of the analyses, and the mutual information was used as the measure in the second. The quantity of interest (QoI) for these analyses was the NO density profile across a 1D shock at ˜8000 m/s (M∞ ≈ 23). This vector QoI was broken into a set of scalar QoIs, each representing the density of NO at a specific point downstream of the shock, and sensitivities were calculated for each scalar QoI based on both measures of sensitivity. Profiles of sensitivity vs. location downstream of the shock were then integrated to determine an overall sensitivity for each reaction. A weighting function was used in the integration in order to emphasize sensitivities in the region of greatest thermal and chemical non-equilibrium. Both sensitivity analysis methods agree on the six reactions which most strongly affect the density of NO. These six reactions are the N2 dissociation reaction N2 + N ⇄ 3N, the O2 dissociation reaction O2 + O ⇄ 3O, the NO dissociation reactions NO + N ⇄ 2N + O and NO + O ⇄ N + 2O, and the exchange reactions N2 + O ⇄ NO + N and NO + O ⇄ O2 + N. This analysis lays the groundwork for the application of Bayesian statistical methods for the calibration of parameters relevant to modeling a hypersonic shock layer with the DSMC method.
Recent advances in steady compressible aerodynamic sensitivity analysis
NASA Technical Reports Server (NTRS)
Taylor, Arthur C., III; Newman, Perry A.; Hou, Gene J.-W.; Jones, Henry E.
1992-01-01
Sensitivity analysis methods are classified as belonging to either of the two broad categories: the discrete (quasi-analytical) approach and the continuous approach. The two approaches differ by the order in which discretization and differentiation of the governing equations and boundary conditions is undertaken. The discussion focuses on the discrete approach. Basic equations are presented, and the major difficulties are reviewed in some detail, as are the proposed solutions. Recent research activity concerned with the continuous approach is also discussed.
Sensitivity Analysis of Inverse Methods in Eddy Current Pit Characterization
NASA Astrophysics Data System (ADS)
Aldrin, John C.; Sabbagh, Harold A.; Murphy, R. Kim; Sabbagh, Elias H.; Knopp, Jeremy S.
2010-02-01
A sensitivity analysis was performed for a pit characterization problem to quantify the impact of potential sources for variation on the performance of inverse methods. Certain data processing steps, including careful feature extraction, background clutter removal and compensation for variation in the scan step size through the tubing, were found to be critical to achieve good estimates of the pit depth and diameter. Variance studied in model probe dimensions did not adversely affect performance.
Computational aspects of sensitivity calculations in transient structural analysis
NASA Technical Reports Server (NTRS)
Greene, William H.; Haftka, Raphael T.
1988-01-01
A key step in the application of formal automated design techniques to structures under transient loading is the calculation of sensitivities of response quantities to the design parameters. This paper considers structures with general forms of damping acted on by general transient loading and addresses issues of computational errors and computational efficiency. The equations of motion are reduced using the traditional basis of vibration modes and then integrated using a highly accurate, explicit integration technique. A critical point constraint formulation is used to place constraints on the magnitude of each response quantity as a function of time. Three different techniques for calculating sensitivities of the critical point constraints are presented. The first two are based on the straightforward application of the forward and central difference operators, respectively. The third is based on explicit differentiation of the equations of motion. Condition errors, finite difference truncation errors, and modal convergence errors for the three techniques are compared by applying them to a simple five-span-beam problem. Sensitivity results are presented for two different transient loading conditions and for both damped and undamped cases.
Global Sensitivity Analysis of Environmental Models: Convergence, Robustness and Accuracy Analysis
NASA Astrophysics Data System (ADS)
Sarrazin, F.; Pianosi, F.; Hartmann, A. J.; Wagener, T.
2014-12-01
Sensitivity analysis aims to characterize the impact that changes in model input factors (e.g. the parameters) have on the model output (e.g. simulated streamflow). It is a valuable diagnostic tool for model understanding and for model improvement, it enhances calibration efficiency, and it supports uncertainty and scenario analysis. It is of particular interest for environmental models because they are often complex, non-linear, non-monotonic and exhibit strong interactions between their parameters. However, sensitivity analysis has to be carefully implemented to produce reliable results at moderate computational cost. For example, sample size can have a strong impact on the results and has to be carefully chosen. Yet, there is little guidance available for this step in environmental modelling. The objective of the present study is to provide guidelines for a robust sensitivity analysis, in order to support modellers in making appropriate choices for its implementation and in interpreting its outcome. We considered hydrological models with increasing level of complexity. We tested four sensitivity analysis methods, Regional Sensitivity Analysis, Method of Morris, a density-based (PAWN) and a variance-based (Sobol) method. The convergence and variability of sensitivity indices were investigated. We used bootstrapping to assess and improve the robustness of sensitivity indices even for limited sample sizes. Finally, we propose a quantitative validation approach for sensitivity analysis based on the Kolmogorov-Smirnov statistics.
NASA Astrophysics Data System (ADS)
Ansari, M. Z.; Cho, C.; Choi, W.; Lee, M.; Lee, S.; Kim, J.
2013-12-01
To improve the surface-stress sensitivity of piezoresistive microcantilever biosensors, this study presents cantilever designs with circular, square and triangular stress concentration regions (SCRs). The cantilever substrates are made of silicon and silicon dioxide and the piezoresistor is p-doped silicon. The cantilevers are characterized using FEM software for their surface-stress sensitivity and thermal vulnerability by applying surface stress, temperature and bias voltage loads. A relation for the surface-stress sensitivity of a silicon cantilever is also presented. Results show that use of SCRs in a silicon cantilever can improve its sensitivity by up to 55% and that a square SCR is the most suitable design to improve the overall sensitivity.
Sensitivity Analysis of Launch Vehicle Debris Risk Model
NASA Technical Reports Server (NTRS)
Gee, Ken; Lawrence, Scott L.
2010-01-01
As part of an analysis of the loss of crew risk associated with an ascent abort system for a manned launch vehicle, a model was developed to predict the impact risk of the debris resulting from an explosion of the launch vehicle on the crew module. The model consisted of a debris catalog describing the number, size and imparted velocity of each piece of debris, a method to compute the trajectories of the debris and a method to calculate the impact risk given the abort trajectory of the crew module. The model provided a point estimate of the strike probability as a function of the debris catalog, the time of abort and the delay time between the abort and destruction of the launch vehicle. A study was conducted to determine the sensitivity of the strike probability to the various model input parameters and to develop a response surface model for use in the sensitivity analysis of the overall ascent abort risk model. The results of the sensitivity analysis and the response surface model are presented in this paper.
Sensitivity analysis of coexistence in ecological communities: theory and application.
Barabás, György; Pásztor, Liz; Meszéna, Géza; Ostling, Annette
2014-12-01
Sensitivity analysis, the study of how ecological variables of interest respond to changes in external conditions, is a theoretically well-developed and widely applied approach in population ecology. Though the application of sensitivity analysis to predicting the response of species-rich communities to disturbances also has a long history, derivation of a mathematical framework for understanding the factors leading to robust coexistence has only been a recent undertaking. Here we suggest that this development opens up a new perspective, providing advances ranging from the applied to the theoretical. First, it yields a framework to be applied in specific cases for assessing the extinction risk of community modules in the face of environmental change. Second, it can be used to determine trait combinations allowing for coexistence that is robust to environmental variation, and limits to diversity in the presence of environmental variation, for specific community types. Third, it offers general insights into the nature of communities that are robust to environmental variation. We apply recent community-level extensions of mathematical sensitivity analysis to example models for illustration. We discuss the advantages and limitations of the method, and some of the empirical questions the theoretical framework could help answer. PMID:25252135
Sensitivity analysis in multiple imputation in effectiveness studies of psychotherapy
Crameri, Aureliano; von Wyl, Agnes; Koemeda, Margit; Schulthess, Peter; Tschuschke, Volker
2015-01-01
The importance of preventing and treating incomplete data in effectiveness studies is nowadays emphasized. However, most of the publications focus on randomized clinical trials (RCT). One flexible technique for statistical inference with missing data is multiple imputation (MI). Since methods such as MI rely on the assumption of missing data being at random (MAR), a sensitivity analysis for testing the robustness against departures from this assumption is required. In this paper we present a sensitivity analysis technique based on posterior predictive checking, which takes into consideration the concept of clinical significance used in the evaluation of intra-individual changes. We demonstrate the possibilities this technique can offer with the example of irregular longitudinal data collected with the Outcome Questionnaire-45 (OQ-45) and the Helping Alliance Questionnaire (HAQ) in a sample of 260 outpatients. The sensitivity analysis can be used to (1) quantify the degree of bias introduced by missing not at random data (MNAR) in a worst reasonable case scenario, (2) compare the performance of different analysis methods for dealing with missing data, or (3) detect the influence of possible violations to the model assumptions (e.g., lack of normality). Moreover, our analysis showed that ratings from the patient's and therapist's version of the HAQ could significantly improve the predictive value of the routine outcome monitoring based on the OQ-45. Since analysis dropouts always occur, repeated measurements with the OQ-45 and the HAQ analyzed with MI are useful to improve the accuracy of outcome estimates in quality assurance assessments and non-randomized effectiveness studies in the field of outpatient psychotherapy. PMID:26283989
Probabilistic constrained load flow based on sensitivity analysis
Karakatsanis, T.S.; Hatziargyriou, N.D. )
1994-11-01
This paper presents a method for network constrained setting of control variables based on probabilistic load flow analysis. The method determines operating constraint violations for a whole planning period together with the probability of each violation. An iterative algorithm is subsequently employed providing adjustments of the control variables based on sensitivity analysis of the constrained variables with respect to the control variables. The method is applied to the IEEE 14 busbar system and to a realistic model of the Hellenic Interconnected system indicating its suitability for short-term operational planning applications.
Sensitivity of Forecast Skill to Different Objective Analysis Schemes
NASA Technical Reports Server (NTRS)
Baker, W. E.
1979-01-01
Numerical weather forecasts are characterized by rapidly declining skill in the first 48 to 72 h. Recent estimates of the sources of forecast error indicate that the inaccurate specification of the initial conditions contributes substantially to this error. The sensitivity of the forecast skill to the initial conditions is examined by comparing a set of real-data experiments whose initial data were obtained with two different analysis schemes. Results are presented to emphasize the importance of the objective analysis techniques used in the assimilation of observational data.
Distributed Design and Analysis of Computer Experiments
Doak, Justin
2002-11-11
DDACE is a C++ object-oriented software library for the design and analysis of computer experiments. DDACE can be used to generate samples from a variety of sampling techniques. These samples may be used as input to a application code. DDACE also contains statistical tools such as response surface models and correlation coefficients to analyze input/output relationships between variables in an application code. DDACE can generate input values for uncertain variables within a user's application. For example, a user might like to vary a temperature variable as well as some material variables in a series of simulations. Through the series of simulations the user might be looking for optimal settings of parameters based on some user criteria. Or the user may be interested in the sensitivity to input variability shown by an output variable. In either case, the user may provide information about the suspected ranges and distributions of a set of input variables, along with a sampling scheme, and DDACE will generate input points based on these specifications. The input values generated by DDACE and the one or more outputs computed through the user's application code can be analyzed with a variety of statistical methods. This can lead to a wealth of information about the relationships between the variables in the problem. While statistical and mathematical packages may be employeed to carry out the analysis on the input/output relationships, DDACE also contains some tools for analyzing the simulation data. DDACE incorporates a software package called MARS (Multivariate Adaptive Regression Splines), developed by Jerome Friedman. MARS is used for generating a spline surface fit of the data. With MARS, a model simplification may be calculated using the input and corresponding output, values for the user's application problem. The MARS grid data may be used for generating 3-dimensional response surface plots of the simulation data. DDACE also contains an implementation of an
Distributed Design and Analysis of Computer Experiments
2002-11-11
DDACE is a C++ object-oriented software library for the design and analysis of computer experiments. DDACE can be used to generate samples from a variety of sampling techniques. These samples may be used as input to a application code. DDACE also contains statistical tools such as response surface models and correlation coefficients to analyze input/output relationships between variables in an application code. DDACE can generate input values for uncertain variables within a user's application. Formore » example, a user might like to vary a temperature variable as well as some material variables in a series of simulations. Through the series of simulations the user might be looking for optimal settings of parameters based on some user criteria. Or the user may be interested in the sensitivity to input variability shown by an output variable. In either case, the user may provide information about the suspected ranges and distributions of a set of input variables, along with a sampling scheme, and DDACE will generate input points based on these specifications. The input values generated by DDACE and the one or more outputs computed through the user's application code can be analyzed with a variety of statistical methods. This can lead to a wealth of information about the relationships between the variables in the problem. While statistical and mathematical packages may be employeed to carry out the analysis on the input/output relationships, DDACE also contains some tools for analyzing the simulation data. DDACE incorporates a software package called MARS (Multivariate Adaptive Regression Splines), developed by Jerome Friedman. MARS is used for generating a spline surface fit of the data. With MARS, a model simplification may be calculated using the input and corresponding output, values for the user's application problem. The MARS grid data may be used for generating 3-dimensional response surface plots of the simulation data. DDACE also contains an implementation
Sensitivity analysis of fine sediment models using heterogeneous data
NASA Astrophysics Data System (ADS)
Kamel, A. M. Yousif; Bhattacharya, B.; El Serafy, G. Y.; van Kessel, T.; Solomatine, D. P.
2012-04-01
Sediments play an important role in many aquatic systems. Their transportation and deposition has significant implication on morphology, navigability and water quality. Understanding the dynamics of sediment transportation in time and space is therefore important in drawing interventions and making management decisions. This research is related to the fine sediment dynamics in the Dutch coastal zone, which is subject to human interference through constructions, fishing, navigation, sand mining, etc. These activities do affect the natural flow of sediments and sometimes lead to environmental concerns or affect the siltation rates in harbours and fairways. Numerical models are widely used in studying fine sediment processes. Accuracy of numerical models depends upon the estimation of model parameters through calibration. Studying the model uncertainty related to these parameters is important in improving the spatio-temporal prediction of suspended particulate matter (SPM) concentrations, and determining the limits of their accuracy. This research deals with the analysis of a 3D numerical model of North Sea covering the Dutch coast using the Delft3D modelling tool (developed at Deltares, The Netherlands). The methodology in this research was divided into three main phases. The first phase focused on analysing the performance of the numerical model in simulating SPM concentrations near the Dutch coast by comparing the model predictions with SPM concentrations estimated from NASA's MODIS sensors at different time scales. The second phase focused on carrying out a sensitivity analysis of model parameters. Four model parameters were identified for the uncertainty and sensitivity analysis: the sedimentation velocity, the critical shear stress above which re-suspension occurs, the shields shear stress for re-suspension pick-up, and the re-suspension pick-up factor. By adopting different values of these parameters the numerical model was run and a comparison between the
Risk analysis of computer system designs
NASA Technical Reports Server (NTRS)
Vallone, A.
1981-01-01
Adverse events during implementation can affect final capabilities, schedule and cost of a computer system even though the system was accurately designed and evaluated. Risk analysis enables the manager to forecast the impact of those events and to timely ask for design revisions or contingency plans before making any decision. This paper presents a structured procedure for an effective risk analysis. The procedure identifies the required activities, separates subjective assessments from objective evaluations, and defines a risk measure to determine the analysis results. The procedure is consistent with the system design evaluation and enables a meaningful comparison among alternative designs.
Integrated Design and Analysis for Heterogeneous Objects
NASA Astrophysics Data System (ADS)
Qian, Xiaoping; Yang, Pinghai
2008-02-01
The recent advancement of solid freeform fabrication, design techniques and fundamental understanding of material properties in functionally graded material objects has made it possible to design and fabricate multifunctional heterogeneous objects. In this paper, we present an integrated design and analysis approach for heterogeneous object realization, which employs a unified design and analysis model based on B-splines and allows for direct interaction between the design and analysis model without a laborious meshing operation. In the design module, a new approach for intuitively modeling multi-material objects, termed heterogeneous lofting, is presented. In the analysis module, a novel graded B-spline finite element solution procedure is described, which gives orders of magnitude better convergence rate in comparison with current methods, as demonstrated in several case studies. Further advantages of this approach include simplified mesh construction, exact geometry/material composition representation and easy extraction of iso-material surface for manufacturing process planning.
Experiment Design and Analysis Guide - Neutronics & Physics
Misti A Lillo
2014-06-01
The purpose of this guide is to provide a consistent, standardized approach to performing neutronics/physics analysis for experiments inserted into the Advanced Test Reactor (ATR). This document provides neutronics/physics analysis guidance to support experiment design and analysis needs for experiments irradiated in the ATR. This guide addresses neutronics/physics analysis in support of experiment design, experiment safety, and experiment program objectives and goals. The intent of this guide is to provide a standardized approach for performing typical neutronics/physics analyses. Deviation from this guide is allowed provided that neutronics/physics analysis details are properly documented in an analysis report.
Structural Analysis in a Conceptual Design Framework
NASA Technical Reports Server (NTRS)
Padula, Sharon L.; Robinson, Jay H.; Eldred, Lloyd B.
2012-01-01
Supersonic aircraft designers must shape the outer mold line of the aircraft to improve multiple objectives, such as mission performance, cruise efficiency, and sonic-boom signatures. Conceptual designers have demonstrated an ability to assess these objectives for a large number of candidate designs. Other critical objectives and constraints, such as weight, fuel volume, aeroelastic effects, and structural soundness, are more difficult to address during the conceptual design process. The present research adds both static structural analysis and sizing to an existing conceptual design framework. The ultimate goal is to include structural analysis in the multidisciplinary optimization of a supersonic aircraft. Progress towards that goal is discussed and demonstrated.
Analysis of frequency characteristics and sensitivity of compliant mechanisms
NASA Astrophysics Data System (ADS)
Liu, Shanzeng; Dai, Jiansheng; Li, Aimin; Sun, Zhaopeng; Feng, Shizhe; Cao, Guohua
2016-03-01
Based on a modified pseudo-rigid-body model, the frequency characteristics and sensitivity of the large-deformation compliant mechanism are studied. Firstly, the pseudo-rigid-body model under the static and kinetic conditions is modified to enable the modified pseudo-rigid-body model to be more suitable for the dynamic analysis of the compliant mechanism. Subsequently, based on the modified pseudo-rigid-body model, the dynamic equations of the ordinary compliant four-bar mechanism are established using the analytical mechanics. Finally, in combination with the finite element analysis software ANSYS, the frequency characteristics and sensitivity of the compliant mechanism are analyzed by taking the compliant parallel-guiding mechanism and the compliant bistable mechanism as examples. From the simulation results, the dynamic characteristics of compliant mechanism are relatively sensitive to the structure size, section parameter, and characteristic parameter of material on mechanisms. The results could provide great theoretical significance and application values for the structural optimization of compliant mechanisms, the improvement of their dynamic properties and the expansion of their application range.
Sensitivity-analysis techniques: self-teaching curriculum
Iman, R.L.; Conover, W.J.
1982-06-01
This self teaching curriculum on sensitivity analysis techniques consists of three parts: (1) Use of the Latin Hypercube Sampling Program (Iman, Davenport and Ziegler, Latin Hypercube Sampling (Program User's Guide), SAND79-1473, January 1980); (2) Use of the Stepwise Regression Program (Iman, et al., Stepwise Regression with PRESS and Rank Regression (Program User's Guide) SAND79-1472, January 1980); and (3) Application of the procedures to sensitivity and uncertainty analyses of the groundwater transport model MWFT/DVM (Campbell, Iman and Reeves, Risk Methodology for Geologic Disposal of Radioactive Waste - Transport Model Sensitivity Analysis; SAND80-0644, NUREG/CR-1377, June 1980: Campbell, Longsine, and Reeves, The Distributed Velocity Method of Solving the Convective-Dispersion Equation, SAND80-0717, NUREG/CR-1376, July 1980). This curriculum is one in a series developed by Sandia National Laboratories for transfer of the capability to use the technology developed under the NRC funded High Level Waste Methodology Development Program.
LSENS, The NASA Lewis Kinetics and Sensitivity Analysis Code
NASA Technical Reports Server (NTRS)
Radhakrishnan, K.
2000-01-01
A general chemical kinetics and sensitivity analysis code for complex, homogeneous, gas-phase reactions is described. The main features of the code, LSENS (the NASA Lewis kinetics and sensitivity analysis code), are its flexibility, efficiency and convenience in treating many different chemical reaction models. The models include: static system; steady, one-dimensional, inviscid flow; incident-shock initiated reaction in a shock tube; and a perfectly stirred reactor. In addition, equilibrium computations can be performed for several assigned states. An implicit numerical integration method (LSODE, the Livermore Solver for Ordinary Differential Equations), which works efficiently for the extremes of very fast and very slow reactions, is used to solve the "stiff" ordinary differential equation systems that arise in chemical kinetics. For static reactions, the code uses the decoupled direct method to calculate sensitivity coefficients of the dependent variables and their temporal derivatives with respect to the initial values of dependent variables and/or the rate coefficient parameters. Solution methods for the equilibrium and post-shock conditions and for perfectly stirred reactor problems are either adapted from or based on the procedures built into the NASA code CEA (Chemical Equilibrium and Applications).
Shipping Cask Design Review Analysis.
1998-01-04
Version 01 SCANS (Shipping Cask ANalysis System) is a microcomputer based system of computer programs and databases for evaluating safety analysis reports on spent fuel shipping casks. SCANS calculates the global response to impact loads, pressure loads, and thermal conditions, providing reviewers with an independent check on analyses submitted by licensees. Analysis options are based on regulatory cases described in the Code of Federal Regulations (1983) and Regulatory Guides published by the NRC in 1977more » and 1978. The system is composed of a series of menus and input entry cask analysis, and output display programs. An analysis is performed by preparing the necessary input data and then selecting the appropriate analysis: impact, thermal (heat transfer), thermally-induced stress, or pressure-induced stress. All data are entered through input screens with descriptive data requests, and, where possible, default values are provided. Output (i.e., impact force, moment and sheer time histories; impact animation; thermal/stress geometry and thermal/stress element outlines; temperature distributions as isocontours or profiles; and temperature time histories) is displayed graphically and can also be printed.« less
Sensitivity of IFMIF-DONES irradiation characteristics to different design parameters
NASA Astrophysics Data System (ADS)
Mota, F.; Ibarra, Á.; García, Á.; Molla, J.
2015-11-01
The DONES (DEMO oriented neutron source) has been conceived as a simplified IFMIF-like plant to provide earlier availability, on a reduced time scale and with a reduced budget—both compared to IFMIF—of basic information on materials damage. The conceptual design of the DONES will consist of a number of changes oriented to reduce the time required for construction and the time required to produce the required damage. This paper evaluates how changes in beam irradiation parameters or geometric issues will affect the damage produced in the specimens located in the experimental irradiation area (test cell) of the DONES. Neutron transport calculations have been performed to evaluate the sensitivity of material radiation effects to different irradiation environments. The neutron transport calculations have been performed using McDeLicious 2011 (based on MCNP5-v1.6), code developed by the KIT research institute to reproduce the IFMIF deuteron-lithium neutron source. With this aim in mind, the materials radiation effects evaluated to analyse the sensitivity to the changes in the irradiation conditions have been the damage dose rate, the He and H production, and the He and H production to damage dose ratio. With the analysis of these parameters it is possible to make an evaluation of the level of primary damage suffered by the specimen placed in the high flux test module (HFTM). Based on results obtained in this paper and taking into account engineering considerations, substantial progress should be possible regarding the optimum design parameters for the irradiation area of the IFMIF-DONES facility.
Hyperspectral data analysis procedures with reduced sensitivity to noise
NASA Technical Reports Server (NTRS)
Landgrebe, David A.
1993-01-01
Multispectral sensor systems have become steadily improved over the years in their ability to deliver increased spectral detail. With the advent of hyperspectral sensors, including imaging spectrometers, this technology is in the process of taking a large leap forward, thus providing the possibility of enabling delivery of much more detailed information. However, this direction of development has drawn even more attention to the matter of noise and other deleterious effects in the data, because reducing the fundamental limitations of spectral detail on information collection raises the limitations presented by noise to even greater importance. Much current effort in remote sensing research is thus being devoted to adjusting the data to mitigate the effects of noise and other deleterious effects. A parallel approach to the problem is to look for analysis approaches and procedures which have reduced sensitivity to such effects. We discuss some of the fundamental principles which define analysis algorithm characteristics providing such reduced sensitivity. One such analysis procedure including an example analysis of a data set is described, illustrating this effect.
Sensitivity analysis for nonrandom dropout: a local influence approach.
Verbeke, G; Molenberghs, G; Thijs, H; Lesaffre, E; Kenward, M G
2001-03-01
Diggle and Kenward (1994, Applied Statistics 43, 49-93) proposed a selection model for continuous longitudinal data subject to nonrandom dropout. It has provoked a large debate about the role for such models. The original enthusiasm was followed by skepticism about the strong but untestable assumptions on which this type of model invariably rests. Since then, the view has emerged that these models should ideally be made part of a sensitivity analysis. This paper presents a formal and flexible approach to such a sensitivity assessment based on local influence (Cook, 1986, Journal of the Royal Statistical Society, Series B 48, 133-169). The influence of perturbing a missing-at-random dropout model in the direction of nonrandom dropout is explored. The method is applied to data from a randomized experiment on the inhibition of testosterone production in rats. PMID:11252620
SENSITIVITY ANALYSIS OF A TPB DEGRADATION RATE MODEL
Crawford, C; Tommy Edwards, T; Bill Wilmarth, B
2006-08-01
A tetraphenylborate (TPB) degradation model for use in aggregating Tank 48 material in Tank 50 is developed in this report. The influential factors for this model are listed as the headings in the table below. A sensitivity study of the predictions of the model over intervals of values for the influential factors affecting the model was conducted. These intervals bound the levels of these factors expected during Tank 50 aggregations. The results from the sensitivity analysis were used to identify settings for the influential factors that yielded the largest predicted TPB degradation rate. Thus, these factor settings are considered as those that yield the ''worst-case'' scenario for TPB degradation rate for Tank 50 aggregation, and, as such they would define the test conditions that should be studied in a waste qualification program whose dual purpose would be the investigation of the introduction of Tank 48 material for aggregation in Tank 50 and the bounding of TPB degradation rates for such aggregations.
An easily implemented static condensation method for structural sensitivity analysis
NASA Technical Reports Server (NTRS)
Gangadharan, S. N.; Haftka, R. T.; Nikolaidis, E.
1990-01-01
A black-box approach to static condensation for sensitivity analysis is presented with illustrative examples of a cube and a car structure. The sensitivity of the structural response with respect to joint stiffness parameter is calculated using the direct method, forward-difference, and central-difference schemes. The efficiency of the various methods for identifying joint stiffness parameters from measured static deflections of these structures is compared. The results indicate that the use of static condensation can reduce computation times significantly and the black-box approach is only slightly less efficient than the standard implementation of static condensation. The ease of implementation of the black-box approach recommends it for use with general-purpose finite element codes that do not have a built-in facility for static condensation.
Sensitive LC MS quantitative analysis of carbohydrates by Cs+ attachment.
Rogatsky, Eduard; Jayatillake, Harsha; Goswami, Gayotri; Tomuta, Vlad; Stein, Daniel
2005-11-01
The development of a sensitive assay for the quantitative analysis of carbohydrates from human plasma using LC/MS/MS is described in this paper. After sample preparation, carbohydrates were cationized by Cs(+) after their separation by normal phase liquid chromatography on an amino based column. Cesium is capable of forming a quasi-molecular ion [M + Cs](+) with neutral carbohydrate molecules in the positive ion mode of electrospray ionization mass spectrometry. The mass spectrometer was operated in multiple reaction monitoring mode, and transitions [M + 133] --> 133 were monitored (M, carbohydrate molecular weight). The new method is robust, highly sensitive, rapid, and does not require postcolumn addition or derivatization. It is useful in clinical research for measurement of carbohydrate molecules by isotope dilution assay. PMID:16182559
Sensitivity Analysis of Hardwired Parameters in GALE Codes
Geelhood, Kenneth J.; Mitchell, Mark R.; Droppo, James G.
2008-12-01
The U.S. Nuclear Regulatory Commission asked Pacific Northwest National Laboratory to provide a data-gathering plan for updating the hardwired data tables and parameters of the Gaseous and Liquid Effluents (GALE) codes to reflect current nuclear reactor performance. This would enable the GALE codes to make more accurate predictions about the normal radioactive release source term applicable to currently operating reactors and to the cohort of reactors planned for construction in the next few years. A sensitivity analysis was conducted to define the importance of hardwired parameters in terms of each parameter’s effect on the emission rate of the nuclides that are most important in computing potential exposures. The results of this study were used to compile a list of parameters that should be updated based on the sensitivity of these parameters to outputs of interest.
Sensitivity analysis for dynamic systems with time-lags
NASA Astrophysics Data System (ADS)
Rihan, Fathalla A.
2003-02-01
Many problems in bioscience for which observations are reported in the literature can be modelled by suitable functional differential equations incorporating time-lags (other terminology: delays) or memory effects, parameterized by scientifically meaningful constant parameters p or/and variable parameters (for example, control functions) u(t). It is often desirable to have information about the effect on the solution of the dynamic system of perturbing the initial data, control functions, time-lags and other parameters appearing in the model. The main purpose of this paper is to derive a general theory for sensitivity analysis of mathematical models that contain time-lags. In this paper, we use adjoint equations and direct methods to estimate the sensitivity functions when the parameters appearing in the model are not only constants but also variables of time. To illustrate the results, the methodology is applied numerically to an example of a delay differential model.
Multi-resolution multi-sensitivity design for parallel-hole SPECT collimators.
Li, Yanzhao; Xiao, Peng; Zhu, Xiaohua; Xie, Qingguo
2016-07-21
Multi-resolution multi-sensitivity (MRMS) collimator offering adjustable trade-off between resolution and sensitivity, can make a SPECT system adaptive. We propose in this paper a new idea for MRMS design based on, for the first time, parallel-hole collimators for clinical SPECT. Multiple collimation states with varied resolution/sensitivity trade-offs can be formed by slightly changing the collimator's inner structure. To validate the idea, the GE LEHR collimator is selected as the design prototype and is modeled using a ray-tracing technique. Point images are generated for several states of the design. Results show that the collimation states of the design can obtain similar point response characteristics to parallel-hole collimators, and can be used just like parallel-hole collimators in clinical SPECT imaging. Ray-tracing modeling also shows that the proposed design can offer varied resolution/sensitivity trade-offs: at 100 mm before the collimator, the highest resolution state provides 6.9 mm full width at a half maximum (FWHM) with a nearly minimum sensitivity of about 96.2 cps MBq(-1), while the lowest resolution state obtains 10.6 mm FWHM with the highest sensitivity of about 167.6 cps MBq(-1). Further comparisons of the states on image qualities are conducted through Monte Carlo simulation of a hot-spot phantom which contains five hot spots with varied sizes. Contrast-to-noise ratios (CNR) of the spots are calculated and compared, showing that different spots can prefer different collimation states: the larger spots obtain better CNRs by using the larger sensitivity states, and the smaller spots prefer the higher resolution states. In conclusion, the proposed idea can be an effective approach for MRMS design for parallel-hole SPECT collimators. PMID:27359049
Multi-resolution multi-sensitivity design for parallel-hole SPECT collimators
NASA Astrophysics Data System (ADS)
Li, Yanzhao; Xiao, Peng; Zhu, Xiaohua; Xie, Qingguo
2016-07-01
Multi-resolution multi-sensitivity (MRMS) collimator offering adjustable trade-off between resolution and sensitivity, can make a SPECT system adaptive. We propose in this paper a new idea for MRMS design based on, for the first time, parallel-hole collimators for clinical SPECT. Multiple collimation states with varied resolution/sensitivity trade-offs can be formed by slightly changing the collimator’s inner structure. To validate the idea, the GE LEHR collimator is selected as the design prototype and is modeled using a ray-tracing technique. Point images are generated for several states of the design. Results show that the collimation states of the design can obtain similar point response characteristics to parallel-hole collimators, and can be used just like parallel-hole collimators in clinical SPECT imaging. Ray-tracing modeling also shows that the proposed design can offer varied resolution/sensitivity trade-offs: at 100 mm before the collimator, the highest resolution state provides 6.9 mm full width at a half maximum (FWHM) with a nearly minimum sensitivity of about 96.2 cps MBq‑1, while the lowest resolution state obtains 10.6 mm FWHM with the highest sensitivity of about 167.6 cps MBq‑1. Further comparisons of the states on image qualities are conducted through Monte Carlo simulation of a hot-spot phantom which contains five hot spots with varied sizes. Contrast-to-noise ratios (CNR) of the spots are calculated and compared, showing that different spots can prefer different collimation states: the larger spots obtain better CNRs by using the larger sensitivity states, and the smaller spots prefer the higher resolution states. In conclusion, the proposed idea can be an effective approach for MRMS design for parallel-hole SPECT collimators.
On options for interdisciplinary analysis and design optimization
NASA Technical Reports Server (NTRS)
Haftka, R. T.; Sobieszczanski-Sobieski, J.; Padula, S. L.
1992-01-01
The interdisciplinary optimization of engineering systems is discussed from the standpoint of the computational alternatives available to the designer. The analysis of such systems typically requires the solution of coupled systems of nonlinear algebraic equations. The solution procedure is necessarily iterative in nature. It is shown that the system can be solved by fixed point iteration, by Newton's method, or by a combination of the two. However, the need for sensitivity analysis may affect the choice of analysis solution method. Similarly, the optimization of the system can be formulated in several ways that are discussed in the paper. It is shown that the effect of the topology of the interaction between disciplines is a key factor in the choice of analysis, sensitivity and optimization methods. Several examples are presented to illustrate the discussion.
Spacecraft design optimization using Taguchi analysis
NASA Technical Reports Server (NTRS)
Unal, Resit
1991-01-01
The quality engineering methods of Dr. Genichi Taguchi, employing design of experiments, are important statistical tools for designing high quality systems at reduced cost. The Taguchi method was utilized to study several simultaneous parameter level variations of a lunar aerobrake structure to arrive at the lightest weight configuration. Finite element analysis was used to analyze the unique experimental aerobrake configurations selected by Taguchi method. Important design parameters affecting weight and global buckling were identified and the lowest weight design configuration was selected.
Design Through Analysis (DTA) roadmap vision.
Blacker, Teddy Dean; Adams, Charles R.; Hoffman, Edward L.; White, David Roger; Sjaardema, Gregory D.
2004-10-01
The Design through Analysis Realization Team (DART) will provide analysts with a complete toolset that reduces the time to create, generate, analyze, and manage the data generated in a computational analysis. The toolset will be both easy to learn and easy to use. The DART Roadmap Vision provides for progressive improvements that will reduce the Design through Analysis (DTA) cycle time by 90-percent over a three-year period while improving both the quality and accountability of the analyses.
Imaging Fourier transform spectrometer (IFTS): parametric sensitivity analysis
NASA Astrophysics Data System (ADS)
Keller, Robert A.; Lomheim, Terrence S.
2005-06-01
Imaging Fourier transform spectrometers (IFTS) allow for very high spectral resolution hyperspectral imaging while using moderate size 2D focal plane arrays in a staring mode. This is not the case for slit scanning dispersive imaging spectrometers where spectral sampling is related to the focal plane pixel count along the spectral dimension of the 2D focal plane used in such an instrument. This can become a major issue in the longwave infrared (LWIR) where the operability and yield of highly sensitivity arrays (i.e.HgCdTe) of large dimension are generally poor. However using an IFTS introduces its own unique set of issues and tradeoffs. In this paper we develop simplified equations for describing the sensitivity of an IFTS, including the effects of data windowing. These equations provide useful insights into the optical, focal plane and operational design trade space that must be considered when examining IFTS concepts aimed at a specific sensitivity and spectral resolution application. The approach is illustrated by computing the LWIR noise-equivalent spectral radiance (NESR) corresponding to the NASA Geosynchronous Imaging Fourier Transform Spectrometer (GIFTS) concept assuming a proven and reasonable noise-equivalent irradiance (NEI) capability for the focal plane.
Biosphere dose conversion Factor Importance and Sensitivity Analysis
M. Wasiolek
2004-10-15
This report presents importance and sensitivity analysis for the environmental radiation model for Yucca Mountain, Nevada (ERMYN). ERMYN is a biosphere model supporting the total system performance assessment (TSPA) for the license application (LA) for the Yucca Mountain repository. This analysis concerns the output of the model, biosphere dose conversion factors (BDCFs) for the groundwater, and the volcanic ash exposure scenarios. It identifies important processes and parameters that influence the BDCF values and distributions, enhances understanding of the relative importance of the physical and environmental processes on the outcome of the biosphere model, includes a detailed pathway analysis for key radionuclides, and evaluates the appropriateness of selected parameter values that are not site-specific or have large uncertainty.
Accuracy analysis and design of A3 parallel spindle head
NASA Astrophysics Data System (ADS)
Ni, Yanbing; Zhang, Biao; Sun, Yupeng; Zhang, Yuan
2016-03-01
As functional components of machine tools, parallel mechanisms are widely used in high efficiency machining of aviation components, and accuracy is one of the critical technical indexes. Lots of researchers have focused on the accuracy problem of parallel mechanisms, but in terms of controlling the errors and improving the accuracy in the stage of design and manufacturing, further efforts are required. Aiming at the accuracy design of a 3-DOF parallel spindle head(A3 head), its error model, sensitivity analysis and tolerance allocation are investigated. Based on the inverse kinematic analysis, the error model of A3 head is established by using the first-order perturbation theory and vector chain method. According to the mapping property of motion and constraint Jacobian matrix, the compensatable and uncompensatable error sources which affect the accuracy in the end-effector are separated. Furthermore, sensitivity analysis is performed on the uncompensatable error sources. The sensitivity probabilistic model is established and the global sensitivity index is proposed to analyze the influence of the uncompensatable error sources on the accuracy in the end-effector of the mechanism. The results show that orientation error sources have bigger effect on the accuracy in the end-effector. Based upon the sensitivity analysis results, the tolerance design is converted into the issue of nonlinearly constrained optimization with the manufacturing cost minimum being the optimization objective. By utilizing the genetic algorithm, the allocation of the tolerances on each component is finally determined. According to the tolerance allocation results, the tolerance ranges of ten kinds of geometric error sources are obtained. These research achievements can provide fundamental guidelines for component manufacturing and assembly of this kind of parallel mechanisms.
Sensitivity Analysis of OECD Benchmark Tests in BISON
Swiler, Laura Painton; Gamble, Kyle; Schmidt, Rodney C.; Williamson, Richard
2015-09-01
This report summarizes a NEAMS (Nuclear Energy Advanced Modeling and Simulation) project focused on sensitivity analysis of a fuels performance benchmark problem. The benchmark problem was defined by the Uncertainty Analysis in Modeling working group of the Nuclear Science Committee, part of the Nuclear Energy Agency of the Organization for Economic Cooperation and Development (OECD ). The benchmark problem involv ed steady - state behavior of a fuel pin in a Pressurized Water Reactor (PWR). The problem was created in the BISON Fuels Performance code. Dakota was used to generate and analyze 300 samples of 17 input parameters defining core boundary conditions, manuf acturing tolerances , and fuel properties. There were 24 responses of interest, including fuel centerline temperatures at a variety of locations and burnup levels, fission gas released, axial elongation of the fuel pin, etc. Pearson and Spearman correlatio n coefficients and Sobol' variance - based indices were used to perform the sensitivity analysis. This report summarizes the process and presents results from this study.
Radiometer Design Analysis Based Upon Measurement Uncertainty
NASA Technical Reports Server (NTRS)
Racette, Paul E.; Lang, Roger H.
2004-01-01
This paper introduces a method for predicting the performance of a radiometer design based on calculating the measurement uncertainty. The variety in radiometer designs and the demand for improved radiometric measurements justify the need for a more general and comprehensive method to assess system performance. Radiometric resolution, or sensitivity, is a figure of merit that has been commonly used to characterize the performance of a radiometer. However when evaluating the performance of a calibration design for a radiometer, the use of radiometric resolution has limited application. These limitations are overcome by considering instead the measurement uncertainty. A method for calculating measurement uncertainty for a generic radiometer design including its calibration algorithm is presented. The result is a generalized technique by which system calibration architectures and design parameters can be studied to optimize instrument performance for given requirements and constraints. Example applications demonstrate the utility of using measurement uncertainty as a figure of merit.
SENSITIVITY ANALYSIS FOR SALTSTONE DISPOSAL UNIT COLUMN DEGRADATION ANALYSES
Flach, G.
2014-10-28
PORFLOW related analyses supporting a Sensitivity Analysis for Saltstone Disposal Unit (SDU) column degradation were performed. Previous analyses, Flach and Taylor 2014, used a model in which the SDU columns degraded in a piecewise manner from the top and bottom simultaneously. The current analyses employs a model in which all pieces of the column degrade at the same time. Information was extracted from the analyses which may be useful in determining the distribution of Tc-99 in the various SDUs throughout time and in determining flow balances for the SDUs.
Path-sensitive analysis for reducing rollback overheads
O'Brien, John K.P.; Wang, Kai-Ting Amy; Yamashita, Mark; Zhuang, Xiaotong
2014-07-22
A mechanism is provided for path-sensitive analysis for reducing rollback overheads. The mechanism receives, in a compiler, program code to be compiled to form compiled code. The mechanism divides the code into basic blocks. The mechanism then determines a restore register set for each of the one or more basic blocks to form one or more restore register sets. The mechanism then stores the one or more register sets such that responsive to a rollback during execution of the compiled code. A rollback routine identifies a restore register set from the one or more restore register sets and restores registers identified in the identified restore register set.
NASA Astrophysics Data System (ADS)
Crucifix, M.; Araya-Melo, P. A.
2013-12-01
The sensitivity of terrestrial climates of the southern hemisphere to astronomical forcing, CO2 and glaciation level is systematically investigated following a global sensitivity analysis. The approach is founded on analysis of about 100 experiments performed with the GCM HadCM3, statistically analysed using a Gaussian Process emulator. The presentation emphasises the importance of the selection of experiments (experiment design) and the validation of the statistical model. At the stage of writing the abstract only preliminary results have been obtained, but following an approach followed by our group for the Indian Monsoon and vegetation feedback analysis we expect to be able to show and discuss amplitude and phase relationships between changes in terrestrial environments of the southern hemisphere and driving factors.
Hu, Fa-Chun; Wang, Sheng-Wei; Planells, Miquel; Robertson, Neil; Padhy, Harihara; Du, Bo-Sian; Chi, Yun; Yang, Po-Fan; Lin, Hao-Wu; Lee, Gene-Hsiang; Chou, Pi-Tai
2013-08-01
A series of Os(II) sensitizers (TFOS-x, in which x=1, 2, or 3) with a single 4,4'-dicarboxy-2,2'-dipyridine (H2 dcbpy) anchor and two chelating 2-pyridyl (or 2-pyrimidyl) triazolate ancillaries was successfully prepared. Single-crystal X-ray structural analysis showed that the core geometry of the Os(II) -based sensitizers consisted of one H2 dcbpy unit and two eclipsed cis-triazolate fragments; this was notably different from the Ru(II) -based counterparts, in which the azolate (both pyrazolate and triazolate) fragments are located at the mutual trans-positions. The basic properties were extensively probed by using spectroscopic and electrochemical methods as well as time-dependent density functional theory (TD-DFT) calculations. Fabrication of dye-sensitized solar cells (DSCs) was then attempted by using the I(-) /I3 (-) -based electrolyte solution. One such DSC device, which utilized TFOS-2 as the sensitizer, showed promising performance characteristics with a short-circuit current density (JSC ) of 15.7 mA cm(-2) , an open-circuit voltage of 610 mV, a fill factor of 0.63, and a power conversion efficiency of 6.08 % under AM 1.5G simulated one-sun irradiation. Importantly, adequate incident photon-to-current conversion efficiency performances were observed for all TFOS derivatives over the wide spectral region of 450 to 950 nm, showing a panchromatic light harvesting capability that extended into the near-infrared regime. Our results underlined a feasible strategy for maximizing JSC and increasing the efficiency of DSCs. PMID:23843354
Design-for-analysis or the unintended role of analysis in the design of piping systems
Antaki, G.A.
1991-12-31
The paper discusses the evolution of piping design in the nuclear industry with its increasing reliance on dynamic analysis. While it is well recognized that the practice has evolved from ``design-by- rule `` to ``design-by-analysis,`` examples are provided of cases where the choice of analysis technique has determined the hardware configuration, which could be called ``design-for-analysis.`` The paper presents practical solutions to some of these cases and summarizes the important recent industry and regulatory developments which, if successful, will reverse the trend towards ``design-for-analysis.`` 14 refs.
Design-for-analysis or the unintended role of analysis in the design of piping systems
Antaki, G.A.
1991-01-01
The paper discusses the evolution of piping design in the nuclear industry with its increasing reliance on dynamic analysis. While it is well recognized that the practice has evolved from design-by- rule '' to design-by-analysis,'' examples are provided of cases where the choice of analysis technique has determined the hardware configuration, which could be called design-for-analysis.'' The paper presents practical solutions to some of these cases and summarizes the important recent industry and regulatory developments which, if successful, will reverse the trend towards design-for-analysis.'' 14 refs.
Sensitivity analysis of state-specific multireference perturbation theory
NASA Astrophysics Data System (ADS)
Szabados, Ágnes
2011-05-01
State-specific multireference perturbation theory (SS-MRPT) developed by Mukherjee et al. [Int. J. Mol. Sci. 3, 733 (2002)] is examined focusing on the dependence of the perturbed energy on the initial model space coefficients. It has been observed earlier, that non-physical kinks may appear on the potential energy surface obtained by SS-MRPT while related coupled-cluster methods may face convergence difficulties. Though exclusion or damping of the division by small coefficients may alleviate the problem, it is demonstrated here that the effect does not originate in an ill-defined division. It is shown that non-negligible model space coefficients may also be linked with the problem. Sensitivity analysis is suggested as a tool for detecting the coefficient responsible. By monitoring the singular values of sensitivity matrices, orders of magnitude increase is found in the largest value, in the vicinity of the problematic geometry point on the potential energy surface. The drastic increase of coefficient sensitivities is found to be linked with a degeneracy of the target root of the effective Hamiltonian. The nature of the one-electron orbitals has a profound influence on the picture: a rotation among active orbitals may screen or worsen the effect.
Post-optimality analysis in aerospace vehicle design
NASA Technical Reports Server (NTRS)
Braun, Robert D.; Kroo, Ilan M.; Gage, Peter J.
1993-01-01
This analysis pertains to the applicability of optimal sensitivity information to aerospace vehicle design. The present analysis demonstrates that post-optimality information generated through first-order computations can be used to accurately predict file effect of constraint and parameter perturbations on the optimal solution. This assessment is based on the solution of an aircraft design problem in which the post-optimality estimates are shown to be within a few percent of the true solution over the practical range of constraint and parameter variations. Through solution of a reusable, single-stage-to-orbit, launch vehicle design problem, this optimal sensitivity information is also shown to improve the efficiency of the design process. For a hierarchically decomposed problem, this computational efficiency is realizable by estimating the main-problem objective gradient through optimal sensitivity calculations. By reducing the need for finite differentiation of a re-optimized subproblem, a significant decrease in the number of objective function evaluations required to reach the optimal solution is obtained.
Parametric sensitivity analysis of an agro-economic model of management of irrigation water
NASA Astrophysics Data System (ADS)
El Ouadi, Ihssan; Ouazar, Driss; El Menyari, Younesse
2015-04-01
The current work aims to build an analysis and decision support tool for policy options concerning the optimal allocation of water resources, while allowing a better reflection on the issue of valuation of water by the agricultural sector in particular. Thus, a model disaggregated by farm type was developed for the rural town of Ait Ben Yacoub located in the east Morocco. This model integrates economic, agronomic and hydraulic data and simulates agricultural gross margin across in this area taking into consideration changes in public policy and climatic conditions, taking into account the competition for collective resources. To identify the model input parameters that influence over the results of the model, a parametric sensitivity analysis is performed by the "One-Factor-At-A-Time" approach within the "Screening Designs" method. Preliminary results of this analysis show that among the 10 parameters analyzed, 6 parameters affect significantly the objective function of the model, it is in order of influence: i) Coefficient of crop yield response to water, ii) Average daily gain in weight of livestock, iii) Exchange of livestock reproduction, iv) maximum yield of crops, v) Supply of irrigation water and vi) precipitation. These 6 parameters register sensitivity indexes ranging between 0.22 and 1.28. Those results show high uncertainties on these parameters that can dramatically skew the results of the model or the need to pay particular attention to their estimates. Keywords: water, agriculture, modeling, optimal allocation, parametric sensitivity analysis, Screening Designs, One-Factor-At-A-Time, agricultural policy, climate change.
Sensitivity Analysis of a process based erosion model using FAST
NASA Astrophysics Data System (ADS)
Gabelmann, Petra; Wienhöfer, Jan; Zehe, Erwin
2015-04-01
deposition are related to overland flow velocity using the equation of Engelund and Hansen and the sinking velocity of grain sizes, respectively. The sensitivity analysis was performed based on virtual hillslopes similar to those in the Weiherbach catchment. We applied the FAST-method (Fourier Amplitude Sensitivity Test), which provides a global sensitivity analysis with comparably few model runs. We varied model parameters in predefined and, for the Weiherbach catchment, physically meaningful parameter ranges. Those parameters included rainfall intensity, surface roughness, hillslope geometry, land use, erosion resistance, and soil hydraulic parameters. The results of this study allow guiding further modelling efforts in the Weiherbach catchment with respect to data collection and model modification.
A global sensitivity analysis of crop virtual water content
NASA Astrophysics Data System (ADS)
Tamea, S.; Tuninetti, M.; D'Odorico, P.; Laio, F.; Ridolfi, L.
2015-12-01
The concepts of virtual water and water footprint are becoming widely used in the scientific literature and they are proving their usefulness in a number of multidisciplinary contexts. With such growing interest a measure of data reliability (and uncertainty) is becoming pressing but, as of today, assessments of data sensitivity to model parameters, performed at the global scale, are not known. This contribution aims at filling this gap. Starting point of this study is the evaluation of the green and blue virtual water content (VWC) of four staple crops (i.e. wheat, rice, maize, and soybean) at a global high resolution scale. In each grid cell, the crop VWC is given by the ratio between the total crop evapotranspiration over the growing season and the crop actual yield, where evapotranspiration is determined with a detailed daily soil water balance and actual yield is estimated using country-based data, adjusted to account for spatial variability. The model provides estimates of the VWC at a 5x5 arc minutes and it improves on previous works by using the newest available data and including multi-cropping practices in the evaluation. The model is then used as the basis for a sensitivity analysis, in order to evaluate the role of model parameters in affecting the VWC and to understand how uncertainties in input data propagate and impact the VWC accounting. In each cell, small changes are exerted to one parameter at a time, and a sensitivity index is determined as the ratio between the relative change of VWC and the relative change of the input parameter with respect to its reference value. At the global scale, VWC is found to be most sensitive to the planting date, with a positive (direct) or negative (inverse) sensitivity index depending on the typical season of crop planting date. VWC is also markedly dependent on the length of the growing period, with an increase in length always producing an increase of VWC, but with higher spatial variability for rice than for
Sensitivity Analysis of a Pharmacokinetic Model of Vaginal Anti-HIV Microbicide Drug Delivery.
Jarrett, Angela M; Gao, Yajing; Hussaini, M Yousuff; Cogan, Nicholas G; Katz, David F
2016-05-01
Uncertainties in parameter values in microbicide pharmacokinetics (PK) models confound the models' use in understanding the determinants of drug delivery and in designing and interpreting dosing and sampling in PK studies. A global sensitivity analysis (Sobol' indices) was performed for a compartmental model of the pharmacokinetics of gel delivery of tenofovir to the vaginal mucosa. The model's parameter space was explored to quantify model output sensitivities to parameters characterizing properties for the gel-drug product (volume, drug transport, initial loading) and host environment (thicknesses of the mucosal epithelium and stroma and the role of ambient vaginal fluid in diluting gel). Greatest sensitivities overall were to the initial drug concentration in gel, gel-epithelium partition coefficient for drug, and rate constant for gel dilution by vaginal fluid. Sensitivities for 3 PK measures of drug concentration values were somewhat different than those for the kinetic PK measure. Sensitivities in the stromal compartment (where tenofovir acts against host cells) and a simulated biopsy also depended on thicknesses of epithelium and stroma. This methodology and results here contribute an approach to help interpret uncertainties in measures of vaginal microbicide gel properties and their host environment. In turn, this will inform rational gel design and optimization. PMID:27012224
Integrated reflector antenna design and analysis
NASA Technical Reports Server (NTRS)
Zimmerman, M. L.; Lee, S. W.; Ni, S.; Christensen, M.; Wang, Y. M.
1993-01-01
Reflector antenna design is a mature field and most aspects were studied. However, of that most previous work is distinguished by the fact that it is narrow in scope, analyzing only a particular problem under certain conditions. Methods of analysis of this type are not useful for working on real-life problems since they can not handle the many and various types of perturbations of basic antenna design. The idea of an integrated design and analysis is proposed. By broadening the scope of the analysis, it becomes possible to deal with the intricacies attendant with modem reflector antenna design problems. The concept of integrated reflector antenna design is put forward. A number of electromagnetic problems related to reflector antenna design are investigated. Some of these show how tools for reflector antenna design are created. In particular, a method for estimating spillover loss for open-ended waveguide feeds is examined. The problem of calculating and optimizing beam efficiency (an important figure of merit in radiometry applications) is also solved. Other chapters deal with applications of this general analysis. The wide angle scan abilities of reflector antennas is examined and a design is proposed for the ATDRSS triband reflector antenna. The development of a general phased-array pattern computation program is discussed and how the concept of integrated design can be extended to other types of antennas is shown. The conclusions are contained in the final chapter.
Principal component analysis for designed experiments
2015-01-01
Background Principal component analysis is used to summarize matrix data, such as found in transcriptome, proteome or metabolome and medical examinations, into fewer dimensions by fitting the matrix to orthogonal axes. Although this methodology is frequently used in multivariate analyses, it has disadvantages when applied to experimental data. First, the identified principal components have poor generality; since the size and directions of the components are dependent on the particular data set, the components are valid only within the data set. Second, the method is sensitive to experimental noise and bias between sample groups. It cannot reflect the experimental design that is planned to manage the noise and bias; rather, it estimates the same weight and independence to all the samples in the matrix. Third, the resulting components are often difficult to interpret. To address these issues, several options were introduced to the methodology. First, the principal axes were identified using training data sets and shared across experiments. These training data reflect the design of experiments, and their preparation allows noise to be reduced and group bias to be removed. Second, the center of the rotation was determined in accordance with the experimental design. Third, the resulting components were scaled to unify their size unit. Results The effects of these options were observed in microarray experiments, and showed an improvement in the separation of groups and robustness to noise. The range of scaled scores was unaffected by the number of items. Additionally, unknown samples were appropriately classified using pre-arranged axes. Furthermore, these axes well reflected the characteristics of groups in the experiments. As was observed, the scaling of the components and sharing of axes enabled comparisons of the components beyond experiments. The use of training data reduced the effects of noise and bias in the data, facilitating the physical interpretation of the
Analysis of Transition-Sensitized Turbulent Transport Equations
NASA Technical Reports Server (NTRS)
Rumsey, Christopher L.; Thacker, William D.; Gatski, Thomas B.; Grosch, Chester E,
2005-01-01
The dynamics of an ensemble of linear disturbances in boundary-layer flows at various Reynolds numbers is studied through an analysis of the transport equations for the mean disturbance kinetic energy and energy dissipation rate. Effects of adverse and favorable pressure-gradients on the disturbance dynamics are also included in the analysis Unlike the fully turbulent regime where nonlinear phase scrambling of the fluctuations affects the flow field even in proximity to the wall, the early stage transition regime fluctuations studied here are influenced cross the boundary layer by the solid boundary. The dominating dynamics in the disturbance kinetic energy and dissipation rate equations are described. These results are then used to formulate transition-sensitized turbulent transport equations, which are solved in a two-step process and applied to zero-pressure-gradient flow over a flat plate. Computed results are in good agreement with experimental data.
Comparative Analysis of State Fish Consumption Advisories Targeting Sensitive Populations
Scherer, Alison C.; Tsuchiya, Ami; Younglove, Lisa R.; Burbacher, Thomas M.; Faustman, Elaine M.
2008-01-01
Objective Fish consumption advisories are issued to warn the public of possible toxicological threats from consuming certain fish species. Although developing fetuses and children are particularly susceptible to toxicants in fish, fish also contain valuable nutrients. Hence, formulating advice for sensitive populations poses challenges. We conducted a comparative analysis of advisory Web sites issued by states to assess health messages that sensitive populations might access. Data sources We evaluated state advisories accessed via the National Listing of Fish Advisories issued by the U.S. Environmental Protection Agency. Data extraction We created criteria to evaluate advisory attributes such as risk and benefit message clarity. Data synthesis All 48 state advisories issued at the time of this analysis targeted children, 90% (43) targeted pregnant women, and 58% (28) targeted women of childbearing age. Only six advisories addressed single contaminants, while the remainder based advice on 2–12 contaminants. Results revealed that advisories associated a dozen contaminants with specific adverse health effects. Beneficial health effects of any kind were specifically associated only with omega-3 fatty acids found in fish. Conclusions These findings highlight the complexity of assessing and communicating information about multiple contaminant exposure from fish consumption. Communication regarding potential health benefits conferred by specific fish nutrients was minimal and focused primarily on omega-3 fatty acids. This overview suggests some lessons learned and highlights a lack of both clarity and consistency in providing the breadth of information that sensitive populations such as pregnant women need to make public health decisions about fish consumption during pregnancy. PMID:19079708
Global sensitivity analysis of the XUV-ABLATOR code
NASA Astrophysics Data System (ADS)
Nevrlý, Václav; Janku, Jaroslav; Dlabka, Jakub; Vašinek, Michal; Juha, Libor; Vyšín, Luděk.; Burian, Tomáš; Lančok, Ján.; Skřínský, Jan; Zelinger, Zdeněk.; Pira, Petr; Wild, Jan
2013-05-01
Availability of numerical model providing reliable estimation of the parameters of ablation processes induced by extreme ultraviolet laser pulses in the range of nanosecond and sub-picosecond timescales is highly desirable for recent experimental research as well as for practical purposes. Performance of the one-dimensional thermodynamic code (XUV-ABLATOR) in predicting the relationship of ablation rate and laser fluence is investigated for three reference materials: (i) silicon, (ii) fused silica and (iii) polymethyl methacrylate. The effect of pulse duration and different material properties on the model predictions is studied in the frame of this contribution for the conditions typical for two compact laser systems operating at 46.9 nm. Software implementation of the XUV-ABLATOR code including graphical user's interface and the set of tools for sensitivity analysis was developed. Global sensitivity analysis using high dimensional model representation in combination with quasi-random sampling was applied in order to identify the most critical input data as well as to explore the uncertainty range of model results.
Simple Sensitivity Analysis for Orion Guidance Navigation and Control
NASA Technical Reports Server (NTRS)
Pressburger, Tom; Hoelscher, Brian; Martin, Rodney; Sricharan, Kumar
2013-01-01
The performance of Orion flight software, especially its GNC software, is being analyzed by running Monte Carlo simulations of Orion spacecraft flights. The simulated performance is analyzed for conformance with flight requirements, expressed as performance constraints. Flight requirements include guidance (e.g. touchdown distance from target) and control (e.g., control saturation) as well as performance (e.g., heat load constraints). The Monte Carlo simulations disperse hundreds of simulation input variables, for everything from mass properties to date of launch. We describe in this paper a sensitivity analysis tool ("Critical Factors Tool" or CFT) developed to find the input variables or pairs of variables which by themselves significantly influence satisfaction of requirements or significantly affect key performance metrics (e.g., touchdown distance from target). Knowing these factors can inform robustness analysis, can inform where engineering resources are most needed, and could even affect operations. The contributions of this paper include the introduction of novel sensitivity measures, such as estimating success probability, and a technique for determining whether pairs of factors are interacting dependently or independently. The tool found that input variables such as moments, mass, thrust dispersions, and date of launch were found to be significant factors for success of various requirements. Examples are shown in this paper as well as a summary and physics discussion of EFT-1 driving factors that the tool found.
Trends in sensitivity analysis practice in the last decade.
Ferretti, Federico; Saltelli, Andrea; Tarantola, Stefano
2016-10-15
The majority of published sensitivity analyses (SAs) are either local or one factor-at-a-time (OAT) analyses, relying on unjustified assumptions of model linearity and additivity. Global approaches to sensitivity analyses (GSA) which would obviate these shortcomings, are applied by a minority of researchers. By reviewing the academic literature on SA, we here present a bibliometric analysis of the trends of different SA practices in last decade. The review has been conducted both on some top ranking journals (Nature and Science) and through an extended analysis in the Elsevier's Scopus database of scientific publications. After correcting for the global growth in publications, the amount of papers performing a generic SA has notably increased over the last decade. Even if OAT is still the most largely used technique in SA, there is a clear increase in the use of GSA with preference respectively for regression and variance-based techniques. Even after adjusting for the growth of publications in the sole modelling field, to which SA and GSA normally apply, the trend is confirmed. Data about regions of origin and discipline are also briefly discussed. The results above are confirmed when zooming on the sole articles published in chemical modelling, a field historically proficient in the use of SA methods. PMID:26934843
Flat heat pipe design, construction, and analysis
Voegler, G.; Boughey, B.; Cerza, M.; Lindler, K.W.
1999-08-02
This paper details the design, construction and partial analysis of a low temperature flat heat pipe in order to determine the feasibility of implementing flat heat pipes into thermophotovoltaic (TPV) energy conversion systems.
Theoretical Noise Analysis on a Position-sensitive Metallic Magnetic Calorimeter
NASA Technical Reports Server (NTRS)
Smith, Stephen J.
2007-01-01
We report on the theoretical noise analysis for a position-sensitive Metallic Magnetic Calorimeter (MMC), consisting of MMC read-out at both ends of a large X-ray absorber. Such devices are under consideration as alternatives to other cryogenic technologies for future X-ray astronomy missions. We use a finite-element model (FEM) to numerically calculate the signal and noise response at the detector outputs and investigate the correlations between the noise measured at each MMC coupled by the absorber. We then calculate, using the optimal filter concept, the theoretical energy and position resolution across the detector and discuss the trade-offs involved in optimizing the detector design for energy resolution, position resolution and count rate. The results show, theoretically, the position-sensitive MMC concept offers impressive spectral and spatial resolving capabilities compared to pixel arrays and similar position-sensitive cryogenic technologies using Transition Edge Sensor (TES) read-out.
NDARC NASA Design and Analysis of Rotorcraft
NASA Technical Reports Server (NTRS)
Johnson, Wayne R.
2009-01-01
The NASA Design and Analysis of Rotorcraft (NDARC) software is an aircraft system analysis tool intended to support both conceptual design efforts and technology impact assessments. The principal tasks are to design (or size) a rotorcraft to meet specified requirements, including vertical takeoff and landing (VTOL) operation, and then analyze the performance of the aircraft for a set of conditions. For broad and lasting utility, it is important that the code have the capability to model general rotorcraft configurations, and estimate the performance and weights of advanced rotor concepts. The architecture of the NDARC code accommodates configuration flexibility; a hierarchy of models; and ultimately multidisciplinary design, analysis, and optimization. Initially the software is implemented with lowfidelity models, typically appropriate for the conceptual design environment. An NDARC job consists of one or more cases, each case optionally performing design and analysis tasks. The design task involves sizing the rotorcraft to satisfy specified design conditions and missions. The analysis tasks can include off-design mission performance calculation, flight performance calculation for point operating conditions, and generation of subsystem or component performance maps. For analysis tasks, the aircraft description can come from the sizing task, from a previous case or a previous NDARC job, or be independently generated (typically the description of an existing aircraft). The aircraft consists of a set of components, including fuselage, rotors, wings, tails, and propulsion. For each component, attributes such as performance, drag, and weight can be calculated; and the aircraft attributes are obtained from the sum of the component attributes. Description and analysis of conventional rotorcraft configurations is facilitated, while retaining the capability to model novel and advanced concepts. Specific rotorcraft configurations considered are single main-rotor and
NDARC - NASA Design and Analysis of Rotorcraft
NASA Technical Reports Server (NTRS)
Johnson, Wayne
2015-01-01
The NASA Design and Analysis of Rotorcraft (NDARC) software is an aircraft system analysis tool that supports both conceptual design efforts and technology impact assessments. The principal tasks are to design (or size) a rotorcraft to meet specified requirements, including vertical takeoff and landing (VTOL) operation, and then analyze the performance of the aircraft for a set of conditions. For broad and lasting utility, it is important that the code have the capability to model general rotorcraft configurations, and estimate the performance and weights of advanced rotor concepts. The architecture of the NDARC code accommodates configuration flexibility, a hierarchy of models, and ultimately multidisciplinary design, analysis, and optimization. Initially the software is implemented with low-fidelity models, typically appropriate for the conceptual design environment. An NDARC job consists of one or more cases, each case optionally performing design and analysis tasks. The design task involves sizing the rotorcraft to satisfy specified design conditions and missions. The analysis tasks can include off-design mission performance calculation, flight performance calculation for point operating conditions, and generation of subsystem or component performance maps. For analysis tasks, the aircraft description can come from the sizing task, from a previous case or a previous NDARC job, or be independently generated (typically the description of an existing aircraft). The aircraft consists of a set of components, including fuselage, rotors, wings, tails, and propulsion. For each component, attributes such as performance, drag, and weight can be calculated; and the aircraft attributes are obtained from the sum of the component attributes. Description and analysis of conventional rotorcraft configurations is facilitated, while retaining the capability to model novel and advanced concepts. Specific rotorcraft configurations considered are single-main-rotor and tail
Sensitivity analysis of infectious disease models: methods, advances and their application.
Wu, Jianyong; Dhingra, Radhika; Gambhir, Manoj; Remais, Justin V
2013-09-01
Sensitivity analysis (SA) can aid in identifying influential model parameters and optimizing model structure, yet infectious disease modelling has yet to adopt advanced SA techniques that are capable of providing considerable insights over traditional methods. We investigate five global SA methods-scatter plots, the Morris and Sobol' methods, Latin hypercube sampling-partial rank correlation coefficient and the sensitivity heat map method-and detail their relative merits and pitfalls when applied to a microparasite (cholera) and macroparasite (schistosomaisis) transmission model. The methods investigated yielded similar results with respect to identifying influential parameters, but offered specific insights that vary by method. The classical methods differed in their ability to provide information on the quantitative relationship between parameters and model output, particularly over time. The heat map approach provides information about the group sensitivity of all model state variables, and the parameter sensitivity spectrum obtained using this method reveals the sensitivity of all state variables to each parameter over the course of the simulation period, especially valuable for expressing the dynamic sensitivity of a microparasite epidemic model to its parameters. A summary comparison is presented to aid infectious disease modellers in selecting appropriate methods, with the goal of improving model performance and design. PMID:23864497
Sensitivity analysis of infectious disease models: methods, advances and their application
Wu, Jianyong; Dhingra, Radhika; Gambhir, Manoj; Remais, Justin V.
2013-01-01
Sensitivity analysis (SA) can aid in identifying influential model parameters and optimizing model structure, yet infectious disease modelling has yet to adopt advanced SA techniques that are capable of providing considerable insights over traditional methods. We investigate five global SA methods—scatter plots, the Morris and Sobol’ methods, Latin hypercube sampling-partial rank correlation coefficient and the sensitivity heat map method—and detail their relative merits and pitfalls when applied to a microparasite (cholera) and macroparasite (schistosomaisis) transmission model. The methods investigated yielded similar results with respect to identifying influential parameters, but offered specific insights that vary by method. The classical methods differed in their ability to provide information on the quantitative relationship between parameters and model output, particularly over time. The heat map approach provides information about the group sensitivity of all model state variables, and the parameter sensitivity spectrum obtained using this method reveals the sensitivity of all state variables to each parameter over the course of the simulation period, especially valuable for expressing the dynamic sensitivity of a microparasite epidemic model to its parameters. A summary comparison is presented to aid infectious disease modellers in selecting appropriate methods, with the goal of improving model performance and design. PMID:23864497
Airbreathing hypersonic vehicle design and analysis methods
NASA Technical Reports Server (NTRS)
Lockwood, Mary Kae; Petley, Dennis H.; Hunt, James L.; Martin, John G.
1996-01-01
The design, analysis, and optimization of airbreathing hypersonic vehicles requires analyses involving many highly coupled disciplines at levels of accuracy exceeding those traditionally considered in a conceptual or preliminary-level design. Discipline analysis methods including propulsion, structures, thermal management, geometry, aerodynamics, performance, synthesis, sizing, closure, and cost are discussed. Also, the on-going integration of these methods into a working environment, known as HOLIST, is described.
GALACSI system design and analysis
NASA Astrophysics Data System (ADS)
Ströbele, S.; La Penna, P.; Arsenault, R.; Conzelmann, R. D.; Delabre, B.; Duchateau, M.; Dorn, R.; Fedrigo, E.; Hubin, N.; Quentin, J.; Jolley, P.; Kiekebusch, M.; Kirchbauer, J. P.; Klein, B.; Kolb, J.; Kuntschner, H.; Le Louarn, M.; Lizon, J. L.; Madec, P.-Y.; Pettazzi, L.; Soenke, C.; Tordo, S.; Vernet, J.; Muradore, R.
2012-07-01
GALACSI is one of the Adaptive Optics (AO) systems part of the ESO Adaptive Optics Facility (AOF). It will use the VLT 4-Laser Guide Stars system, high speed and low noise WaveFront Sensor cameras (<1e-, 1000Hz) the Deformable Secondary Mirror (DSM) and the SPARTA Real Time Computer to sharpen images and enhance faint object detectability of the MUSE Instrument. MUSE is an Integral Field Spectrograph working at wavelengths from 465nm to 930nm. GALACSI implements 2 different AO modes; in Wide Field Mode (WFM) it will perform Ground Layer AO correction and enhance the collected energy in a 0.2" by 0.2" pixel by a factor 2 at 750nm over a Field of View (FoV) of 1' by 1'. The 4 LGSs and one tip tilt reference star (R-mag <17.5) are located outside the MUSE FoV. Key requirements are to provide this performance and a very good image stability for a 1hour long integration time. In Narrow Field Mode (NFM) Laser Tomography AO will be used to reconstruct and correct the turbulence for the center field using the 4 LGSs at 15" off axis and the Near Infra Red (NIR) light of one reference star on axis for tip tilt and focus sensing. In NFM GALACSI will provide a moderate Strehl Ratio of 5% (goal 10%) at 650nm. The NFM hosts several challenges and many subsystems will be pushed to their limits. The opto mechanical design and error budgets of GALACSI is described here.
DESIGN PACKAGE 1D SYSTEM SAFETY ANALYSIS
L.R. Eisler
1995-02-02
The purpose of this analysis is to systematically identify and evaluate hazards related to the Yucca Mountain Project Exploratory Studies Facility (ESF) Design Package 1D, Surface Facilities, (for a list of design items included in the package 1D system safety analysis see section 3). This process is an integral part of the systems engineering process; whereby safety is considered during planning, design, testing, and construction. A largely qualitative approach was used since a radiological System Safety analysis is not required. The risk assessment in this analysis characterizes the accident scenarios associated with the Design Package 1D structures/systems/components in terms of relative risk and includes recommendations for mitigating all identified risks. The priority for recommending and implementing mitigation control features is: (1) Incorporate measures to reduce risks and hazards into the structure/system/component (S/S/C) design, (2) add safety devices and capabilities to the designs that reduce risk, (3) provide devices that detect and warn personnel of hazardous conditions, and (4) develop procedures and conduct training to increase worker awareness of potential hazards, on methods to reduce exposure to hazards, and on the actions required to avoid accidents or correct hazardous conditions. The scope of this analysis is limited to the Design Package 1D structures/systems/components (S/S/Cs) during normal operations excluding hazards occurring during maintenance and ''off normal'' operations.
Design and Analysis Tools for Supersonic Inlets
NASA Technical Reports Server (NTRS)
Slater, John W.; Folk, Thomas C.
2009-01-01
Computational tools are being developed for the design and analysis of supersonic inlets. The objective is to update existing tools and provide design and low-order aerodynamic analysis capability for advanced inlet concepts. The Inlet Tools effort includes aspects of creating an electronic database of inlet design information, a document describing inlet design and analysis methods, a geometry model for describing the shape of inlets, and computer tools that implement the geometry model and methods. The geometry model has a set of basic inlet shapes that include pitot, two-dimensional, axisymmetric, and stream-traced inlet shapes. The inlet model divides the inlet flow field into parts that facilitate the design and analysis methods. The inlet geometry model constructs the inlet surfaces through the generation and transformation of planar entities based on key inlet design factors. Future efforts will focus on developing the inlet geometry model, the inlet design and analysis methods, a Fortran 95 code to implement the model and methods. Other computational platforms, such as Java, will also be explored.
Cruyff, Maarten J L F; Böckenholt, Ulf; van der Heijden, Peter G M
2016-03-01
The conventional randomized response design is unidimensional in the sense that it measures a single dimension of a sensitive attribute, like its prevalence, frequency, magnitude, or duration. This paper introduces a multidimensional design characterized by categorical questions that each measure a different aspect of the same sensitive attribute. The benefits of the multidimensional design are (i) a substantial gain in power and efficiency, and the potential to (i i) evaluate the goodness-of-fit of the model, and (i i i) test hypotheses about evasive response biases in case of a misfit. The method is illustrated for a two-dimensional design measuring both the prevalence and the magnitude of social security fraud. PMID:25877782
Using an Adjoint Approach to Eliminate Mesh Sensitivities in Computational Design
NASA Technical Reports Server (NTRS)
Nielsen, Eric J.; Park, Michael A.
2006-01-01
An algorithm for efficiently incorporating the effects of mesh sensitivities in a computational design framework is introduced. The method is based on an adjoint approach and eliminates the need for explicit linearizations of the mesh movement scheme with respect to the geometric parameterization variables, an expense that has hindered practical large-scale design optimization using discrete adjoint methods. The effects of the mesh sensitivities can be accounted for through the solution of an adjoint problem equivalent in cost to a single mesh movement computation, followed by an explicit matrix-vector product scaling with the number of design variables and the resolution of the parameterized surface grid. The accuracy of the implementation is established and dramatic computational savings obtained using the new approach are demonstrated using several test cases. Sample design optimizations are also shown.
Using an Adjoint Approach to Eliminate Mesh Sensitivities in Computational Design
NASA Technical Reports Server (NTRS)
Nielsen, Eric J.; Park, Michael A.
2005-01-01
An algorithm for efficiently incorporating the effects of mesh sensitivities in a computational design framework is introduced. The method is based on an adjoint approach and eliminates the need for explicit linearizations of the mesh movement scheme with respect to the geometric parameterization variables, an expense that has hindered practical large-scale design optimization using discrete adjoint methods. The effects of the mesh sensitivities can be accounted for through the solution of an adjoint problem equivalent in cost to a single mesh movement computation, followed by an explicit matrix-vector product scaling with the number of design variables and the resolution of the parameterized surface grid. The accuracy of the implementation is established and dramatic computational savings obtained using the new approach are demonstrated using several test cases. Sample design optimizations are also shown.
NASA Technical Reports Server (NTRS)
Edwards, S.; Reuther, J.; Chattot, J. J.
1997-01-01
The objective of this paper is to present a control theory approach for the design of airfoils in the presence of viscous compressible flows. A coupled system of the integral boundary layer and the Euler equations is solved to provide rapid flow simulations. An adjunct approach consistent with the complete coupled state equations is employed to obtain the sensitivities needed to drive a numerical optimization algorithm. Design to target pressure distribution is demonstrated on an RAE 2822 airfoil at transonic speed.
NASA Astrophysics Data System (ADS)
Edwards, S.; Reuther, J.; Chattot, J. J.
The objective of this paper is to present a control theory approach for the design of airfoils in the presence of viscous compressible flows. A coupled system of the integral boundary layer and the Euler equations is solved to provide rapid flow simulations. An adjoint approach consistent with the complete coupled state equations is employed to obtain the sensitivities needed to drive a numerical optimization algorithm. Design to a target pressure distribution is demonstrated on an RAE 2822 airfoil at transonic speeds.
GPU-based Integration with Application in Sensitivity Analysis
NASA Astrophysics Data System (ADS)
Atanassov, Emanouil; Ivanovska, Sofiya; Karaivanova, Aneta; Slavov, Dimitar
2010-05-01
The presented work is an important part of the grid application MCSAES (Monte Carlo Sensitivity Analysis for Environmental Studies) which aim is to develop an efficient Grid implementation of a Monte Carlo based approach for sensitivity studies in the domains of Environmental modelling and Environmental security. The goal is to study the damaging effects that can be caused by high pollution levels (especially effects on human health), when the main modeling tool is the Danish Eulerian Model (DEM). Generally speaking, sensitivity analysis (SA) is the study of how the variation in the output of a mathematical model can be apportioned to, qualitatively or quantitatively, different sources of variation in the input of a model. One of the important classes of methods for Sensitivity Analysis are Monte Carlo based, first proposed by Sobol, and then developed by Saltelli and his group. In MCSAES the general Saltelli procedure has been adapted for SA of the Danish Eulerian model. In our case we consider as factors the constants determining the speeds of the chemical reactions in the DEM and as output a certain aggregated measure of the pollution. Sensitivity simulations lead to huge computational tasks (systems with up to 4 × 109 equations at every time-step, and the number of time-steps can be more than a million) which motivates its grid implementation. MCSAES grid implementation scheme includes two main tasks: (i) Grid implementation of the DEM, (ii) Grid implementation of the Monte Carlo integration. In this work we present our new developments in the integration part of the application. We have developed an algorithm for GPU-based generation of scrambled quasirandom sequences which can be combined with the CPU-based computations related to the SA. Owen first proposed scrambling of Sobol sequence through permutation in a manner that improves the convergence rates. Scrambling is necessary not only for error analysis but for parallel implementations. Good scrambling is
Adjoint sensitivity analysis of hydrodynamic stability in cyclonic flows
NASA Astrophysics Data System (ADS)
Guzman Inigo, Juan; Juniper, Matthew
2015-11-01
Cyclonic separators are used in a variety of industries to efficiently separate mixtures of fluid and solid phases by means of centrifugal forces and gravity. In certain circumstances, the vortex core of cyclonic flows is known to precess due to the instability of the flow, which leads to performance reductions. We aim to characterize the unsteadiness using linear stability analysis of the Reynolds Averaged Navier-Stokes (RANS) equations in a global framework. The system of equations, including the turbulence model, is linearised to obtain an eigenvalue problem. Unstable modes corresponding to the dynamics of the large structures of the turbulent flow are extracted. The analysis shows that the most unstable mode is a helical motion which develops around the axis of the flow. This result is in good agreement with LES and experimental analysis, suggesting the validity of the approach. Finally, an adjoint-based sensitivity analysis is performed to determine the regions of the flow that, when altered, have most influence on the frequency and growth-rate of the unstable eigenvalues.
NASA Astrophysics Data System (ADS)
Benson, James; Ziehn, Tilo; Dixon, Nick S.; Tomlin, Alison S.
In this work global sensitivity studies using Monte Carlo sampling and high dimensional model representations (HDMR) have been carried out on the k- ɛ closure computational fluid dynamic (CFD) model MISKAM, allowing detailed representation of the effects of changing input parameters on the model outputs. The scenario studied is that of a complex street canyon in the city of York, UK. The sensitivity of the turbulence and mean flow fields to the input parameters is detailed both at specific measurement points and in the associated canyon cross-section to aid comparison with field data. This analysis gives insight into how model parameters can influence the predicted outputs. It also shows the relative strength of each parameter in its influence. Four main input parameters are addressed. Three parameters are surface roughness lengths, determining the flow over a surface, and the fourth is the background wind direction. In order to determine the relative importance of each parameter, sensitivity indices are calculated for the canyon cross-section. The sensitivity of the flow structures in and above the canyon to each parameter is found to be very location dependant. In general, at a particular measurement point, it is the closest wall surface that is most influential on the model output. However, due to the complexity of the flow at different wind angles this is not always the case, for example when a re-circulating canyon flow pattern is present. The background wind direction is shown to be an important parameter as it determines the surface features encountered by the flow. The accuracy with which this is specified when modelling a full-scale situation is therefore an important consideration when considering model uncertainty. Overall, the uncertainty due to roughness lengths is small in comparison to the mean outputs, indicating that the model is well defined even with large ranges of input parameter uncertainty.
Least Squares Shadowing Sensitivity Analysis of Chaotic and Turbulent Fluid Flows
NASA Astrophysics Data System (ADS)
Blonigan, Patrick; Wang, Qiqi; Gomez, Steven
2013-11-01
Computational methods for sensitivity analysis are invaluable tools for fluid dynamics research and engineering design. These methods are used in many applications, including aerodynamic shape optimization and adaptive grid refinement. However, traditional sensitivity analysis methods break down when applied to long-time averaged quantities in chaotic fluid flow fields, such as those obtained using high-fidelity turbulence simulations. This break down is due to the ``Butterfly Effect'' the high sensitivity of chaotic dynamical systems to the initial condition. A new sensitivity analysis method developed by the authors, Least Squares Shadowing (LSS), can compute useful and accurate gradients for quantities of interest in chaotic and turbulent fluid flows. LSS computes gradients using the ``shadow trajectory,'' a phase space trajectory (or solution) for which perturbations to the flow field do not grow exponentially in time. This talk will outline Least Squares Shadowing and demonstrate it on several chaotic and turbulent fluid flows, including homogeneous isotropic turbulence, Rayleigh-Bénard convection and turbulent channel flow. We would like to acknowledge AFSOR Award F11B-T06-0007 under Dr. Fariba Fahroo, NASA Award NNH11ZEA001N under Dr. Harold Atkins, as well as financial support from ConocoPhillips, the NDSEG fellowship and the ANSYS Fellowship.
Zhao, Dian; Rao, Xingtang; Yu, Jiancan; Cui, Yuanjing; Yang, Yu; Qian, Guodong
2015-12-01
An important result of research on mixed-lanthanide metal-organic frameworks (M'LnMOFs) is the realization of highly sensitive ratiometric luminescent thermometers. Here, we report the design and synthesis of the new M'LnMOF Tb0.80Eu0.20BPDA with high relative sensitivity in the physiological temperature regime (298-318 K). The emission intensity and luminescence lifetime were investigated and compared to those of existing materials. It was found that the temperature-dependent luminescence properties of Tb0.80Eu0.20BPDA are strongly associated with the distribution of the energy levels of the ligand. Such a property can be useful in the design of highly sensitive M'LnMOF thermometers. PMID:26575207
SAFE(R): A Matlab/Octave Toolbox (and R Package) for Global Sensitivity Analysis
NASA Astrophysics Data System (ADS)
Pianosi, Francesca; Sarrazin, Fanny; Gollini, Isabella; Wagener, Thorsten
2015-04-01
Global Sensitivity Analysis (GSA) is increasingly used in the development and assessment of hydrological models, as well as for dominant control analysis and for scenario discovery to support water resource management under deep uncertainty. Here we present a toolbox for the application of GSA, called SAFE (Sensitivity Analysis For Everybody) that implements several established GSA methods, including method of Morris, Regional Sensitivity Analysis, variance-based sensitivity Analysis (Sobol') and FAST. It also includes new approaches and visualization tools to complement these established methods. The Toolbox is released in two versions, one running under Matlab/Octave (called SAFE) and one running in R (called SAFER). Thanks to its modular structure, SAFE(R) can be easily integrated with other toolbox and packages, and with models running in a different computing environment. Another interesting feature of SAFE(R) is that all the implemented methods include specific functions for assessing the robustness and convergence of the sensitivity estimates. Furthermore, SAFE(R) includes numerous visualisation tools for the effective investigation and communication of GSA results. The toolbox is designed to make GSA accessible to non-specialist users, and to provide a fully commented code for more experienced users to complement their own tools. The documentation includes a set of workflow scripts with practical guidelines on how to apply GSA and how to use the toolbox. SAFE(R) is open source and freely available from the following website: http://bristol.ac.uk/cabot/resources/safe-toolbox/ Ultimately, SAFE(R) aims at improving the diffusion and quality of GSA practice in the hydrological modelling community.
NASA Technical Reports Server (NTRS)
Liberty, S. R.; Maynard, R. A.; Mielke, R. R.
1979-01-01
This brief paper summarizes the approach the authors will take in designing a feedback controller for the F-100 turbofan engine. The technique to be utilized simultaneously realizes dominant closed-loop eigenvalues, approximates specified modal behavior, and achieves low eigensystem sensitivity with respect to certain plant parameter variations.
Design of a portable fluoroquinolone analyzer based on terbium-sensitized luminescence
Technology Transfer Automated Retrieval System (TEKTRAN)
A portable fluoroquinolone (FQ) analyzer is developed in this laboratory based on terbium-sensitized luminescence (TSL). The optical, hardware and software design aspects are described in detail. A 327-nm light emitting diode (LED) is used in pulsed mode as the excitation source; and a photomultip...
Aeroelastic optimization of a helicopter rotor using an efficient sensitivity analysis
NASA Technical Reports Server (NTRS)
Lim, Joon W.; Chopra, Inderjit
1990-01-01
To reduce oscillatory hub loads in forward flight, a structural optimization analysis of a hingeless helicopter rotor has been developed and applied. The aeroelastic analysis of the rotor is based on a finite element method in space and time, and linked with automated optimization algorithms. For the optimization analysis two types of structural representation are used: a generic stiffness-distribution and a single-cell thin-walled beam. For the first type, the design variables are nonstructural mass and its placement, chordwise center of gravity offset from the elastic axis, and stiffness. For the second type, width, height and thickness of spar are used as design variables. For the behavior constraints, frequency placement, autorotational inertia and aeroelastic stability of the blade are included. The required sensitivity derivatives are obtained using a direct analytical approach. An optimum oscillatory hub load shows a 25-77 percent reduction for the generic blade, and 30-50 percent reduction for the box-beam.
Sensitivity analysis of channel-bend hydraulics influenced by vegetation
NASA Astrophysics Data System (ADS)
Bywater-Reyes, S.; Manners, R.; McDonald, R.; Wilcox, A. C.
2015-12-01
Alternating bars influence hydraulics by changing the force balance of channels as part of a morphodynamic feedback loop that dictates channel geometry. Pioneer woody riparian trees recruit on river bars and may steer flow, alter cross-stream and downstream force balances, and ultimately change channel morphology. Quantifying the influence of vegetation on stream hydraulics is difficult, and researchers increasingly rely on two-dimensional hydraulic models. In many cases, channel characteristics (channel drag and lateral eddy viscosity) and vegetation characteristics (density, frontal area, and drag coefficient) are uncertain. This study uses a beta version of FaSTMECH that models vegetation explicitly as a drag force to test the sensitivity of channel-bend hydraulics to riparian vegetation. We use a simplified, scale model of a meandering river with bars and conduct a global sensitivity analysis that ranks the influence of specified channel characteristics (channel drag and lateral eddy viscosity) against vegetation characteristics (density, frontal area, and drag coefficient) on cross-stream hydraulics. The primary influence on cross-stream velocity and shear stress is channel drag (i.e., bed roughness), followed by the near-equal influence of all vegetation parameters and lateral eddy viscosity. To test the implication of the sensitivity indices on bend hydraulics, we hold calibrated channel characteristics constant for a wandering gravel-bed river with bars (Bitterroot River, MT), and vary vegetation parameters on a bar. For a dense vegetation scenario, we find flow to be steered away from the bar, and velocity and shear stress to be reduced within the thalweg. This provides insight into how the morphodynamic evolution of vegetated bars differs from unvegetated bars.
Global sensitivity analysis of the radiative transfer model
NASA Astrophysics Data System (ADS)
Neelam, Maheshwari; Mohanty, Binayak P.
2015-04-01
With the recently launched Soil Moisture Active Passive (SMAP) mission, it is very important to have a complete understanding of the radiative transfer model for better soil moisture retrievals and to direct future research and field campaigns in areas of necessity. Because natural systems show great variability and complexity with respect to soil, land cover, topography, precipitation, there exist large uncertainties and heterogeneities in model input factors. In this paper, we explore the possibility of using global sensitivity analysis (GSA) technique to study the influence of heterogeneity and uncertainties in model inputs on zero order radiative transfer (ZRT) model and to quantify interactions between parameters. GSA technique is based on decomposition of variance and can handle nonlinear and nonmonotonic functions. We direct our analyses toward growing agricultural fields of corn and soybean in two different regions, Iowa, USA (SMEX02) and Winnipeg, Canada (SMAPVEX12). We noticed that, there exists a spatio-temporal variation in parameter interactions under different soil moisture and vegetation conditions. Radiative Transfer Model (RTM) behaves more non-linearly in SMEX02 and linearly in SMAPVEX12, with average parameter interactions of 14% in SMEX02 and 5% in SMAPVEX12. Also, parameter interactions increased with vegetation water content (VWC) and roughness conditions. Interestingly, soil moisture shows an exponentially decreasing sensitivity function whereas parameters such as root mean square height (RMS height) and vegetation water content show increasing sensitivity with 0.05 v/v increase in soil moisture range. Overall, considering the SMAPVEX12 fields to be water rich environment (due to higher observed SM) and SMEX02 fields to be energy rich environment (due to lower SM and wide ranges of TSURF), our results indicate that first order as well as interactions between the parameters change with water and energy rich environments.
Design of a gaze-sensitive virtual social interactive system for children with autism.
Lahiri, Uttama; Warren, Zachary; Sarkar, Nilanjan
2011-08-01
Impairments in social communication skills are thought to be core deficits in children with autism spectrum disorder (ASD). In recent years, several assistive technologies, particularly Virtual Reality (VR), have been investigated to promote social interactions in this population. It is well known that children with ASD demonstrate atypical viewing patterns during social interactions and thus monitoring eye-gaze can be valuable to design intervention strategies. While several studies have used eye-tracking technology to monitor eye-gaze for offline analysis, there exists no real-time system that can monitor eye-gaze dynamically and provide individualized feedback. Given the promise of VR-based social interaction and the usefulness of monitoring eye-gaze in real-time, a novel VR-based dynamic eye-tracking system is developed in this work. This system, called Virtual Interactive system with Gaze-sensitive Adaptive Response Technology (VIGART), is capable of delivering individualized feedback based on a child's dynamic gaze patterns during VR-based interaction. Results from a usability study with six adolescents with ASD are presented that examines the acceptability and usefulness of VIGART. The results in terms of improvement in behavioral viewing and changes in relevant eye physiological indexes of participants while interacting with VIGART indicate the potential of this novel technology. PMID:21609889
Kumar, Pankaj; Ganure, Ashok Laxmanrao; Subudhi, Bharat Bhushan; Shukla, Shubhanjali
2015-01-01
The present investigation deals with the development of controlled release tablets of salbutamol sulphate using graft copolymers (St-g-PMMA and Ast-g-PMMA) of starch and acetylated starch. Drug excipient compatibility was spectroscopically analyzed via FT-IR, which confirmed no interaction between drug and other excipients. Formulations were evaluated for physical characteristics like hardness, friability, weight variations, drug release and drug content analysis which satisfies all the pharmacopoeial requirement of tablet dosage form. Release rate of a model drug from formulated matrix tablets were studied at two different pH namely 1.2 and 6.8, spectrophotometrically. Drug release from the tablets of graft copolymer matrices is profoundly pH-dependent and showed a reduced release rate under acidic conditions as compared to the alkaline conditions. Study of release mechanism by Korsmeyer's model with n values between 0.61-0.67, proved that release was governed by both diffusion and erosion. In comparison to starch and acetylated starch matrix formulations, pharmacokinetic parameters of graft copolymers matrix formulations showed a significant decrease in Cmax with an increase in tmax, indicating the effect of dosage form would last for longer duration. The gastro intestinal transit behavior of the formulation was determined by gamma scintigraphy, using (99m)Tc as a marker in healthy rabbits. The amount of radioactive tracer released from the labelled tablets was minimal when the tablets were in the stomach, whereas it increased as tablets reached to intestine. Thus, in-vitro and in-vivo drug release studies of starch-acrylate graft copolymers proved their controlled release behavior with preferential delivery into alkaline pH environment. PMID:26330856
Kumar, Pankaj; Ganure, Ashok Laxmanrao; Subudhi, Bharat Bhushan; Shukla, Shubhanjali
2015-01-01
The present investigation deals with the development of controlled release tablets of salbutamol sulphate using graft copolymers (St-g-PMMA and Ast-g-PMMA) of starch and acetylated starch. Drug excipient compatibility was spectroscopically analyzed via FT-IR, which confirmed no interaction between drug and other excipients. Formulations were evaluated for physical characteristics like hardness, friability, weight variations, drug release and drug content analysis which satisfies all the pharmacopoeial requirement of tablet dosage form. Release rate of a model drug from formulated matrix tablets were studied at two different pH namely 1.2 and 6.8, spectrophotometrically. Drug release from the tablets of graft copolymer matrices is profoundly pH-dependent and showed a reduced release rate under acidic conditions as compared to the alkaline conditions. Study of release mechanism by Korsmeyer’s model with n values between 0.61-0.67, proved that release was governed by both diffusion and erosion. In comparison to starch and acetylated starch matrix formulations, pharmacokinetic parameters of graft copolymers matrix formulations showed a significant decrease in Cmax with an increase in tmax, indicating the effect of dosage form would last for longer duration. The gastro intestinal transit behavior of the formulation was determined by gamma scintigraphy, using 99mTc as a marker in healthy rabbits. The amount of radioactive tracer released from the labelled tablets was minimal when the tablets were in the stomach, whereas it increased as tablets reached to intestine. Thus, in-vitro and in-vivo drug release studies of starch-acrylate graft copolymers proved their controlled release behavior with preferential delivery into alkaline pH environment. PMID:26330856
Structural analysis at aircraft conceptual design stage
NASA Astrophysics Data System (ADS)
Mansouri, Reza
In the past 50 years, computers have helped by augmenting human efforts with tremendous pace. The aircraft industry is not an exception. Aircraft industry is more than ever dependent on computing because of a high level of complexity and the increasing need for excellence to survive a highly competitive marketplace. Designers choose computers to perform almost every analysis task. But while doing so, existing effective, accurate and easy to use classical analytical methods are often forgotten, which can be very useful especially in the early phases of the aircraft design where concept generation and evaluation demands physical visibility of design parameters to make decisions [39, 2004]. Structural analysis methods have been used by human beings since the very early civilization. Centuries before computers were invented; the pyramids were designed and constructed by Egyptians around 2000 B.C, the Parthenon was built by the Greeks, around 240 B.C, Dujiangyan was built by the Chinese. Persepolis, Hagia Sophia, Taj Mahal, Eiffel tower are only few more examples of historical buildings, bridges and monuments that were constructed before we had any advancement made in computer aided engineering. Aircraft industry is no exception either. In the first half of the 20th century, engineers used classical method and designed civil transport aircraft such as Ford Tri Motor (1926), Lockheed Vega (1927), Lockheed 9 Orion (1931), Douglas DC-3 (1935), Douglas DC-4/C-54 Skymaster (1938), Boeing 307 (1938) and Boeing 314 Clipper (1939) and managed to become airborne without difficulty. Evidencing, while advanced numerical methods such as the finite element analysis is one of the most effective structural analysis methods; classical structural analysis methods can also be as useful especially during the early phase of a fixed wing aircraft design where major decisions are made and concept generation and evaluation demands physical visibility of design parameters to make decisions
Ensemble Solar Forecasting Statistical Quantification and Sensitivity Analysis: Preprint
Cheung, WanYin; Zhang, Jie; Florita, Anthony; Hodge, Bri-Mathias; Lu, Siyuan; Hamann, Hendrik F.; Sun, Qian; Lehman, Brad
2015-12-08
Uncertainties associated with solar forecasts present challenges to maintain grid reliability, especially at high solar penetrations. This study aims to quantify the errors associated with the day-ahead solar forecast parameters and the theoretical solar power output for a 51-kW solar power plant in a utility area in the state of Vermont, U.S. Forecasts were generated by three numerical weather prediction (NWP) models, including the Rapid Refresh, the High Resolution Rapid Refresh, and the North American Model, and a machine-learning ensemble model. A photovoltaic (PV) performance model was adopted to calculate theoretical solar power generation using the forecast parameters (e.g., irradiance, cell temperature, and wind speed). Errors of the power outputs were quantified using statistical moments and a suite of metrics, such as the normalized root mean squared error (NRMSE). In addition, the PV model's sensitivity to different forecast parameters was quantified and analyzed. Results showed that the ensemble model yielded forecasts in all parameters with the smallest NRMSE. The NRMSE of solar irradiance forecasts of the ensemble NWP model was reduced by 28.10% compared to the best of the three NWP models. Further, the sensitivity analysis indicated that the errors of the forecasted cell temperature attributed only approximately 0.12% to the NRMSE of the power output as opposed to 7.44% from the forecasted solar irradiance.
NASA Technical Reports Server (NTRS)
Sliwa, S. M.
1980-01-01
Constrained parameter optimization was used to perform the optimal preliminary design of a medium range transport configuration. The impact of choosing a performance index was studied and the required fare for a 15 percent return-on-investment was proposed as a figure-of-merit. A number of design constants and constraint functions were systematically varied to document the sensitivities of the optimal design to a variety of economic and technological assumptions. Additionally, a comparison is made for each of the parameter variations between the baseline configuration and the optimally redesigned configuration.
NASA Astrophysics Data System (ADS)
Sanjari, M. S.; Chen, X.; Hülsmann, P.; Litvinov, Yu A.; Nolden, F.; Piotrowski, J.; Steck, M.; Stöhlker, Th
2015-11-01
Position sensitive beam monitors are indispensable for the beam diagnostics in storage rings. Apart from their applications in the measurements of beam parameters, they can be used in non-destructive in-ring decay studies of radioactive ion beams as well as enhancing precision in the isochronous mass measurement technique. In this work, we introduce a novel approach based on cavities with elliptical cross-section, in order to compensate the limitations of known designs for the application in ion storage rings. The design is aimed primarily for future heavy ion storage rings of the FAIR project. The conceptual design is discussed together with simulation results.
Neutron activation analysis; A sensitive test for trace elements
Hossain, T.Z. . Ward Lab.)
1992-01-01
This paper discusses neutron activation analysis (NAA), an extremely sensitive technique for determining the elemental constituents of an unknown specimen. Currently, there are some twenty-five moderate-power TRIGA reactors scattered across the United States (fourteen of them at universities), and one of their principal uses is for NAA. NAA is procedurally simple. A small amount of the material to be tested (typically between one and one hundred milligrams) is irradiated for a period that varies from a few minutes to several hours in a neutron flux of around 10{sup 12} neutrons per square centimeter per second. A tiny fraction of the nuclei present (about 10{sup {minus}8}) is transmuted by nuclear reactions into radioactive forms. Subsequently, the nuclei decay, and the energy and intensity of the gamma rays that they emit can be measured in a gamma-ray spectrometer.
Apparatus and Method for Ultra-Sensitive trace Analysis
Lu, Zhengtian; Bailey, Kevin G.; Chen, Chun Yen; Li, Yimin; O'Connor, Thomas P.; Young, Linda
2000-01-03
An apparatus and method for conducting ultra-sensitive trace element and isotope analysis. The apparatus injects a sample through a fine nozzle to form an atomic beam. A DC discharge is used to elevate select atoms to a metastable energy level. These atoms are then acted on by a laser oriented orthogonally to the beam path to reduce the traverse velocity and to decrease the divergence angle of the beam. The beam then enters a Zeeman slower where a counter-propagating laser beam acts to slow the atoms down. Then select atoms are captured in a magneto-optical trap where they undergo fluorescence. A portion of the scattered photons are imaged onto a photo-detector, and the results analyzed to detect the presence of single atoms of the specific trace elements.
Sensitivity analysis and optimization of thin-film thermoelectric coolers
NASA Astrophysics Data System (ADS)
Harsha Choday, Sri; Roy, Kaushik
2013-06-01
The cooling performance of a thermoelectric (TE) material is dependent on the figure-of-merit (ZT = S2σT/κ), where S is the Seebeck coefficient, σ and κ are the electrical and thermal conductivities, respectively. The standard definition of ZT assigns equal importance to power factor (S2σ) and thermal conductivity. In this paper, we analyze the relative importance of each thermoelectric parameter on the cooling performance using the mathematical framework of sensitivity analysis. In addition, the impact of the electrical/thermal contact parasitics on bulk and superlattice Bi2Te3 is also investigated. In the presence of significant contact parasitics, we find that the carrier concentration that results in best cooling is lower than that of the highest ZT. We also establish the level of contact parasitics that are needed such that their impact on TE cooling is negligible.
Neutral density filters with Risley prisms: analysis and design.
Duma, Virgil-Florin; Nicolov, Mirela
2009-05-10
We achieve the analysis and design of optical attenuators with double-prism neutral density filters. A comparative study is performed on three possible device configurations; only two are presented in the literature but without their design calculus. The characteristic parameters of this optical attenuator with Risley translating prisms for each of the three setups are defined and their analytical expressions are derived: adjustment scale (attenuation range) and interval, minimum transmission coefficient and sensitivity. The setups are compared to select the optimal device, and, from this study, the best solution for double-prism neutral density filters, both from a mechanical and an optical point of view, is determined with two identical, symmetrically movable, no mechanical contact prisms. The design calculus of this optimal device is developed in essential steps. The parameters of the prisms, particularly their angles, are studied to improve the design, and we demonstrate the maximum attenuation range that this type of attenuator can provide. PMID:19424388
Worst case estimation of homology design by convex analysis
NASA Technical Reports Server (NTRS)
Yoshikawa, N.; Elishakoff, Isaac; Nakagiri, S.
1998-01-01
The methodology of homology design is investigated for optimum design of advanced structures. for which the achievement of delicate tasks by the aid of active control system is demanded. The proposed formulation of homology design, based on the finite element sensitivity analysis, necessarily requires the specification of external loadings. The formulation to evaluate the worst case for homology design caused by uncertain fluctuation of loadings is presented by means of the convex model of uncertainty, in which uncertainty variables are assigned to discretized nodal forces and are confined within a conceivable convex hull given as a hyperellipse. The worst case of the distortion from objective homologous deformation is estimated by the Lagrange multiplier method searching the point to maximize the error index on the boundary of the convex hull. The validity of the proposed method is demonstrated in a numerical example using the eleven-bar truss structure.
Sensitivity analysis of ecosystem service valuation in a Mediterranean watershed.
Sánchez-Canales, María; López Benito, Alfredo; Passuello, Ana; Terrado, Marta; Ziv, Guy; Acuña, Vicenç; Schuhmacher, Marta; Elorza, F Javier
2012-12-01
The services of natural ecosystems are clearly very important to our societies. In the last years, efforts to conserve and value ecosystem services have been fomented. By way of illustration, the Natural Capital Project integrates ecosystem services into everyday decision making around the world. This project has developed InVEST (a system for Integrated Valuation of Ecosystem Services and Tradeoffs). The InVEST model is a spatially integrated modelling tool that allows us to predict changes in ecosystem services, biodiversity conservation and commodity production levels. Here, InVEST model is applied to a stakeholder-defined scenario of land-use/land-cover change in a Mediterranean region basin (the Llobregat basin, Catalonia, Spain). Of all InVEST modules and sub-modules, only the behaviour of the water provisioning one is investigated in this article. The main novel aspect of this work is the sensitivity analysis (SA) carried out to the InVEST model in order to determine the variability of the model response when the values of three of its main coefficients: Z (seasonal precipitation distribution), prec (annual precipitation) and eto (annual evapotranspiration), change. The SA technique used here is a One-At-a-Time (OAT) screening method known as Morris method, applied over each one of the one hundred and fifty four sub-watersheds in which the Llobregat River basin is divided. As a result, this method provides three sensitivity indices for each one of the sub-watersheds under consideration, which are mapped to study how they are spatially distributed. From their analysis, the study shows that, in the case under consideration and between the limits considered for each factor, the effect of the Z coefficient on the model response is negligible, while the other two need to be accurately determined in order to obtain precise output variables. The results of this study will be applicable to the others watersheds assessed in the Consolider Scarce Project. PMID
DARHT : integration of shielding design and analysis with facility design /
Boudrie, R. L.; Brown, T. H.; Gilmore, W. E.; Downing, J. N. , Jr.; Hack, Alan; McClure, D. A.; Nelson, C. A.; Wadlinger, E. Alan; Zumbro, M. V.
2002-01-01
The design of the interior portions of the Dual Axis Radiographic Hydrodynamic Test (DARHT) Facility incorporated shielding and controls from the beginning of the installation of the Accelerators. The purpose of the design and analysis was to demonstrate the adequacy of shielding or to determine the need for additional shielding or controls. Two classes of events were considered: (1) routine operation defined as the annual production of 10,000 2000-ns pulses of electrons at a nominal energy of 20 MeV, some of which are converted to the x-ray imaging beam consisting of four nominal 60-ns pulses over the 2000-ns time frame, and (2) accident case defined as up to 100 2000-ns pulses of electrons accidentally impinging on some metallic surface, thereby producing x rays. Several locations for both classes of events were considered inside and outside of the accelerator hall buildings. The analysis method consisted of the definition of a source term for each case studied and the definition of a model of the shielding and equipment present between the source and the dose areas. A minimal model of the fixed existing or proposed shielding and equipment structures was used for a first approximation. If the resulting dose from the first approximation was below the design goal (1 rem/yr for routine operations, 5 rem for accident cases), then no further investigations were performed. If the result of the first approximation was above our design goals, the model was refined to include existing or proposed shielding and equipment. In some cases existing shielding and equipment were adequate to meet our goals and in some cases additional shielding was added or administrative controls were imposed to protect the workers. It is expected that the radiation shielding design, exclusion area designations, and access control features, will result in low doses to personnel at the DARHT Facility.
NASA Technical Reports Server (NTRS)
Greene, William H.
1989-01-01
A study has been performed focusing on the calculation of sensitivities of displacements, velocities, accelerations, and stresses in linear, structural, transient response problems. One significant goal was to develop and evaluate sensitivity calculation techniques suitable for large-order finite element analyses. Accordingly, approximation vectors such as vibration mode shapes are used to reduce the dimensionality of the finite element model. Much of the research focused on the accuracy of both response quantities and sensitivities as a function of number of vectors used. Two types of sensitivity calculation techniques were developed and evaluated. The first type of technique is an overall finite difference method where the analysis is repeated for perturbed designs. The second type of technique is termed semianalytical because it involves direct, analytical differentiation of the equations of motion with finite difference approximation of the coefficient matrices. To be computationally practical in large-order problems, the overall finite difference methods must use the approximation vectors from the original design in the analyses of the perturbed models.
Woo, L Y; Glass, R S; Novak, R F; Visser, J H
2009-09-23
Solid-state electrochemical sensors using two different sensing electrode compositions, gold and strontium-doped lanthanum manganite (LSM), were evaluated for gas phase sensing of NO{sub x} (NO and NO{sub 2}) using an impedance-metric technique. An asymmetric cell design utilizing porous YSZ electrolyte exposed both electrodes to the test gas (i.e., no reference gas). Sensitivity to less than 5 ppm NO and response/recovery times (10-90%) less than 10 s were demonstrated. Using an LSM sensing electrode, virtual identical sensitivity towards NO and NO{sub 2} was obtained, indicating that the equilibrium gas concentration was measured by the sensing electrode. In contrast, for cells employing a gold sensing electrode the NO{sub x} sensitivity varied depending on the cell design: increasing the amount of porous YSZ electrolyte on the sensor surface produced higher NO{sub 2} sensitivity compared to NO. In order to achieve comparable sensitivity for both NO and NO{sub 2}, the cell with the LSM sensing electrode required operation at a lower temperature (575 C) than the cell with the gold sensing electrode (650 C). The role of surface reactions are proposed to explain the differences in NO and NO{sub 2} selectivity using the two different electrode materials.
Goal-oriented sensitivity analysis for lattice kinetic Monte Carlo simulations.
Arampatzis, Georgios; Katsoulakis, Markos A
2014-03-28
In this paper we propose a new class of coupling methods for the sensitivity analysis of high dimensional stochastic systems and in particular for lattice Kinetic Monte Carlo (KMC). Sensitivity analysis for stochastic systems is typically based on approximating continuous derivatives with respect to model parameters by the mean value of samples from a finite difference scheme. Instead of using independent samples the proposed algorithm reduces the variance of the estimator by developing a strongly correlated-"coupled"- stochastic process for both the perturbed and unperturbed stochastic processes, defined in a common state space. The novelty of our construction is that the new coupled process depends on the targeted observables, e.g., coverage, Hamiltonian, spatial correlations, surface roughness, etc., hence we refer to the proposed method as goal-oriented sensitivity analysis. In particular, the rates of the coupled Continuous Time Markov Chain are obtained as solutions to a goal-oriented optimization problem, depending on the observable of interest, by considering the minimization functional of the corresponding variance. We show that this functional can be used as a diagnostic tool for the design and evaluation of different classes of couplings. Furthermore, the resulting KMC sensitivity algorithm has an easy implementation that is based on the Bortz-Kalos-Lebowitz algorithm's philosophy, where events are divided in classes depending on level sets of the observable of interest. Finally, we demonstrate in several examples including adsorption, desorption, and diffusion Kinetic Monte Carlo that for the same confidence interval and observable, the proposed goal-oriented algorithm can be two orders of magnitude faster than existing coupling algorithms for spatial KMC such as the Common Random Number approach. We also provide a complete implementation of the proposed sensitivity analysis algorithms, including various spatial KMC examples, in a supplementary MATLAB
Goal-oriented sensitivity analysis for lattice kinetic Monte Carlo simulations
NASA Astrophysics Data System (ADS)
Arampatzis, Georgios; Katsoulakis, Markos A.
2014-03-01
In this paper we propose a new class of coupling methods for the sensitivity analysis of high dimensional stochastic systems and in particular for lattice Kinetic Monte Carlo (KMC). Sensitivity analysis for stochastic systems is typically based on approximating continuous derivatives with respect to model parameters by the mean value of samples from a finite difference scheme. Instead of using independent samples the proposed algorithm reduces the variance of the estimator by developing a strongly correlated-"coupled"- stochastic process for both the perturbed and unperturbed stochastic processes, defined in a common state space. The novelty of our construction is that the new coupled process depends on the targeted observables, e.g., coverage, Hamiltonian, spatial correlations, surface roughness, etc., hence we refer to the proposed method as goal-oriented sensitivity analysis. In particular, the rates of the coupled Continuous Time Markov Chain are obtained as solutions to a goal-oriented optimization problem, depending on the observable of interest, by considering the minimization functional of the corresponding variance. We show that this functional can be used as a diagnostic tool for the design and evaluation of different classes of couplings. Furthermore, the resulting KMC sensitivity algorithm has an easy implementation that is based on the Bortz-Kalos-Lebowitz algorithm's philosophy, where events are divided in classes depending on level sets of the observable of interest. Finally, we demonstrate in several examples including adsorption, desorption, and diffusion Kinetic Monte Carlo that for the same confidence interval and observable, the proposed goal-oriented algorithm can be two orders of magnitude faster than existing coupling algorithms for spatial KMC such as the Common Random Number approach. We also provide a complete implementation of the proposed sensitivity analysis algorithms, including various spatial KMC examples, in a supplementary MATLAB
Goal-oriented sensitivity analysis for lattice kinetic Monte Carlo simulations
Arampatzis, Georgios; Katsoulakis, Markos A.
2014-03-28
In this paper we propose a new class of coupling methods for the sensitivity analysis of high dimensional stochastic systems and in particular for lattice Kinetic Monte Carlo (KMC). Sensitivity analysis for stochastic systems is typically based on approximating continuous derivatives with respect to model parameters by the mean value of samples from a finite difference scheme. Instead of using independent samples the proposed algorithm reduces the variance of the estimator by developing a strongly correlated-“coupled”- stochastic process for both the perturbed and unperturbed stochastic processes, defined in a common state space. The novelty of our construction is that the new coupled process depends on the targeted observables, e.g., coverage, Hamiltonian, spatial correlations, surface roughness, etc., hence we refer to the proposed method as goal-oriented sensitivity analysis. In particular, the rates of the coupled Continuous Time Markov Chain are obtained as solutions to a goal-oriented optimization problem, depending on the observable of interest, by considering the minimization functional of the corresponding variance. We show that this functional can be used as a diagnostic tool for the design and evaluation of different classes of couplings. Furthermore, the resulting KMC sensitivity algorithm has an easy implementation that is based on the Bortz–Kalos–Lebowitz algorithm's philosophy, where events are divided in classes depending on level sets of the observable of interest. Finally, we demonstrate in several examples including adsorption, desorption, and diffusion Kinetic Monte Carlo that for the same confidence interval and observable, the proposed goal-oriented algorithm can be two orders of magnitude faster than existing coupling algorithms for spatial KMC such as the Common Random Number approach. We also provide a complete implementation of the proposed sensitivity analysis algorithms, including various spatial KMC examples, in a supplementary
DESIGN PACKAGE 1E SYSTEM SAFETY ANALYSIS
M. Salem
1995-06-23
The purpose of this analysis is to systematically identify and evaluate hazards related to the Yucca Mountain Project Exploratory Studies Facility (ESF) Design Package 1E, Surface Facilities, (for a list of design items included in the package 1E system safety analysis see section 3). This process is an integral part of the systems engineering process; whereby safety is considered during planning, design, testing, and construction. A largely qualitative approach was used since a radiological System Safety Analysis is not required. The risk assessment in this analysis characterizes the accident scenarios associated with the Design Package 1E structures/systems/components(S/S/Cs) in terms of relative risk and includes recommendations for mitigating all identified risks. The priority for recommending and implementing mitigation control features is: (1) Incorporate measures to reduce risks and hazards into the structure/system/component design, (2) add safety devices and capabilities to the designs that reduce risk, (3) provide devices that detect and warn personnel of hazardous conditions, and (4) develop procedures and conduct training to increase worker awareness of potential hazards, on methods to reduce exposure to hazards, and on the actions required to avoid accidents or correct hazardous conditions.
NASA Astrophysics Data System (ADS)
Notghi, Bahram; Brigham, John C.
2013-12-01
An approach to nondestructive test (NDT) design for material characterization and damage identification in structural components, and more generally in solid continua, is presented and numerically tested. The proposed NDT design approach is based on maximizing a measure of the sensitivity of the test responses to changes in the material properties of the structure while also maximizing a measure of the difference in the response components. As such, the optimally designed NDT provides significant improvement in the ability to solve subsequent inverse characterization problems by extracting the maximum amount of non-redundant information from the system to increase the inverse solution observability. The NDT design approach is theoretically able to include any and all possible design aspects, such as the placement of sensors and actuators and determination of actuation frequency, among others. Through simulated test problems based on the characterization of damage in aluminum structural components utilizing steady-state dynamic surface excitation and localized measurements of displacement, the proposed NDT design approach is shown to provide NDT designs with significantly higher measurement sensitivity as well as lower information redundancy when compared to alternate test approaches. More importantly, the optimized NDT methods are shown to provide consistent and significant improvement in the ability to accurately inversely characterize variations in the Young’s modulus distributions for the simulated test cases considered.
New design of a passive type RADFET reader for enhanced sensitivity
NASA Astrophysics Data System (ADS)
Lee, Dae-Hee
2016-07-01
We present a new design of a passive type RADFET reader with enhanced radiation sensitivity. Using a electostatic plate, we have applied a static electric field to the gate voltage, which impacts a positive biasing on the p-type MOSFET. The resultant effect shows that 1.8 times of radiation sensitivity increased when we measured the threshold voltage shift of the RADFET exposed to 30 krad irradiation. We discuss further about the characteristic changes of a RADFET according to the positive biasing on the gate voltage.
Gutruf, Philipp; Walia, Sumeet; Nur Ali, Md; Sriram, Sharath E-mail: madhu.bhaskaran@gmail.com; Bhaskaran, Madhu E-mail: madhu.bhaskaran@gmail.com
2014-01-13
The functionality of flexible electronics relies on stable performance of thin film micro-electrodes. This letter investigates the behavior of gold thin films on polyimide, a prevalent combination in flexible devices. The dynamic behavior of gold micro-electrodes has been studied by subjecting them to stress while monitoring their resistance in situ. The shape of the electrodes was systematically varied to examine resistive strain sensitivity, while an additional encapsulation was applied to characterize multilayer behavior. The realized designs show remarkable tolerance to repetitive strain, demonstrating that curvature and encapsulation are excellent approaches for minimizing resistive strain sensitivity to enable durable flexible electronics.
Failure analysis helps improve heat exchanger designs
Neel, L.
1993-08-01
Industry's usual approach to evaluating the design of heat transfer equipment is to look for failures that have occurred in existing equipment of the same design, try to determine their causes, and then develop design modifications to prevent them from recurring. This is referred to as failure analysis. There is a tendency in industry to overlook successful designs of existing equipment that have operated for many years without major failures. A manufacturer of this type of equipment that also repairs and retubes heat exchangers is in a unique position to analyze and evaluate the deposition to analyze and evaluate the designs used. Data on the condition of individual components and materials used, and operating conditions, types of fluids and specific design features can all be compiled to develop a complete evaluation. In addition, calculations can be made to verify the validity of currently established design parameters from various sources. Utility heat exchangers cover a wide range of equipment, including feedwater heaters, evaporators, condensate coolers, lube oil coolers, turbine oil coolers, air preheaters, generator coolers, service water exchangers and air ejector condensers. The piece of equipment selected to be analyzed is a horizontal three-zone feedwater heater. Four major areas will be covered: tubeside, desuperheating, condensing, and subcooling zones of heat exchangers. The analysis will look at successful designs and the most common failures associated with each area.
Sensitivity analysis for Probabilistic Tsunami Hazard Assessment (PTHA)
NASA Astrophysics Data System (ADS)
Spada, M.; Basili, R.; Selva, J.; Lorito, S.; Sorensen, M. B.; Zonker, J.; Babeyko, A. Y.; Romano, F.; Piatanesi, A.; Tiberti, M.
2012-12-01
In modern societies, probabilistic hazard assessment of natural disasters is commonly used by decision makers for designing regulatory standards and, more generally, for prioritizing risk mitigation efforts. Systematic formalization of Probabilistic Tsunami Hazard Assessment (PTHA) has started only in recent years, mainly following the giant tsunami disaster of Sumatra in 2004. Typically, PTHA for earthquake sources exploits the long-standing practices developed in probabilistic seismic hazard assessment (PSHA), even though important differences are evident. In PTHA, for example, it is known that far-field sources are more important and that physical models for tsunami propagation are needed for the highly non-isotropic propagation of tsunami waves. However, considering the high impact that PTHA may have on societies, an important effort to quantify the effect of specific assumptions should be performed. Indeed, specific standard hypotheses made in PSHA may prove inappropriate for PTHA, since tsunami waves are sensitive to different aspects of sources (e.g. fault geometry, scaling laws, slip distribution) and propagate differently. In addition, the necessity of running an explicit calculation of wave propagation for every possible event (tsunami scenario) forces analysts to finding strategies for diminishing the computational burden. In this work, we test the sensitivity of hazard results with respect to several assumptions that are peculiar of PTHA and others that are commonly accepted in PSHA. Our case study is located in the central Mediterranean Sea and considers the Western Hellenic Arc as the earthquake source with Crete and Eastern Sicily as near-field and far-field target coasts, respectively. Our suite of sensitivity tests includes: a) comparison of random seismicity distribution within area sources as opposed to systematically distributed ruptures on fault sources; b) effects of statistical and physical parameters (a- and b-value, Mc, Mmax, scaling laws
Computational design of a pH-sensitive IgG binding protein.
Strauch, Eva-Maria; Fleishman, Sarel J; Baker, David
2014-01-14
Computational design provides the opportunity to program protein-protein interactions for desired applications. We used de novo protein interface design to generate a pH-dependent Fc domain binding protein that buries immunoglobulin G (IgG) His-433. Using next-generation sequencing of naïve and selected pools of a library of design variants, we generated a molecular footprint of the designed binding surface, confirming the binding mode and guiding further optimization of the balance between affinity and pH sensitivity. In biolayer interferometry experiments, the optimized design binds IgG with a Kd of ∼ 4 nM at pH 8.2, and approximately 500-fold more weakly at pH 5.5. The protein is extremely stable, heat-resistant and highly expressed in bacteria, and allows pH-based control of binding for IgG affinity purification and diagnostic devices. PMID:24381156
Computational design of a pH-sensitive IgG binding protein
Strauch, Eva-Maria; Fleishman, Sarel J.; Baker, David
2014-01-01
Computational design provides the opportunity to program protein–protein interactions for desired applications. We used de novo protein interface design to generate a pH-dependent Fc domain binding protein that buries immunoglobulin G (IgG) His-433. Using next-generation sequencing of naïve and selected pools of a library of design variants, we generated a molecular footprint of the designed binding surface, confirming the binding mode and guiding further optimization of the balance between affinity and pH sensitivity. In biolayer interferometry experiments, the optimized design binds IgG with a Kd of ∼4 nM at pH 8.2, and approximately 500-fold more weakly at pH 5.5. The protein is extremely stable, heat-resistant and highly expressed in bacteria, and allows pH-based control of binding for IgG affinity purification and diagnostic devices. PMID:24381156
Electric motor designs for attenuating torque disturbance in sensitive space mechanisms
NASA Astrophysics Data System (ADS)
Marks, David B.; Fink, Richard A.
2003-09-01
When a motion control system introduces unwanted torque jitter and motion anomalies into sensitive space flight optical or positioning mechanisms, the pointing accuracy, positioning capability, or scanning resolution of the mission suffers. Special motion control technology must be employed to provide attenuation of the harmful torque disturbances. Brushless DC (BLDC) Motors with low torque disturbance characteristics have been successfully used on such notable missions as the Hubble Space Telescope when conventional approaches to motor design would not work. Motor designs for low disturbance mechanisms can include two and three phase sinusoidal BLDC motors, BLDC motors without iron teeth, and sometimes skewed or non-integral slot designs for motors commutated with Hall effect devices. The principal components of motor torque disturbance, successful BLDC motor designs for attenuating disturbances, and design trade-offs for optimum performance are examined.
Sensitivity Analysis and Uncertainty Propagation in a General-Purpose Thermal Analysis Code
Blackwell, Bennie F.; Dowding, Kevin J.
1999-08-04
Methods are discussed for computing the sensitivity of field variables to changes in material properties and initial/boundary condition parameters for heat transfer problems. The method we focus on is termed the ''Sensitivity Equation Method'' (SEM). It involves deriving field equations for sensitivity coefficients by differentiating the original field equations with respect to the parameters of interest and numerically solving the resulting sensitivity field equations. Uncertainty in the model parameters are then propagated through the computational model using results derived from first-order perturbation theory; this technique is identical to the methodology typically used to propagate experimental uncertainty. Numerical results are presented for the design of an experiment to estimate the thermal conductivity of stainless steel using transient temperature measurements made on prototypical hardware of a companion contact conductance experiment. Comments are made relative to extending the SEM to conjugate heat transfer problems.
Simultaneous analysis and design. [in structural engineering
NASA Technical Reports Server (NTRS)
Haftka, R. T.
1985-01-01
Optimization techniques are increasingly being used for performing nonlinear structural analysis. The development of element by element (EBE) preconditioned conjugate gradient (CG) techniques is expected to extend this trend to linear analysis. Under these circumstances the structural design problem can be viewed as a nested optimization problem. There are computational benefits to treating this nested problem as a large single optimization problem. The response variables (such as displacements) and the structural parameters are all treated as design variables in a unified formulation which performs simultaneously the design and analysis. Two examples are used for demonstration. A seventy-two bar truss is optimized subject to linear stress constraints and a wing box structure is optimized subject to nonlinear collapse constraints. Both examples show substantial computational savings with the unified approach as compared to the traditional nested approach.
NASA Multidisciplinary Design and Analysis Fellowship Program
NASA Technical Reports Server (NTRS)
Schrage, D. P.; Craig, J. I.; Mavris, D. N.; Hale, M. A.; DeLaurentis, D.
1999-01-01
This report summarizes the results of a multi-year training grant for the development and implementation of a Multidisciplinary Design and Analysis (MDA) Fellowship Program at Georgia Tech. The Program funded the creation of graduate MS and PhD degree programs in aerospace systems design, analysis and integration. It also provided prestigious Fellowships with associated Industry Internships for outstanding engineering students. The graduate program has become the foundation for a vigorous and productive research effort and has produced: 20 MS degrees, 7 Ph.D. degrees, and has contributed to 9 ongoing Ph.D. students. The results of the research are documented in 32 publications (23 of which are included on a companion CDROM) and 4 annual student design reports (included on a companion CDROM). The legacy of this critical funding is the Center for Aerospace Systems Analysis at Georgia Tech which is continuing the graduate program, the research, and the industry internships established by this grant.
SENSITIVITY ANALYSIS FOR I-129 WASTES: EFFECT OF HYDRAULIC CONDUCTIVITY
Ades, M; Leonard Collard, L
2007-01-12
Solid low-level radioactive wastes at the Savannah River Site (SRS) are disposed in trenches. In order to determine the permissible radioactive inventory limits for such disposal facilities, it is required to assess the behavior of radioactive waste material over long periods of time. The sensitivity of flow and I-129 (and similar radionuclides) transport in groundwater in the vadose zone to the hydraulic conductivities of the vadose zone subregions and the low-level waste is identified and quantified. A trench configuration and simulation model have been developed to analyze the flow and transport of the radionuclide in the vadose zone as it migrates to the groundwater table. The analysis identifies and quantifies the major dependencies of the flow and radionuclide fractional flux on the subregion hydraulic conductivities. Analysis results indicate the importance of the hydraulic conductivity assigned to the materials modeled, thereby providing the modeler and decision makers with valuable insights on the potential impact of the hydraulic conductivity on flow and radionuclide transport.
Fault sensitivity and wear-out analysis of VLSI systems
NASA Astrophysics Data System (ADS)
Choi, Gwan Seung
1994-07-01
This thesis describes simulation approaches to conduct fault sensitivity and wear-out failure analysis of VLSI systems. A fault-injection approach to study transient impact in VLSI systems is developed. Through simulated fault injection at the device level and, subsequent fault propagation at the gate functional and software levels, it is possible to identify critical bottlenecks in dependability. Techniques to speed up the fault simulation and to perform statistical analysis of fault-impact are developed. A wear-out simulation environment is also developed to closely mimic dynamic sequences of wear-out events in a device through time, to localize weak location/aspect of target chip and to allow generation of TTF (Time-to-failure) distribution of VLSI chip as a whole. First, an accurate simulation of a target chip and its application code is performed to acquire trace data (real workload) on switch activity. Then, using this switch activity information, wear-out of the each component in the entire chip is simulated using Monte Carlo techniques.
SENSITIVITY ANALYSIS AND THE DESIGN OF GAS UPTAKE INHALATION STUDIES
Gas uptake studies analyzed by physiologically based pharmacokinetic (PBPK) models have been used to estimate metabolic parameters for many volatiles. he metabolic constants Vmax, Km, and Kf are typically inferred from the decline in chemical concentration observed in closed cham...
NASA Technical Reports Server (NTRS)
Consoli, Robert David; Sobieszczanski-Sobieski, Jaroslaw
1990-01-01
Advanced multidisciplinary analysis and optimization methods, namely system sensitivity analysis and non-hierarchical system decomposition, are applied to reduce the cost and improve the visibility of an automated vehicle design synthesis process. This process is inherently complex due to the large number of functional disciplines and associated interdisciplinary couplings. Recent developments in system sensitivity analysis as applied to complex non-hierarchic multidisciplinary design optimization problems enable the decomposition of these complex interactions into sub-processes that can be evaluated in parallel. The application of these techniques results in significant cost, accuracy, and visibility benefits for the entire design synthesis process.
On the sensitivity analysis of separated-loop MRS data
NASA Astrophysics Data System (ADS)
Behroozmand, A.; Auken, E.; Fiandaca, G.
2013-12-01
In this study we investigate the sensitivity analysis of separated-loop magnetic resonance sounding (MRS) data and, in light of deploying a separate MRS receiver system from the transmitter system, compare the parameter determination of the separated-loop with the conventional coincident-loop MRS data. MRS has emerged as a promising surface-based geophysical technique for groundwater investigations, as it provides a direct estimate of the water content. The method works based on the physical principle of NMR during which a large volume of protons of the water molecules in the subsurface is excited at the specific Larmor frequency. The measurement consists of a large wire loop (typically 25 - 100 m in side length/diameter) deployed on the surface which typically acts as both a transmitter and a receiver, the so-called coincident-loop configuration. An alternating current is passed through the loop deployed and the superposition of signals from all precessing protons within the investigated volume is measured in a receiver loop; a decaying NMR signal called Free Induction Decay (FID). To provide depth information, the FID signal is measured for a series of pulse moments (Q; product of current amplitude and transmitting pulse length) during which different earth volumes are excited. One of the main and inevitable limitations of MRS measurements is a relatively long measurement dead time, i.e. a non-zero time between the end of the energizing pulse and the beginning of the measurement, which makes it difficult, and in some places impossible, to record SNMR signal from fine-grained geologic units and limits the application of advanced pulse sequences. Therefore, one of the current research activities is the idea of building separate receiver units, which will diminish the dead time. In light of that, the aims of this study are twofold: 1) Using a forward modeling approach, the sensitivity kernels of different separated-loop MRS soundings are studied and compared with
Aviation System Analysis Capability Executive Assistant Design
NASA Technical Reports Server (NTRS)
Roberts, Eileen; Villani, James A.; Osman, Mohammed; Godso, David; King, Brent; Ricciardi, Michael
1998-01-01
In this technical document, we describe the design developed for the Aviation System Analysis Capability (ASAC) Executive Assistant (EA) Proof of Concept (POC). We describe the genesis and role of the ASAC system, discuss the objectives of the ASAC system and provide an overview of components and models within the ASAC system, and describe the design process and the results of the ASAC EA POC system design. We also describe the evaluation process and results for applicable COTS software. The document has six chapters, a bibliography, three appendices and one attachment.
Sensitivity analysis for hydrology and pesticide supply towards the river in SWAT
NASA Astrophysics Data System (ADS)
Holvoet, K.; van Griensven, A.; Seuntjens, P.; Vanrolleghem, P. A.
The dynamic behaviour of pesticides in river systems strongly depends on varying climatological conditions and agricultural management practices. To describe this behaviour at the river-basin scale, integrated hydrological and water quality models are needed. A crucial step in understanding the various processes determining pesticide fate is to perform a sensitivity analysis. Sensitivity analysis for hydrology and pesticide supply in SWAT (Soil and Water Assessment Tool) will provide useful support for the development of a reliable hydrological model and will give insight in which parameters are most sensitive concerning pesticide supply towards rivers. The study was performed on the Nil catchment in Belgium. In this study we utilised an LH-OAT sensitivity analysis. The LH-OAT method combines the One-factor-At-a-Time (OAT) design and Latin Hypercube (LH) sampling by taking the Latin Hypercube samples as initial points for an OAT design. By means of the LH-OAT sensitivity analysis, the dominant hydrological parameters were determined and a reduction of the number of model parameters was performed. Dominant hydrological parameters were the curve number (CN2), the surface runoff lag (surlag), the recharge to deep aquifer (rchrg_dp) and the threshold depth of water in the shallow aquifer (GWQMN). Next, the selected parameters were estimated by manual calibration. Hereby, the Nash-Sutcliffe coefficient of efficiency improved from an initial value of -22.4 to +0.53. In the second part, sensitivity analyses were performed to provide insight in which parameters or model inputs contribute most to variance in pesticide output. The results of this study show that for the Nil catchment, hydrologic parameters are dominant in controlling pesticide predictions. The other parameter that affects pesticide concentrations in surface water is ‘apfp_pest’, which meaning was changed into a parameter that controls direct losses to the river system (e.g., through the clean up of spray
Network interface unit design options performance analysis
NASA Technical Reports Server (NTRS)
Miller, Frank W.
1991-01-01
An analysis is presented of three design options for the Space Station Freedom (SSF) onboard Data Management System (DMS) Network Interface Unit (NIU). The NIU provides the interface from the Fiber Distributed Data Interface (FDDI) local area network (LAN) to the DMS processing elements. The FDDI LAN provides the primary means for command and control and low and medium rate telemetry data transfers on board the SSF. The results of this analysis provide the basis for the implementation of the NIU.
Microgravity isolation system design: A modern control analysis framework
NASA Technical Reports Server (NTRS)
Hampton, R. D.; Knospe, C. R.; Allaire, P. E.; Grodsinsky, C. M.
1994-01-01
Many acceleration-sensitive, microgravity science experiments will require active vibration isolation from the manned orbiters on which they will be mounted. The isolation problem, especially in the case of a tethered payload, is a complex three-dimensional one that is best suited to modern-control design methods. These methods, although more powerful than their classical counterparts, can nonetheless go only so far in meeting the design requirements for practical systems. Once a tentative controller design is available, it must still be evaluated to determine whether or not it is fully acceptable, and to compare it with other possible design candidates. Realistically, such evaluation will be an inherent part of a necessary iterative design process. In this paper, an approach is presented for applying complex mu-analysis methods to a closed-loop vibration isolation system (experiment plus controller). An analysis framework is presented for evaluating nominal stability, nominal performance, robust stability, and robust performance of active microgravity isolation systems, with emphasis on the effective use of mu-analysis methods.
Grid and design variables sensitivity analyses for NACA four-digit wing-sections
NASA Technical Reports Server (NTRS)
Sadrehaghighi, Ideen; Smith, Robert E.; Tiwari, Surendra N.
1993-01-01
Two distinct parameterization procedures are developed for investigating the grid sensitivity with respect to design parameters of a wing-section example. The first procedure is based on traditional (physical) relations defining NACA four-digit wing-sections. The second is advocating a novel (geometrical) parameterization using spline functions such as NURBS (Non-Uniform Rational B-Splines) for defining the wing-section geometry. An interactive algebraic grid generation technique, known as Hermite Cubic Interpolation, is employed to generate C-type grids around wing-sections. The grid sensitivity of the domain with respect to design and grid parameters has been obtained by direct differentiation of the grid equations. A hybrid approach is proposed for more geometrically complex configurations. A comparison of the sensitivity coefficients with those obtained using a finite-difference approach has been made to verify the feasibility of the approach. The aerodynamic sensitivity coefficients are obtained using the compressible two-dimensional thin-layer Navier-Stokes equations.
NASA Technical Reports Server (NTRS)
Huang, Zhao-Feng; Fint, Jeffry A.; Kuck, Frederick M.
2005-01-01
This paper is to address the in-flight reliability of a liquid propulsion engine system for a launch vehicle. We first establish a comprehensive list of system and sub-system reliability drivers for any liquid propulsion engine system. We then build a reliability model to parametrically analyze the impact of some reliability parameters. We present sensitivity analysis results for a selected subset of the key reliability drivers using the model. Reliability drivers identified include: number of engines for the liquid propulsion stage, single engine total reliability, engine operation duration, engine thrust size, reusability, engine de-rating or up-rating, engine-out design (including engine-out switching reliability, catastrophic fraction, preventable failure fraction, unnecessary shutdown fraction), propellant specific hazards, engine start and cutoff transient hazards, engine combustion cycles, vehicle and engine interface and interaction hazards, engine health management system, engine modification, engine ground start hold down with launch commit criteria, engine altitude start (1 in. start), Multiple altitude restart (less than 1 restart), component, subsystem and system design, manufacturing/ground operation support/pre and post flight check outs and inspection, extensiveness of the development program. We present some sensitivity analysis results for the following subset of the drivers: number of engines for the propulsion stage, single engine total reliability, engine operation duration, engine de-rating or up-rating requirements, engine-out design, catastrophic fraction, preventable failure fraction, unnecessary shutdown fraction, and engine health management system implementation (basic redlines and more advanced health management systems).
DAKOTA Design Analysis Kit for Optimization and Terascale
2010-02-24
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes (computational models) and iterative analysis methods. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and analysis of computational models on high performance computers.A user provides a set of DAKOTA commands in an input file andmore » launches DAKOTA. DAKOTA invokes instances of the computational models, collects their results, and performs systems analyses. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, polynomial chaos, stochastic collocation, and epistemic methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as hybrid optimization, surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. Services for parallel computing, simulation interfacing, approximation modeling, fault tolerance, restart, and graphics are also included.« less
DAKOTA Design Analysis Kit for Optimization and Terascale
Adams, Brian M.; Dalbey, Keith R.; Eldred, Michael S.; Gay, David M.; Swiler, Laura P.; Bohnhoff, William J.; Eddy, John P.; Haskell, Karen
2010-02-24
The DAKOTA (Design Analysis Kit for Optimization and Terascale Applications) toolkit provides a flexible and extensible interface between simulation codes (computational models) and iterative analysis methods. By employing object-oriented design to implement abstractions of the key components required for iterative systems analyses, the DAKOTA toolkit provides a flexible and extensible problem-solving environment for design and analysis of computational models on high performance computers.A user provides a set of DAKOTA commands in an input file and launches DAKOTA. DAKOTA invokes instances of the computational models, collects their results, and performs systems analyses. DAKOTA contains algorithms for optimization with gradient and nongradient-based methods; uncertainty quantification with sampling, reliability, polynomial chaos, stochastic collocation, and epistemic methods; parameter estimation with nonlinear least squares methods; and sensitivity/variance analysis with design of experiments and parameter study methods. These capabilities may be used on their own or as components within advanced strategies such as hybrid optimization, surrogate-based optimization, mixed integer nonlinear programming, or optimization under uncertainty. Services for parallel computing, simulation interfacing, approximation modeling, fault tolerance, restart, and graphics are also included.
Accuracy of the domain method for the material derivative approach to shape design sensitivities
NASA Technical Reports Server (NTRS)
Yang, R. J.; Botkin, M. E.
1987-01-01
Numerical accuracy for the boundary and domain methods of the material derivative approach to shape design sensitivities is investigated through the use of mesh refinement. The results show that the domain method is generally more accurate than the boundary method, using the finite element technique. It is also shown that the domain method is equivalent, under certain assumptions, to the implicit differentiation approach not only theoretically but also numerically.
Altair Lander Life Support: Design Analysis Cycles 4 and 5
NASA Technical Reports Server (NTRS)
Anderson, Molly; Curley, Su; Rotter, Henry; Yagoda, Evan
2010-01-01
Life support systems are a critical part of human exploration beyond low earth orbit. NASA s Altair Lunar Lander team is pursuing efficient solutions to the technical challenges of human spaceflight. Life support design efforts up through Design Analysis Cycle (DAC) 4 focused on finding lightweight and reliable solutions for the Sortie and Outpost missions within the Constellation Program. In DAC-4 and later follow on work, changes were made to add functionality for new requirements accepted by the Altair project, and to update the design as knowledge about certain issues or hardware matured. In DAC-5, the Altair project began to consider mission architectures outside the Constellation baseline. Selecting the optimal life support system design is very sensitive to mission duration. When the mission goals and architecture change several trade studies must be conducted to determine the appropriate design. Finally, several areas of work developed through the Altair project may be applicable to other vehicle concepts for microgravity missions. Maturing the Altair life support system related analysis, design, and requirements can provide important information for developers of a wide range of other human vehicles.
Altair Lander Life Support: Design Analysis Cycles 4 and 5
NASA Technical Reports Server (NTRS)
Anderson, Molly; Curley, Su; Rotter, Henry; Stambaugh, Imelda; Yagoda, Evan
2011-01-01
Life support systems are a critical part of human exploration beyond low earth orbit. NASA s Altair Lunar Lander team is pursuing efficient solutions to the technical challenges of human spaceflight. Life support design efforts up through Design Analysis Cycle (DAC) 4 focused on finding lightweight and reliable solutions for the Sortie and Outpost missions within the Constellation Program. In DAC-4 and later follow on work, changes were made to add functionality for new requirements accepted by the Altair project, and to update the design as knowledge about certain issues or hardware matured. In DAC-5, the Altair project began to consider mission architectures outside the Constellation baseline. Selecting the optimal life support system design is very sensitive to mission duration. When the mission goals and architecture change several trade studies must be conducted to determine the appropriate design. Finally, several areas of work developed through the Altair project may be applicable to other vehicle concepts for microgravity missions. Maturing the Altair life support system related analysis, design, and requirements can provide important information for developers of a wide range of other human vehicles.
Sensitivity Analysis Applied to Atomic Data Used for X-ray Spectrum Synthesis
NASA Technical Reports Server (NTRS)
Kallman, T.
2006-01-01
A great deal of work has been devoted to the accumulation of accurate quantities describing atomic processes for use in analysis of astrophysical spectra. But in many situations of interest the interpretation of a quantity which is observed, such as a line flux, depends on the results of a modeling- or spectrum synthesis code. The results of such a code depends in turn on many atomic rates or cross sections, and the sensitivity of the observable quantity on the various rates and cross sections may be non-linear and if so cannot easily be derived analytically. This talk describes simple numerical experiments designed to examine some of these issues.
Meng, Xiawei; Zhao, Yulong
2016-01-01
A piezoresistive pressure sensor with a beam-membrane-dual-island structure is developed for micro-pressure monitoring in the field of aviation, which requires great sensitivity and overload resistance capacity. The design, fabrication, and test of the sensor are presented in this paper. By analyzing the stress distribution of sensitive elements using the finite element method, a novel structure incorporating sensitive beams with a traditional bossed diaphragm is built up. The proposed structure proved to be advantageous in terms of high sensitivity and high overload resistance compared with the conventional bossed diaphragm and flat diaphragm structures. Curve fittings of surface stress and deflection based on ANSYS simulation results are performed to establish the sensor equations. Fabricated on an n-type single crystal silicon wafer, the sensor chips are wire-bonded to a printed circuit board (PCB) and packaged for experiments. The static and dynamic characteristics are tested and discussed. Experimental results show that the sensor has a sensitivity as high as 17.339 μV/V/Pa in the range of 500 Pa at room temperature, and a high overload resistance of 200 times overpressure. Due to the excellent performance, the sensor can be applied in measuring micro-pressure lower than 500 Pa. PMID:27005627
Meng, Xiawei; Zhao, Yulong
2016-01-01
A piezoresistive pressure sensor with a beam-membrane-dual-island structure is developed for micro-pressure monitoring in the field of aviation, which requires great sensitivity and overload resistance capacity. The design, fabrication, and test of the sensor are presented in this paper. By analyzing the stress distribution of sensitive elements using the finite element method, a novel structure incorporating sensitive beams with a traditional bossed diaphragm is built up. The proposed structure proved to be advantageous in terms of high sensitivity and high overload resistance compared with the conventional bossed diaphragm and flat diaphragm structures. Curve fittings of surface stress and deflection based on ANSYS simulation results are performed to establish the sensor equations. Fabricated on an n-type single crystal silicon wafer, the sensor chips are wire-bonded to a printed circuit board (PCB) and packaged for experiments. The static and dynamic characteristics are tested and discussed. Experimental results show that the sensor has a sensitivity as high as 17.339 μV/V/Pa in the range of 500 Pa at room temperature, and a high overload resistance of 200 times overpressure. Due to the excellent performance, the sensor can be applied in measuring micro-pressure lower than 500 Pa. PMID:27005627
Microstructure design of nanoporous TiO2 photoelectrodes for dye-sensitized solar cell modules.
Hu, Linhua; Dai, Songyuan; Weng, Jian; Xiao, Shangfeng; Sui, Yifeng; Huang, Yang; Chen, Shuanghong; Kong, Fantai; Pan, Xu; Liang, Linyun; Wang, Kongjia
2007-01-18
The optimization of dye-sensitized solar cells, especially the design of nanoporous TiO2 film microstructure, is an urgent problem for high efficiency and future commercial applications. However, up to now, little attention has been focused on the design of nanoporous TiO2 microstructure for a high efficiency of dye-sensitized solar cell modules. The optimization and design of TiO2 photoelectrode microstructure are discussed in this paper. TiO2 photoelectrodes with three different layers, including layers of small pore size films, larger pore size films, and light-scattering particles on the conducting glass with the desirable thickness, were designed and investigated. Moreover, the photovoltaic properties showed that the different porosities, pore size distribution, and BET surface area of each layer have a dramatic influence on short-circuit current, open-circuit voltage, and fill factor of the modules. The optimization and design of TiO2 photoelectrode microstructure contribute a high efficiency of DSC modules. The photoelectric conversion efficiency around 6% with 15 x 20 cm2 modules under illumination of simulated AM1.5 sunlight (100 mW/cm2) and 40 x 60 cm2 panels with the same performance tested outdoor have been achieved by our group. PMID:17214486
NASA Astrophysics Data System (ADS)
Razavi, S.; Gupta, H. V.
2014-12-01
Sensitivity analysis (SA) is an important paradigm in the context of Earth System model development and application, and provides a powerful tool that serves several essential functions in modelling practice, including 1) Uncertainty Apportionment - attribution of total uncertainty to different uncertainty sources, 2) Assessment of Similarity - diagnostic testing and evaluation of similarities between the functioning of the model and the real system, 3) Factor and Model Reduction - identification of non-influential factors and/or insensitive components of model structure, and 4) Factor Interdependence - investigation of the nature and strength of interactions between the factors, and the degree to which factors intensify, cancel, or compensate for the effects of each other. A variety of sensitivity analysis approaches have been proposed, each of which formally characterizes a different "intuitive" understanding of what is meant by the "sensitivity" of one or more model responses to its dependent factors (such as model parameters or forcings). These approaches are based on different philosophies and theoretical definitions of sensitivity, and range from simple local derivatives and one-factor-at-a-time procedures to rigorous variance-based (Sobol-type) approaches. In general, each approach focuses on, and identifies, different features and properties of the model response and may therefore lead to different (even conflicting) conclusions about the underlying sensitivity. This presentation revisits the theoretical basis for sensitivity analysis, and critically evaluates existing approaches so as to demonstrate their flaws and shortcomings. With this background, we discuss several important properties of response surfaces that are associated with the understanding and interpretation of sensitivity. Finally, a new approach towards global sensitivity assessment is developed that is consistent with important properties of Earth System model response surfaces.
Analysis to Design: A Technical Training Submethodology.
ERIC Educational Resources Information Center
Garavaglia, Paul L.
1993-01-01
Describes a submethodology for using information from the analysis phase during the design phase when developing technical training. The development of instructional materials is discussed; Keller's ARCS (Attention, Relevance, Confidence, and Satisfaction) motivation model and Gagne's events of instruction are compared; and the development and use…
Orion Orbit Control Design and Analysis
NASA Technical Reports Server (NTRS)
Jackson, Mark; Gonzalez, Rodolfo; Sims, Christopher
2007-01-01
The analysis of candidate thruster configurations for the Crew Exploration Vehicle (CEV) is presented. Six candidate configurations were considered for the prime contractor baseline design. The analysis included analytical assessments of control authority, control precision, efficiency and robustness, as well as simulation assessments of control performance. The principles used in the analytic assessments of controllability, robustness and fuel performance are covered and results provided for the configurations assessed. Simulation analysis was conducted using a pulse width modulated, 6 DOF reaction system control law with a simplex-based thruster selection algorithm. Control laws were automatically derived from hardware configuration parameters including thruster locations, directions, magnitude and specific impulse, as well as vehicle mass properties. This parameterized controller allowed rapid assessment of multiple candidate layouts. Simulation results are presented for final phase rendezvous and docking, as well as low lunar orbit attitude hold. Finally, on-going analysis to consider alternate Service Module designs and to assess the pilot-ability of the baseline design are discussed to provide a status of orbit control design work to date.
ITEM ANALYSIS, TEST DESIGN, AND CLASSIFICATION.
ERIC Educational Resources Information Center
SOLOMON, HERBERT
THE THEME OF THIS PROGRAM WAS THE USE OF PROBABILISTIC MODELS AS MEASUREMENT TOOLS FOR EXPERIMENTAL DESIGN IN TEST CONSTRUCTION AND FOR ANALYSIS OF TEST DATA. SOME CENTRAL ISSUES IN THIS FIELD ARE DISCUSSED AND RESOLVED IN THE FIRST FOUR CHAPTERS OF THE REPORT UNDER THE FOLLOWING TITLES--(1) CONTRIBUTIONS TO ITEM SELECTION, (2) CLUSTERING…
A compact MOST model for design analysis.
NASA Technical Reports Server (NTRS)
Kalinowski, J. J.
1972-01-01
A generalized extension of Kotani's (1970) design analysis model is described that is accurate for a majority of modern MOST structures. The generalized model retains much of the ease of parameter measurement characteristics of the original and enough physical correspondence to accurately represent important steady-state environmental effects, such as those caused by radiation.
High-sensitivity low-noise miniature fluxgate magnetometers using a flip chip conceptual design.
Lu, Chih-Cheng; Huang, Jeff; Chiu, Po-Kai; Chiu, Shih-Liang; Jeng, Jen-Tzong
2014-01-01
This paper presents a novel class of miniature fluxgate magnetometers fabricated on a print circuit board (PCB) substrate and electrically connected to each other similar to the current "flip chip" concept in semiconductor package. This sensor is soldered together by reversely flipping a 5 cm × 3 cm PCB substrate to the other identical one which includes dual magnetic cores, planar pick-up coils, and 3-D excitation coils constructed by planar Cu interconnections patterned on PCB substrates. Principles and analysis of the fluxgate sensor are introduced first, and followed by FEA electromagnetic modeling and simulation for the proposed sensor. Comprehensive characteristic experiments of the miniature fluxgate device exhibit favorable results in terms of sensitivity (or "responsivity" for magnetometers) and field noise spectrum. The sensor is driven and characterized by employing the improved second-harmonic detection technique that enables linear V-B correlation and responsivity verification. In addition, the double magnitude of responsivity measured under very low frequency (1 Hz) magnetic fields is experimentally demonstrated. As a result, the maximum responsivity of 593 V/T occurs at 50 kHz of excitation frequency with the second harmonic wave of excitation; however, the minimum magnetic field noise is found to be 0.05 nT/Hz(1/2) at 1 Hz under the same excitation. In comparison with other miniature planar fluxgates published to date, the fluxgate magnetic sensor with flip chip configuration offers advances in both device functionality and fabrication simplicity. More importantly, the novel design can be further extended to a silicon-based micro-fluxgate chip manufactured by emerging CMOS-MEMS technologies, thus enriching its potential range of applications in modern engineering and the consumer electronics market. PMID:25196107
High-Sensitivity Low-Noise Miniature Fluxgate Magnetometers Using a Flip Chip Conceptual Design
Lu, Chih-Cheng; Huang, Jeff; Chiu, Po-Kai; Chiu, Shih-Liang; Jeng, Jen-Tzong
2014-01-01
This paper presents a novel class of miniature fluxgate magnetometers fabricated on a print circuit board (PCB) substrate and electrically connected to each other similar to the current “flip chip” concept in semiconductor package. This sensor is soldered together by reversely flipping a 5 cm × 3 cm PCB substrate to the other identical one which includes dual magnetic cores, planar pick-up coils, and 3-D excitation coils constructed by planar Cu interconnections patterned on PCB substrates. Principles and analysis of the fluxgate sensor are introduced first, and followed by FEA electromagnetic modeling and simulation for the proposed sensor. Comprehensive characteristic experiments of the miniature fluxgate device exhibit favorable results in terms of sensitivity (or “responsivity” for magnetometers) and field noise spectrum. The sensor is driven and characterized by employing the improved second-harmonic detection technique that enables linear V-B correlation and responsivity verification. In addition, the double magnitude of responsivity measured under very low frequency (1 Hz) magnetic fields is experimentally demonstrated. As a result, the maximum responsivity of 593 V/T occurs at 50 kHz of excitation frequency with the second harmonic wave of excitation; however, the minimum magnetic field noise is found to be 0.05 nT/Hz1/2 at 1 Hz under the same excitation. In comparison with other miniature planar fluxgates published to date, the fluxgate magnetic sensor with flip chip configuration offers advances in both device functionality and fabrication simplicity. More importantly, the novel design can be further extended to a silicon-based micro-fluxgate chip manufactured by emerging CMOS-MEMS technologies, thus enriching its potential range of applications in modern engineering and the consumer electronics market. PMID:25196107
Space shuttle orbiter digital data processing system timing sensitivity analysis OFT ascent phase
NASA Technical Reports Server (NTRS)
Lagas, J. J.; Peterka, J. J.; Becker, D. A.
1977-01-01
Dynamic loads were investigated to provide simulation and analysis of the space shuttle orbiter digital data processing system (DDPS). Segments of the ascent test (OFT) configuration were modeled utilizing the information management system interpretive model (IMSIM) in a computerized simulation modeling of the OFT hardware and software workload. System requirements for simulation of the OFT configuration were defined, and sensitivity analyses determined areas of potential data flow problems in DDPS operation. Based on the defined system requirements and these sensitivity analyses, a test design was developed for adapting, parameterizing, and executing IMSIM, using varying load and stress conditions for model execution. Analyses of the computer simulation runs are documented, including results, conclusions, and recommendations for DDPS improvements.
Dickson, T.L.; Cheverton, R.D.; Bryson, J.W.; Bass, B.R.; Shum, D.K.M.; Keeney, J.A.
1993-08-01
The Nuclear Regulatory Commission (NRC) requested Oak Ridge National Laboratory (ORNL) to perform a pressurized-thermal-shock (PTS) probabilistic fracture mechanics (PFM) sensitivity analysis for the Yankee Rowe reactor pressure vessel, for the fluences corresponding to the end of operating cycle 22, using a specific small-break-loss- of-coolant transient as the loading condition. Regions of the vessel with distinguishing features were to be treated individually -- upper axial weld, lower axial weld, circumferential weld, upper plate spot welds, upper plate regions between the spot welds, lower plate spot welds, and the lower plate regions between the spot welds. The fracture analysis methods used in the analysis of through-clad surface flaws were those contained in the established OCA-P computer code, which was developed during the Integrated Pressurized Thermal Shock (IPTS) Program. The NRC request specified that the OCA-P code be enhanced for this study to also calculate the conditional probabilities of failure for subclad flaws and embedded flaws. The results of this sensitivity analysis provide the NRC with (1) data that could be used to assess the relative influence of a number of key input parameters in the Yankee Rowe PTS analysis and (2) data that can be used for readily determining the probability of vessel failure once a more accurate indication of vessel embrittlement becomes available. This report is designated as HSST report No. 117.
Gan, Yanjun; Duan, Qingyun; Gong, Wei; Tong, Charles; Sun, Yunwei; Chu, Wei; Ye, Aizhong; Miao, Chiyuan; Di, Zhenhua
2014-01-01
Sensitivity analysis (SA) is a commonly used approach for identifying important parameters that dominate model behaviors. We use a newly developed software package, a Problem Solving environment for Uncertainty Analysis and Design Exploration (PSUADE), to evaluate the effectiveness and efficiency of ten widely used SA methods, including seven qualitative and three quantitative ones. All SA methods are tested using a variety of sampling techniques to screen out the most sensitive (i.e., important) parameters from the insensitive ones. The Sacramento Soil Moisture Accounting (SAC-SMA) model, which has thirteen tunable parameters, is used for illustration. The South Branch Potomac River basin near Springfield, West Virginia in the U.S. is chosen as the study area. The key findings from this study are: (1) For qualitative SA methods, Correlation Analysis (CA), Regression Analysis (RA), and Gaussian Process (GP) screening methods are shown to be not effective in this example. Morris One-At-a-Time (MOAT) screening is the most efficient, needing only 280 samples to identify the most important parameters, but it is the least robust method. Multivariate Adaptive Regression Splines (MARS), Delta Test (DT) and Sum-Of-Trees (SOT) screening methods need about 400–600 samples for the same purpose. Monte Carlo (MC), Orthogonal Array (OA) and Orthogonal Array based Latin Hypercube (OALH) are appropriate sampling techniques for them; (2) For quantitative SA methods, at least 2777 samples are needed for Fourier Amplitude Sensitivity Test (FAST) to identity parameter main effect. McKay method needs about 360 samples to evaluate the main effect, more than 1000 samples to assess the two-way interaction effect. OALH and LPτ (LPTAU) sampling techniques are more appropriate for McKay method. For the Sobol' method, the minimum samples needed are 1050 to compute the first-order and total sensitivity indices correctly. These comparisons show that qualitative SA methods are more efficient
NASA Astrophysics Data System (ADS)
Foglia, L.; Mehl, S.; Hill, M. C.
2013-12-01
sensitivity analysis in model evaluation, possibly as a preliminary step to provide insights that can be used to improve the design of more demanding methods. This can be foreseen as a promising direction of future research where frugal and complex methods are combined to provide insights on model development and model results.
Gan, Yanjun; Duan, Qingyun; Gong, Wei; Tong, Charles; Sun, Yunwei; Chu, Wei; Ye, Aizhong; Miao, Chiyuan; Di, Zhenhua
2014-01-01
Sensitivity analysis (SA) is a commonly used approach for identifying important parameters that dominate model behaviors. We use a newly developed software package, a Problem Solving environment for Uncertainty Analysis and Design Exploration (PSUADE), to evaluate the effectiveness and efficiency of ten widely used SA methods, including seven qualitative and three quantitative ones. All SA methods are tested using a variety of sampling techniques to screen out the most sensitive (i.e., important) parameters from the insensitive ones. The Sacramento Soil Moisture Accounting (SAC-SMA) model, which has thirteen tunable parameters, is used for illustration. The South Branch Potomac River basin nearmore » Springfield, West Virginia in the U.S. is chosen as the study area. The key findings from this study are: (1) For qualitative SA methods, Correlation Analysis (CA), Regression Analysis (RA), and Gaussian Process (GP) screening methods are shown to be not effective in this example. Morris One-At-a-Time (MOAT) screening is the most efficient, needing only 280 samples to identify the most important parameters, but it is the least robust method. Multivariate Adaptive Regression Splines (MARS), Delta Test (DT) and Sum-Of-Trees (SOT) screening methods need about 400–600 samples for the same purpose. Monte Carlo (MC), Orthogonal Array (OA) and Orthogonal Array based Latin Hypercube (OALH) are appropriate sampling techniques for them; (2) For quantitative SA methods, at least 2777 samples are needed for Fourier Amplitude Sensitivity Test (FAST) to identity parameter main effect. McKay method needs about 360 samples to evaluate the main effect, more than 1000 samples to assess the two-way interaction effect. OALH and LPτ (LPTAU) sampling techniques are more appropriate for McKay method. For the Sobol' method, the minimum samples needed are 1050 to compute the first-order and total sensitivity indices correctly. These comparisons show that qualitative SA methods are more
Low speed airfoil design and analysis
NASA Technical Reports Server (NTRS)
Eppler, R.; Somers, D. M.
1979-01-01
A low speed airfoil design and analysis program was developed which contains several unique features. In the design mode, the velocity distribution is not specified for one but many different angles of attack. Several iteration options are included which allow the trailing edge angle to be specified while other parameters are iterated. For airfoil analysis, a panel method is available which uses third-order panels having parabolic vorticity distributions. The flow condition is satisfied at the end points of the panels. Both sharp and blunt trailing edges can be analyzed. The integral boundary layer method with its laminar separation bubble analog, empirical transition criterion, and precise turbulent boundary layer equations compares very favorably with other methods, both integral and finite difference. Comparisons with experiment for several airfoils over a very wide Reynolds number range are discussed. Applications to high lift airfoil design are also demonstrated.
Multifidelity Analysis and Optimization for Supersonic Design
NASA Technical Reports Server (NTRS)
Kroo, Ilan; Willcox, Karen; March, Andrew; Haas, Alex; Rajnarayan, Dev; Kays, Cory
2010-01-01
Supersonic aircraft design is a computationally expensive optimization problem and multifidelity approaches over a significant opportunity to reduce design time and computational cost. This report presents tools developed to improve supersonic aircraft design capabilities including: aerodynamic tools for supersonic aircraft configurations; a systematic way to manage model uncertainty; and multifidelity model management concepts that incorporate uncertainty. The aerodynamic analysis tools developed are appropriate for use in a multifidelity optimization framework, and include four analysis routines to estimate the lift and drag of a supersonic airfoil, a multifidelity supersonic drag code that estimates the drag of aircraft configurations with three different methods: an area rule method, a panel method, and an Euler solver. In addition, five multifidelity optimization methods are developed, which include local and global methods as well as gradient-based and gradient-free techniques.
NASA Astrophysics Data System (ADS)
Mack, Yolanda
Computational fluid dynamics (CFD) can be used to improve the design and optimization of rocket engine components that traditionally rely on empirical calculations and limited experimentation. CFD based-design optimization can be made computationally affordable through the use of surrogate modeling which can then facilitate additional parameter sensitivity assessments. The present study investigates surrogate-based adaptive design space refinement (DSR) using estimates of surrogate uncertainty to probe the CFD analyses and to perform sensitivity assessments for complex fluid physics associated with liquid rocket engine components. Three studies were conducted. First, a surrogate-based preliminary design optimization was conducted to improve the efficiency of a compact radial turbine for an expander cycle rocket engine while maintaining low weight. Design space refinement was used to identify function constraints and to obtain a high accuracy surrogate model in the region of interest. A merit function formulation for multi-objective design point selection reduced the number of design points by an order of magnitude while maintaining good surrogate accuracy among the best trade-off points. Second, bluff body-induced flow was investigated to identify the physics and surrogate modeling issues related to the flow's mixing dynamics. Multiple surrogates and DSR were instrumental in identifying designs for which the CFD model was deficient and to help to pinpoint the nature of the deficiency. Next, a three-dimensional computational model was developed to explore the wall heat transfer of a GO2/GH2 shear coaxial single element injector. The interactions between turbulent recirculating flow structures, chemical kinetics, and heat transfer are highlighted. Finally, a simplified computational model of multi-element injector flows was constructed to explore the sensitivity of wall heating and improve combustion efficiency to injector element spacing. Design space refinement
Design analysis, robust methods, and stress classification
Bees, W.J.
1993-01-01
This special edition publication volume is comprised of papers presented at the 1993 ASME Pressure Vessels and Piping Conference, July 25--29, 1993 in Denver, Colorado. The papers were prepared for presentations in technical sessions developed under the auspices of the PVPD Committees on Computer Technology, Design and Analysis, Operations Applications and Components. The topics included are: Analysis of Pressure Vessels and Components; Expansion Joints; Robust Methods; Stress Classification; and Non-Linear Analysis. Individual papers have been processed separately for inclusion in the appropriate data bases.
A comprehensive sensitivity analysis of central-loop MRS data
NASA Astrophysics Data System (ADS)
Behroozmand, Ahmad; Auken, Esben; Dalgaard, Esben; Rejkjaer, Simon
2014-05-01
In this study we investigate the sensitivity analysis of separated-loop magnetic resonance sounding (MRS) data and, in light of deploying a separate MRS receiver system from the transmitter system, compare the parameter determination of the central-loop with the conventional coincident-loop MRS data. MRS, also called surface NMR, has emerged as a promising surface-based geophysical technique for groundwater investigations, as it provides a direct estimate of the water content and, through empirical relations, is linked to hydraulic properties of the subsurface such as hydraulic conductivity. The method works based on the physical principle of NMR during which a large volume of protons of the water molecules in the subsurface is excited at the specific Larmor frequency. The measurement consists of a large wire loop deployed on the surface which typically acts as both a transmitter and a receiver, the so-called coincident-loop configuration. An alternating current is passed through the loop deployed and the superposition of signals from all precessing protons within the investigated volume is measured in a receiver loop; a decaying NMR signal called Free Induction Decay (FID). To provide depth information, the FID signal is measured for a series of pulse moments (Q; product of current amplitude and transmitting pulse length) during which different earth volumes are excited. One of the main and inevitable limitations of MRS measurements is a relatively long measurement dead time, i.e. a non-zero time between the end of the energizing pulse and the beginning of the measurement, which makes it difficult, and in some places impossible, to record MRS signal from fine-grained geologic units and limits the application of advanced pulse sequences. Therefore, one of the current research activities is the idea of building separate receiver units, which will diminish the dead time. In light of that, the aims of this study are twofold: 1) Using a forward modeling approach, the
Sensitivity analysis of surface runoff generation in urban flood forecasting.
Simões, N E; Leitão, J P; Maksimović, C; Sá Marques, A; Pina, R
2010-01-01
Reliable flood forecasting requires hydraulic models capable to estimate pluvial flooding fast enough in order to enable successful operational responses. Increased computational speed can be achieved by using a 1D/1D model, since 2D models are too computationally demanding. Further changes can be made by simplifying 1D network models, removing and by changing some secondary elements. The Urban Water Research Group (UWRG) of Imperial College London developed a tool that automatically analyses, quantifies and generates 1D overland flow network. The overland flow network features (ponds and flow pathways) generated by this methodology are dependent on the number of sewer network manholes and sewer inlets, as some of the overland flow pathways start at manholes (or sewer inlets) locations. Thus, if a simplified version of the sewer network has less manholes (or sewer inlets) than the original one, the overland flow network will be consequently different. This paper compares different overland flow networks generated with different levels of sewer network skeletonisation. Sensitivity analysis is carried out in one catchment area in Coimbra, Portugal, in order to evaluate overland flow network characteristics. PMID:20453333
Potassium Buffering in the Neurovascular Unit: Models and Sensitivity Analysis
Witthoft, Alexandra; Filosa, Jessica A.; Karniadakis, George Em
2013-01-01
Astrocytes are critical regulators of neural and neurovascular network communication. Potassium transport is a central mechanism behind their many functions. Astrocytes encircle synapses with their distal processes, which express two potassium pumps (Na-K and NKCC) and an inward rectifying potassium channel (Kir), whereas the vessel-adjacent endfeet express Kir and BK potassium channels. We provide a detailed model of potassium flow throughout the neurovascular unit (synaptic region, astrocytes, and arteriole) for the cortex of the young brain. Our model reproduces several phenomena observed experimentally: functional hyperemia, in which neural activity triggers astrocytic potassium release at the perivascular endfoot, inducing arteriole dilation; K+ undershoot in the synaptic space after periods of neural activity; neurally induced astrocyte hyperpolarization during Kir blockade. Our results suggest that the dynamics of the vascular response during functional hyperemia are governed by astrocytic Kir for the fast onset and astrocytic BK for maintaining dilation. The model supports the hypothesis that K+ undershoot is caused by excessive astrocytic uptake through Na-K and NKCC pumps, whereas the effect is balanced by Kir. We address parametric uncertainty using high-dimensional stochastic sensitivity analysis and identify possible model limitations. PMID:24209849
Sensitivity analysis and optimization of the nuclear fuel cycle
Passerini, S.; Kazimi, M. S.; Shwageraus, E.
2012-07-01
A sensitivity study has been conducted to assess the robustness of the conclusions presented in the MIT Fuel Cycle Study. The Once Through Cycle (OTC) is considered as the base-line case, while advanced technologies with fuel recycling characterize the alternative fuel cycles. The options include limited recycling in LWRs and full recycling in fast reactors and in high conversion LWRs. Fast reactor technologies studied include both oxide and metal fueled reactors. The analysis allowed optimization of the fast reactor conversion ratio with respect to desired fuel cycle performance characteristics. The following parameters were found to significantly affect the performance of recycling technologies and their penetration over time: Capacity Factors of the fuel cycle facilities, Spent Fuel Cooling Time, Thermal Reprocessing Introduction Date, and in core and Out-of-core TRU Inventory Requirements for recycling technology. An optimization scheme of the nuclear fuel cycle is proposed. Optimization criteria and metrics of interest for different stakeholders in the fuel cycle (economics, waste management, environmental impact, etc.) are utilized for two different optimization techniques (linear and stochastic). Preliminary results covering single and multi-variable and single and multi-objective optimization demonstrate the viability of the optimization scheme. (authors)
Decentred comparative research: Context sensitive analysis of maternal health care.
Wrede, Sirpa; Benoit, Cecilia; Bourgeault, Ivy Lynn; van Teijlingen, Edwin R; Sandall, Jane; De Vries, Raymond G
2006-12-01
Cross-national comparison is an important tool for health care research, but too often those who use this method fail to consider important inter-national differences in the social organisation of health care and in the relationship between health care practices and social experience. In this article we make the case for a context-sensitive and reflexive analysis of health care that allows researchers to understand the important ways that health care systems and practices are situated in time and place. Our approach--decentred comparative research--addresses the often unacknowledged ethnocentrism of traditional comparative research. Decentred cross-national research is a method that draws on the socially situated and distributed expertise of an international research team to develop key concepts and research questions. We used the decentred method to fashion a multilevel framework that used the meso level of organisation (i.e., health care organisations, professional groups and other concrete organisations) as an analytical starting point in our international study of maternity care in eight countries. Our method departs from traditional comparative health systems research that is most often conducted at the macro level. Our approach will help researchers develop new and socially robust knowledge about health care. PMID:16962695