Science.gov

Sample records for experimental design optimization

  1. Optimal Experimental Design for Model Discrimination

    PubMed Central

    Myung, Jay I.; Pitt, Mark A.

    2009-01-01

    Models of a psychological process can be difficult to discriminate experimentally because it is not easy to determine the values of the critical design variables (e.g., presentation schedule, stimulus structure) that will be most informative in differentiating them. Recent developments in sampling-based search methods in statistics make it possible to determine these values, and thereby identify an optimal experimental design. After describing the method, it is demonstrated in two content areas in cognitive psychology in which models are highly competitive: retention (i.e., forgetting) and categorization. The optimal design is compared with the quality of designs used in the literature. The findings demonstrate that design optimization has the potential to increase the informativeness of the experimental method. PMID:19618983

  2. Parameter estimation and optimal experimental design.

    PubMed

    Banga, Julio R; Balsa-Canto, Eva

    2008-01-01

    Mathematical models are central in systems biology and provide new ways to understand the function of biological systems, helping in the generation of novel and testable hypotheses, and supporting a rational framework for possible ways of intervention, like in e.g. genetic engineering, drug development or treatment of diseases. Since the amount and quality of experimental 'omics' data continue to increase rapidly, there is great need for methods for proper model building which can handle this complexity. In the present chapter we review two key steps of the model building process, namely parameter estimation (model calibration) and optimal experimental design. Parameter estimation aims to find the unknown parameters of the model which give the best fit to a set of experimental data. Optimal experimental design aims to devise the dynamic experiments which provide the maximum information content for subsequent non-linear model identification, estimation and/or discrimination. We place emphasis on the need for robust global optimization methods for proper solution of these problems, and we present a motivating example considering a cell signalling model. PMID:18793133

  3. Optimizing Experimental Designs: Finding Hidden Treasure.

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Classical experimental design theory, the predominant treatment in most textbooks, promotes the use of blocking designs for control of spatial variability in field studies and other situations in which there is significant variation among heterogeneity among experimental units. Many blocking design...

  4. Optimal Experimental Design for Model Discrimination

    ERIC Educational Resources Information Center

    Myung, Jay I.; Pitt, Mark A.

    2009-01-01

    Models of a psychological process can be difficult to discriminate experimentally because it is not easy to determine the values of the critical design variables (e.g., presentation schedule, stimulus structure) that will be most informative in differentiating them. Recent developments in sampling-based search methods in statistics make it…

  5. Achieving optimal SERS through enhanced experimental design

    PubMed Central

    Fisk, Heidi; Westley, Chloe; Turner, Nicholas J.

    2016-01-01

    One of the current limitations surrounding surface‐enhanced Raman scattering (SERS) is the perceived lack of reproducibility. SERS is indeed challenging, and for analyte detection, it is vital that the analyte interacts with the metal surface. However, as this is analyte dependent, there is not a single set of SERS conditions that are universal. This means that experimental optimisation for optimum SERS response is vital. Most researchers optimise one factor at a time, where a single parameter is altered first before going onto optimise the next. This is a very inefficient way of searching the experimental landscape. In this review, we explore the use of more powerful multivariate approaches to SERS experimental optimisation based on design of experiments and evolutionary computational methods. We particularly focus on colloidal‐based SERS rather than thin film preparations as a result of their popularity. © 2015 The Authors. Journal of Raman Spectroscopy published by John Wiley & Sons, Ltd.

  6. Ceramic processing: Experimental design and optimization

    NASA Technical Reports Server (NTRS)

    Weiser, Martin W.; Lauben, David N.; Madrid, Philip

    1992-01-01

    The objectives of this paper are to: (1) gain insight into the processing of ceramics and how green processing can affect the properties of ceramics; (2) investigate the technique of slip casting; (3) learn how heat treatment and temperature contribute to density, strength, and effects of under and over firing to ceramic properties; (4) experience some of the problems inherent in testing brittle materials and learn about the statistical nature of the strength of ceramics; (5) investigate orthogonal arrays as tools to examine the effect of many experimental parameters using a minimum number of experiments; (6) recognize appropriate uses for clay based ceramics; and (7) measure several different properties important to ceramic use and optimize them for a given application.

  7. Simultaneous optimal experimental design for in vitro binding parameter estimation.

    PubMed

    Ernest, C Steven; Karlsson, Mats O; Hooker, Andrew C

    2013-10-01

    Simultaneous optimization of in vitro ligand binding studies using an optimal design software package that can incorporate multiple design variables through non-linear mixed effect models and provide a general optimized design regardless of the binding site capacity and relative binding rates for a two binding system. Experimental design optimization was employed with D- and ED-optimality using PopED 2.8 including commonly encountered factors during experimentation (residual error, between experiment variability and non-specific binding) for in vitro ligand binding experiments: association, dissociation, equilibrium and non-specific binding experiments. Moreover, a method for optimizing several design parameters (ligand concentrations, measurement times and total number of samples) was examined. With changes in relative binding site density and relative binding rates, different measurement times and ligand concentrations were needed to provide precise estimation of binding parameters. However, using optimized design variables, significant reductions in number of samples provided as good or better precision of the parameter estimates compared to the original extensive sampling design. Employing ED-optimality led to a general experimental design regardless of the relative binding site density and relative binding rates. Precision of the parameter estimates were as good as the extensive sampling design for most parameters and better for the poorly estimated parameters. Optimized designs for in vitro ligand binding studies provided robust parameter estimation while allowing more efficient and cost effective experimentation by reducing the measurement times and separate ligand concentrations required and in some cases, the total number of samples. PMID:23943088

  8. Optimal experimental design and therapeutic drug monitoring.

    PubMed

    Kulcsár, C; Pronzato, L; Walter, E

    1994-06-01

    A simple example of intravenous theophylline therapy is used to present and compare various drug administration policies based on stochastic control theory. The simplest approach (Heuristic-Certainty-Equivalence (HCE) control) assumes that the model parameters are known. Prior uncertainty on these parameters can be taken into account by using average optimal (AO) control. The available knowledge about the system can be improved by measuring the drug concentration some time after the beginning of the treatment. This corresponds to the notion of feedback and leads to the HCE feedback (HCEF) and AO feedback (AOF) policies. A further step towards optimality consists in choosing the optimal measurement time given that the final purpose is the control of the system and not the estimation of its parameters. Finally, closed-loop optimal (CLO) control optimally chooses both the dosage regimen and measurement time. PMID:7927864

  9. Optimizing experimental design for comparing models of brain function.

    PubMed

    Daunizeau, Jean; Preuschoff, Kerstin; Friston, Karl; Stephan, Klaas

    2011-11-01

    This article presents the first attempt to formalize the optimization of experimental design with the aim of comparing models of brain function based on neuroimaging data. We demonstrate our approach in the context of Dynamic Causal Modelling (DCM), which relates experimental manipulations to observed network dynamics (via hidden neuronal states) and provides an inference framework for selecting among candidate models. Here, we show how to optimize the sensitivity of model selection by choosing among experimental designs according to their respective model selection accuracy. Using Bayesian decision theory, we (i) derive the Laplace-Chernoff risk for model selection, (ii) disclose its relationship with classical design optimality criteria and (iii) assess its sensitivity to basic modelling assumptions. We then evaluate the approach when identifying brain networks using DCM. Monte-Carlo simulations and empirical analyses of fMRI data from a simple bimanual motor task in humans serve to demonstrate the relationship between network identification and the optimal experimental design. For example, we show that deciding whether there is a feedback connection requires shorter epoch durations, relative to asking whether there is experimentally induced change in a connection that is known to be present. Finally, we discuss limitations and potential extensions of this work. PMID:22125485

  10. Decision-oriented Optimal Experimental Design and Data Collection

    NASA Astrophysics Data System (ADS)

    O'Malley, D.; Vesselinov, V. V.

    2015-12-01

    Classical optimal experimental design is a branch of statistics that seeks to construct ("design") a data collection effort ("experiment") that minimizes ("optimal") the uncertainty associated with some quantity of interest. In many real world problems, we are interested in these quantities to help us make a decision. Minimizing the uncertainty associated with the quantity can help inform the decision, but a more holistic approach is possible where the experiment is designed to maximize the information that it provides to the decision-making process. The difference is subtle, but it amounts to focusing on the end-goal (the decision) rather than an intermediary (the quantity). We describe one approach to decision-oriented optimal experimental design that utilizes Bayesian-Information-Gap Decision Theory which combines probabilistic and non-probabilistic methods for uncertainty quantification. In this approach, experimental designs that have a high probability of altering the decision are deemed worthwhile. On the other hand, experimental designs that have little chance or no of altering the decision need not be performed.

  11. Optimal active vibration absorber - Design and experimental results

    NASA Technical Reports Server (NTRS)

    Lee-Glauser, Gina; Juang, Jer-Nan; Sulla, Jeffrey L.

    1993-01-01

    An optimal active vibration absorber can provide guaranteed closed-loop stability and control for large flexible space structures with collocated sensors/actuators. The active vibration absorber is a second-order dynamic system which is designed to suppress any unwanted structural vibration. This can be designed with minimum knowledge of the controlled system. Two methods for optimizing the active vibration absorber parameters are illustrated: minimum resonant amplitude and frequency matched active controllers. The Controls-Structures Interaction Phase-1 Evolutionary Model at NASA LaRC is used to demonstrate the effectiveness of the active vibration absorber for vibration suppression. Performance is compared numerically and experimentally using acceleration feedback.

  12. Optimal active vibration absorber: Design and experimental results

    NASA Technical Reports Server (NTRS)

    Lee-Glauser, Gina; Juang, Jer-Nan; Sulla, Jeffrey L.

    1992-01-01

    An optimal active vibration absorber can provide guaranteed closed-loop stability and control for large flexible space structures with collocated sensors/actuators. The active vibration absorber is a second-order dynamic system which is designed to suppress any unwanted structural vibration. This can be designed with minimum knowledge of the controlled system. Two methods for optimizing the active vibration absorber parameters are illustrated: minimum resonant amplitude and frequency matched active controllers. The Controls-Structures Interaction Phase-1 Evolutionary Model at NASA LaRC is used to demonstrate the effectiveness of the active vibration absorber for vibration suppression. Performance is compared numerically and experimentally using acceleration feedback.

  13. Criteria for the optimal design of experimental tests

    NASA Technical Reports Server (NTRS)

    Canavos, G. C.

    1972-01-01

    Some of the basic concepts are unified that were developed for the problem of finding optimal approximating functions which relate a set of controlled variables to a measurable response. The techniques have the potential for reducing the amount of testing required in experimental investigations. Specifically, two low-order polynomial models are considered as approximations to unknown functionships. For each model, optimal means of designing experimental tests are presented which, for a modest number of measurements, yield prediction equations that minimize the error of an estimated response anywhere inside a selected region of experimentation. Moreover, examples are provided for both models to illustrate their use. Finally, an analysis of a second-order prediction equation is given to illustrate ways of determining maximum or minimum responses inside the experimentation region.

  14. Optimizing an experimental design for an electromagnetic experiment

    NASA Astrophysics Data System (ADS)

    Roux, Estelle; Garcia, Xavier

    2013-04-01

    Most of geophysical studies focus on data acquisition and analysis, but another aspect which is gaining importance is the discussion on acquisition of suitable datasets. This can be done through the design of an optimal experiment. Optimizing an experimental design implies a compromise between maximizing the information we get about the target and reducing the cost of the experiment, considering a wide range of constraints (logistical, financial, experimental …). We are currently developing a method to design an optimal controlled-source electromagnetic (CSEM) experiment to detect a potential CO2 reservoir and monitor this reservoir during and after CO2 injection. Our statistical algorithm combines the use of linearized inverse theory (to evaluate the quality of one given design via the objective function) and stochastic optimization methods like genetic algorithm (to examine a wide range of possible surveys). The particularity of our method is that it uses a multi-objective genetic algorithm that searches for designs that fit several objective functions simultaneously. One main advantage of this kind of technique to design an experiment is that it does not require the acquisition of any data and can thus be easily conducted before any geophysical survey. Our new experimental design algorithm has been tested with a realistic one-dimensional resistivity model of the Earth in the region of study (northern Spain CO2 sequestration test site). We show that a small number of well distributed observations have the potential to resolve the target. This simple test also points out the importance of a well chosen objective function. Finally, in the context of CO2 sequestration that motivates this study, we might be interested in maximizing the information we get about the reservoir layer. In that case, we show how the combination of two different objective functions considerably improve its resolution.

  15. A Hierarchical Adaptive Approach to Optimal Experimental Design

    PubMed Central

    Kim, Woojae; Pitt, Mark A.; Lu, Zhong-Lin; Steyvers, Mark; Myung, Jay I.

    2014-01-01

    Experimentation is at the core of research in the behavioral and neural sciences, yet observations can be expensive and time-consuming to acquire (e.g., MRI scans, responses from infant participants). A major interest of researchers is designing experiments that lead to maximal accumulation of information about the phenomenon under study with the fewest possible number of observations. In addressing this challenge, statisticians have developed adaptive design optimization methods. This letter introduces a hierarchical Bayes extension of adaptive design optimization that provides a judicious way to exploit two complementary schemes of inference (with past and future data) to achieve even greater accuracy and efficiency in information gain. We demonstrate the method in a simulation experiment in the field of visual perception. PMID:25149697

  16. Simulation-based optimal Bayesian experimental design for nonlinear systems

    SciTech Connect

    Huan, Xun; Marzouk, Youssef M.

    2013-01-01

    The optimal selection of experimental conditions is essential to maximizing the value of data for inference and prediction, particularly in situations where experiments are time-consuming and expensive to conduct. We propose a general mathematical framework and an algorithmic approach for optimal experimental design with nonlinear simulation-based models; in particular, we focus on finding sets of experiments that provide the most information about targeted sets of parameters. Our framework employs a Bayesian statistical setting, which provides a foundation for inference from noisy, indirect, and incomplete data, and a natural mechanism for incorporating heterogeneous sources of information. An objective function is constructed from information theoretic measures, reflecting expected information gain from proposed combinations of experiments. Polynomial chaos approximations and a two-stage Monte Carlo sampling method are used to evaluate the expected information gain. Stochastic approximation algorithms are then used to make optimization feasible in computationally intensive and high-dimensional settings. These algorithms are demonstrated on model problems and on nonlinear parameter inference problems arising in detailed combustion kinetics.

  17. Computational procedures for optimal experimental design in biological systems.

    PubMed

    Balsa-Canto, E; Alonso, A A; Banga, J R

    2008-07-01

    Mathematical models of complex biological systems, such as metabolic or cell-signalling pathways, usually consist of sets of nonlinear ordinary differential equations which depend on several non-measurable parameters that can be hopefully estimated by fitting the model to experimental data. However, the success of this fitting is largely conditioned by the quantity and quality of data. Optimal experimental design (OED) aims to design the scheme of actuations and measurements which will result in data sets with the maximum amount and/or quality of information for the subsequent model calibration. New methods and computational procedures for OED in the context of biological systems are presented. The OED problem is formulated as a general dynamic optimisation problem where the time-dependent stimuli profiles, the location of sampling times, the duration of the experiments and the initial conditions are regarded as design variables. Its solution is approached using the control vector parameterisation method. Since the resultant nonlinear optimisation problem is in most of the cases non-convex, the use of a robust global nonlinear programming solver is proposed. For the sake of comparing among different experimental schemes, a Monte-Carlo-based identifiability analysis is then suggested. The applicability and advantages of the proposed techniques are illustrated by considering an example related to a cell-signalling pathway. PMID:18681746

  18. Prediction uncertainty and optimal experimental design for learning dynamical systems

    NASA Astrophysics Data System (ADS)

    Letham, Benjamin; Letham, Portia A.; Rudin, Cynthia; Browne, Edward P.

    2016-06-01

    Dynamical systems are frequently used to model biological systems. When these models are fit to data, it is necessary to ascertain the uncertainty in the model fit. Here, we present prediction deviation, a metric of uncertainty that determines the extent to which observed data have constrained the model's predictions. This is accomplished by solving an optimization problem that searches for a pair of models that each provides a good fit for the observed data, yet has maximally different predictions. We develop a method for estimating a priori the impact that additional experiments would have on the prediction deviation, allowing the experimenter to design a set of experiments that would most reduce uncertainty. We use prediction deviation to assess uncertainty in a model of interferon-alpha inhibition of viral infection, and to select a sequence of experiments that reduces this uncertainty. Finally, we prove a theoretical result which shows that prediction deviation provides bounds on the trajectories of the underlying true model. These results show that prediction deviation is a meaningful metric of uncertainty that can be used for optimal experimental design.

  19. Optimization of model parameters and experimental designs with the Optimal Experimental Design Toolbox (v1.0) exemplified by sedimentation in salt marshes

    NASA Astrophysics Data System (ADS)

    Reimer, J.; Schuerch, M.; Slawig, T.

    2015-03-01

    The geosciences are a highly suitable field of application for optimizing model parameters and experimental designs especially because many data are collected. In this paper, the weighted least squares estimator for optimizing model parameters is presented together with its asymptotic properties. A popular approach to optimize experimental designs called local optimal experimental designs is described together with a lesser known approach which takes into account the potential nonlinearity of the model parameters. These two approaches have been combined with two methods to solve their underlying discrete optimization problem. All presented methods were implemented in an open-source MATLAB toolbox called the Optimal Experimental Design Toolbox whose structure and application is described. In numerical experiments, the model parameters and experimental design were optimized using this toolbox. Two existing models for sediment concentration in seawater and sediment accretion on salt marshes of different complexity served as an application example. The advantages and disadvantages of these approaches were compared based on these models. Thanks to optimized experimental designs, the parameters of these models could be determined very accurately with significantly fewer measurements compared to unoptimized experimental designs. The chosen optimization approach played a minor role for the accuracy; therefore, the approach with the least computational effort is recommended.

  20. Optimizing Experimental Designs Relative to Costs and Effect Sizes.

    ERIC Educational Resources Information Center

    Headrick, Todd C.; Zumbo, Bruno D.

    A general model is derived for the purpose of efficiently allocating integral numbers of units in multi-level designs given prespecified power levels. The derivation of the model is based on a constrained optimization problem that maximizes a general form of a ratio of expected mean squares subject to a budget constraint. This model provides more…

  1. Optimization of fast disintegration tablets using pullulan as diluent by central composite experimental design.

    PubMed

    Patel, Dipil; Chauhan, Musharraf; Patel, Ravi; Patel, Jayvadan

    2012-03-01

    The objective of this work was to apply central composite experimental design to investigate main and interaction effect of formulation parameters in optimizing novel fast disintegration tablets formulation using pullulan as diluents. Face centered central composite experimental design was employed to optimize fast disintegration tablet formulation. The variables studied were concentration of diluents (pullulan, X(1)), superdisintigrant (sodium starch glycolate, X(2)), and direct compression aid (spray dried lactose, X(3)). Tablets were characterized for weight variation, thickness, disintegration time (Y(1)) and hardness (Y(2)). Good correlation between the predicted values and experimental data of the optimized formulation methodology in optimizing fast disintegrating tablets using pullulan as a diluent. PMID:23066220

  2. Optimization of experimental designs and model parameters exemplified by sedimentation in salt marshes

    NASA Astrophysics Data System (ADS)

    Reimer, J.; Schürch, M.; Slawig, T.

    2014-09-01

    The weighted least squares estimator for model parameters was presented together with its asymptotic properties. A popular approach to optimize experimental designs called local optimal experimental designs was described together with a lesser known approach which takes into account a potential nonlinearity of the model parameters. These two approaches were combined with two different methods to solve their underlying discrete optimization problem. All presented methods were implemented in an open source MATLAB toolbox called the Optimal Experimental Design Toolbox whose structure and handling was described. In numerical experiments, the model parameters and experimental design were optimized using this toolbox. Two models for sediment concentration in seawater of different complexity served as application example. The advantages and disadvantages of the different approaches were compared, and an evaluation of the approaches was performed.

  3. Optimization of experimental designs by incorporating NIF facility impacts

    NASA Astrophysics Data System (ADS)

    Eder, D. C.; Whitman, P. K.; Koniges, A. E.; Anderson, R. W.; Wang, P.; Gunney, B. T.; Parham, T. G.; Koerner, J. G.; Dixit, S. N.; Suratwala, T. I.; Blue, B. E.; Hansen, J. F.; Tobin, M. T.; Robey, H. F.; Spaeth, M. L.; MacGowan, B. J.

    2006-06-01

    For experimental campaigns on the National Ignition Facility (NIF) to be successful, they must obtain useful data without causing unacceptable impact on the facility. Of particular concern is excessive damage to optics and diagnostic components. There are 192 fused silica main debris shields (MDS) exposed to the potentially hostile target chamber environment on each shot. Damage in these optics results either from the interaction of laser light with contamination and pre-existing imperfections on the optic surface or from the impact of shrapnel fragments. Mitigation of this second damage source is possible by identifying shrapnel sources and shielding optics from them. It was recently demonstrated that the addition of 1.1-mm thick borosilicate disposable debris shields (DDS) blocks the majority of debris and shrapnel fragments from reaching the relatively expensive MDS's. However, DDS's cannot stop large, fast moving fragments. We have experimentally demonstrated one shrapnel mitigation technique showing that it is possible to direct fast moving fragments by changing the source orientation, in this case a Ta pinhole array. Another mitigation method is to change the source material to one that produces smaller fragments. Simulations and validating experiments are necessary to determine which fragments can penetrate or break 1-3 mm thick DDS's. Three-dimensional modeling of complex target-diagnostic configurations is necessary to predict the size, velocity, and spatial distribution of shrapnel fragments. The tools we are developing will be used to assure that all NIF experimental campaigns meet the requirements on allowed level of debris and shrapnel generation.

  4. Analytical and experimental performance of optimal controller designs for a supersonic inlet

    NASA Technical Reports Server (NTRS)

    Zeller, J. R.; Lehtinen, B.; Geyser, L. C.; Batterton, P. G.

    1973-01-01

    The techniques of modern optimal control theory were applied to the design of a control system for a supersonic inlet. The inlet control problem was approached as a linear stochastic optimal control problem using as the performance index the expected frequency of unstarts. The details of the formulation of the stochastic inlet control problem are presented. The computational procedures required to obtain optimal controller designs are discussed, and the analytically predicted performance of controllers designed for several different inlet conditions is tabulated. The experimental implementation of the optimal control laws is described, and the experimental results obtained in a supersonic wind tunnel are presented. The control laws were implemented with analog and digital computers. Comparisons are made between the experimental and analytically predicted performance results. Comparisons are also made between the results obtained with continuous analog computer controllers and discrete digital computer versions.

  5. Optimal experimental designs for dose-response studies with continuous endpoints.

    PubMed

    Holland-Letz, Tim; Kopp-Schneider, Annette

    2015-11-01

    In most areas of clinical and preclinical research, the required sample size determines the costs and effort for any project, and thus, optimizing sample size is of primary importance. An experimental design of dose-response studies is determined by the number and choice of dose levels as well as the allocation of sample size to each level. The experimental design of toxicological studies tends to be motivated by convention. Statistical optimal design theory, however, allows the setting of experimental conditions (dose levels, measurement times, etc.) in a way which minimizes the number of required measurements and subjects to obtain the desired precision of the results. While the general theory is well established, the mathematical complexity of the problem so far prevents widespread use of these techniques in practical studies. The paper explains the concepts of statistical optimal design theory with a minimum of mathematical terminology and uses these concepts to generate concrete usable D-optimal experimental designs for dose-response studies on the basis of three common dose-response functions in toxicology: log-logistic, log-normal and Weibull functions with four parameters each. The resulting designs usually require control plus only three dose levels and are quite intuitively plausible. The optimal designs are compared to traditional designs such as the typical setup of cytotoxicity studies for 96-well plates. As the optimal design depends on prior estimates of the dose-response function parameters, it is shown what loss of efficiency occurs if the parameters for design determination are misspecified, and how Bayes optimal designs can improve the situation. PMID:25155192

  6. Using a hybrid approach to optimize experimental network design for aquifer parameter identification.

    PubMed

    Chang, Liang-Cheng; Chu, Hone-Jay; Lin, Yu-Pin; Chen, Yu-Wen

    2010-10-01

    This research develops an optimum design model of groundwater network using genetic algorithm (GA) and modified Newton approach, based on the experimental design conception. The goal of experiment design is to minimize parameter uncertainty, represented by the covariance matrix determinant of estimated parameters. The design problem is constrained by a specified cost and solved by GA and a parameter identification model. The latter estimates optimum parameter value and its associated sensitivity matrices. The general problem is simplified into two classes of network design problems: an observation network design problem and a pumping network design problem. Results explore the relationship between the experimental design and the physical processes. The proposed model provides an alternative to solve optimization problems for groundwater experimental design. PMID:19757116

  7. Optimization of experimental design in fMRI: a general framework using a genetic algorithm.

    PubMed

    Wager, Tor D; Nichols, Thomas E

    2003-02-01

    This article describes a method for selecting design parameters and a particular sequence of events in fMRI so as to maximize statistical power and psychological validity. Our approach uses a genetic algorithm (GA), a class of flexible search algorithms that optimize designs with respect to single or multiple measures of fitness. Two strengths of the GA framework are that (1) it operates with any sort of model, allowing for very specific parameterization of experimental conditions, including nonstandard trial types and experimentally observed scanner autocorrelation, and (2) it is flexible with respect to fitness criteria, allowing optimization over known or novel fitness measures. We describe how genetic algorithms may be applied to experimental design for fMRI, and we use the framework to explore the space of possible fMRI design parameters, with the goal of providing information about optimal design choices for several types of designs. In our simulations, we considered three fitness measures: contrast estimation efficiency, hemodynamic response estimation efficiency, and design counterbalancing. Although there are inherent trade-offs between these three fitness measures, GA optimization can produce designs that outperform random designs on all three criteria simultaneously. PMID:12595184

  8. Optimal experimental designs for the estimation of thermal properties of composite materials

    NASA Technical Reports Server (NTRS)

    Scott, Elaine P.; Moncman, Deborah A.

    1994-01-01

    Reliable estimation of thermal properties is extremely important in the utilization of new advanced materials, such as composite materials. The accuracy of these estimates can be increased if the experiments are designed carefully. The objectives of this study are to design optimal experiments to be used in the prediction of these thermal properties and to then utilize these designs in the development of an estimation procedure to determine the effective thermal properties (thermal conductivity and volumetric heat capacity). The experiments were optimized by choosing experimental parameters that maximize the temperature derivatives with respect to all of the unknown thermal properties. This procedure has the effect of minimizing the confidence intervals of the resulting thermal property estimates. Both one-dimensional and two-dimensional experimental designs were optimized. A heat flux boundary condition is required in both analyses for the simultaneous estimation of the thermal properties. For the one-dimensional experiment, the parameters optimized were the heating time of the applied heat flux, the temperature sensor location, and the experimental time. In addition to these parameters, the optimal location of the heat flux was also determined for the two-dimensional experiments. Utilizing the optimal one-dimensional experiment, the effective thermal conductivity perpendicular to the fibers and the effective volumetric heat capacity were then estimated for an IM7-Bismaleimide composite material. The estimation procedure used is based on the minimization of a least squares function which incorporates both calculated and measured temperatures and allows for the parameters to be estimated simultaneously.

  9. Recent developments in optimal experimental designs for functional magnetic resonance imaging

    PubMed Central

    Kao, Ming-Hung; Temkit, M'hamed; Wong, Weng Kee

    2014-01-01

    Functional magnetic resonance imaging (fMRI) is one of the leading brain mapping technologies for studying brain activity in response to mental stimuli. For neuroimaging studies utilizing this pioneering technology, there is a great demand of high-quality experimental designs that help to collect informative data to make precise and valid inference about brain functions. This paper provides a survey on recent developments in experimental designs for fMRI studies. We briefly introduce some analytical and computational tools for obtaining good designs based on a specified design selection criterion. Research results about some commonly considered designs such as blocked designs, and m-sequences are also discussed. Moreover, we present a recently proposed new type of fMRI designs that can be constructed using a certain type of Hadamard matrices. Under certain assumptions, these designs can be shown to be statistically optimal. Some future research directions in design of fMRI experiments are also discussed. PMID:25071884

  10. OPTIMIZING THE PRECISION OF TOXICITY THRESHOLD ESTIMATION USING A TWO-STAGE EXPERIMENTAL DESIGN

    EPA Science Inventory

    An important consideration for risk assessment is the existence of a threshold, i.e., the highest toxicant dose where the response is not distinguishable from background. We have developed methodology for finding an experimental design that optimizes the precision of threshold mo...

  11. Optimization of natural lipstick formulation based on pitaya (Hylocereus polyrhizus) seed oil using D-optimal mixture experimental design.

    PubMed

    Kamairudin, Norsuhaili; Gani, Siti Salwa Abd; Masoumi, Hamid Reza Fard; Hashim, Puziah

    2014-01-01

    The D-optimal mixture experimental design was employed to optimize the melting point of natural lipstick based on pitaya (Hylocereus polyrhizus) seed oil. The influence of the main lipstick components-pitaya seed oil (10%-25% w/w), virgin coconut oil (25%-45% w/w), beeswax (5%-25% w/w), candelilla wax (1%-5% w/w) and carnauba wax (1%-5% w/w)-were investigated with respect to the melting point properties of the lipstick formulation. The D-optimal mixture experimental design was applied to optimize the properties of lipstick by focusing on the melting point with respect to the above influencing components. The D-optimal mixture design analysis showed that the variation in the response (melting point) could be depicted as a quadratic function of the main components of the lipstick. The best combination of each significant factor determined by the D-optimal mixture design was established to be pitaya seed oil (25% w/w), virgin coconut oil (37% w/w), beeswax (17% w/w), candelilla wax (2% w/w) and carnauba wax (2% w/w). With respect to these factors, the 46.0 °C melting point property was observed experimentally, similar to the theoretical prediction of 46.5 °C. Carnauba wax is the most influential factor on this response (melting point) with its function being with respect to heat endurance. The quadratic polynomial model sufficiently fit the experimental data. PMID:25325152

  12. Optimization of Experimental Design for Estimating Groundwater Pumping Using Model Reduction

    NASA Astrophysics Data System (ADS)

    Ushijima, T.; Cheng, W.; Yeh, W. W.

    2012-12-01

    An optimal experimental design algorithm is developed to choose locations for a network of observation wells for estimating unknown groundwater pumping rates in a confined aquifer. The design problem can be expressed as an optimization problem which employs a maximal information criterion to choose among competing designs subject to the specified design constraints. Because of the combinatorial search required in this optimization problem, given a realistic, large-scale groundwater model, the dimensionality of the optimal design problem becomes very large and can be difficult if not impossible to solve using mathematical programming techniques such as integer programming or the Simplex with relaxation. Global search techniques, such as Genetic Algorithms (GAs), can be used to solve this type of combinatorial optimization problem; however, because a GA requires an inordinately large number of calls of a groundwater model, this approach may still be infeasible to use to find the optimal design in a realistic groundwater model. Proper Orthogonal Decomposition (POD) is therefore applied to the groundwater model to reduce the model space and thereby reduce the computational burden of solving the optimization problem. Results for a one-dimensional test case show identical results among using GA, integer programming, and an exhaustive search demonstrating that GA is a valid method for use in a global optimum search and has potential for solving large-scale optimal design problems. Additionally, other results show that the algorithm using GA with POD model reduction is several orders of magnitude faster than an algorithm that employs GA without POD model reduction in terms of time required to find the optimal solution. Application of the proposed methodology is being made to a large-scale, real-world groundwater problem.

  13. Application of experimental design methodology in development and optimization of drug release method.

    PubMed

    Kincl, M; Turk, S; Vrecer, F

    2005-03-01

    The aim of our research was to apply experimental design methodology in the development and optimization of drug release methods. Diclofenac sodium (2-[(2,6-dichlorophenyl)amino]benzeneacetic acid monosodium salt) was selected as a model drug and Naklofen retard prolonged release tablets, containing 100 mg of diclofenac sodium, were chosen as a model prolonged release system. On the basis of previous results, a three-level three-factorial Box-Behnken experimental design was used to characterize and optimize three physicochemical parameters, i.e. rotation speeds of the stirring elements, pH, and ionic strengths of the dissolution medium, affecting the release of diclofenac sodium from the tablets. The chosen dependent variables (responses) were a cumulative percentage of dissolved diclofenac sodium in 2, 6, 12 and 24 h. For estimation of coefficients in the approximating polynomial function, the least square regression method was applied. Afterwards, the information about the model reliability was verified by using the analysis of variance (ANOVA). The estimation of model factors' significance was performed by Student's t-test. For investigation of the shape of the predicted response surfaces and for model optimization, the canonical analysis was applied. Our study proved that experimental design methodology could efficiently be applied for characterization and optimization of analytical parameters affecting drug release and that it is an economical way of obtaining the maximum amount of information in a short period of time and with the fewest number of experiments. PMID:15707730

  14. Experimental validation of optimization-based integrated controls-structures design methodology for flexible space structures

    NASA Technical Reports Server (NTRS)

    Maghami, Peiman G.; Gupta, Sandeep; Joshi, Suresh M.; Walz, Joseph E.

    1993-01-01

    An optimization-based integrated design approach for flexible space structures is experimentally validated using three types of dissipative controllers, including static, dynamic, and LQG dissipative controllers. The nominal phase-0 of the controls structure interaction evolutional model (CEM) structure is redesigned to minimize the average control power required to maintain specified root-mean-square line-of-sight pointing error under persistent disturbances. The redesign structure, phase-1 CEM, was assembled and tested against phase-0 CEM. It is analytically and experimentally demonstrated that integrated controls-structures design is substantially superior to that obtained through the traditional sequential approach. The capability of a software design tool based on an automated design procedure in a unified environment for structural and control designs is demonstrated.

  15. Issues and recent advances in optimal experimental design for site investigation (Invited)

    NASA Astrophysics Data System (ADS)

    Nowak, W.

    2013-12-01

    This presentation provides an overview over issues and recent advances in model-based experimental design for site exploration. The addressed issues and advances are (1) how to provide an adequate envelope to prior uncertainty, (2) how to define the information needs in a task-oriented manner, (3) how to measure the expected impact of a data set that it not yet available but only planned to be collected, and (4) how to perform best the optimization of the data collection plan. Among other shortcomings of the state-of-the-art, it is identified that there is a lack of demonstrator studies where exploration schemes based on expert judgment are compared to exploration schemes obtained by optimal experimental design. Such studies will be necessary do address the often voiced concern that experimental design is an academic exercise with little improvement potential over the well- trained gut feeling of field experts. When addressing this concern, a specific focus has to be given to uncertainty in model structure, parameterizations and parameter values, and to related surprises that data often bring about in field studies, but never in synthetic-data based studies. The background of this concern is that, initially, conceptual uncertainty may be so large that surprises are the rule rather than the exception. In such situations, field experts have a large body of experience in handling the surprises, and expert judgment may be good enough compared to meticulous optimization based on a model that is about to be falsified by the incoming data. In order to meet surprises accordingly and adapt to them, there needs to be a sufficient representation of conceptual uncertainty within the models used. Also, it is useless to optimize an entire design under this initial range of uncertainty. Thus, the goal setting of the optimization should include the objective to reduce conceptual uncertainty. A possible way out is to upgrade experimental design theory towards real-time interaction

  16. Fermilab D-0 Experimental Facility: Energy conservation report and mechanical systems design optimization and cost analysis study

    SciTech Connect

    Krstulovich, S.F.

    1987-10-31

    This report is developed as part of the Fermilab D-0 Experimental Facility Project Title II Design Documentation Update. As such, it concentrates primarily on HVAC mechanical systems design optimization and cost analysis.

  17. Optimal design and experimental analyses of a new micro-vibration control payload-platform

    NASA Astrophysics Data System (ADS)

    Sun, Xiaoqing; Yang, Bintang; Zhao, Long; Sun, Xiaofen

    2016-07-01

    This paper presents a new payload-platform, for precision devices, which possesses the capability of isolating the complex space micro-vibration in low frequency range below 5 Hz. The novel payload-platform equipped with smart material actuators is investigated and designed through optimization strategy based on the minimum energy loss rate, for the aim of achieving high drive efficiency and reducing the effect of the magnetic circuit nonlinearity. Then, the dynamic model of the driving element is established by using the Lagrange method and the performance of the designed payload-platform is further discussed through the combination of the controlled auto regressive moving average (CARMA) model with modified generalized prediction control (MGPC) algorithm. Finally, an experimental prototype is developed and tested. The experimental results demonstrate that the payload-platform has an impressive potential of micro-vibration isolation.

  18. Experimental Investigation of a Point Design Optimized Arrow Wing HSCT Configuration

    NASA Technical Reports Server (NTRS)

    Narducci, Robert P.; Sundaram, P.; Agrawal, Shreekant; Cheung, S.; Arslan, A. E.; Martin, G. L.

    1999-01-01

    The M2.4-7A Arrow Wing HSCT configuration was optimized for straight and level cruise at a Mach number of 2.4 and a lift coefficient of 0.10. A quasi-Newton optimization scheme maximized the lift-to-drag ratio (by minimizing drag-to-lift) using Euler solutions from FL067 to estimate the lift and drag forces. A 1.675% wind-tunnel model of the Opt5 HSCT configuration was built to validate the design methodology. Experimental data gathered at the NASA Langley Unitary Plan Wind Tunnel (UPWT) section #2 facility verified CFL3D Euler and Navier-Stokes predictions of the Opt5 performance at the design point. In turn, CFL3D confirmed the improvement in the lift-to-drag ratio obtained during the optimization, thus validating the design procedure. A data base at off-design conditions was obtained during three wind-tunnel tests. The entry into NASA Langley UPWT section #2 obtained data at a free stream Mach number, M(sub infinity), of 2.55 as well as the design Mach number, M(sub infinity)=2.4. Data from a Mach number range of 1.8 to 2.4 was taken at UPWT section #1. Transonic and low supersonic Mach numbers, M(sub infinity)=0.6 to 1.2, was gathered at the NASA Langley 16 ft. Transonic Wind Tunnel (TWT). In addition to good agreement between CFD and experimental data, highlights from the wind-tunnel tests include a trip dot study suggesting a linear relationship between trip dot drag and Mach number, an aeroelastic study that measured the outboard wing deflection and twist, and a flap scheduling study that identifies the possibility of only one leading-edge and trailing-edge flap setting for transonic cruise and another for low supersonic acceleration.

  19. Optimizing an experimental design for a CSEM experiment: methodology and synthetic tests

    NASA Astrophysics Data System (ADS)

    Roux, E.; Garcia, X.

    2014-04-01

    Optimizing an experimental design is a compromise between maximizing information we get about the target and limiting the cost of the experiment, providing a wide range of constraints. We present a statistical algorithm for experiment design that combines the use of linearized inverse theory and stochastic optimization technique. Linearized inverse theory is used to quantify the quality of one given experiment design while genetic algorithm (GA) enables us to examine a wide range of possible surveys. The particularity of our algorithm is the use of the multi-objective GA NSGA II that searches designs that fit several objective functions (OFs) simultaneously. This ability of NSGA II is helping us in defining an experiment design that focuses on a specified target area. We present a test of our algorithm using a 1-D electrical subsurface structure. The model we use represents a simple but realistic scenario in the context of CO2 sequestration that motivates this study. Our first synthetic test using a single OF shows that a limited number of well-distributed observations from a chosen design have the potential to resolve the given model. This synthetic test also points out the importance of a well-chosen OF, depending on our target. In order to improve these results, we show how the combination of two OFs using a multi-objective GA enables us to determine an experimental design that maximizes information about the reservoir layer. Finally, we present several tests of our statistical algorithm in more challenging environments by exploring the influence of noise, specific site characteristics or its potential for reservoir monitoring.

  20. Optimization of polyvinylidene fluoride (PVDF) membrane fabrication for protein binding using statistical experimental design.

    PubMed

    Ahmad, A L; Ideris, N; Ooi, B S; Low, S C; Ismail, A

    2016-01-01

    Statistical experimental design was employed to optimize the preparation conditions of polyvinylidenefluoride (PVDF) membranes. Three variables considered were polymer concentration, dissolving temperature, and casting thickness, whereby the response variable was membrane-protein binding. The optimum preparation for the PVDF membrane was a polymer concentration of 16.55 wt%, a dissolving temperature of 27.5°C, and a casting thickness of 450 µm. The statistical model exhibits a deviation between the predicted and actual responses of less than 5%. Further characterization of the formed PVDF membrane showed that the morphology of the membrane was in line with the membrane-protein binding performance. PMID:27088961

  1. Formulation and optimization by experimental design of eco-friendly emulsions based on d-limonene.

    PubMed

    Pérez-Mosqueda, Luis M; Trujillo-Cayado, Luis A; Carrillo, Francisco; Ramírez, Pablo; Muñoz, José

    2015-04-01

    d-Limonene is a natural occurring solvent that can replace more pollutant chemicals in agrochemical formulations. In the present work, a comprehensive study of the influence of dispersed phase mass fraction, ϕ, and of the surfactant/oil ratio, R, on the emulsion stability and droplet size distribution of d-limonene-in-water emulsions stabilized by a non-ionic triblock copolymer surfactant has been carried out. An experimental full factorial design 3(2) was conducted in order to optimize the emulsion formulation. The independent variables, ϕ and R were studied in the range 10-50 wt% and 0.02-0.1, respectively. The emulsions studied were mainly destabilized by both creaming and Ostwald ripening. Therefore, initial droplet size and an overall destabilization parameter, the so-called turbiscan stability index, were used as dependent variables. The optimal formulation, comprising minimum droplet size and maximum stability was achieved at ϕ=50 wt%; R=0.062. Furthermore, the surface response methodology allowed us to obtain the formulation yielding sub-micron emulsions by using a single step rotor/stator homogenizer process instead of most commonly used two-step emulsification methods. In addition, the optimal formulation was further improved against Ostwald ripening by adding silicone oil to the dispersed phase. The combination of these experimental findings allowed us to gain a deeper insight into the stability of these emulsions, which can be applied to the rational development of new formulations with potential application in agrochemical formulations. PMID:25734966

  2. Optimizing the spectrofluorimetric determination of cefdinir through a Taguchi experimental design approach.

    PubMed

    Abou-Taleb, Noura Hemdan; El-Wasseef, Dalia Rashad; El-Sherbiny, Dina Tawfik; El-Ashry, Saadia Mohamed

    2016-05-01

    The aim of this work is to optimize a spectrofluorimetric method for the determination of cefdinir (CFN) using the Taguchi method. The proposed method is based on the oxidative coupling reaction of CFN and cerium(IV) sulfate. The quenching effect of CFN on the fluorescence of the produced cerous ions is measured at an emission wavelength (λem ) of 358 nm after excitation (λex ) at 301 nm. The Taguchi orthogonal array L9 (3(4) ) was designed to determine the optimum reaction conditions. The results were analyzed using the signal-to-noise (S/N) ratio and analysis of variance (ANOVA). The optimal experimental conditions obtained from this study were 1 mL of 0.2% MBTH, 0.4 mL of 0.25% Ce(IV), a reaction time of 10 min and methanol as the diluting solvent. The calibration plot displayed a good linear relationship over a range of 0.5-10.0 µg/mL. The proposed method was successfully applied to the determination of CFN in bulk powder and pharmaceutical dosage forms. The results are in good agreement with those obtained using the comparison method. Finally, the Taguchi method provided a systematic and efficient methodology for this optimization, with considerably less effort than would be required for other optimizations techniques. Copyright © 2015 John Wiley & Sons, Ltd. PMID:26456088

  3. Optimization and evaluation of clarithromycin floating tablets using experimental mixture design.

    PubMed

    Uğurlu, Timucin; Karaçiçek, Uğur; Rayaman, Erkan

    2014-01-01

    The purpose of the study was to prepare and evaluate clarithromycin (CLA) floating tablets using experimental mixture design for treatment of Helicobacter pylori provided by prolonged gastric residence time and controlled plasma level. Ten different formulations were generated based on different molecular weight of hypromellose (HPMC K100, K4M, K15M) by using simplex lattice design (a sub-class of mixture design) with Minitab 16 software. Sodium bicarbonate and anhydrous citric acid were used as gas generating agents. Tablets were prepared by wet granulation technique. All of the process variables were fixed. Results of cumulative drug release at 8th h (CDR 8th) were statistically analyzed to get optimized formulation (OF). Optimized formulation, which gave floating lag time lower than 15 s and total floating time more than 10 h, was analyzed and compared with target for CDR 8th (80%). A good agreement was shown between predicted and actual values of CDR 8th with a variation lower than 1%. The activity of clarithromycin contained optimizedformula against H. pylori were quantified using well diffusion agar assay. Diameters of inhibition zones vs. log10 clarithromycin concentrations were plotted in order to obtain a standard curve and clarithromycin activity. PMID:25272652

  4. Doehlert experimental design applied to optimization of light emitting textile structures

    NASA Astrophysics Data System (ADS)

    Oguz, Yesim; Cochrane, Cedric; Koncar, Vladan; Mordon, Serge R.

    2016-07-01

    A light emitting fabric (LEF) has been developed for photodynamic therapy (PDT) for the treatment of dermatologic diseases such as Actinic Keratosis (AK). A successful PDT requires homogenous and reproducible light with controlled power and wavelength on the treated skin area. Due to the shape of the human body, traditional PDT with external light sources is unable to deliver homogenous light everywhere on the skin (head vertex, hand, etc.). For better light delivery homogeneity, plastic optical fibers (POFs) have been woven in textile in order to emit laterally the injected light. The previous studies confirmed that the light power could be locally controlled by modifying the radius of POF macro-bendings within the textile structure. The objective of this study is to optimize the distribution of macro-bendings over the LEF surface in order to increase the light intensity (mW/cm2), and to guarantee the best possible light deliver homogeneity over the LEF which are often contradictory. Fifteen experiments have been carried out with Doehlert experimental design involving Response Surface Methodology (RSM). The proposed models are fitted to the experimental data to enable the optimal set up of the warp yarns tensions.

  5. D-optimal experimental designs to test for departure from additivity in a fixed-ratio mixture ray.

    PubMed

    Coffey, Todd; Gennings, Chris; Simmons, Jane Ellen; Herr, David W

    2005-12-01

    Traditional factorial designs for evaluating interactions among chemicals in a mixture may be prohibitive when the number of chemicals is large. Using a mixture of chemicals with a fixed ratio (mixture ray) results in an economical design that allows estimation of additivity or nonadditive interaction for a mixture of interest. This methodology is extended easily to a mixture with a large number of chemicals. Optimal experimental conditions can be chosen that result in increased power to detect departures from additivity. Although these designs are used widely for linear models, optimal designs for nonlinear threshold models are less well known. In the present work, the use of D-optimal designs is demonstrated for nonlinear threshold models applied to a fixed-ratio mixture ray. For a fixed sample size, this design criterion selects the experimental doses and number of subjects per dose level that result in minimum variance of the model parameters and thus increased power to detect departures from additivity. An optimal design is illustrated for a 2:1 ratio (chlorpyrifos:carbaryl) mixture experiment. For this example, and in general, the optimal designs for the nonlinear threshold model depend on prior specification of the slope and dose threshold parameters. Use of a D-optimal criterion produces experimental designs with increased power, whereas standard nonoptimal designs with equally spaced dose groups may result in low power if the active range or threshold is missed. PMID:16162847

  6. Multi-objective optimization design and experimental investigation of centrifugal fan performance

    NASA Astrophysics Data System (ADS)

    Zhang, Lei; Wang, Songling; Hu, Chenxing; Zhang, Qian

    2013-11-01

    Current studies of fan performance optimization mainly focus on two aspects: one is to improve the blade profile, and another is only to consider the influence of single impeller structural parameter on fan performance. However, there are few studies on the comprehensive effect of the key parameters such as blade number, exit stagger angle of blade and the impeller outlet width on the fan performance. The G4-73 backward centrifugal fan widely used in power plants is selected as the research object. Based on orthogonal design and BP neural network, a model for predicting the centrifugal fan performance parameters is established, and the maximum relative errors of the total pressure and efficiency are 0.974% and 0.333%, respectively. Multi-objective optimization of total pressure and efficiency of the fan is conducted with genetic algorithm, and the optimum combination of impeller structural parameters is proposed. The optimized parameters of blade number, exit stagger angle of blade and the impeller outlet width are seperately 14, 43.9°, and 21 cm. The experiments on centrifugal fan performance and noise are conducted before and after the installation of the new impeller. The experimental results show that with the new impeller, the total pressure of fan increases significantly in total range of the flow rate, and the fan efficiency is improved when the relative flow is above 75%, also the high efficiency area is broadened. Additionally, in 65% -100% relative flow, the fan noise is reduced. Under the design operating condition, total pressure and efficiency of the fan are improved by 6.91% and 0.5%, respectively. This research sheds light on the considering of comprehensive effect of impeller structrual parameters on fan performance, and a new impeller can be designed to satisfy the engineering demand such as energy-saving, noise reduction or solving air pressure insufficiency for power plants.

  7. Optimization of poorly compactable drug tablets manufactured by direct compression using the mixture experimental design.

    PubMed

    Martinello, Tiago; Kaneko, Telma Mary; Velasco, Maria Valéria Robles; Taqueda, Maria Elena Santos; Consiglieri, Vladi O

    2006-09-28

    The poor flowability and bad compressibility characteristics of paracetamol are well known. As a result, the production of paracetamol tablets is almost exclusively by wet granulation, a disadvantageous method when compared to direct compression. The development of a new tablet formulation is still based on a large number of experiments and often relies merely on the experience of the analyst. The purpose of this study was to apply experimental design methodology (DOE) to the development and optimization of tablet formulations containing high amounts of paracetamol (more than 70%) and manufactured by direct compression. Nineteen formulations, screened by DOE methodology, were produced with different proportions of Microcel 102, Kollydon VA 64, Flowlac, Kollydon CL 30, PEG 4000, Aerosil, and magnesium stearate. Tablet properties, except friability, were in accordance with the USP 28th ed. requirements. These results were used to generate plots for optimization, mainly for friability. The physical-chemical data found from the optimized formulation were very close to those from the regression analysis, demonstrating that the mixture project is a great tool for the research and development of new formulations. PMID:16806756

  8. Mixed culture optimization for marigold flower ensilage via experimental design and response surface methodology.

    PubMed

    Navarrete-Bolaños, José Luis; Jiménez-Islas, Hugo; Botello-Alvarez, Enrique; Rico-Martínez, Ramiro

    2003-04-01

    Endogenous microorganisms isolated from the marigold flower (Tagetes erecta) were studied to understand the events taking place during its ensilage. Studies of the cellulase enzymatic activity and the ensilage process were undertaken. In both studies, the use of approximate second-order models and multiple lineal regression, within the context of an experimental mixture design using the response surface methodology as optimization strategy, determined that the microorganisms Flavobacterium IIb, Acinetobacter anitratus, and Rhizopus nigricans are the most significant in marigold flower ensilage and exhibit high cellulase activity. A mixed culture comprised of 9.8% Flavobacterium IIb, 41% A. anitratus, and 49.2% R. nigricans used during ensilage resulted in an increased yield of total xanthophylls extracted of 24.94 g/kg of dry weight compared with 12.92 for the uninoculated control ensilage. PMID:12670157

  9. Optimization of ultrasonic-assisted extraction of cordycepin from Cordyceps militaris using orthogonal experimental design.

    PubMed

    Wang, Hsiu-Ju; Pan, Meng-Chun; Chang, Chao-Kai; Chang, Shu-Wei; Hsieh, Chang-Wei

    2014-01-01

    This study reports on the optimization of the extraction conditions of cordycepin from Cordyceps militaris by using ultrasonication. For this purpose, the orthogonal experimental design was used to investigate the effects of factors on the ultrasonic-assisted extraction (UAE). Four factors: extraction time (min), ethanol concentration (%), extraction temperature (°C) and extraction frequency (kHz), were studied. The results showed that the highest cordycepin yield of 7.04 mg/g (86.98% ± 0.23%) was obtained with an extraction time of 60 min, ethanol concentration of 50%, extraction temperature of 65 °C and extraction frequency of 56 kHz. It was found that the cordycepin extraction yield increased with the effect of ultrasonication during the extraction process. Therefore, UAE can be used as an alternative to conventional immersion extraction with respect to the recovery of cordycepin from C. militaris, with the advantages of shorter extraction time and reduced solvent consumption. PMID:25514223

  10. Parameter estimation and uncertainty quantification in a biogeochemical model using optimal experimental design methods

    NASA Astrophysics Data System (ADS)

    Reimer, Joscha; Piwonski, Jaroslaw; Slawig, Thomas

    2016-04-01

    The statistical significance of any model-data comparison strongly depends on the quality of the used data and the criterion used to measure the model-to-data misfit. The statistical properties (such as mean values, variances and covariances) of the data should be taken into account by choosing a criterion as, e.g., ordinary, weighted or generalized least squares. Moreover, the criterion can be restricted onto regions or model quantities which are of special interest. This choice influences the quality of the model output (also for not measured quantities) and the results of a parameter estimation or optimization process. We have estimated the parameters of a three-dimensional and time-dependent marine biogeochemical model describing the phosphorus cycle in the ocean. For this purpose, we have developed a statistical model for measurements of phosphate and dissolved organic phosphorus. This statistical model includes variances and correlations varying with time and location of the measurements. We compared the obtained estimations of model output and parameters for different criteria. Another question is if (and which) further measurements would increase the model's quality at all. Using experimental design criteria, the information content of measurements can be quantified. This may refer to the uncertainty in unknown model parameters as well as the uncertainty regarding which model is closer to reality. By (another) optimization, optimal measurement properties such as locations, time instants and quantities to be measured can be identified. We have optimized such properties for additional measurement for the parameter estimation of the marine biogeochemical model. For this purpose, we have quantified the uncertainty in the optimal model parameters and the model output itself regarding the uncertainty in the measurement data using the (Fisher) information matrix. Furthermore, we have calculated the uncertainty reduction by additional measurements depending on time

  11. Optimizing indomethacin-loaded chitosan nanoparticle size, encapsulation, and release using Box-Behnken experimental design.

    PubMed

    Abul Kalam, Mohd; Khan, Abdul Arif; Khan, Shahanavaj; Almalik, Abdulaziz; Alshamsan, Aws

    2016-06-01

    Indomethacin chitosan nanoparticles (NPs) were developed by ionotropic gelation and optimized by concentrations of chitosan and tripolyphosphate (TPP) and stirring time by 3-factor 3-level Box-Behnken experimental design. Optimal concentration of chitosan (A) and TPP (B) were found 0.6mg/mL and 0.4mg/mL with 120min stirring time (C), with applied constraints of minimizing particle size (R1) and maximizing encapsulation efficiency (R2) and drug release (R3). Based on obtained 3D response surface plots, factors A, B and C were found to give synergistic effect on R1, while factor A has a negative impact on R2 and R3. Interaction of AB was negative on R1 and R2 but positive on R3. The factor AC was having synergistic effect on R1 and on R3, while the same combination had a negative effect on R2. The interaction BC was positive on the all responses. NPs were found in the size range of 321-675nm with zeta potentials (+25 to +32mV) after 6 months storage. Encapsulation, drug release, and content were in the range of 56-79%, 48-73% and 98-99%, respectively. In vitro drug release data were fitted in different kinetic models and pattern of drug release followed Higuchi-matrix type. PMID:26893052

  12. Experimental design and optimization of raloxifene hydrochloride loaded nanotransfersomes for transdermal application

    PubMed Central

    Mahmood, Syed; Taher, Muhammad; Mandal, Uttam Kumar

    2014-01-01

    Raloxifene hydrochloride, a highly effective drug for the treatment of invasive breast cancer and osteoporosis in post-menopausal women, shows poor oral bioavailability of 2%. The aim of this study was to develop, statistically optimize, and characterize raloxifene hydrochloride-loaded transfersomes for transdermal delivery, in order to overcome the poor bioavailability issue with the drug. A response surface methodology experimental design was applied for the optimization of transfersomes, using Box-Behnken experimental design. Phospholipon® 90G, sodium deoxycholate, and sonication time, each at three levels, were selected as independent variables, while entrapment efficiency, vesicle size, and transdermal flux were identified as dependent variables. The formulation was characterized by surface morphology and shape, particle size, and zeta potential. Ex vivo transdermal flux was determined using a Hanson diffusion cell assembly, with rat skin as a barrier medium. Transfersomes from the optimized formulation were found to have spherical, unilamellar structures, with a homogeneous distribution and low polydispersity index (0.08). They had a particle size of 134±9 nM, with an entrapment efficiency of 91.00%±4.90%, and transdermal flux of 6.5±1.1 μg/cm2/hour. Raloxifene hydrochloride-loaded transfersomes proved significantly superior in terms of amount of drug permeated and deposited in the skin, with enhancement ratios of 6.25±1.50 and 9.25±2.40, respectively, when compared with drug-loaded conventional liposomes, and an ethanolic phosphate buffer saline. Differential scanning calorimetry study revealed a greater change in skin structure, compared with a control sample, during the ex vivo drug diffusion study. Further, confocal laser scanning microscopy proved an enhanced permeation of coumarin-6-loaded transfersomes, to a depth of approximately160 μM, as compared with rigid liposomes. These ex vivo findings proved that a raloxifene hydrochloride

  13. Experimental design and optimization of raloxifene hydrochloride loaded nanotransfersomes for transdermal application.

    PubMed

    Mahmood, Syed; Taher, Muhammad; Mandal, Uttam Kumar

    2014-01-01

    Raloxifene hydrochloride, a highly effective drug for the treatment of invasive breast cancer and osteoporosis in post-menopausal women, shows poor oral bioavailability of 2%. The aim of this study was to develop, statistically optimize, and characterize raloxifene hydrochloride-loaded transfersomes for transdermal delivery, in order to overcome the poor bioavailability issue with the drug. A response surface methodology experimental design was applied for the optimization of transfersomes, using Box-Behnken experimental design. Phospholipon(®) 90G, sodium deoxycholate, and sonication time, each at three levels, were selected as independent variables, while entrapment efficiency, vesicle size, and transdermal flux were identified as dependent variables. The formulation was characterized by surface morphology and shape, particle size, and zeta potential. Ex vivo transdermal flux was determined using a Hanson diffusion cell assembly, with rat skin as a barrier medium. Transfersomes from the optimized formulation were found to have spherical, unilamellar structures, with a homogeneous distribution and low polydispersity index (0.08). They had a particle size of 134±9 nM, with an entrapment efficiency of 91.00%±4.90%, and transdermal flux of 6.5±1.1 μg/cm(2)/hour. Raloxifene hydrochloride-loaded transfersomes proved significantly superior in terms of amount of drug permeated and deposited in the skin, with enhancement ratios of 6.25±1.50 and 9.25±2.40, respectively, when compared with drug-loaded conventional liposomes, and an ethanolic phosphate buffer saline. Differential scanning calorimetry study revealed a greater change in skin structure, compared with a control sample, during the ex vivo drug diffusion study. Further, confocal laser scanning microscopy proved an enhanced permeation of coumarin-6-loaded transfersomes, to a depth of approximately160 μM, as compared with rigid liposomes. These ex vivo findings proved that a raloxifene hydrochloride

  14. Optimizing the experimental design of soil columns in saturated and unsaturated transport experiments

    NASA Astrophysics Data System (ADS)

    Lewis, Jeffrey; Sjöstrom, Jan

    2010-06-01

    Soil column experiments in both the saturated and unsaturated regimes are widely used for applied and theoretical studies in such diverse fields as transport model evaluation, fate and transport of pesticides, explosives, microbes, heavy metals and non aqueous phase liquids, and for evapotranspiration studies. The apparent simplicity of constructing soil columns conceals a number of technical issues which can seriously affect the outcome of an experiment, such as the presence or absence of macropores, artificial preferential flow paths, non-ideal infiltrate injection and unrealistic moisture regimes. This review examines the literature to provide an analysis of the state of the art for constructing both saturated and unsaturated soil columns. Common design challenges are discussed and best practices for potential solutions are presented. This article discusses both basic principles and the practical advantages and disadvantages of various experimental approaches. Both repacked and monolith-type columns are discussed. The information in this review will assist soil scientists, hydrogeologists and environmental professionals in optimizing the construction and operation of soil column experiments in order to achieve their objectives, while avoiding serious design flaws which can compromise the integrity of their results.

  15. Design and optimization of an experimental bioregenerative life support system with higher plants and silkworms

    NASA Astrophysics Data System (ADS)

    Hu, Enzhu; Bartsev, Sergey I.; Zhao, Ming; Liu, Professor Hong

    The conceptual scheme of an experimental bioregenerative life support system (BLSS) for planetary exploration was designed, which consisted of four elements - human metabolism, higher plants, silkworms and waste treatment. 15 kinds of higher plants, such as wheat, rice, soybean, lettuce, mulberry, et al., were selected as regenerative component of BLSS providing the crew with air, water, and vegetable food. Silkworms, which producing animal nutrition for crews, were fed by mulberry-leaves during the first three instars, and lettuce leaves last two instars. The inedible biomass of higher plants, human wastes and silkworm feces were composted into soil like substrate, which can be reused by higher plants cultivation. Salt, sugar and some household material such as soap, shampoo would be provided from outside. To support the steady state of BLSS the same amount and elementary composition of dehydrated wastes were removed periodically. The balance of matter flows between BLSS components was described by the system of algebraic equations. The mass flows between the components were optimized by EXCEL spreadsheets and using Solver. The numerical method used in this study was Newton's method.

  16. Degradation of caffeine by photo-Fenton process: optimization of treatment conditions using experimental design.

    PubMed

    Trovó, Alam G; Silva, Tatiane F S; Gomes, Oswaldo; Machado, Antonio E H; Neto, Waldomiro Borges; Muller, Paulo S; Daniel, Daniela

    2013-01-01

    The degradation of caffeine in different kind of effluents, via photo-Fenton process, was investigated in lab-scale and in a solar pilot plant. The treatment conditions (caffeine, Fe(2+) and H(2)O(2) concentrations) were defined by experimental design. The optimized conditions for each variable, obtained using the response factor (% mineralization), were: 52.0 mg L(-1)caffeine, 10.0 mg L(-1)Fe(2+) and 42.0 mg L(-1)H(2)O(2) (replaced in kinetic experiments). Under these conditions, in ultrapure water (UW), the caffeine concentration reached the quantitation limit (0.76 mg L(-1)) after 20 min, and 78% of mineralization was obtained respectively after 120 min of reaction. Using the same conditions, the matrix influence (surface water - SW and sewage treatment plant effluent - STP) on caffeine degradation was also evaluated. The total removal of caffeine in SW was reached at the same time in UW (after 20 min), while 40 min were necessary in STP. Although lower mineralization rates were verified for high organic load, under the same operational conditions, less H(2)O(2) was necessary to mineralize the dissolved organic carbon as the initial organic load increases. A high efficiency of the photo-Fenton process was also observed in caffeine degradation by solar photocatalysis using a CPC reactor, as well as intermediates of low toxicity, demonstrating that photo-Fenton process can be a viable alternative for caffeine removal in wastewater. PMID:22795305

  17. Molecular assay optimized by Taguchi experimental design method for venous thromboembolism investigation.

    PubMed

    Celani de Souza, Helder Jose; Moyses, Cinthia B; Pontes, Fabrício J; Duarte, Roberto N; Sanches da Silva, Carlos Eduardo; Alberto, Fernando Lopes; Ferreira, Ubirajara R; Silva, Messias Borges

    2011-01-01

    Two mutations - Factor V Leiden (1691G > A) and the 20210G > A on the Prothrombin gene - are key risk factors for a frequent and potentially fatal disorder called Venous Thromboembolism. These molecular alterations can be investigated using real-time Polymerase Chain Reaction (PCR) with Fluorescence Resonance Energy Transfer (FRET) probes and distinct DNA pools for both factors. The objective of this paper is to present an application of Taguchi Experimental Design Method to determine the best parameters adjustment of a Molecular Assays Process in order to obtain the best diagnostic result for Venous Thromboembolism investigation. The complete process contains six three-level factors which usually demands 729 experiments to obtain the final result, if using a Full Factorial Array. In this research, a Taguchi L27 Orthogonal Array is chosen to optimize the analysis and reduce the number of experiments to 27 without degrading the final result accuracy. The application of this method can lessen the time and cost necessary to achieve the best operation condition for a required performance. The results is proven in practice and confirmed that the Taguchi method can really offer a good approach for clinical assay efficiency and effectiveness improvement even though the clinical diagnostics can be based on the use of qualitative techniques. PMID:21867748

  18. Numerical and Experimental Approach for the Optimal Design of a Dual Plate Under Ballistic Impact

    NASA Astrophysics Data System (ADS)

    Yoo, Jeonghoon; Chung, Dong-Teak; Park, Myung Soo

    To predict the behavior of a dual plate composed of 5052-aluminum and 1002-cold rolled steel under ballistic impact, numerical and experimental approaches are attempted. For the accurate numerical simulation of the impact phenomena, the appropriate selection of the key parameter values based on numerical or experimental tests are critical. This study is focused on not only the optimization technique using the numerical simulation but also numerical and experimental procedures to obtain the required parameter values in the simulation. The Johnson-Cook model is used to simulate the mechanical behaviors, and the simplified experimental and the numerical approaches are performed to obtain the material properties of the model. The element erosion scheme for the robust simulation of the ballistic impact problem is applied by adjusting the element erosion criteria of each material based on numerical and experimental results. The adequate mesh size and the aspect ratio are chosen based on parametric studies. Plastic energy is suggested as a response representing the strength of the plate for the optimization under dynamic loading. Optimized thickness of the dual plate is obtained to resist the ballistic impact without penetration as well as to minimize the total weight.

  19. Vitamin B12 production from crude glycerol by Propionibacterium freudenreichii ssp. shermanii: optimization of medium composition through statistical experimental designs.

    PubMed

    Kośmider, Alicja; Białas, Wojciech; Kubiak, Piotr; Drożdżyńska, Agnieszka; Czaczyk, Katarzyna

    2012-02-01

    A two-step statistical experimental design was employed to optimize the medium for vitamin B(12) production from crude glycerol by Propionibacterium freudenreichii ssp. shermanii. In the first step, using Plackett-Burman design, five of 13 tested medium components (calcium pantothenate, NaH(2)PO(4)·2H(2)O, casein hydrolysate, glycerol and FeSO(4)·7H(2)O) were identified as factors having significant influence on vitamin production. In the second step, a central composite design was used to optimize levels of medium components selected in the first step. Valid statistical models describing the influence of significant factors on vitamin B(12) production were established for each optimization phase. The optimized medium provided a 93% increase in final vitamin concentration compared to the original medium. PMID:22178491

  20. Optimal experimental design for nano-particle atom-counting from high-resolution STEM images.

    PubMed

    De Backer, A; De Wael, A; Gonnissen, J; Van Aert, S

    2015-04-01

    In the present paper, the principles of detection theory are used to quantify the probability of error for atom-counting from high resolution scanning transmission electron microscopy (HR STEM) images. Binary and multiple hypothesis testing have been investigated in order to determine the limits to the precision with which the number of atoms in a projected atomic column can be estimated. The probability of error has been calculated when using STEM images, scattering cross-sections or peak intensities as a criterion to count atoms. Based on this analysis, we conclude that scattering cross-sections perform almost equally well as images and perform better than peak intensities. Furthermore, the optimal STEM detector design can be derived for atom-counting using the expression for the probability of error. We show that for very thin objects LAADF is optimal and that for thicker objects the optimal inner detector angle increases. PMID:25499018

  1. Optimization of α-amylase production by Bacillus subtilis RSKK96: using the Taguchi experimental design approach.

    PubMed

    Uysal, Ersin; Akcan, Nurullah; Baysal, Zübeyde; Uyar, Fikret

    2011-01-01

    In this study, the Taguchi experimental design was applied to optimize the conditions for α-amylase production by Bacillus subtilis RSKK96, which was purchased from Refik Saydam Hifzissihha Industry (RSHM). Four factors, namely, carbon source, nitrogen source, amino acid, and fermentation time, each at four levels, were selected, and an orthogonal array layout of L(16) (4(5)) was performed. The model equation obtained was validated experimentally at maximum casein (1%), corn meal (1%), and glutamic acid (0.01%) concentrations with incubation time to 72 h in the presence of 1% inoculum density. Point prediction of the design showed that maximum α-amylase production of 503.26 U/mg was achieved under optimal experimental conditions. PMID:21229466

  2. Design optimization and experimental testing of the High-Flux Test Module of IFMIF

    NASA Astrophysics Data System (ADS)

    Leichtle, D.; Arbeiter, F.; Dolensky, B.; Fischer, U.; Gordeev, S.; Heinzel, V.; Ihli, T.; Moeslang, A.; Simakov, S. P.; Slobodchuk, V.; Stratmanns, E.

    2009-04-01

    The design of the High-Flux Test Module of the International Fusion Material Irradiation Facility has been developed continuously in the past few years. The present paper highlights recent design achievements, including a thorough state-of-the-art validation assessment of CFD tools and models. Along with design related analyses exercises on manufacturing procedures have been performed. Recommendations for the use of container, rig, and capsule materials as well as recent progress in brazing of electrical heaters are discussed. A test matrix starting from High-Flux Test Module compartments, i.e. segments of the full module, with heated dummy rigs up to the full-scale module with instrumented irradiation rigs has been developed and the appropriate helium gas loop has been designed conceptually. A roadmap of the envisaged experimental activities is presented in accordance with the test loop facility construction and mock-up design and fabrication schedules.

  3. A Resampling Based Approach to Optimal Experimental Design for Computer Analysis of a Complex System

    SciTech Connect

    Rutherford, Brian

    1999-08-04

    The investigation of a complex system is often performed using computer generated response data supplemented by system and component test results where possible. Analysts rely on an efficient use of limited experimental resources to test the physical system, evaluate the models and to assure (to the extent possible) that the models accurately simulate the system order investigation. The general problem considered here is one where only a restricted number of system simulations (or physical tests) can be performed to provide additional data necessary to accomplish the project objectives. The levels of variables used for defining input scenarios, for setting system parameters and for initializing other experimental options must be selected in an efficient way. The use of computer algorithms to support experimental design in complex problems has been a topic of recent research in the areas of statistics and engineering. This paper describes a resampling based approach to form dating this design. An example is provided illustrating in two dimensions how the algorithm works and indicating its potential on larger problems. The results show that the proposed approach has characteristics desirable of an algorithmic approach on the simple examples. Further experimentation is needed to evaluate its performance on larger problems.

  4. An experimental evaluation of a helicopter rotor section designed by numerical optimization

    NASA Technical Reports Server (NTRS)

    Hicks, R. M.; Mccroskey, W. J.

    1980-01-01

    The wind tunnel performance of a 10-percent thick helicopter rotor section design by numerical optimization is presented. The model was tested at Mach number from 0.2 to 0.84 with Reynolds number ranging from 1,900,000 at Mach 0.2 to 4,000,000 at Mach numbers above 0.5. The airfoil section exhibited maximum lift coefficients greater than 1.3 at Mach numbers below 0.45 and a drag divergence Mach number of 0.82 for lift coefficients near 0. A moderate 'drag creep' is observed at low lift coefficients for Mach numbers greater than 0.6.

  5. Design and optimization of an experimental test bench for the study of impulsive fluid-structure interactions

    NASA Astrophysics Data System (ADS)

    Russo, S.; Krastev, V. K.; Jannelli, E.; Falcucci, G.

    2016-06-01

    In this work, the design and the optimization of an experimental test bench for the experimental characterization of impulsive water-entry problems are presented. Currently, the majority of the experimental apparatus allow impact test only in specific conditions. Our test bench allows for testing of rigid and compliant bodies and allows performing experiments on floating or sinking structures, in free-fall, or under dynamic motion control. The experimental apparatus is characterized by the adoption of accelerometers, encoders, position sensors and, above all, FBG (fiber Bragg grating) sensors that, together with a high speed camera, provide accurate and fast data acquisitions for the dissection of structural deformations and hydrodynamic loadings under a broad set of experimental conditions.

  6. Optimal design of disc-type magneto-rheological brake for mid-sized motorcycle: experimental evaluation

    NASA Astrophysics Data System (ADS)

    Sohn, Jung Woo; Jeon, Juncheol; Nguyen, Quoc Hung; Choi, Seung-Bok

    2015-08-01

    In this paper, a disc-type magneto-rheological (MR) brake is designed for a mid-sized motorcycle and its performance is experimentally evaluated. The proposed MR brake consists of an outer housing, a rotating disc immersed in MR fluid, and a copper wire coiled around a bobbin to generate a magnetic field. The structural configuration of the MR brake is first presented with consideration of the installation space for the conventional hydraulic brake of a mid-sized motorcycle. The design parameters of the proposed MR brake are optimized to satisfy design requirements such as the braking torque, total mass of the MR brake, and cruising temperature caused by the magnetic-field friction of the MR fluid. In the optimization procedure, the braking torque is calculated based on the Herschel-Bulkley rheological model, which predicts MR fluid behavior well at high shear rate. An optimization tool based on finite element analysis is used to obtain the optimized dimensions of the MR brake. After manufacturing the MR brake, mechanical performances regarding the response time, braking torque and cruising temperature are experimentally evaluated.

  7. Experimental validation of a magnetorheological energy absorber design optimized for shock and impact loads

    NASA Astrophysics Data System (ADS)

    Singh, Harinder J.; Hu, Wei; Wereley, Norman M.; Glass, William

    2014-12-01

    A linear stroke adaptive magnetorheological energy absorber (MREA) was designed, fabricated and tested for intense impact conditions with piston velocities up to 8 m s-1. The performance of the MREA was characterized using dynamic range, which is defined as the ratio of maximum on-state MREA force to the off-state MREA force. Design optimization techniques were employed in order to maximize the dynamic range at high impact velocities such that MREA maintained good control authority. Geometrical parameters of the MREA were optimized by evaluating MREA performance on the basis of a Bingham-plastic analysis incorporating minor losses (BPM analysis). Computational fluid dynamics and magnetic FE analysis were conducted to verify the performance of passive and controllable MREA force, respectively. Subsequently, high-speed drop testing (0-4.5 m s-1 at 0 A) was conducted for quantitative comparison with the numerical simulations. Refinements to the nonlinear BPM analysis were carried out to improve prediction of MREA performance.

  8. Optimization of the preparation of sonogenic phospholipids-based microbubbles by using central composite experimental design and response surface methodology.

    PubMed

    Zhao, Ying-Zheng; Lu, Cui-Tao; Mei, Xin-Guo

    2008-08-01

    Sonogenic microbubble agent is a newly developed drug targeting delivery system, which uses ultrasonic beam to enhance the delivery of drug and gene to targeted cells and tissues. In this paper, the preparation of sonogenic phospholipids-based microbubbles was optimized by using central composite experimental design (CCD) and response surface methodology (RSM). Hydrogenated egg phosphatidylcholine (EPC), Tween 80 and polyethylene glycol 1500 (PEG 1500) were important components affecting the concentration of 2 - 8 microm microbubbles in the preparation. The combined effects of these three factors were analyzed by CCD and optimized by RSM. Evaluation variable was the concentration of 2 - 8 microm microbubbles. Overall desirability was fitted to a second-order polynomial equation, through which three dimensional response surface graphs were produced. Optimal experimental conditions were selected from the stationary point of the response surfaces. The stability of the sonogenic phospholipids-based microbubbles by the optimal formulation was investigated by accelerated experiment. The contrast effect in vivo of the optimal formulation was investigated. Foreign market product SonoVue was used as the control. From the results, all the three factors had positive effects on the concentration of 2 - 8 microm microbubbles. The optimal condition in the preparation of phospholipids-based microbubbles was obtained as following: EPC 8.35 mg, Tween 80 21.68 mg and PEG 1500 201 mg. The mean value of the concentration of 2 - 8 microm microbubbles in rechecking experiment reached 8.60 x 10(9) x mL(-1). From the accelerated experiment, phospholipids-based microbubbles showed good physical stability. The intensity (relative unit) and duration of the contrast effect by the optimal formulation were 4.47 +/- 0.15 and (302 +/- 7) s respectively, which showed little difference with foreign market product SonoVue [4.28 +/- 0.13, (309 +/- 8) s]. The optimal formulation selected by CCD and

  9. Active SAmpling Protocol (ASAP) to Optimize Individual Neurocognitive Hypothesis Testing: A BCI-Inspired Dynamic Experimental Design

    PubMed Central

    Sanchez, Gaëtan; Lecaignard, Françoise; Otman, Anatole; Maby, Emmanuel; Mattout, Jérémie

    2016-01-01

    The relatively young field of Brain-Computer Interfaces has promoted the use of electrophysiology and neuroimaging in real-time. In the meantime, cognitive neuroscience studies, which make extensive use of functional exploration techniques, have evolved toward model-based experiments and fine hypothesis testing protocols. Although these two developments are mostly unrelated, we argue that, brought together, they may trigger an important shift in the way experimental paradigms are being designed, which should prove fruitful to both endeavors. This change simply consists in using real-time neuroimaging in order to optimize advanced neurocognitive hypothesis testing. We refer to this new approach as the instantiation of an Active SAmpling Protocol (ASAP). As opposed to classical (static) experimental protocols, ASAP implements online model comparison, enabling the optimization of design parameters (e.g., stimuli) during the course of data acquisition. This follows the well-known principle of sequential hypothesis testing. What is radically new, however, is our ability to perform online processing of the huge amount of complex data that brain imaging techniques provide. This is all the more relevant at a time when physiological and psychological processes are beginning to be approached using more realistic, generative models which may be difficult to tease apart empirically. Based upon Bayesian inference, ASAP proposes a generic and principled way to optimize experimental design adaptively. In this perspective paper, we summarize the main steps in ASAP. Using synthetic data we illustrate its superiority in selecting the right perceptual model compared to a classical design. Finally, we briefly discuss its future potential for basic and clinical neuroscience as well as some remaining challenges. PMID:27458364

  10. Molecular identification of potential denitrifying bacteria and use of D-optimal mixture experimental design for the optimization of denitrification process.

    PubMed

    Ben Taheur, Fadia; Fdhila, Kais; Elabed, Hamouda; Bouguerra, Amel; Kouidhi, Bochra; Bakhrouf, Amina; Chaieb, Kamel

    2016-04-01

    Three bacterial strains (TE1, TD3 and FB2) were isolated from date palm (degla), pistachio and barley. The presence of nitrate reductase (narG) and nitrite reductase (nirS and nirK) genes in the selected strains was detected by PCR technique. Molecular identification based on 16S rDNA sequencing method was applied to identify positive strains. In addition, the D-optimal mixture experimental design was used to optimize the optimal formulation of probiotic bacteria for denitrification process. Strains harboring denitrification genes were identified as: TE1, Agrococcus sp LN828197; TD3, Cronobacter sakazakii LN828198 and FB2, Pedicoccus pentosaceus LN828199. PCR results revealed that all strains carried the nirS gene. However only C. sakazakii LN828198 and Agrococcus sp LN828197 harbored the nirK and the narG genes respectively. Moreover, the studied bacteria were able to form biofilm on abiotic surfaces with different degree. Process optimization showed that the most significant reduction of nitrate was 100% with 14.98% of COD consumption and 5.57 mg/l nitrite accumulation. Meanwhile, the response values were optimized and showed that the most optimal combination was 78.79% of C. sakazakii LN828198 (curve value), 21.21% of P. pentosaceus LN828199 (curve value) and absence (0%) of Agrococcus sp LN828197 (curve value). PMID:26893037

  11. Optimization of the simultaneous determination of imatinib and its major metabolite, CGP74588, in human plasma by a rapid HPLC method using D-optimal experimental design.

    PubMed

    Golabchifar, Ali-Akbar; Rouini, Mohammad-Reza; Shafaghi, Bijan; Rezaee, Saeed; Foroumadi, Alireza; Khoshayand, Mohammad-Reza

    2011-10-15

    A simple, rapid and specific HPLC method has been developed and validated for the simultaneous determination of imatinib, a tyrosine kinase inhibitor, and its major metabolite, CGP74588, in human plasma. The optimization of the HPLC procedure involved several variables, of which the influences of each was studied. After a series of preliminary-screening experiments, the composition of the mobile phase and the pH of the added buffer solution were set as the investigated variables, while the resolution between imatinib and CGP74588 peaks, the retention time and the imatinib peak width were chosen as the dependent variables. Applying D-optimal design, the optimal chromatographic conditions for the separation were defined. The method proved to show good agreement between the experimental data and predictive values throughout the studied parameter range. The optimum assay conditions were achieved with a Chromolith™ Performance RP-8e 100 mm × 4.6 mm column and a mixture of methanol/acetonitrile/triethylamine/diammonium hydrogen phosphate (pH 6.25, 0.048 mol L(-1)) (20:20:0.1:59.9, v/v/v/v) as the mobile phase at a flow rate of 2 mL min(-1) and detection wavelength of 261 nm. The run time was less than 5 min, which is much shorter than the previously optimized methods. The optimized method was validated according to FDA guidelines to confirm specificity, linearity, accuracy and precision. PMID:21962649

  12. Application of a mixture experimental design in the optimization of a self-emulsifying formulation with a high drug load.

    PubMed

    Gao, Ping; Witt, Martha J; Haskell, Roy J; Zamora, Kathryn M; Shifflett, John R

    2004-08-01

    Response surface methodology (RSM) was applied to optimize the self-emulsifying drug delivery system (SEDDS) containing 25% (w/w) Drug A, a model drug with a high lipophilicity and low water solubility. The key objective of this study was to identify an optimal SEDDS formulation that: 1) possesses a minimum concentration of the surfactant and a maximum concentration of lipid and 2) generates a fine emulsion and eliminates large size droplets (> or = 1 microm) upon dilution with an aqueous medium. Three ingredient variables [PEG 400, Cremophor EL, and a mixture of glycerol dioleate (GDO), and glycerol monooleate (GMO)] were included in the experimental design, while keeping the other ingredients at a fixed level (25% Drug A, 6% ethanol, 3% propylene glycol, 4% water, and 2% tromethamine) in the SEDDS formulation. Dispersion performance of these formulations upon dilution with a simulated gastrointestinal fluid was measured, and the population of the large droplets was used as the primary response for statistical modeling. The results of this mixture study revealed significant interactions among the three ingredients, and their individual levels in the formulation collectively dictated the dispersion performance. The fitted response surface model predicted an optimal region of the SEDDS formulation compositions that generate fine emulsions and essentially eliminates large droplets upon dilution. The predicted optimal 25% Drug A-SEDDS formulations with the levels of Cremophor EL ranging from 40-44%, GDO/GMO ranging from 10-13%, and PEG 400 ranging from 2.7-9.0% were selected and prepared. The dispersion experiment results confirmed the prediction of this model and identified potential optimal formulations for further development. This work demonstrates that RSM is an efficient approach for optimization of the SEDDS formulation. PMID:15458235

  13. Experimental design and multiple response optimization. Using the desirability function in analytical methods development.

    PubMed

    Candioti, Luciana Vera; De Zan, María M; Cámara, María S; Goicoechea, Héctor C

    2014-06-01

    A review about the application of response surface methodology (RSM) when several responses have to be simultaneously optimized in the field of analytical methods development is presented. Several critical issues like response transformation, multiple response optimization and modeling with least squares and artificial neural networks are discussed. Most recent analytical applications are presented in the context of analytLaboratorio de Control de Calidad de Medicamentos (LCCM), Facultad de Bioquímica y Ciencias Biológicas, Universidad Nacional del Litoral, C.C. 242, S3000ZAA Santa Fe, ArgentinaLaboratorio de Control de Calidad de Medicamentos (LCCM), Facultad de Bioquímica y Ciencias Biológicas, Universidad Nacional del Litoral, C.C. 242, S3000ZAA Santa Fe, Argentinaical methods development, especially in multiple response optimization procedures using the desirability function. PMID:24767454

  14. Limit of detection of 15{sub N} by gas-chromatography atomic emission detection: Optimization using an experimental design

    SciTech Connect

    Deruaz, D.; Bannier, A.; Pionchon, C.

    1995-08-01

    This paper deals with the optimal conditions for the detection of {sup 15}N determined using a four-factor experimental design from [2{sup 13}C,-1,3 {sup 15}N] caffeine measured with an atomic emission detector (AED) coupled to gas chromatography (GC). Owing to the capability of a photodiodes array, AED can simultaneously detect several elements using their specific emission lines within a wavelength range of 50 nm. So, the emissions of {sup 15}N and {sup 14}N are simultaneously detected at 420.17 nm and 421.46 nm respectively. Four independent experimental factors were tested (1) helium flow rate (plasma gas); (2) methane pressure (reactant gas); (3) oxygen pressure; (4) hydrogen pressure. It has been shown that these four gases had a significant influence on the analytical response of {sup 15}N. The linearity of the detection was determined using {sup 15}N amounts ranging from 1.52 pg to 19 ng under the optimal conditions obtained from the experimental design. The limit of detection was studied using different methods. The limits of detection of {sup 15}N was 1.9 pg/s according to the IUPAC method (International-Union of Pure and Applied Chemistry). The method proposed by Quimby and Sullivan gave a value of 2.3 pg/s and that of Oppenheimer gave a limit of 29 pg/s. For each determination, and internal standard: 1-isobutyl-3.7 dimethylxanthine was used. The results clearly demonstrate that GC AED is sensitive and selective enough to detect and measure {sup 15}N-labelled molecules after gas chromatographic separation.

  15. Conceptual design optimization study

    NASA Technical Reports Server (NTRS)

    Hollowell, S. J.; Beeman, E. R., II; Hiyama, R. M.

    1990-01-01

    The feasibility of applying multilevel functional decomposition and optimization techniques to conceptual design of advanced fighter aircraft was investigated. Applying the functional decomposition techniques to the conceptual design phase appears to be feasible. The initial implementation of the modified design process will optimize wing design variables. A hybrid approach, combining functional decomposition techniques for generation of aerodynamic and mass properties linear sensitivity derivatives with existing techniques for sizing mission performance and optimization, is proposed.

  16. Experimental validation of structural optimization methods

    NASA Technical Reports Server (NTRS)

    Adelman, Howard M.

    1992-01-01

    The topic of validating structural optimization methods by use of experimental results is addressed. The need for validating the methods as a way of effecting a greater and an accelerated acceptance of formal optimization methods by practicing engineering designers is described. The range of validation strategies is defined which includes comparison of optimization results with more traditional design approaches, establishing the accuracy of analyses used, and finally experimental validation of the optimization results. Examples of the use of experimental results to validate optimization techniques are described. The examples include experimental validation of the following: optimum design of a trussed beam; combined control-structure design of a cable-supported beam simulating an actively controlled space structure; minimum weight design of a beam with frequency constraints; minimization of the vibration response of helicopter rotor blade; minimum weight design of a turbine blade disk; aeroelastic optimization of an aircraft vertical fin; airfoil shape optimization for drag minimization; optimization of the shape of a hole in a plate for stress minimization; optimization to minimize beam dynamic response; and structural optimization of a low vibration helicopter rotor.

  17. Optimization of Magnetosome Production and Growth by the Magnetotactic Vibrio Magnetovibrio blakemorei Strain MV-1 through a Statistics-Based Experimental Design

    PubMed Central

    Silva, Karen T.; Leão, Pedro E.; Abreu, Fernanda; López, Jimmy A.; Gutarra, Melissa L.; Farina, Marcos; Bazylinski, Dennis A.; Freire, Denise M. G.

    2013-01-01

    The growth and magnetosome production of the marine magnetotactic vibrio Magnetovibrio blakemorei strain MV-1 were optimized through a statistics-based experimental factorial design. In the optimized growth medium, maximum magnetite yields of 64.3 mg/liter in batch cultures and 26 mg/liter in a bioreactor were obtained. PMID:23396329

  18. An experimental design approach to optimize an amperometric immunoassay on a screen printed electrode for Clostridium tetani antibody determination.

    PubMed

    Patris, Stéphanie; Vandeput, Marie; Kenfack, Gersonie Momo; Mertens, Dominique; Dejaegher, Bieke; Kauffmann, Jean-Michel

    2016-03-15

    An immunoassay for the determination of anti-tetani antibodies has been developed using a screen printed electrode (SPE) as solid support for toxoid (antigen) immobilization. The assay was performed in guinea pig serum. The immunoreaction and the subsequent amperometric detection occurred directly onto the SPE surface. The assay consisted of spiking the anti-tetani sample directly onto the toxoid modified SPE, and then a second antibody, i.e. a HRP-labeled anti-immunoglobulin G, was deposited onto the biosensor. Subsequent amperometric detection was realized by spiking 10 µL of a hydroquinone (HQ) solution into 40 µL of buffer solution containing hydrogen peroxide. An experimental design approach was implemented for the optimization of the immunoassay. The variables of interest, such as bovine serum albumin (BSA) concentration, incubation times and labeled antibody dilution, were optimized with the aid of the response surface methodology using a circumscribed central composite design (CCCD). It was observed that two factors exhibited the greatest impact on the response, i.e. the anti-tetani incubation time and the dilution factor of the labeled antibody. It was discovered that in order to maximize the response, the dilution factor should be small, while the anti-tetani antibody incubation time should be long. The BSA concentration and the HRP-anti-IgG incubation had very limited influence. Under the optimized conditions, the immunoassay had a limit of detection of 0.011 IU/mL and a limit of quantification of 0.012 IU/mL. These values were below the protective human antibody limit of 0.06 IU/mL. PMID:26454827

  19. Heterogeneity of the gut microbiome in mice: guidelines for optimizing experimental design

    PubMed Central

    Laukens, Debby; Brinkman, Brigitta M.; Raes, Jeroen; De Vos, Martine; Vandenabeele, Peter

    2015-01-01

    Targeted manipulation of the gut flora is increasingly being recognized as a means to improve human health. Yet, the temporal dynamics and intra- and interindividual heterogeneity of the microbiome represent experimental limitations, especially in human cross-sectional studies. Therefore, rodent models represent an invaluable tool to study the host–microbiota interface. Progress in technical and computational tools to investigate the composition and function of the microbiome has opened a new era of research and we gradually begin to understand the parameters that influence variation of host-associated microbial communities. To isolate true effects from confounding factors, it is essential to include such parameters in model intervention studies. Also, explicit journal instructions to include essential information on animal experiments are mandatory. The purpose of this review is to summarize the factors that influence microbiota composition in mice and to provide guidelines to improve the reproducibility of animal experiments. PMID:26323480

  20. Quantitative and qualitative optimization of allergen extraction from peanut and selected tree nuts. Part 1. Screening of optimal extraction conditions using a D-optimal experimental design.

    PubMed

    L'Hocine, Lamia; Pitre, Mélanie

    2016-03-01

    A D-optimal design was constructed to optimize allergen extraction efficiency simultaneously from roasted, non-roasted, defatted, and non-defatted almond, hazelnut, peanut, and pistachio flours using three non-denaturing aqueous (phosphate, borate, and carbonate) buffers at various conditions of ionic strength, buffer-to-protein ratio, extraction temperature, and extraction duration. Statistical analysis showed that roasting and non-defatting significantly lowered protein recovery for all nuts. Increasing the temperature and the buffer-to-protein ratio during extraction significantly increased protein recovery, whereas increasing the extraction time had no significant impact. The impact of the three buffers on protein recovery varied significantly among the nuts. Depending on the extraction conditions, protein recovery varied from 19% to 95% for peanut, 31% to 73% for almond, 17% to 64% for pistachio, and 27% to 88% for hazelnut. A modulation by the buffer type and ionic strength of protein and immunoglobuline E binding profiles of extracts was evidenced, where high protein recovery levels did not always correlate with high immunoreactivity. PMID:26471618

  1. Numerical and experimental analysis of a ducted propeller designed by a fully automated optimization process under open water condition

    NASA Astrophysics Data System (ADS)

    Yu, Long; Druckenbrod, Markus; Greve, Martin; Wang, Ke-qi; Abdel-Maksoud, Moustafa

    2015-10-01

    A fully automated optimization process is provided for the design of ducted propellers under open water conditions, including 3D geometry modeling, meshing, optimization algorithm and CFD analysis techniques. The developed process allows the direct integration of a RANSE solver in the design stage. A practical ducted propeller design case study is carried out for validation. Numerical simulations and open water tests are fulfilled and proved that the optimum ducted propeller improves hydrodynamic performance as predicted.

  2. D-OPTIMAL EXPERIMENTAL DESIGNS TO TEST FOR DEPARTURE FROM ADDITIVITY IN A FIXED-RATIO MIXTURE RAY.

    EPA Science Inventory

    Humans are exposed to mixtures of environmental compounds. A regulatory assumption is that the mixtures of chemicals act in an additive manner. However, this assumption requires experimental validation. Traditional experimental designs (full factorial) require a large number of e...

  3. Optimization of Wear Behavior of Magnesium Alloy AZ91 Hybrid Composites Using Taguchi Experimental Design

    NASA Astrophysics Data System (ADS)

    Girish, B. M.; Satish, B. M.; Sarapure, Sadanand; Basawaraj

    2016-06-01

    In the present paper, the statistical investigation on wear behavior of magnesium alloy (AZ91) hybrid metal matrix composites using Taguchi technique has been reported. The composites were reinforced with SiC and graphite particles of average size 37 μm. The specimens were processed by stir casting route. Dry sliding wear of the hybrid composites were tested on a pin-on-disk tribometer under dry conditions at different normal loads (20, 40, and 60 N), sliding speeds (1.047, 1.57, and 2.09 m/s), and composition (1, 2, and 3 wt pct of each of SiC and graphite). The design of experiments approach using Taguchi technique was employed to statistically analyze the wear behavior of hybrid composites. Signal-to-noise ratio and analysis of variance were used to investigate the influence of the parameters on the wear rate.

  4. Optimization of protein solution by a novel experimental design method using thermodynamic properties.

    PubMed

    Kim, Nam Ah; An, In Bok; Lee, Sang Yeol; Park, Eun-Seok; Jeong, Seong Hoon

    2012-09-01

    In this study, the structural stability of hen egg white lysozyme in solution at various pH levels and in different types of buffers, including acetate, phosphate, histidine, and Tris, was investigated by means of differential scanning calorimetry (DSC). Reasonable pH values were selected from the buffer ranges and were analyzed statistically through design of experiment (DoE). Four factors were used to characterize the thermograms: calorimetric enthalpy (ΔH), temperature at maximum heat flux (T( m )), van't Hoff enthalpy (ΔH( V )), and apparent activation energy of protein solution (E(app)). It was possible to calculate E(app) through mathematical elaboration from the Lumry-Eyring model by changing the scan rate. The transition temperature of protein solution, T( m ), increased when the scan rate was faster. When comparing the T( m ), ΔH( V ), ΔH, and E(app) of lysozyme in various pH ranges and buffers with different priorities, lysozyme in acetate buffer at pH 4.767 (scenario 9) to pH 4.969 (scenario 11) exhibited the highest thermodynamic stability. Through this experiment, we found a significant difference in the thermal stability of lysozyme in various pH ranges and buffers and also a new approach to investigate the physical stability of protein by DoE. PMID:23054718

  5. D-OPTIMAL EXPERIMENTAL DESIGNS TO TEST FOR DEPARTURE FROM ADDITIVITY IN A FIXED-RATIO MIXTURE RAY.

    EPA Science Inventory

    Traditional factorial designs for evaluating interactions among chemicals in a mixture are prohibitive when the number of chemicals is large. However, recent advances in statistically-based experimental design have made it easier to evaluate interactions involving many chemicals...

  6. Propeller design by optimization

    NASA Technical Reports Server (NTRS)

    Rizk, M. H.; Jou, W.-H.

    1986-01-01

    The feasibility of designing propellers by an optimization procedure is investigated. A scheme, which solves the full potential flow equation about a propeller by line relaxation, is modified so that the iterative solutions of the flow equation and the design parameters are updated simultaneously. Some technical problems in using optimization for designing propellers with maximum efficiency are identified. Approaches for overcoming these problems are presented.

  7. Optimized design for PIGMI

    SciTech Connect

    Hansborough, L.; Hamm, R.; Stovall, J.; Swenson, D.

    1980-01-01

    PIGMI (Pion Generator for Medical Irradiations) is a compact linear proton accelerator design, optimized for pion production and cancer treatment use in a hospital environment. Technology developed during a four-year PIGMI Prototype experimental program allows the design of smaller, less expensive, and more reliable proton linacs. A new type of low-energy accelerating structure, the radio-frequency quadrupole (RFQ) has been tested; it produces an exceptionally good-quality beam and allows the use of a simple 30-kV injector. Average axial electric-field gradients of over 9 MV/m have been demonstrated in a drift-tube linac (DTL) structure. Experimental work is underway to test the disk-and-washer (DAW) structure, another new type of accelerating structure for use in the high-energy coupled-cavity linac (CCL). Sufficient experimental and developmental progress has been made to closely define an actual PIGMI. It will consist of a 30-kV injector, and RFQ linac to a proton energy of 2.5 MeV, a DTL linac to 125 MeV, and a CCL linac to the final energy of 650 MeV. The total length of the accelerator is 133 meters. The RFQ and DTL will be driven by a single 440-MHz klystron; the CCL will be driven by six 1320-MHz klystrons. The peak beam current is 28 mA. The beam pulse length is 60 ..mu..s at a 60-Hz repetition rate, resulting in a 100-..mu..A average beam current. The total cost of the accelerator is estimated to be approx. $10 million.

  8. Optimization of a Three-Component Green Corrosion Inhibitor Mixture for Using in Cooling Water by Experimental Design

    NASA Astrophysics Data System (ADS)

    Asghari, E.; Ashassi-Sorkhabi, H.; Ahangari, M.; Bagheri, R.

    2016-04-01

    Factors such as inhibitor concentration, solution hydrodynamics, and temperature influence the performance of corrosion inhibitor mixtures. The simultaneous studying of the impact of different factors is a time- and cost-consuming process. The use of experimental design methods can be useful in minimizing the number of experiments and finding local optimized conditions for factors under the investigation. In the present work, the inhibition performance of a three-component inhibitor mixture against corrosion of St37 steel rotating disk electrode, RDE, was studied. The mixture was composed of citric acid, lanthanum(III) nitrate, and tetrabutylammonium perchlorate. In order to decrease the number of experiments, the L16 Taguchi orthogonal array was used. The "control factors" were the concentration of each component and the rotation rate of RDE and the "response factor" was the inhibition efficiency. The scanning electron microscopy and energy dispersive x-ray spectroscopy techniques verified the formation of islands of adsorbed citrate complexes with lanthanum ions and insoluble lanthanum(III) hydroxide. From the Taguchi analysis results the mixture of 0.50 mM lanthanum(III) nitrate, 0.50 mM citric acid, and 2.0 mM tetrabutylammonium perchlorate under the electrode rotation rate of 1000 rpm was found as optimum conditions.

  9. Experimental design applied to the optimization of pyrolysis and atomization temperatures for As measurement in water samples by GFAAS

    NASA Astrophysics Data System (ADS)

    Ávila, Akie K.; Araujo, Thiago O.; Couto, Paulo R. G.; Borges, Renata M. H.

    2005-10-01

    In general, research experimentation is often used mainly when new methodologies are being developed or existing ones are being improved. The characteristics of any method depend on its factors or components. The planning techniques and analysis of experiments are basically used to improve the analytical conditions of methods, to reduce experimental labour with the minimum of tests and to optimize the use of resources (reagents, time of analysis, availability of the equipment, operator time, etc). These techniques are applied by the identification of variables (control factors) of a process that have the most influence on the response of the parameters of interest, by attributing values to the influential variables of the process in order that the variability of response can be minimum, or the obtained value (quality parameter) be very close to the nominal value, and by attributing values to the influential variables of the process so that the effects of uncontrollable variables can be reduced. In this central composite design (CCD), four permanent modifiers (Pd, Ir, W and Rh) and one combined permanent modifier W+Ir were studied. The study selected two factors: pyrolysis and atomization temperatures at five different levels for all the possible combinations. The pyrolysis temperatures with different permanent modifiers varied from 600 °C to 1600 °C with hold times of 25 s, while atomization temperatures ranged between 1900 °C and 2280 °C. The characteristic masses for As were in the range of 31 pg to 81 pg. Assuming the best conditions obtained on CCD, it was possible to estimate the measurement uncertainty of As determination in water samples. The results showed that considering the main uncertainty sources such as the repetitivity of measurement inherent in the equipment, the calibration curve which evaluates the adjustment of the mathematical model to the results and the calibration standards concentrations, the values obtained were similar to international

  10. Optimal Flow Control Design

    NASA Technical Reports Server (NTRS)

    Allan, Brian; Owens, Lewis

    2010-01-01

    In support of the Blended-Wing-Body aircraft concept, a new flow control hybrid vane/jet design has been developed for use in a boundary-layer-ingesting (BLI) offset inlet in transonic flows. This inlet flow control is designed to minimize the engine fan-face distortion levels and the first five Fourier harmonic half amplitudes while maximizing the inlet pressure recovery. This concept represents a potentially enabling technology for quieter and more environmentally friendly transport aircraft. An optimum vane design was found by minimizing the engine fan-face distortion, DC60, and the first five Fourier harmonic half amplitudes, while maximizing the total pressure recovery. The optimal vane design was then used in a BLI inlet wind tunnel experiment at NASA Langley's 0.3-meter transonic cryogenic tunnel. The experimental results demonstrated an 80-percent decrease in DPCPavg, the reduction in the circumferential distortion levels, at an inlet mass flow rate corresponding to the middle of the operational range at the cruise condition. Even though the vanes were designed at a single inlet mass flow rate, they performed very well over the entire inlet mass flow range tested in the wind tunnel experiment with the addition of a small amount of jet flow control. While the circumferential distortion was decreased, the radial distortion on the outer rings at the aerodynamic interface plane (AIP) increased. This was a result of the large boundary layer being distributed from the bottom of the AIP in the baseline case to the outer edges of the AIP when using the vortex generator (VG) vane flow control. Experimental results, as already mentioned, showed an 80-percent reduction of DPCPavg, the circumferential distortion level at the engine fan-face. The hybrid approach leverages strengths of vane and jet flow control devices, increasing inlet performance over a broader operational range with significant reduction in mass flow requirements. Minimal distortion level requirements

  11. An experimental design approach for hydrothermal synthesis of NaYF4: Yb3+, Tm3+ upconversion microcrystal: UV emission optimization

    NASA Astrophysics Data System (ADS)

    Kaviani Darani, Masoume; Bastani, Saeed; Ghahari, Mehdi; Kardar, Pooneh

    2015-11-01

    Ultraviolet (UV) emissions of hydrothermally synthesized NaYF4: Yb3+, Tm3+ upconversion crystals were optimized using the response surface methodology experimental design. In these experimental designs, 9 runs, two factors namely (1) Tm3+ ion concentration, and (2) pH value were investigated using 3 different ligands. Introducing UV upconversion emissions as responses, their intensity were separately maximized. Analytical methods such as XRD, SEM, and FTIR could be used to study crystal structure, morphology, and fluorescent spectroscopy in order to obtain luminescence properties. From the photo-luminescence spectra, emissions centered at 347, 364, 452, 478, 648 and 803 nm were observed. Some results show that increasing each DOE factor up to an optimum value resulted in an increase in emission intensity, followed by reduction. To optimize UV emission, as a final result to the UV emission optimization, each design had a suggestion.

  12. Evidence evaluation: measure Z corresponds to human utility judgments better than measure L and optimal-experimental-design models.

    PubMed

    Rusconi, Patrice; Marelli, Marco; D'Addario, Marco; Russo, Selena; Cherubini, Paolo

    2014-05-01

    Evidence evaluation is a crucial process in many human activities, spanning from medical diagnosis to impression formation. The present experiments investigated which, if any, normative model best conforms to people's intuition about the value of the obtained evidence. Psychologists, epistemologists, and philosophers of science have proposed several models to account for people's intuition about the utility of the obtained evidence with respect either to a focal hypothesis or to a constellation of hypotheses. We pitted against each other the so-called optimal-experimental-design models (i.e., Bayesian diagnosticity, log₁₀ diagnosticity, information gain, Kullback-Leibler distance, probability gain, and impact) and measures L and Z to compare their ability to describe humans' intuition about the value of the obtained evidence. Participants received words-and-numbers scenarios concerning 2 hypotheses and binary features. They were asked to evaluate the utility of "yes" and "no" answers to questions about some features possessed in different proportions (i.e., the likelihoods) by 2 types of extraterrestrial creatures (corresponding to 2 mutually exclusive and exhaustive hypotheses). Participants evaluated either how an answer was helpful or how an answer decreased/increased their beliefs with respect either to a single hypothesis or to both hypotheses. We fitted mixed-effects models and used the Akaike information criterion and the Bayesian information criterion values to compare the competing models of the value of the obtained evidence. Overall, the experiments showed that measure Z was the best fitting model of participants' judgments of the value of obtained answers. We discussed the implications for the human hypothesis-evaluation process. PMID:24446753

  13. Optimization of digital designs

    NASA Technical Reports Server (NTRS)

    Whitaker, Sterling R. (Inventor); Miles, Lowell H. (Inventor)

    2009-01-01

    An application specific integrated circuit is optimized by translating a first representation of its digital design to a second representation. The second representation includes multiple syntactic expressions that admit a representation of a higher-order function of base Boolean values. The syntactic expressions are manipulated to form a third representation of the digital design.

  14. Improvement of production of citric acid from oil palm empty fruit bunches: optimization of media by statistical experimental designs.

    PubMed

    Bari, Md Niamul; Alam, Md Zahangir; Muyibi, Suleyman A; Jamal, Parveen; Abdullah-Al-Mamun

    2009-06-01

    A sequential optimization based on statistical design and one-factor-at-a-time (OFAT) method was employed to optimize the media constituents for the improvement of citric acid production from oil palm empty fruit bunches (EFB) through solid state bioconversion using Aspergillus niger IBO-103MNB. The results obtained from the Plackett-Burman design indicated that the co-substrate (sucrose), stimulator (methanol) and minerals (Zn, Cu, Mn and Mg) were found to be the major factors for further optimization. Based on the OFAT method, the selected medium constituents and inoculum concentration were optimized by the central composite design (CCD) under the response surface methodology (RSM). The statistical analysis showed that the optimum media containing 6.4% (w/w) of sucrose, 9% (v/w) of minerals and 15.5% (v/w) of inoculum gave the maximum production of citric acid (337.94 g/kg of dry EFB). The analysis showed that sucrose (p<0.0011) and mineral solution (p<0.0061) were more significant compared to inoculum concentration (p<0.0127) for the citric acid production. PMID:19231166

  15. Optimal control model predictions of system performance and attention allocation and their experimental validation in a display design study

    NASA Technical Reports Server (NTRS)

    Johannsen, G.; Govindaraj, T.

    1980-01-01

    The influence of different types of predictor displays in a longitudinal vertical takeoff and landing (VTOL) hover task is analyzed in a theoretical study. Several cases with differing amounts of predictive and rate information are compared. The optimal control model of the human operator is used to estimate human and system performance in terms of root-mean-square (rms) values and to compute optimized attention allocation. The only part of the model which is varied to predict these data is the observation matrix. Typical cases are selected for a subsequent experimental validation. The rms values as well as eye-movement data are recorded. The results agree favorably with those of the theoretical study in terms of relative differences. Better matching is achieved by revised model input data.

  16. Optimal experimental design for the detection of light atoms from high-resolution scanning transmission electron microscopy images

    SciTech Connect

    Gonnissen, J.; De Backer, A.; Martinez, G. T.; Van Aert, S.; Dekker, A. J. den; Rosenauer, A.; Sijbers, J.

    2014-08-11

    We report an innovative method to explore the optimal experimental settings to detect light atoms from scanning transmission electron microscopy (STEM) images. Since light elements play a key role in many technologically important materials, such as lithium-battery devices or hydrogen storage applications, much effort has been made to optimize the STEM technique in order to detect light elements. Therefore, classical performance criteria, such as contrast or signal-to-noise ratio, are often discussed hereby aiming at improvements of the direct visual interpretability. However, when images are interpreted quantitatively, one needs an alternative criterion, which we derive based on statistical detection theory. Using realistic simulations of technologically important materials, we demonstrate the benefits of the proposed method and compare the results with existing approaches.

  17. Optimal designs for copula models

    PubMed Central

    Perrone, E.; Müller, W.G.

    2016-01-01

    Copula modelling has in the past decade become a standard tool in many areas of applied statistics. However, a largely neglected aspect concerns the design of related experiments. Particularly the issue of whether the estimation of copula parameters can be enhanced by optimizing experimental conditions and how robust all the parameter estimates for the model are with respect to the type of copula employed. In this paper an equivalence theorem for (bivariate) copula models is provided that allows formulation of efficient design algorithms and quick checks of whether designs are optimal or at least efficient. Some examples illustrate that in practical situations considerable gains in design efficiency can be achieved. A natural comparison between different copula models with respect to design efficiency is provided as well. PMID:27453616

  18. "Real-time" disintegration analysis and D-optimal experimental design for the optimization of diclofenac sodium fast-dissolving films.

    PubMed

    El-Malah, Yasser; Nazzal, Sami

    2013-01-01

    The objective of this work was to study the dissolution and mechanical properties of fast-dissolving films prepared from a tertiary mixture of pullulan, polyvinylpyrrolidone and hypromellose. Disintegration studies were performed in real-time by probe spectroscopy to detect the onset of film disintegration. Tensile strength and elastic modulus of the films were measured by texture analysis. Disintegration time of the films ranged from 21 to 105 seconds whereas their mechanical properties ranged from approximately 2 to 49 MPa for tensile strength and 1 to 21 MPa% for young's modulus. After generating polynomial models correlating the variables using a D-Optimal mixture design, an optimal formulation with desired responses was proposed by the statistical package. For validation, a new film formulation loaded with diclofenac sodium based on the optimized composition was prepared and tested for dissolution and tensile strength. Dissolution of the optimized film was found to commence almost immediately with 50% of the drug released within one minute. Tensile strength and young's modulus of the film were 11.21 MPa and 6, 78 MPa%, respectively. Real-time spectroscopy in conjunction with statistical design were shown to be very efficient for the optimization and development of non-conventional intraoral delivery system such as fast dissolving films. PMID:22780708

  19. Greater enhancement of Bacillus subtilis spore yields in submerged cultures by optimization of medium composition through statistical experimental designs.

    PubMed

    Chen, Zhen-Min; Li, Qing; Liu, Hua-Mei; Yu, Na; Xie, Tian-Jian; Yang, Ming-Yuan; Shen, Ping; Chen, Xiang-Dong

    2010-02-01

    Bacillus subtilis spore preparations are promising probiotics and biocontrol agents, which can be used in plants, animals, and humans. The aim of this work was to optimize the nutritional conditions using a statistical approach for the production of B. subtilis (WHK-Z12) spores. Our preliminary experiments show that corn starch, corn flour, and wheat bran were the best carbon sources. Using Plackett-Burman design, corn steep liquor, soybean flour, and yeast extract were found to be the best nitrogen source ingredients for enhancing spore production and were studied for further optimization using central composite design. The key medium components in our optimization medium were 16.18 g/l of corn steep liquor, 17.53 g/l of soybean flour, and 8.14 g/l of yeast extract. The improved medium produced spores as high as 1.52 +/- 0.06 x 10(10) spores/ml under flask cultivation conditions, and 1.56 +/- 0.07 x 10(10) spores/ml could be achieved in a 30-l fermenter after 40 h of cultivation. To the best of our knowledge, these results compared favorably to the documented spore yields produced by B. subtilis strains. PMID:19697022

  20. Optimization of a liquid chromatography ion mobility-mass spectrometry method for untargeted metabolomics using experimental design and multivariate data analysis.

    PubMed

    Tebani, Abdellah; Schmitz-Afonso, Isabelle; Rutledge, Douglas N; Gonzalez, Bruno J; Bekri, Soumeya; Afonso, Carlos

    2016-03-24

    High-resolution mass spectrometry coupled with pattern recognition techniques is an established tool to perform comprehensive metabolite profiling of biological datasets. This paves the way for new, powerful and innovative diagnostic approaches in the post-genomic era and molecular medicine. However, interpreting untargeted metabolomic data requires robust, reproducible and reliable analytical methods to translate results into biologically relevant and actionable knowledge. The analyses of biological samples were developed based on ultra-high performance liquid chromatography (UHPLC) coupled to ion mobility - mass spectrometry (IM-MS). A strategy for optimizing the analytical conditions for untargeted UHPLC-IM-MS methods is proposed using an experimental design approach. Optimization experiments were conducted through a screening process designed to identify the factors that have significant effects on the selected responses (total number of peaks and number of reliable peaks). For this purpose, full and fractional factorial designs were used while partial least squares regression was used for experimental design modeling and optimization of parameter values. The total number of peaks yielded the best predictive model and is used for optimization of parameters setting. PMID:26944989

  1. The Automatic Neuroscientist: A framework for optimizing experimental design with closed-loop real-time fMRI

    PubMed Central

    Lorenz, Romy; Monti, Ricardo Pio; Violante, Inês R.; Anagnostopoulos, Christoforos; Faisal, Aldo A.; Montana, Giovanni; Leech, Robert

    2016-01-01

    Functional neuroimaging typically explores how a particular task activates a set of brain regions. Importantly though, the same neural system can be activated by inherently different tasks. To date, there is no approach available that systematically explores whether and how distinct tasks probe the same neural system. Here, we propose and validate an alternative framework, the Automatic Neuroscientist, which turns the standard fMRI approach on its head. We use real-time fMRI in combination with modern machine-learning techniques to automatically design the optimal experiment to evoke a desired target brain state. In this work, we present two proof-of-principle studies involving perceptual stimuli. In both studies optimization algorithms of varying complexity were employed; the first involved a stochastic approximation method while the second incorporated a more sophisticated Bayesian optimization technique. In the first study, we achieved convergence for the hypothesized optimum in 11 out of 14 runs in less than 10 min. Results of the second study showed how our closed-loop framework accurately and with high efficiency estimated the underlying relationship between stimuli and neural responses for each subject in one to two runs: with each run lasting 6.3 min. Moreover, we demonstrate that using only the first run produced a reliable solution at a group-level. Supporting simulation analyses provided evidence on the robustness of the Bayesian optimization approach for scenarios with low contrast-to-noise ratio. This framework is generalizable to numerous applications, ranging from optimizing stimuli in neuroimaging pilot studies to tailoring clinical rehabilitation therapy to patients and can be used with multiple imaging modalities in humans and animals. PMID:26804778

  2. The Automatic Neuroscientist: A framework for optimizing experimental design with closed-loop real-time fMRI.

    PubMed

    Lorenz, Romy; Monti, Ricardo Pio; Violante, Inês R; Anagnostopoulos, Christoforos; Faisal, Aldo A; Montana, Giovanni; Leech, Robert

    2016-04-01

    Functional neuroimaging typically explores how a particular task activates a set of brain regions. Importantly though, the same neural system can be activated by inherently different tasks. To date, there is no approach available that systematically explores whether and how distinct tasks probe the same neural system. Here, we propose and validate an alternative framework, the Automatic Neuroscientist, which turns the standard fMRI approach on its head. We use real-time fMRI in combination with modern machine-learning techniques to automatically design the optimal experiment to evoke a desired target brain state. In this work, we present two proof-of-principle studies involving perceptual stimuli. In both studies optimization algorithms of varying complexity were employed; the first involved a stochastic approximation method while the second incorporated a more sophisticated Bayesian optimization technique. In the first study, we achieved convergence for the hypothesized optimum in 11 out of 14 runs in less than 10min. Results of the second study showed how our closed-loop framework accurately and with high efficiency estimated the underlying relationship between stimuli and neural responses for each subject in one to two runs: with each run lasting 6.3min. Moreover, we demonstrate that using only the first run produced a reliable solution at a group-level. Supporting simulation analyses provided evidence on the robustness of the Bayesian optimization approach for scenarios with low contrast-to-noise ratio. This framework is generalizable to numerous applications, ranging from optimizing stimuli in neuroimaging pilot studies to tailoring clinical rehabilitation therapy to patients and can be used with multiple imaging modalities in humans and animals. PMID:26804778

  3. Ethanol production from orange peels: two-stage hydrolysis and fermentation studies using optimized parameters through experimental design.

    PubMed

    Oberoi, Harinder Singh; Vadlani, Praveen Venkata; Madl, Ronald L; Saida, Lavudi; Abeykoon, Jithma P

    2010-03-24

    Orange peels were evaluated as a fermentation feedstock, and process conditions for enhanced ethanol production were determined. Primary hydrolysis of orange peel powder (OPP) was carried out at acid concentrations from 0 to 1.0% (w/v) at 121 degrees C and 15 psi for 15 min. High-performance liquid chromatography analysis of sugars and inhibitory compounds showed a higher production of hydroxymethyfurfural and acetic acid and a decrease in sugar concentration when the acid level was beyond 0.5% (w/v). Secondary hydrolysis of pretreated biomass obtained from primary hydrolysis was carried out at 0.5% (w/v) acid. Response surface methodology using three factors and a two-level central composite design was employed to optimize the effect of pH, temperature, and fermentation time on ethanol production from OPP hydrolysate at the shake flask level. On the basis of results obtained from the optimization experiment and numerical optimization software, a validation study was carried out in a 2 L batch fermenter at pH 5.4 and a temperature of 34 degrees C for 15 h. The hydrolysate obtained from primary and secondary hydrolysis processes was fermented separately employing parameters optimized through RSM. Ethanol yields of 0.25 g/g on a biomass basis (YP/X) and 0.46 g/g on a substrate-consumed basis (YP/S) and a promising volumetric ethanol productivity of 3.37 g/L/h were attained using this process at the fermenter level, which shows promise for further scale-up studies. PMID:20158208

  4. Study of Viscoelasticity and stability Effects on an Optimal Design of ULD Scientific Balloon: Simulation and Experimental Verification

    NASA Astrophysics Data System (ADS)

    Singh, Balbir; Radhakrishnan, Jayakrishnan

    The need for uniform stress distribution of early super pressure balloons led a long period of the development of several optimal designs of scientific balloons for significant benefits in public outreach and education for the next generation of scientists, engineers and technicians. Scientific ballooning is continuing to evolve. The paper is divided into two sections. First section deals with the design, development and optimization of a cost effective high altitude ultra-long duration scientific balloon which are made up of stiff meridional tendons of a specific material. The numerical calculation along with the analysis of elastic buckling pressure, anisotropic viscoelstic effects is taken into consideration. The second section deals with the stability of the ULD balloon constructed by joining together a number of gores so that the structure forms a series of bulges and lobes. Under certain condition at high altitudes the nominal configuration of these structures is unstable and hence the analysis for development scenario must be envisaged so that the structure will never assume its intended shape which may end up in distorted but more stable configuration. The stability of the structure is shown depending upon the geometry and calculations using different softwares like MATLAB 2012b, ANSYS 14.0 (APDL).

  5. Optimal design of antireflection coating and experimental verification by plasma enhanced chemical vapor deposition in small displays

    SciTech Connect

    Yang, S. M.; Hsieh, Y. C.; Jeng, C. A.

    2009-03-15

    Conventional antireflection coating by thin films of quarter-wavelength thickness is limited by material selections and these films' refractive indices. The optimal design by non-quarter-wavelength thickness is presented in this study. A multilayer thin-film model is developed by the admittance loci to show that the two-layer thin film of SiN{sub x}/SiO{sub y} at 124/87 nm and three layer of SiN{sub x}/SiN{sub y}/SiO{sub z} at 58/84/83 nm can achieve average transmittances of 94.4% and 94.9%, respectively, on polymer, glass, and silicon substrates. The optimal design is validated by plasma enhanced chemical vapor deposition of N{sub 2}O/SiH{sub 4} and NH{sub 3}/SiH{sub 4} to achieve the desired optical constants. Application of the antireflection coating to a 4 in. liquid crystal display demonstrates that the transmittance is over 94%, the mean luminance can be increased by 25%, and the total reflection angle increased from 41 deg. to 58 deg.

  6. Optimization of ultrasound assisted dispersive liquid-liquid microextraction of six antidepressants in human plasma using experimental design.

    PubMed

    Fernández, P; Taboada, V; Regenjo, M; Morales, L; Alvarez, I; Carro, A M; Lorenzo, R A

    2016-05-30

    A simple Ultrasounds Assisted-Dispersive Liquid Liquid Microextraction (UA-DLLME) method is presented for the simultaneous determination of six second-generation antidepressants in plasma by Ultra Performance Liquid Chromatography with Photodiode Array Detector (UPLC-PDA). The main factors that potentially affect to DLLME were optimized by a screening design followed by a response surface design and desirability functions. The optimal conditions were 2.5mL of acetonitrile as dispersant solvent, 0.2mL of chloroform as extractant solvent, 3min of ultrasounds stirring and extraction pH 9.8.Under optimized conditions, the UPLC-PDA method showed good separation of antidepressants in 2.5min and good linearity in the range of 0.02-4μgmL(-1), with determination coefficients higher than 0.998. The limits of detection were in the range 4-5ngmL(-1). The method precision (n=5) was evaluated showing relative standard deviations (RSD) lower than 8.1% for all compounds. The average recoveries ranged from 92.5% for fluoxetine to 110% for mirtazapine. The applicability of DLLME/UPLC-PDA was successfully tested in twenty nine plasma samples from antidepressant consumers. Real samples were analyzed by the proposed method and the results were successfully submitted to comparison with those obtained by a Liquid Liquid Extraction-Gas Chromatography - Mass Spectrometry (LLE-GC-MS) method. The results confirmed the presence of venlafaxine in most cases (19 cases), followed by sertraline (3 cases) and fluoxetine (3 cases) at concentrations below toxic levels. PMID:26955756

  7. Elements of Bayesian experimental design

    SciTech Connect

    Sivia, D.S.

    1997-09-01

    We consider some elements of the Bayesian approach that are important for optimal experimental design. While the underlying principles used are very general, and are explained in detail in a recent tutorial text, they are applied here to the specific case of characterising the inferential value of different resolution peakshapes. This particular issue was considered earlier by Silver, Sivia and Pynn (1989, 1990a, 1990b), and the following presentation confirms and extends the conclusions of their analysis.

  8. An orbital angular momentum radio communication system optimized by intensity controlled masks effectively: Theoretical design and experimental verification

    NASA Astrophysics Data System (ADS)

    Gao, Xinlu; Huang, Shanguo; Wei, Yongfeng; Zhai, Wensheng; Xu, Wenjing; Yin, Shan; Zhou, Jing; Gu, Wanyi

    2014-12-01

    A system of generating and receiving orbital angular momentum (OAM) radio beams, which are collectively formed by two circular array antennas (CAAs) and effectively optimized by two intensity controlled masks, is proposed and experimentally investigated. The scheme is effective in blocking of the unwanted OAM modes and enhancing the power of received radio signals, which results in the capacity gain of system and extended transmission distance of the OAM radio beams. The operation principle of the intensity controlled masks, which can be regarded as both collimator and filter, is feasible and simple to realize. Numerical simulations of intensity and phase distributions at each key cross-sectional plane of the radio beams demonstrate the collimated results. The experimental results match well with the theoretical analysis and the receive distance of the OAM radio beam at radio frequency (RF) 20 GHz is extended up to 200 times of the wavelength of the RF signals, the measured distance is 5 times of the original measured distance. The presented proof-of-concept experiment demonstrates the feasibility of the system.

  9. An orbital angular momentum radio communication system optimized by intensity controlled masks effectively: Theoretical design and experimental verification

    SciTech Connect

    Gao, Xinlu; Huang, Shanguo Wei, Yongfeng; Zhai, Wensheng; Xu, Wenjing; Yin, Shan; Gu, Wanyi; Zhou, Jing

    2014-12-15

    A system of generating and receiving orbital angular momentum (OAM) radio beams, which are collectively formed by two circular array antennas (CAAs) and effectively optimized by two intensity controlled masks, is proposed and experimentally investigated. The scheme is effective in blocking of the unwanted OAM modes and enhancing the power of received radio signals, which results in the capacity gain of system and extended transmission distance of the OAM radio beams. The operation principle of the intensity controlled masks, which can be regarded as both collimator and filter, is feasible and simple to realize. Numerical simulations of intensity and phase distributions at each key cross-sectional plane of the radio beams demonstrate the collimated results. The experimental results match well with the theoretical analysis and the receive distance of the OAM radio beam at radio frequency (RF) 20 GHz is extended up to 200 times of the wavelength of the RF signals, the measured distance is 5 times of the original measured distance. The presented proof-of-concept experiment demonstrates the feasibility of the system.

  10. A simple way to obtain high saturation magnetization for superparamagnetic iron oxide nanoparticles synthesized in air atmosphere: Optimization by experimental design

    NASA Astrophysics Data System (ADS)

    Karaagac, Oznur; Kockar, Hakan

    2016-07-01

    Orthogonal design technique was applied to obtain superparamagnetic iron oxide nanoparticles with high saturation magnetization, Ms. Synthesis of the nanoparticles were done in air atmosphere according to the orthogonal table L934. Magnetic properties of the synthesized nanoparticles were measured by a vibrating sample magnetometer. Structural analysis of the nanoparticles was also carried out by X-ray diffraction technique (XRD), Fourier transform infrared spectroscopy (FTIR) and transmission electron microscopy (TEM). After the analysis of magnetic data, the optimized experimental parameters were determined as [Fe+2]/[Fe+3]=6/6, iron ion concentration=1500 mM, base concentration=6.7 M and reaction time=2 min. Magnetic results showed that the synthesis carried out according to the optimized conditions gave the highest Ms of 69.83 emu/g for the nanoparticles synthesized in air atmosphere. Magnetic measurements at 10 K and 300 K showed the sample is superparamagnetic at room temperature. Structural analysis by XRD, FTIR and selected area electron diffraction showed that the sample had the inverse spinel crystal structure of iron oxide. The particle size of the optimized sample determined from the TEM image is 7.0±2.2 nm. The results indicated that the Ms of superparamagnetic iron oxide nanoparticles can be optimized by experimental design with the suitable choice of the synthesis parameters.

  11. Teaching experimental design.

    PubMed

    Fry, Derek J

    2014-01-01

    Awareness of poor design and published concerns over study quality stimulated the development of courses on experimental design intended to improve matters. This article describes some of the thinking behind these courses and how the topics can be presented in a variety of formats. The premises are that education in experimental design should be undertaken with an awareness of educational principles, of how adults learn, and of the particular topics in the subject that need emphasis. For those using laboratory animals, it should include ethical considerations, particularly severity issues, and accommodate learners not confident with mathematics. Basic principles, explanation of fully randomized, randomized block, and factorial designs, and discussion of how to size an experiment form the minimum set of topics. A problem-solving approach can help develop the skills of deciding what are correct experimental units and suitable controls in different experimental scenarios, identifying when an experiment has not been properly randomized or blinded, and selecting the most efficient design for particular experimental situations. Content, pace, and presentation should suit the audience and time available, and variety both within a presentation and in ways of interacting with those being taught is likely to be effective. Details are given of a three-day course based on these ideas, which has been rated informative, educational, and enjoyable, and can form a postgraduate module. It has oral presentations reinforced by group exercises and discussions based on realistic problems, and computer exercises which include some analysis. Other case studies consider a half-day format and a module for animal technicians. PMID:25541547

  12. Dynamic modeling, experimental evaluation, optimal design and control of integrated fuel cell system and hybrid energy systems for building demands

    NASA Astrophysics Data System (ADS)

    Nguyen, Gia Luong Huu

    obtained experimental data, the research studied the control of airflow to regulate the temperature of reactors within the fuel processor. The dynamic model provided a platform to test the dynamic response for different control gains. With sufficient sensing and appropriate control, a rapid response to maintain the temperature of the reactor despite an increase in power was possible. The third part of the research studied the use of a fuel cell in conjunction with photovoltaic panels, and energy storage to provide electricity for buildings. This research developed an optimization framework to determine the size of each device in the hybrid energy system to satisfy the electrical demands of buildings and yield the lowest cost. The advantage of having the fuel cell with photovoltaic and energy storage was the ability to operate the fuel cell at baseload at night, thus reducing the need for large battery systems to shift the solar power produced in the day to the night. In addition, the dispatchability of the fuel cell provided an extra degree of freedom necessary for unforeseen disturbances. An operation framework based on model predictive control showed that the method is suitable for optimizing the dispatch of the hybrid energy system.

  13. Improved Titanium Billet Inspection Sensitivity through Optimized Phased Array Design, Part II: Experimental Validation and Comparative Study with Multizone

    SciTech Connect

    Hassan, W.; Vensel, F.; Knowles, B.

    2006-03-06

    The inspection of critical rotating components of aircraft engines has made important advances over the last decade. The development of Phased Array (PA) inspection capability for billet and forging materials used in the manufacturing of critical engine rotating components has been a priority for Honeywell Aerospace. The demonstration of improved PA inspection system sensitivity over what is currently used at the inspection houses is a critical step in the development of this technology and its introduction to the supply base as a production inspection. As described in Part I (in these proceedings), a new phased array transducer was designed and manufactured for optimal inspection of eight inch diameter Ti-6Al-4V billets. After confirming that the transducer was manufactured in accordance with the design specifications a validation study was conducted to assess the sensitivity improvement of the PAI over the current capability of Multi-zone (MZ) inspection. The results of this study confirm the significant ({approx_equal} 6 dB in FBH number sign sensitivity) improvement of the PAI sensitivity over that of MZI.

  14. Optimized quadrature surface coil designs

    PubMed Central

    Kumar, Ananda; Bottomley, Paul A.

    2008-01-01

    Background Quadrature surface MRI/MRS detectors comprised of circular loop and figure-8 or butterfly-shaped coils offer improved signal-to-noise-ratios (SNR) compared to single surface coils, and reduced power and specific absorption rates (SAR) when used for MRI excitation. While the radius of the optimum loop coil for performing MRI at depth d in a sample is known, the optimum geometry for figure-8 and butterfly coils is not. Materials and methods The geometries of figure-8 and square butterfly detector coils that deliver the optimum SNR are determined numerically by the electromagnetic method of moments. Figure-8 and loop detectors are then combined to create SNR-optimized quadrature detectors whose theoretical and experimental SNR performance are compared with a novel quadrature detector comprised of a strip and a loop, and with two overlapped loops optimized for the same depth at 3 T. The quadrature detection efficiency and local SAR during transmission for the three quadrature configurations are analyzed and compared. Results The SNR-optimized figure-8 detector has loop radius r8 ∼ 0.6d, so r8/r0 ∼ 1.3 in an optimized quadrature detector at 3 T. The optimized butterfly coil has side length ∼ d and crossover angle of ≥ 150° at the center. Conclusions These new design rules for figure-8 and butterfly coils optimize their performance as linear and quadrature detectors. PMID:18057975

  15. Experimental design approach to the optimization of ultrasonic degradation of alachlor and enhancement of treated water biodegradability.

    PubMed

    Torres, Ricardo A; Mosteo, Rosa; Pétrier, Christian; Pulgarin, Cesar

    2009-03-01

    This work presents the application of experimental design for the ultrasonic degradation of alachlor which is pesticide classified as priority substance by the European Commission within the scope of the Water Framework Directive. The effect of electrical power (20-80W), pH (3-10) and substrate concentration (10-50mgL(-1)) was evaluated. For a confidential level of 90%, pH showed a low effect on the initial degradation rate of alachlor; whereas electrical power, pollutant concentration and the interaction of these two parameters were significant. A reduced model taking into account the significant variables and interactions between variables has shown a good correlation with the experimental results. Additional experiments conducted in natural and deionised water indicated that the alachlor degradation by ultrasound is practically unaffected by the presence of potential *OH radical scavengers: bicarbonate, sulphate, chloride and oxalic acid. In both cases, alachlor was readily eliminated ( approximately 75min). However, after 4h of treatment only 20% of the initial TOC was removed, showing that alachlor by-products are recalcitrant to the ultrasonic action. Biodegradability test (BOD5/COD) carried out during the course of the treatment indicated that the ultrasonic system noticeably increases the biodegradability of the initial solution. PMID:18930694

  16. OPTIMAL NETWORK TOPOLOGY DESIGN

    NASA Technical Reports Server (NTRS)

    Yuen, J. H.

    1994-01-01

    This program was developed as part of a research study on the topology design and performance analysis for the Space Station Information System (SSIS) network. It uses an efficient algorithm to generate candidate network designs (consisting of subsets of the set of all network components) in increasing order of their total costs, and checks each design to see if it forms an acceptable network. This technique gives the true cost-optimal network, and is particularly useful when the network has many constraints and not too many components. It is intended that this new design technique consider all important performance measures explicitly and take into account the constraints due to various technical feasibilities. In the current program, technical constraints are taken care of by the user properly forming the starting set of candidate components (e.g. nonfeasible links are not included). As subsets are generated, they are tested to see if they form an acceptable network by checking that all requirements are satisfied. Thus the first acceptable subset encountered gives the cost-optimal topology satisfying all given constraints. The user must sort the set of "feasible" link elements in increasing order of their costs. The program prompts the user for the following information for each link: 1) cost, 2) connectivity (number of stations connected by the link), and 3) the stations connected by that link. Unless instructed to stop, the program generates all possible acceptable networks in increasing order of their total costs. The program is written only to generate topologies that are simply connected. Tests on reliability, delay, and other performance measures are discussed in the documentation, but have not been incorporated into the program. This program is written in PASCAL for interactive execution and has been implemented on an IBM PC series computer operating under PC DOS. The disk contains source code only. This program was developed in 1985.

  17. Optimization of a pharmaceutical freeze-dried product and its process using an experimental design approach and innovative process analyzers.

    PubMed

    De Beer, T R M; Wiggenhorn, M; Hawe, A; Kasper, J C; Almeida, A; Quinten, T; Friess, W; Winter, G; Vervaet, C; Remon, J P

    2011-02-15

    The aim of the present study was to examine the possibilities/advantages of using recently introduced in-line spectroscopic process analyzers (Raman, NIR and plasma emission spectroscopy), within well-designed experiments, for the optimization of a pharmaceutical formulation and its freeze-drying process. The formulation under investigation was a mannitol (crystalline bulking agent)-sucrose (lyo- and cryoprotector) excipient system. The effects of two formulation variables (mannitol/sucrose ratio and amount of NaCl) and three process variables (freezing rate, annealing temperature and secondary drying temperature) upon several critical process and product responses (onset and duration of ice crystallization, onset and duration of mannitol crystallization, duration of primary drying, residual moisture content and amount of mannitol hemi-hydrate in end product) were examined using a design of experiments (DOE) methodology. A 2-level fractional factorial design (2(5-1)=16 experiments+3 center points=19 experiments) was employed. All experiments were monitored in-line using Raman, NIR and plasma emission spectroscopy, which supply continuous process and product information during freeze-drying. Off-line X-ray powder diffraction analysis and Karl-Fisher titration were performed to determine the morphology and residual moisture content of the end product, respectively. In first instance, the results showed that - besides the previous described findings in De Beer et al., Anal. Chem. 81 (2009) 7639-7649 - Raman and NIR spectroscopy are able to monitor the product behavior throughout the complete annealing step during freeze-drying. The DOE approach allowed predicting the optimum combination of process and formulation parameters leading to the desired responses. Applying a mannitol/sucrose ratio of 4, without adding NaCl and processing the formulation without an annealing step, using a freezing rate of 0.9°C/min and a secondary drying temperature of 40°C resulted in

  18. Thermal and optical design analyses, optimizations, and experimental verification for a novel glare-free LED lamp for household applications.

    PubMed

    Khan, M Nisa

    2015-07-20

    Light-emitting diode (LED) technologies are undergoing very fast developments to enable household lamp products with improved energy efficiency and lighting properties at lower cost. Although many LED replacement lamps are claimed to provide similar or better lighting quality at lower electrical wattage compared with general-purpose incumbent lamps, certain lighting characteristics important to human vision are neglected in this comparison, which include glare-free illumination and omnidirectional or sufficiently broad light distribution with adequate homogeneity. In this paper, we comprehensively investigate the thermal and lighting performance and trade-offs for several commercial LED replacement lamps for the most popular Edison incandescent bulb. We present simulations and analyses for thermal and optical performance trade-offs for various LED lamps at the chip and module granularity levels. In addition, we present a novel, glare-free, and production-friendly LED lamp design optimized to produce very desirable light distribution properties as demonstrated by our simulation results, some of which are verified by experiments. PMID:26367844

  19. Single-crystal nickel-based superalloys developed by numerical multi-criteria optimization techniques: design based on thermodynamic calculations and experimental validation

    NASA Astrophysics Data System (ADS)

    Rettig, Ralf; Ritter, Nils C.; Helmer, Harald E.; Neumeier, Steffen; Singer, Robert F.

    2015-04-01

    A method for finding the optimum alloy compositions considering a large number of property requirements and constraints by systematic exploration of large composition spaces is proposed. It is based on a numerical multi-criteria global optimization algorithm (multistart solver using Sequential Quadratic Programming), which delivers the exact optimum considering all constraints. The CALPHAD method is used to provide the thermodynamic equilibrium properties, and the creep strength of the alloys is predicted based on a qualitative numerical model considering the solid solution strengthening of the matrix by the elements Re, Mo and W and the optimum morphology and fraction of the γ‧-phase. The calculated alloy properties which are required as an input for the optimization algorithm are provided via very fast Kriging surrogate models. This greatly reduces the total calculation time of the optimization to the order of minutes on a personal computer. The capability of the multi-criteria optimization method developed was experimentally verified with two new single crystal superalloys. Their compositions were designed such that the content of expensive elements was reduced. One of the newly designed alloys, termed ERBO/13, is found to possess creep strength of only 14 K below CMSX-4 in the high-temperature/low-stress regime although it is a Re-free alloy.

  20. Cr(VI) transport via a supported ionic liquid membrane containing CYPHOS IL101 as carrier: system analysis and optimization through experimental design strategies.

    PubMed

    Rodríguez de San Miguel, Eduardo; Vital, Xóchitl; de Gyves, Josefina

    2014-05-30

    Chromium(VI) transport through a supported liquid membrane (SLM) system containing the commercial ionic liquid CYPHOS IL101 as carrier was studied. A reducing stripping phase was used as a mean to increase recovery and to simultaneously transform Cr(VI) into a less toxic residue for disposal or reuse. General functions which describe the time-depending evolution of the metal fractions in the cell compartments were defined and used in data evaluation. An experimental design strategy, using factorial and central-composite design matrices, was applied to assess the influence of the extractant, NaOH and citrate concentrations in the different phases, while a desirability function scheme allowed the synchronized optimization of depletion and recovery of the analyte. The mechanism for chromium permeation was analyzed and discussed to contribute to the understanding of the transfer process. The influence of metal concentration was evaluated as well. The presence of different interfering ions (Ca(2+), Al(3+), NO3(-), SO4(2-), and Cl(-)) at several Cr(VI): interfering ion ratios was studied through the use of a Plackett and Burman experimental design matrix. Under optimized conditions 90% of recovery was obtained from a feed solution containing 7mgL(-1) of Cr(VI) in 0.01moldm(-3) HCl medium after 5h of pertraction. PMID:24751491

  1. Optimization of the marinating conditions of cassava fish (Pseudotolithus sp.) fillet for Lanhouin production through application of Doehlert experimental design.

    PubMed

    Kindossi, Janvier Mêlégnonfan; Anihouvi, Victor Bienvenu; Vieira-Dalodé, Générose; Akissoé, Noël Houédougbé; Hounhouigan, Djidjoho Joseph

    2016-03-01

    Lanhouin is a traditional fermented salted fish made from the spontaneous and uncontrolled fermentation of whole salted cassava fish (Pseudotolithus senegalensis) mainly produced in the coastal regions of West Africa. The combined effects of NaCl, citric acid concentration, and marination time on the physicochemical and microbiological characteristics of the fish fillet used for Lanhouin production were studied using a Doehlert experimental design with the objective of preserving its quality and safety. The marination time has significant effects on total viable and lactic acid bacteria counts, and NaCl content of the marinated fish fillet while the pH was significantly affected by citric acid concentration and marination duration with high regression coefficient R (2) of 0.83. The experiment showed that the best conditions for marination process of fish fillet were salt ratio 10 g/100 g, acid citric concentration 2.5 g/100 g, and marination time 6 h. These optimum marinating conditions obtained present the best quality of marinated flesh fish leading to the safety of the final fermented product. This pretreatment is necessary in Lanhouin production processes to ensure its safety quality. PMID:27004115

  2. An experimental design approach to optimization of the liquid chromatographic separation conditions for the determination of metformin and glibenclamide in pharmaceutical formulation.

    PubMed

    Demiralay, Ebru Çubuk

    2012-06-01

    An optimization methodology is introduced for investigating the retention behavior and the separation factor of metformin, gliclazide (I.S.) and glibenclamide. This investigation has been focused on studying the influence of pH value of the mobile phase, concentration of acetonitrile and column temperature, which affect a complete separation of the chromatographic peaks of these compounds. The significant factors were optimized using full factorial design. Retention factor and separation factor were chosen as dependent variable. Optimum RP-LC chromatographic conditions for the separation of metformin, glibenclamide and gliclazide were obtained using X Terra column (150 mm × 4.6 mm I.D., 5 µm). The results show that the percentage of acetonitrile are the most important to investigate and sspH of the mobile phase and column temperature do not significantly affect the experimental results. The procedure was validated for linearity, accuracy, precision and recovery. Quantitation was accomplished using internal standard method. PMID:24061246

  3. Aerodynamic design using numerical optimization

    NASA Technical Reports Server (NTRS)

    Murman, E. M.; Chapman, G. T.

    1983-01-01

    The procedure of using numerical optimization methods coupled with computational fluid dynamic (CFD) codes for the development of an aerodynamic design is examined. Several approaches that replace wind tunnel tests, develop pressure distributions and derive designs, or fulfill preset design criteria are presented. The method of Aerodynamic Design by Numerical Optimization (ADNO) is described and illustrated with examples.

  4. Structural Optimization in automotive design

    NASA Technical Reports Server (NTRS)

    Bennett, J. A.; Botkin, M. E.

    1984-01-01

    Although mathematical structural optimization has been an active research area for twenty years, there has been relatively little penetration into the design process. Experience indicates that often this is due to the traditional layout-analysis design process. In many cases, optimization efforts have been outgrowths of analysis groups which are themselves appendages to the traditional design process. As a result, optimization is often introduced into the design process too late to have a significant effect because many potential design variables have already been fixed. A series of examples are given to indicate how structural optimization has been effectively integrated into the design process.

  5. Optimization of the combined ultrasonic assisted/adsorption method for the removal of malachite green by gold nanoparticles loaded on activated carbon: Experimental design

    NASA Astrophysics Data System (ADS)

    Roosta, M.; Ghaedi, M.; Shokri, N.; Daneshfar, A.; Sahraei, R.; Asghari, A.

    2014-01-01

    The present study was aimed to experimental design optimization applied to removal of malachite green (MG) from aqueous solution by ultrasound-assisted removal onto the gold nanoparticles loaded on activated carbon (Au-NP-AC). This nanomaterial was characterized using different techniques such as FESEM, TEM, BET, and UV-vis measurements. The effects of variables such as pH, initial dye concentration, adsorbent dosage (g), temperature and sonication time on MG removal were studied using central composite design (CCD) and the optimum experimental conditions were found with desirability function (DF) combined response surface methodology (RSM). Fitting the experimental equilibrium data to various isotherm models such as Langmuir, Freundlich, Tempkin and Dubinin-Radushkevich models show the suitability and applicability of the Langmuir model. Kinetic models such as pseudo -first order, pseudo-second order, Elovich and intraparticle diffusion models applicability was tested for experimental data and the second-order equation and intraparticle diffusion models control the kinetic of the adsorption process. The small amount of proposed adsorbent (0.015 g) is applicable for successful removal of MG (RE > 99%) in short time (4.4 min) with high adsorption capacity (140-172 mg g-1).

  6. Removal of Mefenamic acid from aqueous solutions by oxidative process: Optimization through experimental design and HPLC/UV analysis.

    PubMed

    Colombo, Renata; Ferreira, Tanare C R; Ferreira, Renato A; Lanza, Marcos R V

    2016-02-01

    Mefenamic acid (MEF) is a non-steroidal anti-inflammatory drug indicated for relief of mild to moderate pain, and for the treatment of primary dysmenorrhea. The presence of MEF in raw and sewage waters has been detected worldwide at concentrations exceeding the predicted no-effect concentration. In this study, using experimental designs, different oxidative processes (H2O2, H2O2/UV, fenton and Photo-fenton) were simultaneously evaluated for MEF degradation efficiency. The influence and interaction effects of the most important variables in the oxidative process (concentration and addition mode of hydrogen peroxide, concentration and type of catalyst, pH, reaction period and presence/absence of light) were investigated. The parameters were determined based on the maximum efficiency to save time and minimize the consumption of reagents. According to the results, the photo-Fenton process is the best procedure to remove the drug from water. A reaction mixture containing 1.005 mmol L(-1) of ferrioxalate and 17.5 mmol L(-1) of hydrogen peroxide, added at the initial reaction period, pH of 6.1 and 60 min of degradation indicated the most efficient degradation, promoting 95% of MEF removal. The development and validation of a rapid and efficient qualitative and quantitative HPLC/UV methodology for detecting this pollutant in aqueous solution is also reported. The method can be applied in water quality control that is generated and/or treated in municipal or industrial wastewater treatment plants. PMID:26686073

  7. Quantitative and qualitative optimization of allergen extraction from peanut and selected tree nuts. Part 2. Optimization of buffer and ionic strength using a full factorial experimental design.

    PubMed

    L'Hocine, Lamia; Pitre, Mélanie

    2016-03-01

    A full factorial design was used to assess the single and interactive effects of three non-denaturing aqueous (phosphate, borate, and carbonate) buffers at various ionic strengths (I) on allergen extractability from and immunoglobulin E (IgE) immunoreactivity of peanut, almond, hazelnut, and pistachio. The results indicated that the type and ionic strength of the buffer had different effects on protein recovery from the nuts under study. Substantial differences in protein profiles, abundance, and IgE-binding intensity with different combinations of pH and ionic strength were found. A significant interaction between pH and ionic strength was observed for pistachio and almond. The optimal buffer system conditions, which maximized the IgE-binding efficiency of allergens and provided satisfactory to superior protein recovery yield and profiles, were carbonate buffer at an ionic strength of I=0.075 for peanut, carbonate buffer at I=0.15 for almond, phosphate buffer at I=0.5 for hazelnut, and borate at I=0.15 for pistachio. The buffer type and its ionic strength could be manipulated to achieve the selective solubility of desired allergens. PMID:26471623

  8. Propeller design by numerical optimization

    NASA Technical Reports Server (NTRS)

    Mendoza, J. P.

    1977-01-01

    A computer program designed to optimize propeller characteristics was developed by combining two main programs: the first is the optimization program based on the gradient algorithm; the second is based on a propeller blade element theory and uses an aerodynamics subprogram to approximate the lift and drag characteristics of the NACA 16-series airfoil section. To evaluate the propeller program alone (with its aerodynamics subprogram), propeller characteristics were computed and compared to those from wind tunnel investigations conducted on three different NACA propellers. Although the thrust and power coefficients which were computed using the blade element theory were generally higher than the experimental results for two of the three propellers, the corresponding efficiencies showed good agreement for all three propellers. The propeller optimization program was then used to study the NACA 4-(5)(08)-03 propeller at various Mach numbers from 0.175 to 0.60. Improvements in propeller efficiency and thrust were obtained through the use of the propeller optimization program.

  9. Phenazine-1-carboxylic acid production in a chromosomally non-scar triple-deleted mutant Pseudomonas aeruginosa using statistical experimental designs to optimize yield.

    PubMed

    Du, Xilin; Li, Yaqian; Zhou, Wanping; Zhou, Quan; Liu, Haiming; Xu, Yuquan

    2013-09-01

    We constructed a non-scar triple-deleted mutant Pseudomonas aeruginosa to improve phenazine-1-carboxylic acid (PCA) yield and then optimized the culture conditions for PCA production. Using a non-scar deletion strategy, the 5'-untranslated region of the phz1 gene cluster and two genes, phzM and phzS, were knocked out of the P. aeruginosa strain M18 genome. The potential ability for high-yield PCA production in this triple-deleted mutant M18MSU1 was successfully realized by using statistical experimental designs. A 2(5-1) fractional factorial design was used to show that the three culture components of soybean meal, corn steep liquor and ethanol had the most significant effect on PCA production. Using a central composite design, the concentration of the three components was optimized. The maximum PCA production was predicted to be 4,725.1 mg/L. With the optimal medium containing soybean meal 74.25 g/L, corn steep liquor 13.01 g/L and ethanol 21.84 ml/L, a PCA production of 4,771.2 mg/L was obtained in the validation experiments, which was nearly twofold of that before optimization and tenfold of that in the wild-type strain. This non-scar triple-deleted mutant M18MSU1 may be a suitable strain for industrial production of this biologically synthesized fungicide due to its high PCA production, presumed safety, thermal adaptability and cost-effectiveness. PMID:23636695

  10. Integrated controls design optimization

    DOEpatents

    Lou, Xinsheng; Neuschaefer, Carl H.

    2015-09-01

    A control system (207) for optimizing a chemical looping process of a power plant includes an optimizer (420), an income algorithm (230) and a cost algorithm (225) and a chemical looping process models. The process models are used to predict the process outputs from process input variables. Some of the process in puts and output variables are related to the income of the plant; and some others are related to the cost of the plant operations. The income algorithm (230) provides an income input to the optimizer (420) based on a plurality of input parameters (215) of the power plant. The cost algorithm (225) provides a cost input to the optimizer (420) based on a plurality of output parameters (220) of the power plant. The optimizer (420) determines an optimized operating parameter solution based on at least one of the income input and the cost input, and supplies the optimized operating parameter solution to the power plant.

  11. Design Optimization Toolkit: Users' Manual

    SciTech Connect

    Aguilo Valentin, Miguel Alejandro

    2014-07-01

    The Design Optimization Toolkit (DOTk) is a stand-alone C++ software package intended to solve complex design optimization problems. DOTk software package provides a range of solution methods that are suited for gradient/nongradient-based optimization, large scale constrained optimization, and topology optimization. DOTk was design to have a flexible user interface to allow easy access to DOTk solution methods from external engineering software packages. This inherent flexibility makes DOTk barely intrusive to other engineering software packages. As part of this inherent flexibility, DOTk software package provides an easy-to-use MATLAB interface that enables users to call DOTk solution methods directly from the MATLAB command window.

  12. Optimization of phase feeding of starter, grower, and finisher diets for male broilers by mixture experimental design: forty-eight-day production period.

    PubMed

    Roush, W B; Boykin, D; Branton, S L

    2004-08-01

    A mixture experiment, a variant of response surface methodology, was designed to determine the proportion of time to feed broiler starter (23% protein), grower (20% protein), and finisher (18% protein) diets to optimize production and processing variables based on a total production time of 48 d. Mixture designs are useful for proportion problems where the components of the experiment (i.e., length of time the diets were fed) add up to a unity (48 d). The experiment was conducted with day-old male Ross x Ross broiler chicks. The birds were placed 50 birds per pen in each of 60 pens. The experimental design was a 10-point augmented simplex-centroid (ASC) design with 6 replicates of each point. Each design point represented the portion(s) of the 48 d that each of the diets was fed. Formulation of the diets was based on NRC standards. At 49 d, each pen of birds was evaluated for production data including BW, feed conversion, and cost of feed consumed. Then, 6 birds were randomly selected from each pen for processing data. Processing variables included live weight, hot carcass weight, dressing percentage, fat pad percentage, and breast yield (pectoralis major and pectoralis minor weights). Production and processing data were fit to simplex regression models. Model terms determined not to be significant (P > 0.05) were removed. The models were found to be statistically adequate for analysis of the response surfaces. A compromise solution was calculated based on optimal constraints designated for the production and processing data. The results indicated that broilers fed a starter and finisher diet for 30 and 18 d, respectively, would meet the production and processing constraints. Trace plots showed that the production and processing variables were not very sensitive to the grower diet. PMID:15339000

  13. Optimization and validation of a HPLC method for simultaneous determination of aflatoxin B1, B2, G1, G2, ochratoxin A and zearalenone using an experimental design.

    PubMed

    Rahmani, Anosheh; Selamat, Jinap; Soleimany, Farhang

    2011-01-01

    A reversed-phase HPLC optimization strategy is presented for investigating the separation and retention behavior of aflatoxin B1, B2, G1, G2, ochratoxin A and zearalenone, simultaneously. A fractional factorial design (FFD) was used to screen the significance effect of seven independent variables on chromatographic responses. The independent variables used were: (X1) column oven temperature (20-40°C), (X2) flow rate (0.8-1.2 ml/min), (X3) acid concentration in aqueous phase (0-2%), (X4) organic solvent percentage at the beginning (40-50%), and (X5) at the end (50-60%) of the gradient mobile phase, as well as (X6) ratio of methanol/acetonitrile at the beginning (1-4) and (X7) at the end (0-1) of gradient mobile phase. Responses of chromatographic analysis were resolution of mycotoxin peaks and HPLC run time. A central composite design (CCD) using response surface methodology (RSM) was then carried out for optimization of the most significant factors by multiple regression models for response variables. The proposed optimal method using 40°C oven temperature, 1 ml/min flow rate, 0.1% acetic acid concentration in aqueous phase, 41% organic phase (beginning), 60% organic phase (end), 1.92 ratio of methanol to acetonitrile (beginning) and 0.2 ratio (end) for X1-X7, respectively, showed good prediction ability between the experimental data and predictive values throughout the studied parameter space. Finally, the optimized method was validated by measuring the linearity, sensitivity, accuracy and precision parameters, and has been applied successfully to the analysis of spiked cereal samples. PMID:21598138

  14. D-OPTIMAL EXPERIMENTAL DESIGNS TO TEST FOR DEPARTURE FROM ADDITIVITY IN A FIXED-RATIO RAY MIXTURE.

    EPA Science Inventory

    Risk assessors are becoming increasingly aware of the importance of assessing interactions between chemicals in a mixture. Most traditional designs for evaluating interactions are prohibitive when the number of chemicals in the mixture is large. However, evaluation of interacti...

  15. A Case Study on the Application of a Structured Experimental Method for Optimal Parameter Design of a Complex Control System

    NASA Technical Reports Server (NTRS)

    Torres-Pomales, Wilfredo

    2015-01-01

    This report documents a case study on the application of Reliability Engineering techniques to achieve an optimal balance between performance and robustness by tuning the functional parameters of a complex non-linear control system. For complex systems with intricate and non-linear patterns of interaction between system components, analytical derivation of a mathematical model of system performance and robustness in terms of functional parameters may not be feasible or cost-effective. The demonstrated approach is simple, structured, effective, repeatable, and cost and time efficient. This general approach is suitable for a wide range of systems.

  16. [Optimization of the composition of a multicomponent drug form by mathematical investigational designing and experimental testing (author's transl)].

    PubMed

    Jontschev, C; Welikowa, E

    1981-01-01

    Experiments have been performed to optimize the composition of tablets with 250 mg of noramidopyrine methanesulphonate sodium, 250 mg of aminophenazone and 100 mg of caffeine + sodium benzoate. The effects of the adjuvants Micricell, Aerosil 200, Kollidon CE 5050 and wheat starch on different tablet characteristics (disintegration, mechanical strength, friability and rheological properties) were studied. The results obtained permitted to prepare tablets with good properties: disintegration time, up to 194 s; mechanical strength, greater than 12 kg; friability, less than 0.250%. PMID:7220583

  17. Experimental and Quasi-Experimental Design.

    ERIC Educational Resources Information Center

    Cottrell, Edward B.

    With an emphasis on the problems of control of extraneous variables and threats to internal and external validity, the arrangement or design of experiments is discussed. The purpose of experimentation in an educational institution, and the principles governing true experimentation (randomization, replication, and control) are presented, as are…

  18. Grafting of copolymer styrene maleic anhydride on poly(ethylene terephthalate) film by chemical reaction and by plasma method: Optimization of the grafting reaction using experimental design

    NASA Astrophysics Data System (ADS)

    Bigan, Muriel; Bigot, Julien; Mutel, Brigitte; Coqueret, Xavier

    2008-02-01

    This work deals with the chemical grafting of a styrene maleic anhydride copolymer on the surface of a previously hydrolyzed polyethylene terephthalate (PET) film 12 μm thick via covalent bond. Two different ways are studied. The first one involves an activation of the hydrolyzed PET by the triethylamine before the grafting step. In the second one, the copolymer reacts with the 4-dimethylaminopyridine in order to form maleinyl pyridinium salt which reacts with alcohol function of the hydrolyzed PET. Characterization and quantification of the grafting are performed by Fourier transform infrared spectroscopy. Factorial experiment designs are used to optimize the process and to estimate experimental parameters effects. The opportunity to associate the chemical process to a cold remote nitrogen plasma one is also examined.

  19. Optimization of experimental parameters based on the Taguchi robust design for the formation of zinc oxide nanocrystals by solvothermal method

    SciTech Connect

    Yiamsawas, Doungporn; Boonpavanitchakul, Kanittha; Kangwansupamonkon, Wiyong

    2011-05-15

    Research highlights: {yields} Taguchi robust design can be applied to study ZnO nanocrystal growth. {yields} Spherical-like and rod-like shaped of ZnO nanocrystals can be obtained from solvothermal method. {yields} [NaOH]/[Zn{sup 2+}] ratio plays the most important factor on the aspect ratio of prepared ZnO. -- Abstract: Zinc oxide (ZnO) nanoparticles and nanorods were successfully synthesized by a solvothermal process. Taguchi robust design was applied to study the factors which result in stronger ZnO nanocrystal growth. The factors which have been studied are molar concentration ratio of sodium hydroxide and zinc acetate, amount of polymer templates and molecular weight of polymer templates. Transmission electron microscopy and X-ray diffraction technique were used to analyze the experiment results. The results show that the concentration ratio of sodium hydroxide and zinc acetate ratio has the greatest effect on ZnO nanocrystal growth.

  20. Experimental design for the formulation and optimization of novel cross-linked oilispheres developed for in vitro site-specific release of Mentha piperita oil.

    PubMed

    Sibanda, Wilbert; Pillay, Viness; Danckwerts, Michael P; Viljoen, Alvaro M; van Vuuren, Sandy; Khan, Riaz A

    2004-03-12

    A Plackett-Burman design was employed to develop and optimize a novel crosslinked calcium-aluminum-alginate-pectinate oilisphere complex as a potential system for the in vitro site-specific release of Mentha piperita, an essential oil used for the treatment of irritable bowel syndrome. The physicochemical and textural properties (dependent variables) of this complex were found to be highly sensitive to changes in the concentration of the polymers (0%-1.5% wt/vol), crosslinkers (0%-4% wt/vol), and crosslinking reaction times (0.5-6 hours) (independent variables). Particle size analysis indicated both unimodal and bimodal populations with the highest frequency of 2 mm oilispheres. Oil encapsulation ranged from 6 to 35 mg/100 mg oilispheres. Gravimetric changes of the crosslinked matrix indicated significant ion sequestration and loss in an exponential manner, while matrix erosion followed Higuchi's cube root law. Among the various measured responses, the total fracture energy was the most suitable optimization objective (R2 = 0.88, Durbin-Watson Index = 1.21%, Coefficient of Variation (CV) = 33.21%). The Lagrangian technique produced no significant differences (P > .05) between the experimental and predicted total fracture energy values (0.0150 vs 0.0107 J). Artificial Neural Networks, as an alternative predictive tool of the total fracture energy, was highly accurate (final mean square error of optimal network epoch approximately 0.02). Fused-coated optimized oilispheres produced a 4-hour lag phase followed by zero-order kinetics (n > 0.99), whereby analysis of release data indicated that diffusion (Fickian constant k1 = 0.74 vs relaxation constant k2 = 0.02) was the predominant release mechanism. PMID:15198539

  1. Optimization of ultrasound-assisted dispersive solid-phase microextraction based on nanoparticles followed by spectrophotometry for the simultaneous determination of dyes using experimental design.

    PubMed

    Asfaram, Arash; Ghaedi, Mehrorang; Goudarzi, Alireza

    2016-09-01

    A simple, low cost and ultrasensitive method for the simultaneous preconcentration and determination of trace amount of auramine-O and malachite green in aqueous media following accumulation on novel and lower toxicity nanomaterials by ultrasound-assisted dispersive solid phase micro-extraction (UA-DSPME) procedure combined with spectrophotometric has been described. The Mn doped ZnS nanoparticles loaded on activated carbon were characterized by Field emission scanning electron microscopy (FE-SEM), particle size distribution, X-ray diffraction (XRD), Fourier transform infrared spectroscopy (FT-IR) analyses and subsequently were used as green and efficient material for dyes accumulation. Contribution of experimental variables such as ultrasonic time, ultrasonic temperature, adsorbent mass, vortex time, ionic strength, pH and elution volume were optimized through experimental design, and while the preconcentrated analytes were efficiently eluted by acetone. Preliminary Plackett-Burman design was applied for selection of most significant factors and giving useful information about their main and interaction part of significant variables like ultrasonic time, adsorbent mass, elution volume and pH were obtained by central composite design combined with response surface analysis and optimum experimental conditions was set at pH of 8.0, 1.2mg of adsorbent, 150μL eluent and 3.7min sonication. Under optimized conditions, the average recoveries (five replicates) for two dyes (spiked at 500.0ngmL(-1)) changes in the range of 92.80-97.70% with acceptable RSD% less than 4.0% over a linear range of 3.0-5000.0ngmL(-1) for the AO and MG in water samples with regression coefficients (R(2)) of 0.9975 and 0.9977, respectively. Acceptable limits of detection of 0.91 and 0.61ngmL(-1) for AO and MG, respectively and high accuracy and repeatability are unique advantages of present method to improve the figures of merit for their accurate determination at trace level in complicated

  2. Optimization of low-frequency low-intensity ultrasound-mediated microvessel disruption on prostate cancer xenografts in nude mice using an orthogonal experimental design

    PubMed Central

    YANG, YU; BAI, WENKUN; CHEN, YINI; LIN, YANDUAN; HU, BING

    2015-01-01

    The present study aimed to provide a complete exploration of the effect of sound intensity, frequency, duty cycle, microbubble volume and irradiation time on low-frequency low-intensity ultrasound (US)-mediated microvessel disruption, and to identify an optimal combination of the five factors that maximize the blockage effect. An orthogonal experimental design approach was used. Enhanced US imaging and acoustic quantification were performed to assess tumor blood perfusion. In the confirmatory test, in addition to acoustic quantification, the specimens of the tumor were stained with hematoxylin and eosin and observed using light microscopy. The results revealed that sound intensity, frequency, duty cycle, microbubble volume and irradiation time had a significant effect on the average peak intensity (API). The extent of the impact of the variables on the API was in the following order: Sound intensity; frequency; duty cycle; microbubble volume; and irradiation time. The optimum conditions were found to be as follows: Sound intensity, 1.00 W/cm2; frequency, 20 Hz; duty cycle, 40%; microbubble volume, 0.20 ml; and irradiation time, 3 min. In the confirmatory test, the API was 19.97±2.66 immediately subsequent to treatment, and histological examination revealed signs of tumor blood vessel injury in the optimum parameter combination group. In conclusion, the Taguchi L18 (3)6 orthogonal array design was successfully applied for determining the optimal parameter combination of API following treatment. Under the optimum orthogonal design condition, a minimum API of 19.97±2.66 subsequent to low-frequency and low-intensity mediated blood perfusion blockage was obtained. PMID:26722279

  3. Nanoparticle-Laden Contact Lens for Controlled Ocular Delivery of Prednisolone: Formulation Optimization Using Statistical Experimental Design.

    PubMed

    ElShaer, Amr; Mustafa, Shelan; Kasar, Mohamad; Thapa, Sapana; Ghatora, Baljit; Alany, Raid G

    2016-01-01

    Human eye is one of the most accessible organs in the body, nonetheless, its physiology and associated precorneal factors such as nasolacrimal drainage, blinking, tear film, tear turnover, and induced lacrimation has significantly decreased the residence time of any foreign substances including pharmaceutical dosage forms. Soft contact lenses are promising delivery devices that can sustain the drug release and prolong residence time by acting as a geometric barrier to drug diffusion to tear fluid. This study investigates experimental parameters such as composition of polymer mixtures, stabilizer and the amount of active pharmaceutical ingredient on the preparation of a polymeric drug delivery system for the topical ocular administration of Prednisolone. To achieve this goal, prednisolone-loaded poly (lactic-co-glycolic acid) (PLGA) nanoparticles were prepared by single emulsion solvent evaporation method. Prednisolone was quantified using a validated high performance liquid chromatography (HPLC) method. Nanoparticle size was mostly affected by the amount of co-polymer (PLGA) used whereas drug load was mostly affected by amount of prednisolone (API) used. Longer homogenization time along with higher amount of API yielded the smallest size nanoparticles. The nanoparticles prepared had an average particle size of 347.1 ± 11.9 nm with a polydispersity index of 0.081. The nanoparticles were then incorporated in the contact lens mixture before preparing them. Clear and transparent contact lenses were successfully prepared. When the nanoparticle (NP)-loaded contact lenses were compared with control contact lenses (unloaded NP contact lenses), a decrease in hydration by 2% (31.2% ± 1.25% hydration for the 0.2 g loaded NP contact lenses) and light transmission by 8% (unloaded NP contact lenses 94.5% NP 0.2 g incorporated contact lenses 86.23%). The wettability of the contact lenses remained within the desired value (<90 °C) even upon incorporation of the NP. NP alone and

  4. Nanoparticle-Laden Contact Lens for Controlled Ocular Delivery of Prednisolone: Formulation Optimization Using Statistical Experimental Design

    PubMed Central

    ElShaer, Amr; Mustafa, Shelan; Kasar, Mohamad; Thapa, Sapana; Ghatora, Baljit; Alany, Raid G.

    2016-01-01

    Human eye is one of the most accessible organs in the body, nonetheless, its physiology and associated precorneal factors such as nasolacrimal drainage, blinking, tear film, tear turnover, and induced lacrimation has significantly decreased the residence time of any foreign substances including pharmaceutical dosage forms. Soft contact lenses are promising delivery devices that can sustain the drug release and prolong residence time by acting as a geometric barrier to drug diffusion to tear fluid. This study investigates experimental parameters such as composition of polymer mixtures, stabilizer and the amount of active pharmaceutical ingredient on the preparation of a polymeric drug delivery system for the topical ocular administration of Prednisolone. To achieve this goal, prednisolone-loaded poly (lactic-co-glycolic acid) (PLGA) nanoparticles were prepared by single emulsion solvent evaporation method. Prednisolone was quantified using a validated high performance liquid chromatography (HPLC) method. Nanoparticle size was mostly affected by the amount of co-polymer (PLGA) used whereas drug load was mostly affected by amount of prednisolone (API) used. Longer homogenization time along with higher amount of API yielded the smallest size nanoparticles. The nanoparticles prepared had an average particle size of 347.1 ± 11.9 nm with a polydispersity index of 0.081. The nanoparticles were then incorporated in the contact lens mixture before preparing them. Clear and transparent contact lenses were successfully prepared. When the nanoparticle (NP)-loaded contact lenses were compared with control contact lenses (unloaded NP contact lenses), a decrease in hydration by 2% (31.2% ± 1.25% hydration for the 0.2 g loaded NP contact lenses) and light transmission by 8% (unloaded NP contact lenses 94.5% NP 0.2 g incorporated contact lenses 86.23%). The wettability of the contact lenses remained within the desired value (<90 °C) even upon incorporation of the NP. NP alone and

  5. Optimization by experimental design and artificial neural networks of the ion-interaction reversed-phase liquid chromatographic separation of twenty cosmetic preservatives.

    PubMed

    Marengo, E; Gianotti, V; Angioi, S; Gennaro, M C

    2004-03-12

    Particular attention are recently receiving antimicrobial agents added as preservatives in hygiene and cosmetics commercial products, since some of them are suspected to be harmful to the human health. The preservatives used belong to different classes of chemical species and are generally used in their mixtures. Multi-component methods able to simultaneously determinate species with different chemical structure are therefore highly required in quality control analysis. This paper presents an ion interaction RP-HPLC method for the simultaneous separation of the 20 typical antimicrobial agents most used in cosmetics and hygiene products, that are: benzoic acid, salicylic acid, 4-hydroxybenzoic acid, methyl-, ethyl-, propyl-, butyl-, benzyl-benzoate, methyl-, ethyl-, propyl-, butyl-, benzyl-paraben, o-phenyl-phenol, 4-chloro-m-cresol, triclocarban, dehydroacetic acid, bronopol, sodium pyrithione and chlorhexidine. For the development of the method and the optimization of the chromatographic conditions, an experimental design was planned and models were built by the use of artificial neural network to correlate the retention time of each analyte to the variables and their interactions. The neuronal models developed showed good predictive ability and were used, by a grid search algorithm, to optimize the chromatographic conditions for the separation of the mixture. PMID:15032350

  6. Optimal Network-Topology Design

    NASA Technical Reports Server (NTRS)

    Li, Victor O. K.; Yuen, Joseph H.; Hou, Ting-Chao; Lam, Yuen Fung

    1987-01-01

    Candidate network designs tested for acceptability and cost. Optimal Network Topology Design computer program developed as part of study on topology design and analysis of performance of Space Station Information System (SSIS) network. Uses efficient algorithm to generate candidate network designs consisting of subsets of set of all network components, in increasing order of total costs and checks each design to see whether it forms acceptable network. Technique gives true cost-optimal network and particularly useful when network has many constraints and not too many components. Program written in PASCAL.

  7. Design optimization of transonic airfoils

    NASA Technical Reports Server (NTRS)

    Joh, C.-Y.; Grossman, B.; Haftka, R. T.

    1991-01-01

    Numerical optimization procedures were considered for the design of airfoils in transonic flow based on the transonic small disturbance (TSD) and Euler equations. A sequential approximation optimization technique was implemented with an accurate approximation of the wave drag based on the Nixon's coordinate straining approach. A modification of the Euler surface boundary conditions was implemented in order to efficiently compute design sensitivities without remeshing the grid. Two effective design procedures producing converged designs in approximately 10 global iterations were developed: interchanging the role of the objective function and constraint and the direct lift maximization with move limits which were fixed absolute values of the design variables.

  8. A Surrogate Approach to the Experimental Optimization of Multielement Airfoils

    NASA Technical Reports Server (NTRS)

    Otto, John C.; Landman, Drew; Patera, Anthony T.

    1996-01-01

    The incorporation of experimental test data into the optimization process is accomplished through the use of Bayesian-validated surrogates. In the surrogate approach, a surrogate for the experiment (e.g., a response surface) serves in the optimization process. The validation step of the framework provides a qualitative assessment of the surrogate quality, and bounds the surrogate-for-experiment error on designs "near" surrogate-predicted optimal designs. The utility of the framework is demonstrated through its application to the experimental selection of the trailing edge ap position to achieve a design lift coefficient for a three-element airfoil.

  9. Designing an Experimental "Accident"

    ERIC Educational Resources Information Center

    Picker, Lester

    1974-01-01

    Describes an experimental "accident" that resulted in much student learning, seeks help in the identification of nematodes, and suggests biology teachers introduce similar accidents into their teaching to stimulate student interest. (PEB)

  10. Habitat Design Optimization and Analysis

    NASA Technical Reports Server (NTRS)

    SanSoucie, Michael P.; Hull, Patrick V.; Tinker, Michael L.

    2006-01-01

    Long-duration surface missions to the Moon and Mars will require habitats for the astronauts. The materials chosen for the habitat walls play a direct role in the protection against the harsh environments found on the surface. Choosing the best materials, their configuration, and the amount required is extremely difficult due to the immense size of the design region. Advanced optimization techniques are necessary for habitat wall design. Standard optimization techniques are not suitable for problems with such large search spaces; therefore, a habitat design optimization tool utilizing genetic algorithms has been developed. Genetic algorithms use a "survival of the fittest" philosophy, where the most fit individuals are more likely to survive and reproduce. This habitat design optimization tool is a multi-objective formulation of structural analysis, heat loss, radiation protection, and meteoroid protection. This paper presents the research and development of this tool.

  11. Optimization of ejector design and operation

    NASA Astrophysics Data System (ADS)

    Kuzmenko, Konstantin; Yurchenko, Nina; Vynogradskyy, Pavlo; Paramonov, Yuriy

    2016-03-01

    The investigation aims at optimization of gas ejector operation. The goal consists in the improvement of the inflator design so that to enable 50 liters of gas inflation within ~30 milliseconds. For that, an experimental facility was developed and fabricated together with the measurement system to study pressure patterns in the inflator path.

  12. Computational design and optimization of energy materials

    NASA Astrophysics Data System (ADS)

    Chan, Maria

    The use of density functional theory (DFT) to understand and improve energy materials for diverse applications - including energy storage, thermal management, catalysis, and photovoltaics - is widespread. The further step of using high throughput DFT calculations to design materials and has led to an acceleration in materials discovery and development. Due to various limitations in DFT, including accuracy and computational cost, however, it is important to leverage effective models and, in some cases, experimental information to aid the design process. In this talk, I will discuss efforts in design and optimization of energy materials using a combination of effective models, DFT, machine learning, and experimental information.

  13. Experimental design of a waste glass study

    SciTech Connect

    Piepel, G.F.; Redgate, P.E.; Hrma, P.

    1995-04-01

    A Composition Variation Study (CVS) is being performed to support a future high-level waste glass plant at Hanford. A total of 147 glasses, covering a broad region of compositions melting at approximately 1150{degrees}C, were tested in five statistically designed experimental phases. This paper focuses on the goals, strategies, and techniques used in designing the five phases. The overall strategy was to investigate glass compositions on the boundary and interior of an experimental region defined by single- component, multiple-component, and property constraints. Statistical optimal experimental design techniques were used to cover various subregions of the experimental region in each phase. Empirical mixture models for glass properties (as functions of glass composition) from previous phases wee used in designing subsequent CVS phases.

  14. Design optimization of space structures

    NASA Technical Reports Server (NTRS)

    Felippa, Carlos

    1991-01-01

    The topology-shape-size optimization of space structures is investigated through Kikuchi's homogenization method. The method starts from a 'design domain block,' which is a region of space into which the structure is to materialize. This domain is initially filled with a finite element mesh, typically regular. Force and displacement boundary conditions corresponding to applied loads and supports are applied at specific points in the domain. An optimal structure is to be 'carved out' of the design under two conditions: (1) a cost function is to be minimized, and (2) equality or inequality constraints are to be satisfied. The 'carving' process is accomplished by letting microstructure holes develop and grow in elements during the optimization process. These holes have a rectangular shape in two dimensions and a cubical shape in three dimensions, and may also rotate with respect to the reference axes. The properties of the perforated element are obtained through an homogenization procedure. Once a hole reaches the volume of the element, that element effectively disappears. The project has two phases. In the first phase the method was implemented as the combination of two computer programs: a finite element module, and an optimization driver. In the second part, focus is on the application of this technique to planetary structures. The finite element part of the method was programmed for the two-dimensional case using four-node quadrilateral elements to cover the design domain. An element homogenization technique different from that of Kikuchi and coworkers was implemented. The optimization driver is based on an augmented Lagrangian optimizer, with the volume constraint treated as a Courant penalty function. The optimizer has to be especially tuned to this type of optimization because the number of design variables can reach into the thousands. The driver is presently under development.

  15. Design of optimal systolic arrays

    SciTech Connect

    Li, G.J.; Wah, B.W.

    1985-01-01

    Conventional design of systolic arrays is based on the mapping of an algorithm onto an interconnection of processing elements in a VLSI chip. This mapping is done in an ad hoc manner, and the resulting configuration usually represents a feasible but suboptimal design. In this paper, systolic arrays are characterized by three classes of parameters: the velocities of data flows, the spatial distributions of data, and the periods of computation. By relating these parameters in constraint equations that govern the correctness of the design, the design is formulated into an optimization problem. The size of the search space is a polynomial of the problem size, and a methodology to systematically search and reduce this space and to obtain the optimal design is proposed. Some examples of applying the method, including matrix multiplication, finite impulse response filtering, deconvolution, and triangular-matrix inversion, are given. 30 references.

  16. Spin bearing retainer design optimization

    NASA Technical Reports Server (NTRS)

    Boesiger, Edward A.; Warner, Mark H.

    1991-01-01

    The dynamics behavior of spin bearings for momentum wheels (control-moment gyroscope, reaction wheel assembly) is critical to satellite stability and life. Repeated bearing retainer instabilities hasten lubricant deterioration and can lead to premature bearing failure and/or unacceptable vibration. These instabilities are typically distinguished by increases in torque, temperature, audible noise, and vibration induced by increases into the bearing cartridge. Ball retainer design can be optimized to minimize these occurrences. A retainer was designed using a previously successful smaller retainer as an example. Analytical methods were then employed to predict its behavior and optimize its configuration.

  17. Optimization process in helicopter design

    NASA Technical Reports Server (NTRS)

    Logan, A. H.; Banerjee, D.

    1984-01-01

    In optimizing a helicopter configuration, Hughes Helicopters uses a program called Computer Aided Sizing of Helicopters (CASH), written and updated over the past ten years, and used as an important part of the preliminary design process of the AH-64. First, measures of effectiveness must be supplied to define the mission characteristics of the helicopter to be designed. Then CASH allows the designer to rapidly and automatically develop the basic size of the helicopter (or other rotorcraft) for the given mission. This enables the designer and management to assess the various tradeoffs and to quickly determine the optimum configuration.

  18. Global Design Optimization for Fluid Machinery Applications

    NASA Technical Reports Server (NTRS)

    Shyy, Wei; Papila, Nilay; Tucker, Kevin; Vaidyanathan, Raj; Griffin, Lisa

    2000-01-01

    Recent experiences in utilizing the global optimization methodology, based on polynomial and neural network techniques for fluid machinery design are summarized. Global optimization methods can utilize the information collected from various sources and by different tools. These methods offer multi-criterion optimization, handle the existence of multiple design points and trade-offs via insight into the entire design space can easily perform tasks in parallel, and are often effective in filtering the noise intrinsic to numerical and experimental data. Another advantage is that these methods do not need to calculate the sensitivity of each design variable locally. However, a successful application of the global optimization method needs to address issues related to data requirements with an increase in the number of design variables and methods for predicting the model performance. Examples of applications selected from rocket propulsion components including a supersonic turbine and an injector element and a turbulent flow diffuser are used to illustrate the usefulness of the global optimization method.

  19. Experimental Investigation of Optimal Adhesion of Mushroomlike Elastomer Microfibrillar Adhesives.

    PubMed

    Marvi, Hamidreza; Song, Sukho; Sitti, Metin

    2015-09-22

    Optimal fiber designs for the maximal pull-off force have been indispensable for increasing the attachment performance of recently introduced gecko-inspired reversible micro/nanofibrillar adhesives. There are several theoretical studies on such optimal designs; however, due to the lack of three-dimensional (3D) fabrication techniques that can fabricate such optimal designs in 3D, there have not been many experimental investigations on this challenge. In this study, we benefitted from recent advances in two-photon lithography techniques to fabricate mushroomlike polyurethane elastomer fibers with different aspect ratios of tip to stalk diameter (β) and tip wedge angles (θ) to investigate the effect of these two parameters on the pull-off force. We found similar trends to those predicted theoretically. We found that β has an impact on the slope of the force-displacement curve while both β and θ play a role in the stress distribution and crack propagation. We found that these effects are coupled and the optimal set of parameters also depends on the fiber material. This is the first experimental verification of such optimal designs proposed for mushroomlike microfibers. This experimental approach could be used to evaluate a wide range of complex microstructured adhesive designs suggested in the literature and optimize them. PMID:26322396

  20. A Tutorial on Adaptive Design Optimization

    PubMed Central

    Myung, Jay I.; Cavagnaro, Daniel R.; Pitt, Mark A.

    2013-01-01

    Experimentation is ubiquitous in the field of psychology and fundamental to the advancement of its science, and one of the biggest challenges for researchers is designing experiments that can conclusively discriminate the theoretical hypotheses or models under investigation. The recognition of this challenge has led to the development of sophisticated statistical methods that aid in the design of experiments and that are within the reach of everyday experimental scientists. This tutorial paper introduces the reader to an implementable experimentation methodology, dubbed Adaptive Design Optimization, that can help scientists to conduct “smart” experiments that are maximally informative and highly efficient, which in turn should accelerate scientific discovery in psychology and beyond. PMID:23997275

  1. Acoustic design by topology optimization

    NASA Astrophysics Data System (ADS)

    Dühring, Maria B.; Jensen, Jakob S.; Sigmund, Ole

    2008-11-01

    To bring down noise levels in human surroundings is an important issue and a method to reduce noise by means of topology optimization is presented here. The acoustic field is modeled by Helmholtz equation and the topology optimization method is based on continuous material interpolation functions in the density and bulk modulus. The objective function is the squared sound pressure amplitude. First, room acoustic problems are considered and it is shown that the sound level can be reduced in a certain part of the room by an optimized distribution of reflecting material in a design domain along the ceiling or by distribution of absorbing and reflecting material along the walls. We obtain well defined optimized designs for a single frequency or a frequency interval for both 2D and 3D problems when considering low frequencies. Second, it is shown that the method can be applied to design outdoor sound barriers in order to reduce the sound level in the shadow zone behind the barrier. A reduction of up to 10 dB for a single barrier and almost 30 dB when using two barriers are achieved compared to utilizing conventional sound barriers.

  2. Parameters optimization using experimental design for headspace solid phase micro-extraction analysis of short-chain chlorinated paraffins in waters under the European water framework directive.

    PubMed

    Gandolfi, F; Malleret, L; Sergent, M; Doumenq, P

    2015-08-01

    The water framework directives (WFD 2000/60/EC and 2013/39/EU) force European countries to monitor the quality of their aquatic environment. Among the priority hazardous substances targeted by the WFD, short chain chlorinated paraffins C10-C13 (SCCPs), still represent an analytical challenge, because few laboratories are nowadays able to analyze them. Moreover, an annual average quality standards as low as 0.4μgL(-1) was set for SCCPs in surface water. Therefore, to test for compliance, the implementation of sensitive and reliable analysis method of SCCPs in water are required. The aim of this work was to address this issue by evaluating automated solid phase micro-extraction (SPME) combined on line with gas chromatography-electron capture negative ionization mass spectrometry (GC/ECNI-MS). Fiber polymer, extraction mode, ionic strength, extraction temperature and time were the most significant thermodynamic and kinetic parameters studied. To determine the suitable factors working ranges, the study of the extraction conditions was first carried out by using a classical one factor-at-a-time approach. Then a mixed level factorial 3×2(3) design was performed, in order to give rise to the most influent parameters and to estimate potential interactions effects between them. The most influent factors, i.e. extraction temperature and duration, were optimized by using a second experimental design, in order to maximize the chromatographic response. At the close of the study, a method involving headspace SPME (HS-SPME) coupled to GC/ECNI-MS is proposed. The optimum extraction conditions were sample temperature 90°C, extraction time 80min, with the PDMS 100μm fiber and desorption at 250°C during 2min. Linear response from 0.2ngmL(-1) to 10ngmL(-1) with r(2)=0.99 and limits of detection and quantification, respectively of 4pgmL(-1) and 120pgmL(-1) in MilliQ water, were achieved. The method proved to be applicable in different types of waters and show key advantages, such

  3. Experimental design in chemistry: A tutorial.

    PubMed

    Leardi, Riccardo

    2009-10-12

    In this tutorial the main concepts and applications of experimental design in chemistry will be explained. Unfortunately, nowadays experimental design is not as known and applied as it should be, and many papers can be found in which the "optimization" of a procedure is performed one variable at a time. Goal of this paper is to show the real advantages in terms of reduced experimental effort and of increased quality of information that can be obtained if this approach is followed. To do that, three real examples will be shown. Rather than on the mathematical aspects, this paper will focus on the mental attitude required by experimental design. The readers being interested to deepen their knowledge of the mathematical and algorithmical part can find very good books and tutorials in the references [G.E.P. Box, W.G. Hunter, J.S. Hunter, Statistics for Experimenters: An Introduction to Design, Data Analysis, and Model Building, John Wiley & Sons, New York, 1978; R. Brereton, Chemometrics: Data Analysis for the Laboratory and Chemical Plant, John Wiley & Sons, New York, 1978; R. Carlson, J.E. Carlson, Design and Optimization in Organic Synthesis: Second Revised and Enlarged Edition, in: Data Handling in Science and Technology, vol. 24, Elsevier, Amsterdam, 2005; J.A. Cornell, Experiments with Mixtures: Designs, Models and the Analysis of Mixture Data, in: Series in Probability and Statistics, John Wiley & Sons, New York, 1991; R.E. Bruns, I.S. Scarminio, B. de Barros Neto, Statistical Design-Chemometrics, in: Data Handling in Science and Technology, vol. 25, Elsevier, Amsterdam, 2006; D.C. Montgomery, Design and Analysis of Experiments, 7th edition, John Wiley & Sons, Inc., 2009; T. Lundstedt, E. Seifert, L. Abramo, B. Thelin, A. Nyström, J. Pettersen, R. Bergman, Chemolab 42 (1998) 3; Y. Vander Heyden, LC-GC Europe 19 (9) (2006) 469]. PMID:19786177

  4. Optimal design of solidification processes

    NASA Technical Reports Server (NTRS)

    Dantzig, Jonathan A.; Tortorelli, Daniel A.

    1991-01-01

    An optimal design algorithm is presented for the analysis of general solidification processes, and is demonstrated for the growth of GaAs crystals in a Bridgman furnace. The system is optimal in the sense that the prespecified temperature distribution in the solidifying materials is obtained to maximize product quality. The optimization uses traditional numerical programming techniques which require the evaluation of cost and constraint functions and their sensitivities. The finite element method is incorporated to analyze the crystal solidification problem, evaluate the cost and constraint functions, and compute the sensitivities. These techniques are demonstrated in the crystal growth application by determining an optimal furnace wall temperature distribution to obtain the desired temperature profile in the crystal, and hence to maximize the crystal's quality. Several numerical optimization algorithms are studied to determine the proper convergence criteria, effective 1-D search strategies, appropriate forms of the cost and constraint functions, etc. In particular, we incorporate the conjugate gradient and quasi-Newton methods for unconstrained problems. The efficiency and effectiveness of each algorithm is presented in the example problem.

  5. Research on optimization-based design

    NASA Technical Reports Server (NTRS)

    Balling, R. J.; Parkinson, A. R.; Free, J. C.

    1989-01-01

    Research on optimization-based design is discussed. Illustrative examples are given for cases involving continuous optimization with discrete variables and optimization with tolerances. Approximation of computationally expensive and noisy functions, electromechanical actuator/control system design using decomposition and application of knowledge-based systems and optimization for the design of a valve anti-cavitation device are among the topics covered.

  6. Optimal design of capacitor-driven coilgun

    NASA Astrophysics Data System (ADS)

    Kim, Seog-Whan; Jung, Hyun-Kyo; Hahn, Song-Yop

    1994-03-01

    This paper presents an analysis and optimal design of a capacitor-driven inductive coilgun. An equivalent circuit is used for a launch simulation of the coilgun. The circuit equations are solved together with the equation of motion of the projectile by using the Runge-Kutta method. The numerical results are compared with the experimental values to verify the usefulness of the developed simulation program. It is shown that the numerical and the experimental results are in a good agreement. In the design of the system the optimization is achieved by employing the genetic algorithm. The resultant specifications of the coilgun optimally designed by the proposed algorithm are tested by experiment. Finally the obtained results are compared with those designed by approximate equations and by linear search methods as well. It is found that the proposed algorithm gives a better result in the energy efficiency of the system, namely it enables one to obtain a higher muzzle velocity of the projectile with the same amount of energy.

  7. Optimization of confocal scanning laser ophthalmoscope design

    PubMed Central

    Dhalla, Al-Hafeez; Kelly, Michael P.; Farsiu, Sina; Izatt, Joseph A.

    2013-01-01

    Abstract. Confocal scanning laser ophthalmoscopy (cSLO) enables high-resolution and high-contrast imaging of the retina by employing spatial filtering for scattered light rejection. However, to obtain optimized image quality, one must design the cSLO around scanner technology limitations and minimize the effects of ocular aberrations and imaging artifacts. We describe a cSLO design methodology resulting in a simple, relatively inexpensive, and compact lens-based cSLO design optimized to balance resolution and throughput for a 20-deg field of view (FOV) with minimal imaging artifacts. We tested the imaging capabilities of our cSLO design with an experimental setup from which we obtained fast and high signal-to-noise ratio (SNR) retinal images. At lower FOVs, we were able to visualize parafoveal cone photoreceptors and nerve fiber bundles even without the use of adaptive optics. Through an experiment comparing our optimized cSLO design to a commercial cSLO system, we show that our design demonstrates a significant improvement in both image quality and resolution. PMID:23864013

  8. In vitro and in vivo studies on chitosan beads of losartan Duolite AP143 complex, optimized by using statistical experimental design.

    PubMed

    Madgulkar, Ashwini; Bhalekar, Mangesh; Swami, Megha

    2009-01-01

    The aim of the present research work was to develop release modulated beads of losartan potassium complexed with anion exchange resin, Duolite AP143 (cholestyramine). Chitosan was selected as a hydrophilic polymer for the formation of beads which could sustain the release of the drug up to 12 h, along with drug resin complex (DRC). Chitosan beads were prepared using an in-liquid curing method by ionotropic cross-linking or interpolymer linkage with sodium tripolyphosphate (TPP). The formulation of the beads was optimized for entrapment efficiency and drug release using 3(2) full factorial design. The independent variables selected were DRC/chitosan and percent of TPP. The optimization model was validated for its performance characteristics. Studies revealed that as the concentration of chitosan and TPP was increased, entrapment efficiency and the drug release were found to increase and decrease, respectively. The swelling capacity of chitosan-TPP beads decreased with increasing concentration of TPP. The effect of chitosan concentration and percentage of TPP solution used for cross-linking on entrapment efficiency and drug release rate was extensively investigated. Optimized beads were subjected to in vivo studies in Wistar albino rats to determine the mean arterial blood pressure and compared with marketed formulation. The pharmacodynamic study demonstrates steady blood pressure control for optimized formulation as compared to fluctuated blood pressure for the marketed formulation. PMID:19495989

  9. Design Of Theoretically Optimal Thermoacoustic Cooling Device

    NASA Astrophysics Data System (ADS)

    Tisovský, Tomáš; Vít, Tomáš

    2016-03-01

    The aim of this article is to design theoretically optimal thermoacoustic cooling device. The opening chapter gives the reader brief introduction to thermoacoustic, specializing in the thermoacoustic principle in refrigerator regime. Subsequent part of the article aims to explain the principle on which thermoacoustic is simulated in DeltaEC. Numbers of executed numerical simulations are listed and the resulting thermoacoustic cooling device design is presented along with its main operation characteristics. In conclusion, recommendations for future experimental work are given and the results are discussed.

  10. Design optimization of LiNi0.6Co0.2Mn0.2O2/graphite lithium-ion cells based on simulation and experimental data

    NASA Astrophysics Data System (ADS)

    Appiah, Williams Agyei; Park, Joonam; Song, Seonghyun; Byun, Seoungwoo; Ryou, Myung-Hyun; Lee, Yong Min

    2016-07-01

    LiNi0.6Co0.2Mn0.2O2 cathodes of different thicknesses and porosities are prepared and tested, in order to optimize the design of lithium-ion cells. A mathematical model for simulating multiple types of particles with different contact resistances in a single electrode is adopted to study the effects of the different cathode thicknesses and porosities on lithium-ion transport using the nonlinear least squares technique. The model is used to optimize the design of LiNi0.6Co0.2Mn0.2O2/graphite lithium-ion cells by employing it to generate a number of Ragone plots. The cells are optimized for cathode porosity and thickness, while the anode porosity, anode-to-cathode capacity ratio, thickness and porosity of separator, and electrolyte salt concentration are held constant. Optimization is performed for discharge times ranging from 10 h to 5 min. Using the Levenberg-Marquardt method as a fitting technique, accounting for multiple particles with different contact resistances, and employing a rate-dependent solid-phase diffusion coefficient results in there being good agreement between the simulated and experimentally determined discharge curves. The optimized parameters obtained from this study should serve as a guide for the battery industry as well as for researchers for determining the optimal cell design for different applications.

  11. Optimization of the enantioseparation of a diaryl-pyrazole sulfonamide derivative by capillary electrophoresis in a dual CD mode using experimental design.

    PubMed

    Rogez-Florent, Tiphaine; Foulon, Catherine; Six, Perrine; Goossens, Laurence; Danel, Cécile; Goossens, Jean-François

    2014-10-01

    A CE method using dual cationic and neutral cyclodextrins (CD) was optimized for the enantiomeric separation of a compound presenting a diaryl sulfonamide group. Preliminary studies were made to select the optimal CDs and pH of the BGE. Two CDs (amino-β-CD and β-CD) were selected to separate the enantiomers in a 67 mM phosphate buffer at pH 7.4. However, the repeatability of the analyses obtained on bare-fused silica capillary was not acceptable owing to the adsorption of the amino-β-CD to the capillary. To prevent this, a dynamic coating of the capillary was used employing five layers of ionic-polymer (poly(diallyldimethylammonium) chloride (PDADMAC) and poly(sodium 4-styrenesulfonate). The efficiency of the coating was assessed by measuring the EOF stability. Repeatability of the injections was obtained when intermediate coating with PDADMAC was performed between each run. Secondly, this enantioseparation method was optimized using a central composite circumscribed design including three factors: amino-β-CD and β-CD concentrations and the percentage of methanol. Under the optimal conditions (i.e. 16.6 mM of amino-β-CD, 2.6 mM of β-CD, 0% MeOH in 67 mM phosphate buffer (pH 7.4) as BGE, cathodic injection 0.5 psi, 5 s, separation voltage 15 kV and a temperature of 15°C), complete enantioresolution of the analyte was obtained. It is worth mentioning that the design of experiments (DOE) protocol employed showed a significant interaction between CDs, highlighting the utility of DOE in method development. Finally, small variations in the ionic-polymer concentrations did not significantly influence the EOF, confirming the robustness of the coating method. PMID:24723380

  12. Optimal design of airlift fermenters

    SciTech Connect

    Moresi, M.

    1981-11-01

    In this article a modeling of a draft-tube airlift fermenter (ALF) based on perfect back-mixing of liquid and plugflow for gas bubbles has been carried out to optimize the design and operation of fermentation units at different working capacities. With reference to a whey fermentation by yeasts the economic optimization has led to a slim ALF with an aspect ratio of about 15. As far as power expended per unit of oxygen transfer is concerned, the responses of the model are highly influenced by kLa. However, a safer use of the model has been suggested in order to assess the feasibility of the fermentation process under study. (Refs. 39).

  13. Development of a novel pH sensor based upon Janus Green B immobilized on triacetyl cellulose membrane: Experimental design and optimization

    NASA Astrophysics Data System (ADS)

    Chamkouri, Narges; Niazi, Ali; Zare-Shahabadi, Vali

    2016-03-01

    A novel pH optical sensor was prepared by immobilizing an azo dye called Janus Green B on the triacetylcellulose membrane. Condition of the dye solution used in the immobilization step, including concentration of the dye, pH, and duration were considered and optimized using the Box-Behnken design. The proposed sensor showed good behavior and precision (RSD < 5%) in the pH range of 2.0-10.0. Advantages of this optical sensor include on-line applicability, no leakage, long-term stability (more than 6 months), fast response time (less than 1 min), high selectivity and sensitivity as well as good reversibility and reproducibility.

  14. Computational Optimization of a Natural Laminar Flow Experimental Wing Glove

    NASA Technical Reports Server (NTRS)

    Hartshom, Fletcher

    2012-01-01

    Computational optimization of a natural laminar flow experimental wing glove that is mounted on a business jet is presented and discussed. The process of designing a laminar flow wing glove starts with creating a two-dimensional optimized airfoil and then lofting it into a three-dimensional wing glove section. The airfoil design process does not consider the three dimensional flow effects such as cross flow due wing sweep as well as engine and body interference. Therefore, once an initial glove geometry is created from the airfoil, the three dimensional wing glove has to be optimized to ensure that the desired extent of laminar flow is maintained over the entire glove. TRANAIR, a non-linear full potential solver with a coupled boundary layer code was used as the main tool in the design and optimization process of the three-dimensional glove shape. The optimization process uses the Class-Shape-Transformation method to perturb the geometry with geometric constraints that allow for a 2-in clearance from the main wing. The three-dimensional glove shape was optimized with the objective of having a spanwise uniform pressure distribution that matches the optimized two-dimensional pressure distribution as closely as possible. Results show that with the appropriate inputs, the optimizer is able to match the two dimensional pressure distributions practically across the entire span of the wing glove. This allows for the experiment to have a much higher probability of having a large extent of natural laminar flow in flight.

  15. Parametric Design Optimization By Integrating CAD Systems And Optimization Tools

    NASA Astrophysics Data System (ADS)

    Rehan, M.; Olabi, A. G.

    2009-11-01

    Designing a cost effective product in minimum time is a complex process. In order to achieve this goal the requirement of optimum designs are becoming more important. One of the time consuming factor in the design optimization cycle is the modifications of Computer Aided Design (CAD) model after optimization. In conventional design optimization techniques the design engineer has to update the CAD model after receiving optimum design from optimization tools. It is worthwhile using parametric design optimization process to minimize the optimization cycle time. This paper presents a comprehensive study to integrate the optimization parameters between CAD system and optimization tools which were driven from a single user environment. Finally, design optimization of a Compressed Natural Gas (CNG) cylinder was implemented as case study. In this case study the optimization tools were fully integrated with CAD system, therefore, all the deliverables including; part design, drawings and assembly can be automatically updated after achieving the optimum geometry having minimum volume and satisfying all imposed constraints.

  16. Experimental design optimization for the ICP-AES determination of Li, Na, K, Al, Fe, Mn and Zn in human serum.

    PubMed

    Bianchi, F; Maffini, M; Mangia, A; Marengo, E; Mucchino, C

    2007-01-17

    A chemometric approach based on experimental design and desirability functions was used to develop and validated a method for the determination of some metals of biological concern by a fast sequential ICP-AES. The elements considered are Al, Fe, Mn, Zn, Li, Na and K. The experimental design was used to investigate the effects of three instrumental most crucial parameters, such as sheath gas flow rate, pump speed and auxiliary gas flow rate. In order to improve the multielemental analysis speed, although a sequential instrument allows the use of a separate parameter set for each wavelength, regression models and desirability functions were applied to find the experimental conditions providing the highest global sensitivity. Validation was performed in terms of limits of detection (LOD), limits of quantitation (LOQ), linearity, precision and recovery. By using the 167.02 nm wavelength, aluminium LOD was 0.5 microg L(-1) while the highest LOD was found for K (65 microg L(-1)). A linear range of at least three orders of magnitude was statistically demonstrated for each element. Precision was evaluated by testing two concentration levels, and good results in terms of intra-day repeatability were obtained, with R.S.D. values lower than 4.1% at the lowest concentration level. Lacking a suitable certified reference material, trueness was estimated using the recovery rate on fortified samples. The validated method was then used in the quantification of the elements considered in a serum sample. PMID:17079109

  17. Probabilistic Finite Element Analysis & Design Optimization for Structural Designs

    NASA Astrophysics Data System (ADS)

    Deivanayagam, Arumugam

    This study focuses on implementing probabilistic nature of material properties (Kevlar® 49) to the existing deterministic finite element analysis (FEA) of fabric based engine containment system through Monte Carlo simulations (MCS) and implementation of probabilistic analysis in engineering designs through Reliability Based Design Optimization (RBDO). First, the emphasis is on experimental data analysis focusing on probabilistic distribution models which characterize the randomness associated with the experimental data. The material properties of Kevlar® 49 are modeled using experimental data analysis and implemented along with an existing spiral modeling scheme (SMS) and user defined constitutive model (UMAT) for fabric based engine containment simulations in LS-DYNA. MCS of the model are performed to observe the failure pattern and exit velocities of the models. Then the solutions are compared with NASA experimental tests and deterministic results. MCS with probabilistic material data give a good prospective on results rather than a single deterministic simulation results. The next part of research is to implement the probabilistic material properties in engineering designs. The main aim of structural design is to obtain optimal solutions. In any case, in a deterministic optimization problem even though the structures are cost effective, it becomes highly unreliable if the uncertainty that may be associated with the system (material properties, loading etc.) is not represented or considered in the solution process. Reliable and optimal solution can be obtained by performing reliability optimization along with the deterministic optimization, which is RBDO. In RBDO problem formulation, in addition to structural performance constraints, reliability constraints are also considered. This part of research starts with introduction to reliability analysis such as first order reliability analysis, second order reliability analysis followed by simulation technique that

  18. An optimal structural design algorithm using optimality criteria

    NASA Technical Reports Server (NTRS)

    Taylor, J. E.; Rossow, M. P.

    1976-01-01

    An algorithm for optimal design is given which incorporates several of the desirable features of both mathematical programming and optimality criteria, while avoiding some of the undesirable features. The algorithm proceeds by approaching the optimal solution through the solutions of an associated set of constrained optimal design problems. The solutions of the constrained problems are recognized at each stage through the application of optimality criteria based on energy concepts. Two examples are described in which the optimal member size and layout of a truss is predicted, given the joint locations and loads.

  19. Teaching experimental design to biologists.

    PubMed

    Zolman, J F

    1999-12-01

    The teaching of research design and data analysis to our graduate students has been a persistent problem. A course is described in which students, early in their graduate training, obtain extensive practice in designing experiments and interpreting data. Lecture-discussions on the essentials of biostatistics are given, and then these essentials are repeatedly reviewed by illustrating their applications and misapplications in numerous research design problems. Students critique these designs and prepare similar problems for peer evaluation. In most problems the treatments are confounded by extraneous variables, proper controls may be absent, or data analysis may be incorrect. For each problem, students must decide whether the researchers' conclusions are valid and, if not, must identify a fatal experimental flaw. Students learn that an experiment is a well-conceived plan for data collection, analysis, and interpretation. They enjoy the interactive evaluations of research designs and appreciate the repetitive review of common flaws in different experiments. They also benefit from their practice in scientific writing and in critically evaluating their peers' designs. PMID:10644236

  20. Optimization Algorithm for Designing Diffractive Optical Elements

    NASA Astrophysics Data System (ADS)

    Agudelo, Viviana A.; Orozco, Ricardo Amézquita

    2008-04-01

    Diffractive Optical Elements (DOEs) are commonly used in many applications such as laser beam shaping, recording of micro reliefs, wave front analysis, metrology and many others where they can replace single or multiple conventional optical elements (diffractive or refractive). One of the most versatile way to produce them, is to use computer assisted techniques for their design and optimization, as well as optical or electron beam micro-lithography techniques for the final fabrication. The fundamental figures of merit involved in the optimization of such devices are both the diffraction efficiency and the signal to noise ratio evaluated in the reconstructed wave front at the image plane. A design and optimization algorithm based on the error—reduction method (Gerchberg and Saxton) is proposed to obtain binary discrete phase-only Fresnel DOEs that will be used to produce specific intensity patterns. Some experimental results were obtained using a spatial light modulator acting as a binary programmable diffractive phase element. Although the DOEs optimized here are discrete in phase, they present an acceptable signal noise relation and diffraction efficiency.

  1. Ultra-high performance liquid chromatographic determination of levofloxacin in human plasma and prostate tissue with use of experimental design optimization procedures.

    PubMed

    Szerkus, O; Jacyna, J; Wiczling, P; Gibas, A; Sieczkowski, M; Siluk, D; Matuszewski, M; Kaliszan, R; Markuszewski, M J

    2016-09-01

    Fluoroquinolones are considered as gold standard for the prevention of bacterial infections after transrectal ultrasound guided prostate biopsy. However, recent studies reported that fluoroquinolone- resistant bacterial strains are responsible for gradually increasing number of infections after transrectal prostate biopsy. In daily clinical practice, antibacterial efficacy is evaluated only in vitro, by measuring the reaction of bacteria with an antimicrobial agent in culture media (i.e. calculation of minimal inhibitory concentration). Such approach, however, has no relation to the treated tissue characteristics and might be highly misleading. Thus, the objective of this study was to develop, with the use of Design of Experiments approach, a reliable, specific and sensitive ultra-high performance liquid chromatography- diode array detection method for the quantitative analysis of levofloxacin in plasma and prostate tissue samples obtained from patients undergoing prostate biopsy. Moreover, correlation study between concentrations observed in plasma samples vs prostatic tissue samples was performed, resulting in better understanding, evaluation and optimization of the fluoroquinolone-based antimicrobial prophylaxis during transrectal ultrasound guided prostate biopsy. Box-Behnken design was employed to optimize chromatographic conditions of the isocratic elution program in order to obtain desirable retention time, peak symmetry and resolution of levofloxacine and ciprofloxacine (internal standard) peaks. Fractional Factorial design 2(4-1) with four center points was used for screening of significant factors affecting levofloxacin extraction from the prostatic tissue. Due to the limited number of tissue samples the prostatic sample preparation procedure was further optimized using Central Composite design. Design of Experiments approach was also utilized for evaluation of parameter robustness. The method was found linear over the range of 0.030-10μg/mL for human

  2. An Evolutionary Optimization System for Spacecraft Design

    NASA Technical Reports Server (NTRS)

    Fukunaga, A.; Stechert, A.

    1997-01-01

    Spacecraft design optimization is a domian that can benefit from the application of optimization algorithms such as genetic algorithms. In this paper, we describe DEVO, an evolutionary optimization system that addresses these issues and provides a tool that can be applied to a number of real-world spacecraft design applications. We describe two current applications of DEVO: physical design if a Mars Microprobe Soil Penetrator, and system configuration optimization for a Neptune Orbiter.

  3. Optimal design of compact spur gear reductions

    NASA Technical Reports Server (NTRS)

    Savage, M.; Lattime, S. B.; Kimmel, J. A.; Coe, H. H.

    1992-01-01

    The optimal design of compact spur gear reductions includes the selection of bearing and shaft proportions in addition to gear mesh parameters. Designs for single mesh spur gear reductions are based on optimization of system life, system volume, and system weight including gears, support shafts, and the four bearings. The overall optimization allows component properties to interact, yielding the best composite design. A modified feasible directions search algorithm directs the optimization through a continuous design space. Interpolated polynomials expand the discrete bearing properties and proportions into continuous variables for optimization. After finding the continuous optimum, the designer can analyze near optimal designs for comparison and selection. Design examples show the influence of the bearings on the optimal configurations.

  4. Experimental Validation of an Integrated Controls-Structures Design Methodology

    NASA Technical Reports Server (NTRS)

    Maghami, Peiman G.; Gupta, Sandeep; Elliot, Kenny B.; Walz, Joseph E.

    1996-01-01

    The first experimental validation of an integrated controls-structures design methodology for a class of large order, flexible space structures is described. Integrated redesign of the controls-structures-interaction evolutionary model, a laboratory testbed at NASA Langley, was described earlier. The redesigned structure was fabricated, assembled in the laboratory, and experimentally tested against the original structure. Experimental results indicate that the structure redesigned using the integrated design methodology requires significantly less average control power than the nominal structure with control-optimized designs, while maintaining the required line-of-sight pointing performance. Thus, the superiority of the integrated design methodology over the conventional design approach is experimentally demonstrated. Furthermore, amenability of the integrated design structure to other control strategies is evaluated, both analytically and experimentally. Using Linear-Quadratic-Guassian optimal dissipative controllers, it is observed that the redesigned structure leads to significantly improved performance with alternate controllers as well.

  5. Rational Experimental Design for Electrical Resistivity Imaging

    NASA Astrophysics Data System (ADS)

    Mitchell, V.; Pidlisecky, A.; Knight, R.

    2008-12-01

    Over the past several decades advances in the acquisition and processing of electrical resistivity data, through multi-channel acquisition systems and new inversion algorithms, have greatly increased the value of these data to near-surface environmental and hydrological problems. There has, however, been relatively little advancement in the design of actual surveys. Data acquisition still typically involves using a small number of traditional arrays (e.g. Wenner, Schlumberger) despite a demonstrated improvement in data quality from the use of non-standard arrays. While optimized experimental design has been widely studied in applied mathematics and the physical and biological sciences, it is rarely implemented for non-linear problems, such as electrical resistivity imaging (ERI). We focus specifically on using ERI in the field for monitoring changes in the subsurface electrical resistivity structure. For this application we seek an experimental design method that can be used in the field to modify the data acquisition scheme (spatial and temporal sampling) based on prior knowledge of the site and/or knowledge gained during the imaging experiment. Some recent studies have investigated optimized design of electrical resistivity surveys by linearizing the problem or with computationally-intensive search algorithms. We propose a method for rational experimental design based on the concept of informed imaging, the use of prior information regarding subsurface properties and processes to develop problem-specific data acquisition and inversion schemes. Specifically, we use realistic subsurface resistivity models to aid in choosing source configurations that maximize the information content of our data. Our approach is based on first assessing the current density within a region of interest, in order to provide sufficient energy to the region of interest to overcome a noise threshold, and then evaluating the direction of current vectors, in order to maximize the

  6. Piezoelectric transducer design via multiobjective optimization.

    PubMed

    Fu, B; Hemsel, T; Wallaschek, J

    2006-12-22

    The design of piezoelectric transducers is usually based on single-objective optimization only. In most practical applications of piezoelectric transducers, however, there exist multiple design objectives that often are contradictory to each other by their very nature. It is impossible to find a solution at which each objective function gets its optimal value simultaneously. Our design approach is to first find a set of Pareto-optimal solutions, which can be considered to be best compromises among multiple design objectives. Among these Pareto-optimal solutions, the designer can then select the one solution which he considers to be the best one. In this paper we investigate the optimal design of a Langevin transducer. The design problem is formulated mathematically as a constrained multiobjective optimization problem. The maximum vibration amplitude and the minimum electrical input power are considered as optimization objectives. Design variables involve continuous variables (dimensions of the transducer) and discrete variables (the number of piezoelectric rings and material types). In order to formulate the optimization problem, the behavior of piezoelectric transducers is modeled using the transfer matrix method based on analytical models. Multiobjective evolutionary algorithms are applied in the optimization process and a set of Pareto-optimal designs is calculated. The optimized results are analyzed and the preferred design is determined. PMID:16814826

  7. Design optimization method for Francis turbine

    NASA Astrophysics Data System (ADS)

    Kawajiri, H.; Enomoto, Y.; Kurosawa, S.

    2014-03-01

    This paper presents a design optimization system coupled CFD. Optimization algorithm of the system employs particle swarm optimization (PSO). Blade shape design is carried out in one kind of NURBS curve defined by a series of control points. The system was applied for designing the stationary vanes and the runner of higher specific speed francis turbine. As the first step, single objective optimization was performed on stay vane profile, and second step was multi-objective optimization for runner in wide operating range. As a result, it was confirmed that the design system is useful for developing of hydro turbine.

  8. A novel homocystine-agarose adsorbent for separation and preconcentration of nickel in table salt and baking soda using factorial design optimization of the experimental conditions.

    PubMed

    Hashemi, Payman; Rahmani, Zohreh

    2006-02-28

    Homocystine was for the first time, chemically linked to a highly cross-linked agarose support (Novarose) to be employed as a chelating adsorbent for preconcentration and AAS determination of nickel in table salt and baking soda. Nickel is quantitatively adsorbed on a small column packed with 0.25ml of the adsorbent, in a pH range of 5.5-6.5 and simply eluted with 5ml of a 1moll(-1) hydrochloric acid solution. A factorial design was used for optimization of the effects of five different variables on the recovery of nickel. The results indicated that the factors of flow rate and column length, and the interactions between pH and sample volume are significant. In the optimized conditions, the column could tolerate salt concentrations up to 0.5moll(-1) and sample volumes beyond 500ml. Matrix ions of Mg(2+) and Ca(2+), with a concentration of 200mgl(-1), and potentially interfering ions of Cd(2+), Cu(2+), Zn(2+) and Mn(2+), with a concentration of 10mgl(-1), did not have significant effect on the analyte's signal. Preconcentration factors up to 100 and a detection limit of 0.49mugl(-1), corresponding to an enrichment volume of 500ml, were obtained for the determination of the analyte by flame AAS. Application of the method to the determination of natural and spiked nickel in table salt and baking soda solutions resulted in quantitative recoveries. Direct ETAAS determination of nickel in the same samples was not possible because of a high background observed. PMID:18970514

  9. Animal husbandry and experimental design.

    PubMed

    Nevalainen, Timo

    2014-01-01

    If the scientist needs to contact the animal facility after any study to inquire about husbandry details, this represents a lost opportunity, which can ultimately interfere with the study results and their interpretation. There is a clear tendency for authors to describe methodological procedures down to the smallest detail, but at the same time to provide minimal information on animals and their husbandry. Controlling all major variables as far as possible is the key issue when establishing an experimental design. The other common mechanism affecting study results is a change in the variation. Factors causing bias or variation changes are also detectable within husbandry. Our lives and the lives of animals are governed by cycles: the seasons, the reproductive cycle, the weekend-working days, the cage change/room sanitation cycle, and the diurnal rhythm. Some of these may be attributable to routine husbandry, and the rest are cycles, which may be affected by husbandry procedures. Other issues to be considered are consequences of in-house transport, restrictions caused by caging, randomization of cage location, the physical environment inside the cage, the acoustic environment audible to animals, olfactory environment, materials in the cage, cage complexity, feeding regimens, kinship, and humans. Laboratory animal husbandry issues are an integral but underappreciated part of investigators' experimental design, which if ignored can cause major interference with the results. All researchers should familiarize themselves with the current routine animal care of the facility serving them, including their capabilities for the monitoring of biological and physicochemical environment. PMID:25541541

  10. Optimizing experimental parameters for tracking of diffusing particles

    NASA Astrophysics Data System (ADS)

    Vestergaard, Christian L.

    2016-08-01

    We describe how a single-particle tracking experiment should be designed in order for its recorded trajectories to contain the most information about a tracked particle's diffusion coefficient. The precision of estimators for the diffusion coefficient is affected by motion blur, limited photon statistics, and the length of recorded time series. We demonstrate for a particle undergoing free diffusion that precision is negligibly affected by motion blur in typical experiments, while optimizing photon counts and the number of recorded frames is the key to precision. Building on these results, we describe for a wide range of experimental scenarios how to choose experimental parameters in order to optimize the precision. Generally, one should choose quantity over quality: experiments should be designed to maximize the number of frames recorded in a time series, even if this means lower information content in individual frames.

  11. Optimal design of biaxial tensile cruciform specimens

    NASA Astrophysics Data System (ADS)

    Demmerle, S.; Boehler, J. P.

    1993-01-01

    F OR EXPERIMENTAL investigations concerning the mechanical behaviour under biaxial stress states of rolled sheet metals, mostly cruciform flat specimens are used. By means of empirical methods, different specimen geometries have been proposed in the literature. In order to evaluate the suitability of a specimen design, a mathematically well defined criterion is developed, based on the standard deviations of the values of the stresses in the test section. Applied to the finite element method, the criterion is employed to realize the shape optimization of biaxial cruciform specimens for isotropic elastic materials. Furthermore, the performance of the obtained optimized specimen design is investigated in the case of off-axes tests on anisotropic materials. Therefore, for the first time, an original testing device, consisting of hinged fixtures with knife edges at each arm of the specimen, is applied to the biaxial test. The obtained results indicate the decisive superiority of the optimized specimens for the proper performance on isotropic materials, as well as the paramount importance of the proposed off-axes testing technique for biaxial tests on anisotropic materials.

  12. Experimental validation of a topology optimized acoustic cavity.

    PubMed

    Christiansen, Rasmus E; Sigmund, Ole; Fernandez-Grande, Efren

    2015-12-01

    This paper presents the experimental validation of an acoustic cavity designed using topology optimization with the goal of minimizing the sound pressure locally for monochromatic excitation. The presented results show good agreement between simulations and measurements. The effect of damping, errors in the production of the cavity, and variations in operating frequency is discussed and the importance of taking these factors into account in the modeling process is highlighted. PMID:26723304

  13. Experimental Validation of Passive Safety System Models: Application to Design and Optimization of Fluoride-Salt-Cooled, High-Temperature Reactors

    NASA Astrophysics Data System (ADS)

    Zweibaum, Nicolas

    The development of advanced nuclear reactor technology requires understanding of complex, integrated systems that exhibit novel phenomenology under normal and accident conditions. The advent of passive safety systems and enhanced modular construction methods requires the development and use of new frameworks to predict the behavior of advanced nuclear reactors, both from a safety standpoint and from an environmental impact perspective. This dissertation introduces such frameworks for scaling of integral effects tests for natural circulation in fluoride-salt-cooled, high-temperature reactors (FHRs) to validate evaluation models (EMs) for system behavior; subsequent reliability assessment of passive, natural- circulation-driven decay heat removal systems, using these validated models; evaluation of life cycle carbon dioxide emissions as a key environmental impact metric; and recommendations for further work to apply these frameworks in the development and optimization of advanced nuclear reactor designs. In this study, the developed frameworks are applied to the analysis of the Mark 1 pebble-bed FHR (Mk1 PB-FHR) under current investigation at the University of California, Berkeley (UCB). (Abstract shortened by UMI.).

  14. Program Aids Analysis And Optimization Of Design

    NASA Technical Reports Server (NTRS)

    Rogers, James L., Jr.; Lamarsh, William J., II

    1994-01-01

    NETS/ PROSSS (NETS Coupled With Programming System for Structural Synthesis) computer program developed to provide system for combining NETS (MSC-21588), neural-network application program and CONMIN (Constrained Function Minimization, ARC-10836), optimization program. Enables user to reach nearly optimal design. Design then used as starting point in normal optimization process, possibly enabling user to converge to optimal solution in significantly fewer iterations. NEWT/PROSSS written in C language and FORTRAN 77.

  15. Sequential experimental design based generalised ANOVA

    NASA Astrophysics Data System (ADS)

    Chakraborty, Souvik; Chowdhury, Rajib

    2016-07-01

    Over the last decade, surrogate modelling technique has gained wide popularity in the field of uncertainty quantification, optimization, model exploration and sensitivity analysis. This approach relies on experimental design to generate training points and regression/interpolation for generating the surrogate. In this work, it is argued that conventional experimental design may render a surrogate model inefficient. In order to address this issue, this paper presents a novel distribution adaptive sequential experimental design (DA-SED). The proposed DA-SED has been coupled with a variant of generalised analysis of variance (G-ANOVA), developed by representing the component function using the generalised polynomial chaos expansion. Moreover, generalised analytical expressions for calculating the first two statistical moments of the response, which are utilized in predicting the probability of failure, have also been developed. The proposed approach has been utilized in predicting probability of failure of three structural mechanics problems. It is observed that the proposed approach yields accurate and computationally efficient estimate of the failure probability.

  16. Vehicle systems design optimization study

    SciTech Connect

    Gilmour, J. L.

    1980-04-01

    The optimization of an electric vehicle layout requires a weight distribution in the range of 53/47 to 62/38 in order to assure dynamic handling characteristics comparable to current production internal combustion engine vehicles. It is possible to achieve this goal and also provide passenger and cargo space comparable to a selected current production sub-compact car either in a unique new design or by utilizing the production vehicle as a base. Necessary modification of the base vehicle can be accomplished without major modification of the structure or running gear. As long as batteries are as heavy and require as much space as they currently do, they must be divided into two packages - one at front under the hood and a second at the rear under the cargo area - in order to achieve the desired weight distribution. The weight distribution criteria requires the placement of batteries at the front of the vehicle even when the central tunnel is used for the location of some batteries. The optimum layout has a front motor and front wheel drive. This configuration provides the optimum vehicle dynamic handling characteristics and the maximum passsenger and cargo space for a given size vehicle.

  17. Integrated multidisciplinary design optimization of rotorcraft

    NASA Technical Reports Server (NTRS)

    Adelman, Howard M.; Mantay, Wayne R.

    1989-01-01

    The NASA/Army research plan for developing the logic elements for helicopter rotor design optimization by integrating appropriate disciplines and accounting for important interactions among the disciplines is discussed. The paper describes the optimization formulation in terms of the objective function, design variables, and constraints. The analysis aspects are discussed, and an initial effort at defining the interdisciplinary coupling is summarized. Results are presented on the achievements made in the rotor aerodynamic performance optimization for minimum hover horsepower, rotor dynamic optimization for vibration reduction, rotor structural optimization for minimum weight, and integrated aerodynamic load/dynamics optimization for minimum vibration and weight.

  18. Comparison of optimal design methods in inverse problems

    NASA Astrophysics Data System (ADS)

    Banks, H. T.; Holm, K.; Kappel, F.

    2011-07-01

    Typical optimal design methods for inverse or parameter estimation problems are designed to choose optimal sampling distributions through minimization of a specific cost function related to the resulting error in parameter estimates. It is hoped that the inverse problem will produce parameter estimates with increased accuracy using data collected according to the optimal sampling distribution. Here we formulate the classical optimal design problem in the context of general optimization problems over distributions of sampling times. We present a new Prohorov metric-based theoretical framework that permits one to treat succinctly and rigorously any optimal design criteria based on the Fisher information matrix. A fundamental approximation theory is also included in this framework. A new optimal design, SE-optimal design (standard error optimal design), is then introduced in the context of this framework. We compare this new design criterion with the more traditional D-optimal and E-optimal designs. The optimal sampling distributions from each design are used to compute and compare standard errors; the standard errors for parameters are computed using asymptotic theory or bootstrapping and the optimal mesh. We use three examples to illustrate ideas: the Verhulst-Pearl logistic population model (Banks H T and Tran H T 2009 Mathematical and Experimental Modeling of Physical and Biological Processes (Boca Raton, FL: Chapman and Hall/CRC)), the standard harmonic oscillator model (Banks H T and Tran H T 2009) and a popular glucose regulation model (Bergman R N, Ider Y Z, Bowden C R and Cobelli C 1979 Am. J. Physiol. 236 E667-77 De Gaetano A and Arino O 2000 J. Math. Biol. 40 136-68 Toffolo G, Bergman R N, Finegood D T, Bowden C R and Cobelli C 1980 Diabetes 29 979-90).

  19. Flat-plate photovoltaic array design optimization

    NASA Technical Reports Server (NTRS)

    Ross, R. G., Jr.

    1980-01-01

    An analysis is presented which integrates the results of specific studies in the areas of photovoltaic structural design optimization, optimization of array series/parallel circuit design, thermal design optimization, and optimization of environmental protection features. The analysis is based on minimizing the total photovoltaic system life-cycle energy cost including repair and replacement of failed cells and modules. This approach is shown to be a useful technique for array optimization, particularly when time-dependent parameters such as array degradation and maintenance are involved.

  20. Design Optimization of Composite Structures under Uncertainty

    NASA Technical Reports Server (NTRS)

    Haftka, Raphael T.

    2003-01-01

    Design optimization under uncertainty is computationally expensive and is also challenging in terms of alternative formulation. The work under the grant focused on developing methods for design against uncertainty that are applicable to composite structural design with emphasis on response surface techniques. Applications included design of stiffened composite plates for improved damage tolerance, the use of response surfaces for fitting weights obtained by structural optimization, and simultaneous design of structure and inspection periods for fail-safe structures.

  1. A Bayesian experimental design approach to structural health monitoring

    SciTech Connect

    Farrar, Charles; Flynn, Eric; Todd, Michael

    2010-01-01

    Optimal system design for SHM involves two primarily challenges. The first is the derivation of a proper performance function for a given system design. The second is the development of an efficient optimization algorithm for choosing a design that maximizes, or nearly maximizes the performance function. In this paper we will outline how an SHM practitioner can construct the proper performance function by casting the entire design problem into a framework of Bayesian experimental design. The approach demonstrates how the design problem necessarily ties together all steps of the SHM process.

  2. Rationally reduced libraries for combinatorial pathway optimization minimizing experimental effort.

    PubMed

    Jeschek, Markus; Gerngross, Daniel; Panke, Sven

    2016-01-01

    Rational flux design in metabolic engineering approaches remains difficult since important pathway information is frequently not available. Therefore empirical methods are applied that randomly change absolute and relative pathway enzyme levels and subsequently screen for variants with improved performance. However, screening is often limited on the analytical side, generating a strong incentive to construct small but smart libraries. Here we introduce RedLibs (Reduced Libraries), an algorithm that allows for the rational design of smart combinatorial libraries for pathway optimization thereby minimizing the use of experimental resources. We demonstrate the utility of RedLibs for the design of ribosome-binding site libraries by in silico and in vivo screening with fluorescent proteins and perform a simple two-step optimization of the product selectivity in the branched multistep pathway for violacein biosynthesis, indicating a general applicability for the algorithm and the proposed heuristics. We expect that RedLibs will substantially simplify the refactoring of synthetic metabolic pathways. PMID:27029461

  3. Rationally reduced libraries for combinatorial pathway optimization minimizing experimental effort

    PubMed Central

    Jeschek, Markus; Gerngross, Daniel; Panke, Sven

    2016-01-01

    Rational flux design in metabolic engineering approaches remains difficult since important pathway information is frequently not available. Therefore empirical methods are applied that randomly change absolute and relative pathway enzyme levels and subsequently screen for variants with improved performance. However, screening is often limited on the analytical side, generating a strong incentive to construct small but smart libraries. Here we introduce RedLibs (Reduced Libraries), an algorithm that allows for the rational design of smart combinatorial libraries for pathway optimization thereby minimizing the use of experimental resources. We demonstrate the utility of RedLibs for the design of ribosome-binding site libraries by in silico and in vivo screening with fluorescent proteins and perform a simple two-step optimization of the product selectivity in the branched multistep pathway for violacein biosynthesis, indicating a general applicability for the algorithm and the proposed heuristics. We expect that RedLibs will substantially simplify the refactoring of synthetic metabolic pathways. PMID:27029461

  4. Genetic algorithms for the construction of D-optimal designs

    SciTech Connect

    Heredia-Langner, Alejandro; Carlyle, W M.; Montgomery, D C.; Borror, Connie M.; Runger, George C.

    2003-01-01

    Computer-generated designs are useful for situations where standard factorial, fractional factorial or response surface designs cannot be easily employed. Alphabetically-optimal designs are the most widely used type of computer-generated designs, and of these, the D-optimal (or D-efficient) class of designs are extremely popular. D-optimal designs are usually constructed by algorithms that sequentially add and delete points from a potential design based using a candidate set of points spaced over the region of interest. We present a technique to generate D-efficient designs using genetic algorithms (GA). This approach eliminates the need to explicitly consider a candidate set of experimental points and it can handle highly constrained regions while maintaining a level of performance comparable to more traditional design construction techniques.

  5. Solar-Diesel Hybrid Power System Optimization and Experimental Validation

    NASA Astrophysics Data System (ADS)

    Jacobus, Headley Stewart

    As of 2008 1.46 billion people, or 22 percent of the World's population, were without electricity. Many of these people live in remote areas where decentralized generation is the only method of electrification. Most mini-grids are powered by diesel generators, but new hybrid power systems are becoming a reliable method to incorporate renewable energy while also reducing total system cost. This thesis quantifies the measurable Operational Costs for an experimental hybrid power system in Sierra Leone. Two software programs, Hybrid2 and HOMER, are used during the system design and subsequent analysis. Experimental data from the installed system is used to validate the two programs and to quantify the savings created by each component within the hybrid system. This thesis bridges the gap between design optimization studies that frequently lack subsequent validation and experimental hybrid system performance studies.

  6. Optimal pupil design for confocal microscopy

    NASA Astrophysics Data System (ADS)

    Patel, Yogesh G.; Rajadhyaksha, Milind; DiMarzio, Charles A.

    2010-02-01

    Confocal reflectance microscopy may enable screening and diagnosis of skin cancers noninvasively and in real-time, as an adjunct to biopsy and pathology. Current instruments are large, complex, and expensive. A simpler, confocal line-scanning microscope may accelerate the translation of confocal microscopy in clinical and surgical dermatology. A confocal reflectance microscope may use a beamsplitter, transmitting and detecting through the pupil, or a divided pupil, or theta configuration, with half used for transmission and half for detection. The divided pupil may offer better sectioning and contrast. We present a Fourier optics model and compare the on-axis irradiance of a confocal point-scanning microscope in both pupil configurations, optimizing the profile of a Gaussian beam in a circular or semicircular aperture. We repeat both calculations with a cylindrical lens which focuses the source to a line. The variable parameter is the fillfactor, h, the ratio of the 1/e2 diameter of the Gaussian beam to the diameter of the full aperture. The optimal values of h, for point scanning are 0.90 (full) and 0.66 for the half-aperture. For line-scanning, the fill-factors are 1.02 (full) and 0.52 (half). Additional parameters to consider are the optimal location of the point-source beam in the divided-pupil configuration, the optimal line width for the line-source, and the width of the aperture in the divided-pupil configuration. Additional figures of merit are field-of-view and sectioning. Use of optimal designs is critical in comparing the experimental performance of the different configurations.

  7. Topology Optimization for Architected Materials Design

    NASA Astrophysics Data System (ADS)

    Osanov, Mikhail; Guest, James K.

    2016-07-01

    Advanced manufacturing processes provide a tremendous opportunity to fabricate materials with precisely defined architectures. To fully leverage these capabilities, however, materials architectures must be optimally designed according to the target application, base material used, and specifics of the fabrication process. Computational topology optimization offers a systematic, mathematically driven framework for navigating this new design challenge. The design problem is posed and solved formally as an optimization problem with unit cell and upscaling mechanics embedded within this formulation. This article briefly reviews the key requirements to apply topology optimization to materials architecture design and discusses several fundamental findings related to optimization of elastic, thermal, and fluidic properties in periodic materials. Emerging areas related to topology optimization for manufacturability and manufacturing variations, nonlinear mechanics, and multiscale design are also discussed.

  8. Control design for the SERC experimental testbeds

    NASA Technical Reports Server (NTRS)

    Jacques, Robert; Blackwood, Gary; Macmartin, Douglas G.; How, Jonathan; Anderson, Eric

    1992-01-01

    Viewgraphs on control design for the Space Engineering Research Center experimental testbeds are presented. Topics covered include: SISO control design and results; sensor and actuator location; model identification; control design; experimental results; preliminary LAC experimental results; active vibration isolation problem statement; base flexibility coupling into isolation feedback loop; cantilever beam testbed; and closed loop results.

  9. Design optimization of a portable, micro-hydrokinetic turbine

    NASA Astrophysics Data System (ADS)

    Schleicher, W. Chris

    Marine and hydrokinetic (MHK) technology is a growing field that encompasses many different types of turbomachinery that operate on the kinetic energy of water. Micro hydrokinetics are a subset of MHK technology comprised of units designed to produce less than 100 kW of power. A propeller-type hydrokinetic turbine is investigated as a solution for a portable micro-hydrokinetic turbine with the needs of the United States Marine Corps in mind, as well as future commercial applications. This dissertation investigates using a response surface optimization methodology to create optimal turbine blade designs under many operating conditions. The field of hydrokinetics is introduced. The finite volume method is used to solve the Reynolds-Averaged Navier-Stokes equations with the k ω Shear Stress Transport model, for different propeller-type hydrokinetic turbines. The adaptive response surface optimization methodology is introduced as related to hydrokinetic turbines, and is benchmarked with complex algebraic functions. The optimization method is further studied to characterize the size of the experimental design on its ability to find optimum conditions. It was found that a large deviation between experimental design points was preferential. Different propeller hydrokinetic turbines were designed and compared with other forms of turbomachinery. It was found that the rapid simulations usually under predict performance compare to the refined simulations, and for some other designs it drastically over predicted performance. The optimization method was used to optimize a modular pump-turbine, verifying that the optimization work for other hydro turbine designs.

  10. Integrated multidisciplinary design optimization of rotorcraft

    NASA Technical Reports Server (NTRS)

    Adelman, Howard M.; Mantay, Wayne R.

    1989-01-01

    The NASA/Army research plan for developing the logic elements for helicopter rotor design optimization by integrating appropriate disciplines and accounting for important interactions among the disciplines is discussed. The optimization formulation is described in terms of the objective function, design variables, and constraints. The analysis aspects are discussed, and an initial effort at defining the interdisciplinary coupling is summarized. Results are presented on the achievements made in the rotor dynamic optimization for vibration reduction, rotor structural optimization for minimum weight, and integrated aerodynamic load/dynamics optimization for minimum vibration and weight.

  11. Optimal Designs of Staggered Dean Vortex Micromixers

    PubMed Central

    Chen, Jyh Jian; Chen, Chun Huei; Shie, Shian Ruei

    2011-01-01

    A novel parallel laminar micromixer with a two-dimensional staggered Dean Vortex micromixer is optimized and fabricated in our study. Dean vortices induced by centrifugal forces in curved rectangular channels cause fluids to produce secondary flows. The split-and-recombination (SAR) structures of the flow channels and the impinging effects result in the reduction of the diffusion distance of two fluids. Three different designs of a curved channel micromixer are introduced to evaluate the mixing performance of the designed micromixer. Mixing performances are demonstrated by means of a pH indicator using an optical microscope and fluorescent particles via a confocal microscope at different flow rates corresponding to Reynolds numbers (Re) ranging from 0.5 to 50. The comparison between the experimental data and numerical results shows a very reasonable agreement. At a Re of 50, the mixing length at the sixth segment, corresponding to the downstream distance of 21.0 mm, can be achieved in a distance 4 times shorter than when the Re equals 1. An optimization of this micromixer is performed with two geometric parameters. These are the angle between the lines from the center to two intersections of two consecutive curved channels, θ, and the angle between two lines of the centers of three consecutive curved channels, ϕ. It can be found that the maximal mixing index is related to the maximal value of the sum of θ and ϕ, which is equal to 139.82°. PMID:21747691

  12. Optimality of a Fully Stressed Design

    NASA Technical Reports Server (NTRS)

    Patnaik, Surya N.; Hopkins, Dale A.

    1998-01-01

    For a truss a fully stressed state is reached and when all its members are utilized to their full strength capacity. Historically, engineers considered such a design optimum. But recently this optimality has been questioned, especially since the weight of the structure is not explicitly used in fully stressed design calculations. This paper examines optimality of the full stressed design (FSD) with analytical and graphical illustrations. Solutions for a set of examples obtained by using the FSD method and optimization methods numerically confirm the optimality of the FSD. The FSD, which can be obtained with a small amount of calculation, can be extended to displacement constraints and to nontruss-type structures.

  13. Design optimization studies using COSMIC NASTRAN

    NASA Technical Reports Server (NTRS)

    Pitrof, Stephen M.; Bharatram, G.; Venkayya, Vipperla B.

    1993-01-01

    The purpose of this study is to create, test and document a procedure to integrate mathematical optimization algorithms with COSMIC NASTRAN. This procedure is very important to structural design engineers who wish to capitalize on optimization methods to ensure that their design is optimized for its intended application. The OPTNAST computer program was created to link NASTRAN and design optimization codes into one package. This implementation was tested using two truss structure models and optimizing their designs for minimum weight, subject to multiple loading conditions and displacement and stress constraints. However, the process is generalized so that an engineer could design other types of elements by adding to or modifying some parts of the code.

  14. Optimal multiobjective design of digital filters using spiral optimization technique.

    PubMed

    Ouadi, Abderrahmane; Bentarzi, Hamid; Recioui, Abdelmadjid

    2013-01-01

    The multiobjective design of digital filters using spiral optimization technique is considered in this paper. This new optimization tool is a metaheuristic technique inspired by the dynamics of spirals. It is characterized by its robustness, immunity to local optima trapping, relative fast convergence and ease of implementation. The objectives of filter design include matching some desired frequency response while having minimum linear phase; hence, reducing the time response. The results demonstrate that the proposed problem solving approach blended with the use of the spiral optimization technique produced filters which fulfill the desired characteristics and are of practical use. PMID:24083108

  15. Web-based tools for finding optimal designs in biomedical studies

    PubMed Central

    Wong, Weng Kee

    2013-01-01

    Experimental costs are rising and applications of optimal design ideas are increasingly applied in many disciplines. However, the theory for constructing optimal designs can be esoteric and its implementation can be difficult. To help practitioners have easier access to optimal designs and better appreciate design issues, we present a web site at http://optimal-design.biostat.ucla.edu/optimal/ capable of generating different types of tailor-made optimal designs for popular models in the biological sciences. This site also evaluates various efficiencies of a user-specified design and so enables practitioners to appreciate robustness properties of the design before implementation. PMID:23806678

  16. Optimization, an Important Stage of Engineering Design

    ERIC Educational Resources Information Center

    Kelley, Todd R.

    2010-01-01

    A number of leaders in technology education have indicated that a major difference between the technological design process and the engineering design process is analysis and optimization. The analysis stage of the engineering design process is when mathematical models and scientific principles are employed to help the designer predict design…

  17. Computational design optimization for microfluidic magnetophoresis

    PubMed Central

    Plouffe, Brian D.; Lewis, Laura H.; Murthy, Shashi K.

    2011-01-01

    Current macro- and microfluidic approaches for the isolation of mammalian cells are limited in both efficiency and purity. In order to design a robust platform for the enumeration of a target cell population, high collection efficiencies are required. Additionally, the ability to isolate pure populations with minimal biological perturbation and efficient off-chip recovery will enable subcellular analyses of these cells for applications in personalized medicine. Here, a rational design approach for a simple and efficient device that isolates target cell populations via magnetic tagging is presented. In this work, two magnetophoretic microfluidic device designs are described, with optimized dimensions and operating conditions determined from a force balance equation that considers two dominant and opposing driving forces exerted on a magnetic-particle-tagged cell, namely, magnetic and viscous drag. Quantitative design criteria for an electromagnetic field displacement-based approach are presented, wherein target cells labeled with commercial magnetic microparticles flowing in a central sample stream are shifted laterally into a collection stream. Furthermore, the final device design is constrained to fit on standard rectangular glass coverslip (60 (L)×24 (W)×0.15 (H) mm3) to accommodate small sample volume and point-of-care design considerations. The anticipated performance of the device is examined via a parametric analysis of several key variables within the model. It is observed that minimal currents (<500 mA) are required to generate magnetic fields sufficient to separate cells from the sample streams flowing at rate as high as 7 ml∕h, comparable to the performance of current state-of-the-art magnet-activated cell sorting systems currently used in clinical settings. Experimental validation of the presented model illustrates that a device designed according to the derived rational optimization can effectively isolate (∼100%) a magnetic-particle-tagged cell

  18. Computational design optimization for microfluidic magnetophoresis.

    PubMed

    Plouffe, Brian D; Lewis, Laura H; Murthy, Shashi K

    2011-01-01

    Current macro- and microfluidic approaches for the isolation of mammalian cells are limited in both efficiency and purity. In order to design a robust platform for the enumeration of a target cell population, high collection efficiencies are required. Additionally, the ability to isolate pure populations with minimal biological perturbation and efficient off-chip recovery will enable subcellular analyses of these cells for applications in personalized medicine. Here, a rational design approach for a simple and efficient device that isolates target cell populations via magnetic tagging is presented. In this work, two magnetophoretic microfluidic device designs are described, with optimized dimensions and operating conditions determined from a force balance equation that considers two dominant and opposing driving forces exerted on a magnetic-particle-tagged cell, namely, magnetic and viscous drag. Quantitative design criteria for an electromagnetic field displacement-based approach are presented, wherein target cells labeled with commercial magnetic microparticles flowing in a central sample stream are shifted laterally into a collection stream. Furthermore, the final device design is constrained to fit on standard rectangular glass coverslip (60 (L)×24 (W)×0.15 (H) mm(3)) to accommodate small sample volume and point-of-care design considerations. The anticipated performance of the device is examined via a parametric analysis of several key variables within the model. It is observed that minimal currents (<500 mA) are required to generate magnetic fields sufficient to separate cells from the sample streams flowing at rate as high as 7 ml∕h, comparable to the performance of current state-of-the-art magnet-activated cell sorting systems currently used in clinical settings. Experimental validation of the presented model illustrates that a device designed according to the derived rational optimization can effectively isolate (∼100%) a magnetic-particle-tagged cell

  19. Singularities in Optimal Structural Design

    NASA Technical Reports Server (NTRS)

    Patnaik, S. N.; Guptill, J. D.; Berke, L.

    1992-01-01

    Singularity conditions that arise during structural optimization can seriously degrade the performance of the optimizer. The singularities are intrinsic to the formulation of the structural optimization problem and are not associated with the method of analysis. Certain conditions that give rise to singularities have been identified in earlier papers, encompassing the entire structure. Further examination revealed more complex sets of conditions in which singularities occur. Some of these singularities are local in nature, being associated with only a segment of the structure. Moreover, the likelihood that one of these local singularities may arise during an optimization procedure can be much greater than that of the global singularity identified earlier. Examples are provided of these additional forms of singularities. A framework is also given in which these singularities can be recognized. In particular, the singularities can be identified by examination of the stress displacement relations along with the compatibility conditions and/or the displacement stress relations derived in the integrated force method of structural analysis.

  20. Singularities in optimal structural design

    NASA Technical Reports Server (NTRS)

    Patnaik, S. N.; Guptill, J. D.; Berke, L.

    1992-01-01

    Singularity conditions that arise during structural optimization can seriously degrade the performance of the optimizer. The singularities are intrinsic to the formulation of the structural optimization problem and are not associated with the method of analysis. Certain conditions that give rise to singularities have been identified in earlier papers, encompassing the entire structure. Further examination revealed more complex sets of conditions in which singularities occur. Some of these singularities are local in nature, being associated with only a segment of the structure. Moreover, the likelihood that one of these local singularities may arise during an optimization procedure can be much greater than that of the global singularity identified earlier. Examples are provided of these additional forms of singularities. A framework is also given in which these singularities can be recognized. In particular, the singularities can be identified by examination of the stress displacement relations along with the compatibility conditions and/or the displacement stress relations derived in the integrated force method of structural analysis.

  1. Multidisciplinary Optimization Methods for Aircraft Preliminary Design

    NASA Technical Reports Server (NTRS)

    Kroo, Ilan; Altus, Steve; Braun, Robert; Gage, Peter; Sobieski, Ian

    1994-01-01

    This paper describes a research program aimed at improved methods for multidisciplinary design and optimization of large-scale aeronautical systems. The research involves new approaches to system decomposition, interdisciplinary communication, and methods of exploiting coarse-grained parallelism for analysis and optimization. A new architecture, that involves a tight coupling between optimization and analysis, is intended to improve efficiency while simplifying the structure of multidisciplinary, computation-intensive design problems involving many analysis disciplines and perhaps hundreds of design variables. Work in two areas is described here: system decomposition using compatibility constraints to simplify the analysis structure and take advantage of coarse-grained parallelism; and collaborative optimization, a decomposition of the optimization process to permit parallel design and to simplify interdisciplinary communication requirements.

  2. Magnetic design optimization using variable metrics

    SciTech Connect

    Davey, K.R.

    1995-11-01

    The optimal design of a magnet assembly for a magnetic levitated train is approached using a three step process. First, the key parameters within the objective performance index are computed for the variation range of the problem. Second, the performance index is fitted to a smooth polynomial involving products of the powers of all variables. Third, a constrained optimization algorithm is employed to predict the optimal choice of the variables. An assessment of the integrity of the optimization program is obtained by comparing the final optimized solution with that predicted by the field analysis in the final configuration. Additional field analysis is recommended around the final solution to fine tune the solution.

  3. Cold Climates Heat Pump Design Optimization

    SciTech Connect

    Abdelaziz, Omar; Shen, Bo

    2012-01-01

    Heat pumps provide an efficient heating method; however they suffer from sever capacity and performance degradation at low ambient conditions. This has deterred market penetration in cold climates. There is a continuing effort to find an efficient air source cold climate heat pump that maintains acceptable capacity and performance at low ambient conditions. Systematic optimization techniques provide a reliable approach for the design of such systems. This paper presents a step-by-step approach for the design optimization of cold climate heat pumps. We first start by describing the optimization problem: objective function, constraints, and design space. Then we illustrate how to perform this design optimization using an open source publically available optimization toolbox. The response of the heat pump design was evaluated using a validated component based vapor compression model. This model was treated as a black box model within the optimization framework. Optimum designs for different system configurations are presented. These optimum results were further analyzed to understand the performance tradeoff and selection criteria. The paper ends with a discussion on the use of systematic optimization for the cold climate heat pump design.

  4. Exponential approximations in optimal design

    NASA Technical Reports Server (NTRS)

    Belegundu, A. D.; Rajan, S. D.; Rajgopal, J.

    1990-01-01

    One-point and two-point exponential functions have been developed and proved to be very effective approximations of structural response. The exponential has been compared to the linear, reciprocal and quadratic fit methods. Four test problems in structural analysis have been selected. The use of such approximations is attractive in structural optimization to reduce the numbers of exact analyses which involve computationally expensive finite element analysis.

  5. Optimization criteria for SRAM design: lithography contribution

    NASA Astrophysics Data System (ADS)

    Cole, Daniel C.; Bula, Orest; Conrad, Edward W.; Coops, Daniel S.; Leipold, William C.; Mann, Randy W.; Oppold, Jeffrey H.

    1999-07-01

    Here we discuss the use of well calibrated resist and etch bias models, in conjunction with a fast microlithography aerial image simulator, to predict and 'optimize' the printed shapes through all critical levels in a dense SRAM design. Our key emphasis here is on 'optimization criteria', namely, having achieved good predictability for printability with lithography models, how to use this capability in conjunction of best electrical performance, yield, and density. The key lithography/design optimization issues discussed here are: (1) tightening of gate width variation by reducing spatial curvature in the source and drain regions, (2) achieving sufficient contact areas, (3) maximizing process window for overlay, (4) reducing leakage mechanisms by reducing contributions of stress and strain due to the printed shape of oxide isolation regions, (5) examining topological differences in design during the optimization process, (6) accounting for mask corner rounding, and (7) designing for scalability to smaller dimensions to achieve optical design reusability issues without hardware.

  6. D-optimal experimental design coupled with parallel factor analysis 2 decomposition a useful tool in the determination of triazines in oranges by programmed temperature vaporization-gas chromatography-mass spectrometry when using dispersive-solid phase extraction.

    PubMed

    Herrero, A; Ortiz, M C; Sarabia, L A

    2013-05-01

    The determination of triazines in oranges using a GC-MS system coupled to a programmed temperature vaporizer (PTV) inlet in the context of legislation is performed. Both pretreatment (using a Quick Easy Cheap Effective Rugged and Safe (QuEChERS) procedure) and injection steps are optimized using D-optimal experimental designs for reducing the experimental effort. The relative dirty extracts obtained and the elution time shifts make it necessary to use a PARAFAC2 decomposition to solve these two usual problems in the chromatographic determinations. The "second-order advantage" of the PARAFAC2 decomposition allows unequivocal identification according to document SANCO/12495/2011 (taking into account the tolerances for relative retention time and the relative abundance for the diagnostic ions), avoiding false negatives even in the presence of unknown co-eluents. The detection limits (CCα) found, from 0.51 to 1.05μgkg(-1), are far below the maximum residue levels (MRLs) established by the European Union for simazine, atrazine, terbuthylazine, ametryn, simetryn, prometryn and terbutryn in oranges. No MRL violations were found in the commercial oranges analyzed. PMID:23522618

  7. Optimization design of electromagnetic shielding composites

    NASA Astrophysics Data System (ADS)

    Qu, Zhaoming; Wang, Qingguo; Qin, Siliang; Hu, Xiaofeng

    2013-03-01

    The effective electromagnetic parameters physical model of composites and prediction formulas of composites' shielding effectiveness and reflectivity were derived based on micromechanics, variational principle and electromagnetic wave transmission theory. The multi-objective optimization design of multilayer composites was carried out using genetic algorithm. The optimized results indicate that material parameter proportioning of biggest absorption ability can be acquired under the condition of the minimum shielding effectiveness can be satisfied in certain frequency band. The validity of optimization design model was verified and the scheme has certain theoretical value and directive significance to the design of high efficiency shielding composites.

  8. Material design using surrogate optimization algorithm

    NASA Astrophysics Data System (ADS)

    Khadke, Kunal R.

    Nanocomposite ceramics have been widely studied in order to tailor desired properties at high temperatures. Methodologies for development of material design are still under effect . While finite element modeling (FEM) provides significant insight on material behavior, few design researchers have addressed the design paradox that accompanies this rapid design space expansion. A surrogate optimization model management framework has been proposed to make this design process tractable. In the surrogate optimization material design tool, the analysis cost is reduced by performing simulations on the surrogate model instead of high density finite element model. The methodology is incorporated to find the optimal number of silicon carbide (SiC) particles, in a silicon-nitride Si3N 4 composite with maximum fracture energy [2]. Along with a deterministic optimization algorithm, model uncertainties have also been considered with the use of robust design optimization (RDO) method ensuring a design of minimum sensitivity to changes in the parameters. These methodologies applied to nanocomposites design have a signicant impact on cost and design cycle time reduced.

  9. Optimization methods applied to hybrid vehicle design

    NASA Technical Reports Server (NTRS)

    Donoghue, J. F.; Burghart, J. H.

    1983-01-01

    The use of optimization methods as an effective design tool in the design of hybrid vehicle propulsion systems is demonstrated. Optimization techniques were used to select values for three design parameters (battery weight, heat engine power rating and power split between the two on-board energy sources) such that various measures of vehicle performance (acquisition cost, life cycle cost and petroleum consumption) were optimized. The apporach produced designs which were often significant improvements over hybrid designs already reported on in the literature. The principal conclusions are as follows. First, it was found that the strategy used to split the required power between the two on-board energy sources can have a significant effect on life cycle cost and petroleum consumption. Second, the optimization program should be constructed so that performance measures and design variables can be easily changed. Third, the vehicle simulation program has a significant effect on the computer run time of the overall optimization program; run time can be significantly reduced by proper design of the types of trips the vehicle takes in a one year period. Fourth, care must be taken in designing the cost and constraint expressions which are used in the optimization so that they are relatively smooth functions of the design variables. Fifth, proper handling of constraints on battery weight and heat engine rating, variables which must be large enough to meet power demands, is particularly important for the success of an optimization study. Finally, the principal conclusion is that optimization methods provide a practical tool for carrying out the design of a hybrid vehicle propulsion system.

  10. Design Optimization Programmable Calculators versus Campus Computers.

    ERIC Educational Resources Information Center

    Savage, Michael

    1982-01-01

    A hypothetical design optimization problem and technical information on the three design parameters are presented. Although this nested iteration problem can be solved on a computer (flow diagram provided), this article suggests that several hand held calculators can be used to perform the same design iteration. (SK)

  11. Optimality criteria solution strategies in multiple constraint design optimization

    NASA Technical Reports Server (NTRS)

    Levy, R.; Parzynski, W.

    1981-01-01

    Procedures and solution strategies are described to solve the conventional structural optimization problem using the Lagrange multiplier technique. The multipliers, obtained through solution of an auxiliary nonlinear optimization problem, lead to optimality criteria to determine the design variables. It is shown that this procedure is essentially equivalent to an alternative formulation using a dual method Lagrangian function objective. Although mathematical formulations are straight-forward, successful applications and computational efficiency depend upon execution procedure strategies. Strategies examined, with application examples, include selection of active constraints, move limits, line search procedures, and side constraint boundaries.

  12. Interaction Prediction Optimization in Multidisciplinary Design Optimization Problems

    PubMed Central

    Zhang, Xiaoling; Huang, Hong-Zhong; Wang, Zhonglai; Xu, Huanwei

    2014-01-01

    The distributed strategy of Collaborative Optimization (CO) is suitable for large-scale engineering systems. However, it is hard for CO to converge when there is a high level coupled dimension. Furthermore, the discipline objectives cannot be considered in each discipline optimization problem. In this paper, one large-scale systems control strategy, the interaction prediction method (IPM), is introduced to enhance CO. IPM is utilized for controlling subsystems and coordinating the produce process in large-scale systems originally. We combine the strategy of IPM with CO and propose the Interaction Prediction Optimization (IPO) method to solve MDO problems. As a hierarchical strategy, there are a system level and a subsystem level in IPO. The interaction design variables (including shared design variables and linking design variables) are operated at the system level and assigned to the subsystem level as design parameters. Each discipline objective is considered and optimized at the subsystem level simultaneously. The values of design variables are transported between system level and subsystem level. The compatibility constraints are replaced with the enhanced compatibility constraints to reduce the dimension of design variables in compatibility constraints. Two examples are presented to show the potential application of IPO for MDO. PMID:24744685

  13. DESIGN AND OPTIMIZATION OF A REFRIGERATION SYSTEM

    EPA Science Inventory

    The paper discusses the design and optimization of a refrigeration system, using a mathematical model of a refrigeration system modified to allow its use with the optimization program. he model was developed using only algebraic equations so that it could be used with the optimiz...

  14. Turbomachinery Airfoil Design Optimization Using Differential Evolution

    NASA Technical Reports Server (NTRS)

    Madavan, Nateri K.; Biegel, Bryan (Technical Monitor)

    2002-01-01

    An aerodynamic design optimization procedure that is based on a evolutionary algorithm known at Differential Evolution is described. Differential Evolution is a simple, fast, and robust evolutionary strategy that has been proven effective in determining the global optimum for several difficult optimization problems, including highly nonlinear systems with discontinuities and multiple local optima. The method is combined with a Navier-Stokes solver that evaluates the various intermediate designs and provides inputs to the optimization procedure. An efficient constraint handling mechanism is also incorporated. Results are presented for the inverse design of a turbine airfoil from a modern jet engine and compared to earlier methods. The capability of the method to search large design spaces and obtain the optimal airfoils in an automatic fashion is demonstrated. Substantial reductions in the overall computing time requirements are achieved by using the algorithm in conjunction with neural networks.

  15. Turbomachinery Airfoil Design Optimization Using Differential Evolution

    NASA Technical Reports Server (NTRS)

    Madavan, Nateri K.; Biegel, Bryan A. (Technical Monitor)

    2002-01-01

    An aerodynamic design optimization procedure that is based on a evolutionary algorithm known at Differential Evolution is described. Differential Evolution is a simple, fast, and robust evolutionary strategy that has been proven effective in determining the global optimum for several difficult optimization problems, including highly nonlinear systems with discontinuities and multiple local optima. The method is combined with a Navier-Stokes solver that evaluates the various intermediate designs and provides inputs to the optimization procedure. An efficient constraint handling mechanism is also incorporated. Results are presented for the inverse design of a turbine airfoil from a modern jet engine. The capability of the method to search large design spaces and obtain the optimal airfoils in an automatic fashion is demonstrated. Substantial reductions in the overall computing time requirements are achieved by using the algorithm in conjunction with neural networks.

  16. Experimental design for improved ceramic processing, emphasizing the Taguchi Method

    SciTech Connect

    Weiser, M.W. . Mechanical Engineering Dept.); Fong, K.B. )

    1993-12-01

    Ceramic processing often requires substantial experimentation to produce acceptable product quality and performance. This is a consequence of ceramic processes depending upon a multitude of factors, some of which can be controlled and others that are beyond the control of the manufacturer. Statistical design of experiments is a procedure that allows quick, economical, and accurate evaluation of processes and products that depend upon several variables. Designed experiments are sets of tests in which the variables are adjusted methodically. A well-designed experiment yields unambiguous results at minimal cost. A poorly designed experiment may reveal little information of value even with complex analysis, wasting valuable time and resources. This article will review the most common experimental designs. This will include both nonstatistical designs and the much more powerful statistical experimental designs. The Taguchi Method developed by Grenichi Taguchi will be discussed in some detail. The Taguchi method, based upon fractional factorial experiments, is a powerful tool for optimizing product and process performance.

  17. Optimal design of artificial reefs for sturgeon

    NASA Astrophysics Data System (ADS)

    Yarbrough, Cody; Cotel, Aline; Kleinheksel, Abby

    2015-11-01

    The Detroit River, part of a busy corridor between Lakes Huron and Erie, was extensively modified to create deep shipping channels, resulting in a loss of spawning habitat for lake sturgeon and other native fish (Caswell et al. 2004, Bennion and Manny 2011). Under the U.S.- Canada Great Lakes Water Quality Agreement, there are remediation plans to construct fish spawning reefs to help with historic habitat losses and degraded fish populations, specifically sturgeon. To determine optimal reef design, experimental work has been undertaken. Different sizes and shapes of reefs are tested for a given set of physical conditions, such as flow depth and flow velocity, matching the relevant dimensionless parameters dominating the flow physics. The physical conditions are matched with the natural conditions encountered in the Detroit River. Using Particle Image Velocimetry, Acoustic Doppler Velocimetry and dye studies, flow structures, vorticity and velocity gradients at selected locations have been identified and quantified to allow comparison with field observations and numerical model results. Preliminary results are helping identify the design features to be implemented in the next phase of reef construction. Sponsored by NOAA.

  18. A More Rigorous Quasi-Experimental Alternative to the One-Group Pretest-Posttest Design.

    ERIC Educational Resources Information Center

    Johnson, Craig W.

    1986-01-01

    A simple quasi-experimental design is described which may have utility in a variety of applied and laboratory research settings where ordinarily the one-group pretest-posttest pre-experimental design might otherwise be the procedure of choice. The design approaches the internal validity of true experimental designs while optimizing external…

  19. Multidisciplinary design optimization using genetic algorithms

    NASA Astrophysics Data System (ADS)

    Unal, Resit

    1994-12-01

    Multidisciplinary design optimization (MDO) is an important step in the conceptual design and evaluation of launch vehicles since it can have a significant impact on performance and life cycle cost. The objective is to search the system design space to determine values of design variables that optimize the performance characteristic subject to system constraints. Gradient-based optimization routines have been used extensively for aerospace design optimization. However, one limitation of gradient based optimizers is their need for gradient information. Therefore, design problems which include discrete variables can not be studied. Such problems are common in launch vehicle design. For example, the number of engines and material choices must be integer values or assume only a few discrete values. In this study, genetic algorithms are investigated as an approach to MDO problems involving discrete variables and discontinuous domains. Optimization by genetic algorithms (GA) uses a search procedure which is fundamentally different from those gradient based methods. Genetic algorithms seek to find good solutions in an efficient and timely manner rather than finding the best solution. GA are designed to mimic evolutionary selection. A population of candidate designs is evaluated at each iteration, and each individual's probability of reproduction (existence in the next generation) depends on its fitness value (related to the value of the objective function). Progress toward the optimum is achieved by the crossover and mutation operations. GA is attractive since it uses only objective function values in the search process, so gradient calculations are avoided. Hence, GA are able to deal with discrete variables. Studies report success in the use of GA for aircraft design optimization studies, trajectory analysis, space structure design and control systems design. In these studies reliable convergence was achieved, but the number of function evaluations was large compared

  20. Multidisciplinary design optimization using genetic algorithms

    NASA Technical Reports Server (NTRS)

    Unal, Resit

    1994-01-01

    Multidisciplinary design optimization (MDO) is an important step in the conceptual design and evaluation of launch vehicles since it can have a significant impact on performance and life cycle cost. The objective is to search the system design space to determine values of design variables that optimize the performance characteristic subject to system constraints. Gradient-based optimization routines have been used extensively for aerospace design optimization. However, one limitation of gradient based optimizers is their need for gradient information. Therefore, design problems which include discrete variables can not be studied. Such problems are common in launch vehicle design. For example, the number of engines and material choices must be integer values or assume only a few discrete values. In this study, genetic algorithms are investigated as an approach to MDO problems involving discrete variables and discontinuous domains. Optimization by genetic algorithms (GA) uses a search procedure which is fundamentally different from those gradient based methods. Genetic algorithms seek to find good solutions in an efficient and timely manner rather than finding the best solution. GA are designed to mimic evolutionary selection. A population of candidate designs is evaluated at each iteration, and each individual's probability of reproduction (existence in the next generation) depends on its fitness value (related to the value of the objective function). Progress toward the optimum is achieved by the crossover and mutation operations. GA is attractive since it uses only objective function values in the search process, so gradient calculations are avoided. Hence, GA are able to deal with discrete variables. Studies report success in the use of GA for aircraft design optimization studies, trajectory analysis, space structure design and control systems design. In these studies reliable convergence was achieved, but the number of function evaluations was large compared

  1. Quasi experimental designs in pharmacist intervention research.

    PubMed

    Krass, Ines

    2016-06-01

    Background In the field of pharmacist intervention research it is often difficult to conform to the rigorous requirements of the "true experimental" models, especially the requirement of randomization. When randomization is not feasible, a practice based researcher can choose from a range of "quasi-experimental designs" i.e., non-randomised and at time non controlled. Objective The aim of this article was to provide an overview of quasi-experimental designs, discuss their strengths and weaknesses and to investigate their application in pharmacist intervention research over the previous decade. Results In the literature quasi experimental studies may be classified into five broad categories: quasi-experimental design without control groups; quasi-experimental design that use control groups with no pre-test; quasi-experimental design that use control groups and pre-tests; interrupted time series and stepped wedge designs. Quasi-experimental study design has consistently featured in the evolution of pharmacist intervention research. The most commonly applied of all quasi experimental designs in the practice based research literature are the one group pre-post-test design and the non-equivalent control group design i.e., (untreated control group with dependent pre-tests and post-tests) and have been used to test the impact of pharmacist interventions in general medications management as well as in specific disease states. Conclusion Quasi experimental studies have a role to play as proof of concept, in the pilot phases of interventions when testing different intervention components, especially in complex interventions. They serve to develop an understanding of possible intervention effects: while in isolation they yield weak evidence of clinical efficacy, taken collectively, they help build a body of evidence in support of the value of pharmacist interventions across different practice settings and countries. However, when a traditional RCT is not feasible for

  2. Three-mirror telescopes: design and optimization.

    PubMed

    Robb, P N

    1978-09-01

    A set of equations is developed which yields the constructional parameters of three-mirror all-reflecting optical systems. An equation whose factors allow the shape of the image surface to be controlled is also derived. A method of optimizing the performance of three-mirror systems by varying the inputs to the design equations is described, and the results are compared with those obtained through a conventional numerical design optimization. The technique described is shown to be markedly superior to the usual optimization method of varying the constructional parameters of the system. PMID:20203850

  3. GCFR shielding design and supporting experimental programs

    SciTech Connect

    Perkins, R.G.; Hamilton, C.J.; Bartine, D.

    1980-05-01

    The shielding for the conceptual design of the gas-cooled fast breeder reactor (GCFR) is described, and the component exposure design criteria which determine the shield design are presented. The experimental programs for validating the GCFR shielding design methods and data (which have been in existence since 1976) are also discussed.

  4. Optimal design of reverse osmosis module networks

    SciTech Connect

    Maskan, F.; Wiley, D.E.; Johnston, L.P.M.; Clements, D.J.

    2000-05-01

    The structure of individual reverse osmosis modules, the configuration of the module network, and the operating conditions were optimized for seawater and brackish water desalination. The system model included simple mathematical equations to predict the performance of the reverse osmosis modules. The optimization problem was formulated as a constrained multivariable nonlinear optimization. The objective function was the annual profit for the system, consisting of the profit obtained from the permeate, capital cost for the process units, and operating costs associated with energy consumption and maintenance. Optimization of several dual-stage reverse osmosis systems were investigated and compared. It was found that optimal network designs are the ones that produce the most permeate. It may be possible to achieve economic improvements by refining current membrane module designs and their operating pressures.

  5. Vehicle systems design optimization study

    NASA Technical Reports Server (NTRS)

    Gilmour, J. L.

    1980-01-01

    The optimum vehicle configuration and component locations are determined for an electric drive vehicle based on using the basic structure of a current production subcompact vehicle. The optimization of an electric vehicle layout requires a weight distribution in the range of 53/47 to 62/38 in order to assure dynamic handling characteristics comparable to current internal combustion engine vehicles. Necessary modification of the base vehicle can be accomplished without major modification of the structure or running gear. As long as batteries are as heavy and require as much space as they currently do, they must be divided into two packages, one at front under the hood and a second at the rear under the cargo area, in order to achieve the desired weight distribution. The weight distribution criteria requires the placement of batteries at the front of the vehicle even when the central tunnel is used for the location of some batteries. The optimum layout has a front motor and front wheel drive. This configuration provides the optimum vehicle dynamic handling characteristics and the maximum passenger and cargo space for a given size vehicle.

  6. Optimal needle design for minimal insertion force and bevel length.

    PubMed

    Wang, Yancheng; Chen, Roland K; Tai, Bruce L; McLaughlin, Patrick W; Shih, Albert J

    2014-09-01

    This research presents a methodology for optimal design of the needle geometry to minimize the insertion force and bevel length based on mathematical models of cutting edge inclination and rake angles and the insertion force. In brachytherapy, the needle with lower insertion force typically is easier for guidance and has less deflection. In this study, the needle with lancet point (denoted as lancet needle) is applied to demonstrate the model-based optimization for needle design. Mathematical models to calculate the bevel length and inclination and rake angles for lancet needle are presented. A needle insertion force model is developed to predict the insertion force for lancet needle. The genetic algorithm is utilized to optimize the needle geometry for two cases. One is to minimize the needle insertion force. Using the geometry of a commercial lancet needle as the baseline, the optimized needle has 11% lower insertion force with the same bevel length. The other case is to minimize the bevel length under the same needle insertion force. The optimized design can reduce the bevel length by 46%. Both optimized needle designs were validated experimentally in ex vivo porcine liver needle insertion tests and demonstrated the methodology of the model-based optimal needle design. PMID:24957487

  7. Torsional ultrasonic transducer computational design optimization.

    PubMed

    Melchor, J; Rus, G

    2014-09-01

    A torsional piezoelectric ultrasonic sensor design is proposed in this paper and computationally tested and optimized to measure shear stiffness properties of soft tissue. These are correlated with a number of pathologies like tumors, hepatic lesions and others. The reason is that, whereas compressibility is predominantly governed by the fluid phase of the tissue, the shear stiffness is dependent on the stroma micro-architecture, which is directly affected by those pathologies. However, diagnostic tools to quantify them are currently not well developed. The first contribution is a new typology of design adapted to quasifluids. A second contribution is the procedure for design optimization, for which an analytical estimate of the Robust Probability Of Detection, called RPOD, is presented for use as optimality criteria. The RPOD is formulated probabilistically to maximize the probability of detecting the least possible pathology while minimizing the effect of noise. The resulting optimal transducer has a resonance frequency of 28 kHz. PMID:24882020

  8. Design optimization for active twist rotor blades

    NASA Astrophysics Data System (ADS)

    Mok, Ji Won

    This dissertation introduces the process of optimizing active twist rotor blades in the presence of embedded anisotropic piezo-composite actuators. Optimum design of active twist blades is a complex task, since it involves a rich design space with tightly coupled design variables. The study presents the development of an optimization framework for active helicopter rotor blade cross-sectional design. This optimization framework allows for exploring a rich and highly nonlinear design space in order to optimize the active twist rotor blades. Different analytical components are combined in the framework: cross-sectional analysis (UM/VABS), an automated mesh generator, a beam solver (DYMORE), a three-dimensional local strain recovery module, and a gradient based optimizer within MATLAB. Through the mathematical optimization problem, the static twist actuation performance of a blade is maximized while satisfying a series of blade constraints. These constraints are associated with locations of the center of gravity and elastic axis, blade mass per unit span, fundamental rotating blade frequencies, and the blade strength based on local three-dimensional strain fields under worst loading conditions. Through pre-processing, limitations of the proposed process have been studied. When limitations were detected, resolution strategies were proposed. These include mesh overlapping, element distortion, trailing edge tab modeling, electrode modeling and foam implementation of the mesh generator, and the initial point sensibility of the current optimization scheme. Examples demonstrate the effectiveness of this process. Optimization studies were performed on the NASA/Army/MIT ATR blade case. Even though that design was built and shown significant impact in vibration reduction, the proposed optimization process showed that the design could be improved significantly. The second example, based on a model scale of the AH-64D Apache blade, emphasized the capability of this framework to

  9. OSHA and Experimental Safety Design.

    ERIC Educational Resources Information Center

    Sichak, Stephen, Jr.

    1983-01-01

    Suggests that a governmental agency, most likely Occupational Safety and Health Administration (OSHA) be considered in the safety design stage of any experiment. Focusing on OSHA's role, discusses such topics as occupational health hazards of toxic chemicals in laboratories, occupational exposure to benzene, and role/regulations of other agencies.…

  10. Acoustic control in enclosures using optimally designed Helmholtz resonators

    NASA Astrophysics Data System (ADS)

    Driesch, Patricia Lynne

    A virtual design methodology is developed to minimize the noise in enclosures with optimally designed, passive, acoustic absorbers (Helmholtz resonators). A series expansion of eigen functions is used to represent the acoustic absorbers as external volume velocities, eliminating the need for a solution of large matrix eigen value problems. A determination of this type (efficient model/reevaluation approach) significantly increases the design possibilities when optimization techniques are implemented. As a benchmarking exercise, this novel methodology was experimentally validated for a narrowband acoustic assessment of two optimally designed Helmholtz resonators coupled to a 2D enclosure. The resonators were tuned to the two lowest resonance frequencies of a 30.5 by 40.6 by 2.5 cm (12 x 16 x 1 inch) cavity with the resonator volume occupying only 2% of the enclosure volume. A maximum potential energy reduction of 12.4 dB was obtained at the second resonance of the cavity. As a full-scale demonstration of the efficacy of the proposed design method, the acoustic response from 90--190 Hz of a John Deere 7000 Ten series tractor cabin was investigated. The lowest cabin mode, referred to as a "boom" mode, proposes a significant challenge to a noise control engineer since its anti-node is located near the head of the operator and often generates unacceptable sound pressure levels. Exploiting the low frequency capability of Helmholtz resonators, lumped parameter models of these resonators were coupled to the enclosure via an experimentally determined acoustic model of the tractor cabin. The virtual design methodology uses gradient optimization techniques as a post processor for the modeling and analysis of the unmodified acoustic interior to determine optimal resonator characteristics. Using two optimally designed Helmholtz resonators; potential energy was experimentally reduced by 3.4 and 10.3 dB at 117 and 167 Hz, respectively.

  11. Stress constraints in optimality criteria design

    NASA Technical Reports Server (NTRS)

    Levy, R.

    1982-01-01

    Procedures described emphasize the processing of stress constraints within optimality criteria designs for low structural weight with stress and compliance constraints. Prescreening criteria are used to partition stress constraints into either potentially active primary sets or passive secondary sets that require minimal processing. Side constraint boundaries for passive constraints are derived by projections from design histories to modify conventional stress-ratio boundaries. Other procedures described apply partial structural modification reanalysis to design variable groups to correct stress constraint violations of unfeasible designs. Sample problem results show effective design convergence and, in particular, advantages for reanalysis in obtaining lower feasible design weights.

  12. Antimicrobial Peptides Design by Evolutionary Multiobjective Optimization

    PubMed Central

    Maccari, Giuseppe; Di Luca, Mariagrazia; Nifosí, Riccardo; Cardarelli, Francesco; Signore, Giovanni; Boccardi, Claudia; Bifone, Angelo

    2013-01-01

    Antimicrobial peptides (AMPs) are an abundant and wide class of molecules produced by many tissues and cell types in a variety of mammals, plant and animal species. Linear alpha-helical antimicrobial peptides are among the most widespread membrane-disruptive AMPs in nature, representing a particularly successful structural arrangement in innate defense. Recently, AMPs have received increasing attention as potential therapeutic agents, owing to their broad activity spectrum and their reduced tendency to induce resistance. The introduction of non-natural amino acids will be a key requisite in order to contrast host resistance and increase compound's life. In this work, the possibility to design novel AMP sequences with non-natural amino acids was achieved through a flexible computational approach, based on chemophysical profiles of peptide sequences. Quantitative structure-activity relationship (QSAR) descriptors were employed to code each peptide and train two statistical models in order to account for structural and functional properties of alpha-helical amphipathic AMPs. These models were then used as fitness functions for a multi-objective evolutional algorithm, together with a set of constraints for the design of a series of candidate AMPs. Two ab-initio natural peptides were synthesized and experimentally validated for antimicrobial activity, together with a series of control peptides. Furthermore, a well-known Cecropin-Mellitin alpha helical antimicrobial hybrid (CM18) was optimized by shortening its amino acid sequence while maintaining its activity and a peptide with non-natural amino acids was designed and tested, demonstrating the higher activity achievable with artificial residues. PMID:24039565

  13. Global Design Optimization for Aerodynamics and Rocket Propulsion Components

    NASA Technical Reports Server (NTRS)

    Shyy, Wei; Papila, Nilay; Vaidyanathan, Rajkumar; Tucker, Kevin; Turner, James E. (Technical Monitor)

    2000-01-01

    Modern computational and experimental tools for aerodynamics and propulsion applications have matured to a stage where they can provide substantial insight into engineering processes involving fluid flows, and can be fruitfully utilized to help improve the design of practical devices. In particular, rapid and continuous development in aerospace engineering demands that new design concepts be regularly proposed to meet goals for increased performance, robustness and safety while concurrently decreasing cost. To date, the majority of the effort in design optimization of fluid dynamics has relied on gradient-based search algorithms. Global optimization methods can utilize the information collected from various sources and by different tools. These methods offer multi-criterion optimization, handle the existence of multiple design points and trade-offs via insight into the entire design space, can easily perform tasks in parallel, and are often effective in filtering the noise intrinsic to numerical and experimental data. However, a successful application of the global optimization method needs to address issues related to data requirements with an increase in the number of design variables, and methods for predicting the model performance. In this article, we review recent progress made in establishing suitable global optimization techniques employing neural network and polynomial-based response surface methodologies. Issues addressed include techniques for construction of the response surface, design of experiment techniques for supplying information in an economical manner, optimization procedures and multi-level techniques, and assessment of relative performance between polynomials and neural networks. Examples drawn from wing aerodynamics, turbulent diffuser flows, gas-gas injectors, and supersonic turbines are employed to help demonstrate the issues involved in an engineering design context. Both the usefulness of the existing knowledge to aid current design

  14. Applications of Chemiluminescence in the Teaching of Experimental Design

    ERIC Educational Resources Information Center

    Krawczyk, Tomasz; Slupska, Roksana; Baj, Stefan

    2015-01-01

    This work describes a single-session laboratory experiment devoted to teaching the principles of factorial experimental design. Students undertook the rational optimization of a luminol oxidation reaction, using a two-level experiment that aimed to create a long-lasting bright emission. During the session students used only simple glassware and…

  15. Experimental design for single point diamond turning of silicon optics

    SciTech Connect

    Krulewich, D.A.

    1996-06-16

    The goal of these experiments is to determine optimum cutting factors for the machining of silicon optics. This report describes experimental design, a systematic method of selecting optimal settings for a limited set of experiments, and its use in the silcon-optics turning experiments. 1 fig., 11 tabs.

  16. Multiobjective optimization techniques for structural design

    NASA Technical Reports Server (NTRS)

    Rao, S. S.

    1984-01-01

    The multiobjective programming techniques are important in the design of complex structural systems whose quality depends generally on a number of different and often conflicting objective functions which cannot be combined into a single design objective. The applicability of multiobjective optimization techniques is studied with reference to simple design problems. Specifically, the parameter optimization of a cantilever beam with a tip mass and a three-degree-of-freedom vabration isolation system and the trajectory optimization of a cantilever beam are considered. The solutions of these multicriteria design problems are attempted by using global criterion, utility function, game theory, goal programming, goal attainment, bounded objective function, and lexicographic methods. It has been observed that the game theory approach required the maximum computational effort, but it yielded better optimum solutions with proper balance of the various objective functions in all the cases.

  17. Autonomous entropy-based intelligent experimental design

    NASA Astrophysics Data System (ADS)

    Malakar, Nabin Kumar

    2011-07-01

    The aim of this thesis is to explore the application of probability and information theory in experimental design, and to do so in a way that combines what we know about inference and inquiry in a comprehensive and consistent manner. Present day scientific frontiers involve data collection at an ever-increasing rate. This requires that we find a way to collect the most relevant data in an automated fashion. By following the logic of the scientific method, we couple an inference engine with an inquiry engine to automate the iterative process of scientific learning. The inference engine involves Bayesian machine learning techniques to estimate model parameters based upon both prior information and previously collected data, while the inquiry engine implements data-driven exploration. By choosing an experiment whose distribution of expected results has the maximum entropy, the inquiry engine selects the experiment that maximizes the expected information gain. The coupled inference and inquiry engines constitute an autonomous learning method for scientific exploration. We apply it to a robotic arm to demonstrate the efficacy of the method. Optimizing inquiry involves searching for an experiment that promises, on average, to be maximally informative. If the set of potential experiments is described by many parameters, the search involves a high-dimensional entropy space. In such cases, a brute force search method will be slow and computationally expensive. We develop an entropy-based search algorithm, called nested entropy sampling, to select the most informative experiment. This helps to reduce the number of computations necessary to find the optimal experiment. We also extended the method of maximizing entropy, and developed a method of maximizing joint entropy so that it could be used as a principle of collaboration between two robots. This is a major achievement of this thesis, as it allows the information-based collaboration between two robotic units towards a same

  18. Response Surface Model Building and Multidisciplinary Optimization Using D-Optimal Designs

    NASA Technical Reports Server (NTRS)

    Unal, Resit; Lepsch, Roger A.; McMillin, Mark L.

    1998-01-01

    This paper discusses response surface methods for approximation model building and multidisciplinary design optimization. The response surface methods discussed are central composite designs, Bayesian methods and D-optimal designs. An over-determined D-optimal design is applied to a configuration design and optimization study of a wing-body, launch vehicle. Results suggest that over determined D-optimal designs may provide an efficient approach for approximation model building and for multidisciplinary design optimization.

  19. Spacecraft design optimization using Taguchi analysis

    NASA Technical Reports Server (NTRS)

    Unal, Resit

    1991-01-01

    The quality engineering methods of Dr. Genichi Taguchi, employing design of experiments, are important statistical tools for designing high quality systems at reduced cost. The Taguchi method was utilized to study several simultaneous parameter level variations of a lunar aerobrake structure to arrive at the lightest weight configuration. Finite element analysis was used to analyze the unique experimental aerobrake configurations selected by Taguchi method. Important design parameters affecting weight and global buckling were identified and the lowest weight design configuration was selected.

  20. Noise constraints effecting optimal propeller designs

    NASA Technical Reports Server (NTRS)

    Miller, C. J.; Sullivan, J. P.

    1985-01-01

    A preliminary design tool for advanced propellers was developed combining a fast vortex lattice aerodynamic analysis, a fast subsonic point source noise analysis, and an optimization scheme using a conjugate directions method. Twist, chord and sweep distributions are optimized to simultaneously improve both the aerodynamic performance and the noise observed at a fixed relative position. The optimal noise/performance tradeoffs for straight and advanced concept blades are presented. The techniques used include increasing the blade number, blade sweep, reducing the rotational speed, shifting the spanwise loading and diameter changes.

  1. Dynamic optimization and adaptive controller design

    NASA Astrophysics Data System (ADS)

    Inamdar, S. R.

    2010-10-01

    In this work I present a new type of controller which is an adaptive tracking controller which employs dynamic optimization for optimizing current value of controller action for the temperature control of nonisothermal continuously stirred tank reactor (CSTR). We begin with a two-state model of nonisothermal CSTR which are mass and heat balance equations and then add cooling system dynamics to eliminate input multiplicity. The initial design value is obtained using local stability of steady states where approach temperature for cooling action is specified as a steady state and a design specification. Later we make a correction in the dynamics where material balance is manipulated to use feed concentration as a system parameter as an adaptive control measure in order to avoid actuator saturation for the main control loop. The analysis leading to design of dynamic optimization based parameter adaptive controller is presented. The important component of this mathematical framework is reference trajectory generation to form an adaptive control measure.

  2. Multifidelity Analysis and Optimization for Supersonic Design

    NASA Technical Reports Server (NTRS)

    Kroo, Ilan; Willcox, Karen; March, Andrew; Haas, Alex; Rajnarayan, Dev; Kays, Cory

    2010-01-01

    Supersonic aircraft design is a computationally expensive optimization problem and multifidelity approaches over a significant opportunity to reduce design time and computational cost. This report presents tools developed to improve supersonic aircraft design capabilities including: aerodynamic tools for supersonic aircraft configurations; a systematic way to manage model uncertainty; and multifidelity model management concepts that incorporate uncertainty. The aerodynamic analysis tools developed are appropriate for use in a multifidelity optimization framework, and include four analysis routines to estimate the lift and drag of a supersonic airfoil, a multifidelity supersonic drag code that estimates the drag of aircraft configurations with three different methods: an area rule method, a panel method, and an Euler solver. In addition, five multifidelity optimization methods are developed, which include local and global methods as well as gradient-based and gradient-free techniques.

  3. Application of an optimization method to high performance propeller designs

    NASA Technical Reports Server (NTRS)

    Li, K. C.; Stefko, G. L.

    1984-01-01

    The application of an optimization method to determine the propeller blade twist distribution which maximizes propeller efficiency is presented. The optimization employs a previously developed method which has been improved to include the effects of blade drag, camber and thickness. Before the optimization portion of the computer code is used, comparisons of calculated propeller efficiencies and power coefficients are made with experimental data for one NACA propeller at Mach numbers in the range of 0.24 to 0.50 and another NACA propeller at a Mach number of 0.71 to validate the propeller aerodynamic analysis portion of the computer code. Then comparisons of calculated propeller efficiencies for the optimized and the original propellers show the benefits of the optimization method in improving propeller performance. This method can be applied to the aerodynamic design of propellers having straight, swept, or nonplanar propeller blades.

  4. Multidisciplinary design optimization using multiobjective formulation techniques

    NASA Technical Reports Server (NTRS)

    Chattopadhyay, Aditi; Pagaldipti, Narayanan S.

    1995-01-01

    This report addresses the development of a multidisciplinary optimization procedure using an efficient semi-analytical sensitivity analysis technique and multilevel decomposition for the design of aerospace vehicles. A semi-analytical sensitivity analysis procedure is developed for calculating computational grid sensitivities and aerodynamic design sensitivities. Accuracy and efficiency of the sensitivity analysis procedure is established through comparison of the results with those obtained using a finite difference technique. The developed sensitivity analysis technique are then used within a multidisciplinary optimization procedure for designing aerospace vehicles. The optimization problem, with the integration of aerodynamics and structures, is decomposed into two levels. Optimization is performed for improved aerodynamic performance at the first level and improved structural performance at the second level. Aerodynamic analysis is performed by solving the three-dimensional parabolized Navier Stokes equations. A nonlinear programming technique and an approximate analysis procedure are used for optimization. The proceduredeveloped is applied to design the wing of a high speed aircraft. Results obtained show significant improvements in the aircraft aerodynamic and structural performance when compared to a reference or baseline configuration. The use of the semi-analytical sensitivity technique provides significant computational savings.

  5. Design optimization for cost and quality: The robust design approach

    NASA Technical Reports Server (NTRS)

    Unal, Resit

    1990-01-01

    Designing reliable, low cost, and operable space systems has become the key to future space operations. Designing high quality space systems at low cost is an economic and technological challenge to the designer. A systematic and efficient way to meet this challenge is a new method of design optimization for performance, quality, and cost, called Robust Design. Robust Design is an approach for design optimization. It consists of: making system performance insensitive to material and subsystem variation, thus allowing the use of less costly materials and components; making designs less sensitive to the variations in the operating environment, thus improving reliability and reducing operating costs; and using a new structured development process so that engineering time is used most productively. The objective in Robust Design is to select the best combination of controllable design parameters so that the system is most robust to uncontrollable noise factors. The robust design methodology uses a mathematical tool called an orthogonal array, from design of experiments theory, to study a large number of decision variables with a significantly small number of experiments. Robust design also uses a statistical measure of performance, called a signal-to-noise ratio, from electrical control theory, to evaluate the level of performance and the effect of noise factors. The purpose is to investigate the Robust Design methodology for improving quality and cost, demonstrate its application by the use of an example, and suggest its use as an integral part of space system design process.

  6. Using Approximations to Accelerate Engineering Design Optimization

    NASA Technical Reports Server (NTRS)

    Torczon, Virginia; Trosset, Michael W.

    1998-01-01

    Optimization problems that arise in engineering design are often characterized by several features that hinder the use of standard nonlinear optimization techniques. Foremost among these features is that the functions used to define the engineering optimization problem often are computationally intensive. Within a standard nonlinear optimization algorithm, the computational expense of evaluating the functions that define the problem would necessarily be incurred for each iteration of the optimization algorithm. Faced with such prohibitive computational costs, an attractive alternative is to make use of surrogates within an optimization context since surrogates can be chosen or constructed so that they are typically much less expensive to compute. For the purposes of this paper, we will focus on the use of algebraic approximations as surrogates for the objective. In this paper we introduce the use of so-called merit functions that explicitly recognize the desirability of improving the current approximation to the objective during the course of the optimization. We define and experiment with the use of merit functions chosen to simultaneously improve both the solution to the optimization problem (the objective) and the quality of the approximation. Our goal is to further improve the effectiveness of our general approach without sacrificing any of its rigor.

  7. Application of Optimal Designs to Item Calibration

    PubMed Central

    Lu, Hung-Yi

    2014-01-01

    In computerized adaptive testing (CAT), examinees are presented with various sets of items chosen from a precalibrated item pool. Consequently, the attrition speed of the items is extremely fast, and replenishing the item pool is essential. Therefore, item calibration has become a crucial concern in maintaining item banks. In this study, a two-parameter logistic model is used. We applied optimal designs and adaptive sequential analysis to solve this item calibration problem. The results indicated that the proposed optimal designs are cost effective and time efficient. PMID:25188318

  8. Evaluation of Frameworks for HSCT Design Optimization

    NASA Technical Reports Server (NTRS)

    Krishnan, Ramki

    1998-01-01

    This report is an evaluation of engineering frameworks that could be used to augment, supplement, or replace the existing FIDO 3.5 (Framework for Interdisciplinary Design and Optimization Version 3.5) framework. The report begins with the motivation for this effort, followed by a description of an "ideal" multidisciplinary design and optimization (MDO) framework. The discussion then turns to how each candidate framework stacks up against this ideal. This report ends with recommendations as to the "best" frameworks that should be down-selected for detailed review.

  9. Experimental Design for Vector Output Systems

    PubMed Central

    Banks, H.T.; Rehm, K.L.

    2013-01-01

    We formulate an optimal design problem for the selection of best states to observe and optimal sampling times for parameter estimation or inverse problems involving complex nonlinear dynamical systems. An iterative algorithm for implementation of the resulting methodology is proposed. Its use and efficacy is illustrated on two applied problems of practical interest: (i) dynamic models of HIV progression and (ii) modeling of the Calvin cycle in plant metabolism and growth. PMID:24563655

  10. Optimal Experiment Design for Thermal Characterization of Functionally Graded Materials

    NASA Technical Reports Server (NTRS)

    Cole, Kevin D.

    2003-01-01

    The purpose of the project was to investigate methods to accurately verify that designed , materials meet thermal specifications. The project involved heat transfer calculations and optimization studies, and no laboratory experiments were performed. One part of the research involved study of materials in which conduction heat transfer predominates. Results include techniques to choose among several experimental designs, and protocols for determining the optimum experimental conditions for determination of thermal properties. Metal foam materials were also studied in which both conduction and radiation heat transfer are present. Results of this work include procedures to optimize the design of experiments to accurately measure both conductive and radiative thermal properties. Detailed results in the form of three journal papers have been appended to this report.

  11. Three-dimensional shape optimization of a cemented hip stem and experimental validations.

    PubMed

    Higa, Masaru; Tanino, Hiromasa; Nishimura, Ikuya; Mitamura, Yoshinori; Matsuno, Takeo; Ito, Hiroshi

    2015-03-01

    This study proposes novel optimized stem geometry with low stress values in the cement using a finite element (FE) analysis combined with an optimization procedure and experimental measurements of cement stress in vitro. We first optimized an existing stem geometry using a three-dimensional FE analysis combined with a shape optimization technique. One of the most important factors in the cemented stem design is to reduce stress in the cement. Hence, in the optimization study, we minimized the largest tensile principal stress in the cement mantle under a physiological loading condition by changing the stem geometry. As the next step, the optimized stem and the existing stem were manufactured to validate the usefulness of the numerical models and the results of the optimization in vitro. In the experimental study, strain gauges were embedded in the cement mantle to measure the strain in the cement mantle adjacent to the stems. The overall trend of the experimental study was in good agreement with the results of the numerical study, and we were able to reduce the largest stress by more than 50% in both shape optimization and strain gauge measurements. Thus, we could validate the usefulness of the numerical models and the results of the optimization using the experimental models. The optimization employed in this study is a useful approach for developing new stem designs. PMID:25320015

  12. Fatigue reliability based optimal design of planar compliant micropositioning stages

    NASA Astrophysics Data System (ADS)

    Wang, Qiliang; Zhang, Xianmin

    2015-10-01

    Conventional compliant micropositioning stages are usually developed based on static strength and deterministic methods, which may lead to either unsafe or excessive designs. This paper presents a fatigue reliability analysis and optimal design of a three-degree-of-freedom (3 DOF) flexure-based micropositioning stage. Kinematic, modal, static, and fatigue stress modelling of the stage were conducted using the finite element method. The maximum equivalent fatigue stress in the hinges was derived using sequential quadratic programming. The fatigue strength of the hinges was obtained by considering various influencing factors. On this basis, the fatigue reliability of the hinges was analysed using the stress-strength interference method. Fatigue-reliability-based optimal design of the stage was then conducted using the genetic algorithm and MATLAB. To make fatigue life testing easier, a 1 DOF stage was then optimized and manufactured. Experimental results demonstrate the validity of the approach.

  13. Instrument design and optimization using genetic algorithms

    SciTech Connect

    Hoelzel, Robert; Bentley, Phillip M.; Fouquet, Peter

    2006-10-15

    This article describes the design of highly complex physical instruments by using a canonical genetic algorithm (GA). The procedure can be applied to all instrument designs where performance goals can be quantified. It is particularly suited to the optimization of instrument design where local optima in the performance figure of merit are prevalent. Here, a GA is used to evolve the design of the neutron spin-echo spectrometer WASP which is presently being constructed at the Institut Laue-Langevin, Grenoble, France. A comparison is made between this artificial intelligence approach and the traditional manual design methods. We demonstrate that the search of parameter space is more efficient when applying the genetic algorithm, and the GA produces a significantly better instrument design. Furthermore, it is found that the GA increases flexibility, by facilitating the reoptimization of the design after changes in boundary conditions during the design phase. The GA also allows the exploration of 'nonstandard' magnet coil geometries. We conclude that this technique constitutes a powerful complementary tool for the design and optimization of complex scientific apparatus, without replacing the careful thought processes employed in traditional design methods.

  14. Branch target buffer design and optimization

    NASA Technical Reports Server (NTRS)

    Perleberg, Chris H.; Smith, Alan J.

    1993-01-01

    Consideration is given to two major issues in the design of branch target buffers (BTBs), with the goal of achieving maximum performance for a given number of bits allocated to the BTB design. The first issue is BTB management; the second is what information to keep in the BTB. A number of solutions to these problems are reviewed, and various optimizations in the design of BTBs are discussed. Design target miss ratios for BTBs are developed, making it possible to estimate the performance of BTBs for real workloads.

  15. A comparison of controller designs for an experimental flexible structure

    NASA Technical Reports Server (NTRS)

    Lim, K. B.; Maghami, P. G.; Joshi, S. M.

    1991-01-01

    Control systems design and hardware testing are addressed for an experimental structure that displays the characteristics of a typical flexible spacecraft. The results of designing and implementing various control design methodologies are described. The design methodologies under investigation include linear quadratic Gaussian control, static and dynamic dissipative controls, and H-infinity optimal control. Among the three controllers considered, it is shown, through computer simulation and laboratory experiments on the evolutionary structure, that the dynamic dissipative controller gave the best results in terms of vibration suppression and robustness with respect to modeling errors.

  16. Global optimization of bilinear engineering design models

    SciTech Connect

    Grossmann, I.; Quesada, I.

    1994-12-31

    Recently Quesada and Grossmann have proposed a global optimization algorithm for solving NLP problems involving linear fractional and bilinear terms. This model has been motivated by a number of applications in process design. The proposed method relies on the derivation of a convex NLP underestimator problem that is used within a spatial branch and bound search. This paper explores the use of alternative bounding approximations for constructing the underestimator problem. These are applied in the global optimization of problems arising in different engineering areas and for which different relaxations are proposed depending on the mathematical structure of the models. These relaxations include linear and nonlinear underestimator problems. Reformulations that generate additional estimator functions are also employed. Examples from process design, structural design, portfolio investment and layout design are presented.

  17. Integrated structural-aerodynamic design optimization

    NASA Technical Reports Server (NTRS)

    Haftka, R. T.; Kao, P. J.; Grossman, B.; Polen, D.; Sobieszczanski-Sobieski, J.

    1988-01-01

    This paper focuses on the processes of simultaneous aerodynamic and structural wing design as a prototype for design integration, with emphasis on the major difficulty associated with multidisciplinary design optimization processes, their enormous computational costs. Methods are presented for reducing this computational burden through the development of efficient methods for cross-sensitivity calculations and the implementation of approximate optimization procedures. Utilizing a modular sensitivity analysis approach, it is shown that the sensitivities can be computed without the expensive calculation of the derivatives of the aerodynamic influence coefficient matrix, and the derivatives of the structural flexibility matrix. The same process is used to efficiently evaluate the sensitivities of the wing divergence constraint, which should be particularly useful, not only in problems of complete integrated aircraft design, but also in aeroelastic tailoring applications.

  18. Optimal design of crossflow heat exchangers

    SciTech Connect

    Van den Bulck, E. )

    1991-05-01

    The design of plate-fin and tube-fin crossflow heat exchangers is discussed. The transfer surface area of crossflow heat exchangers is used ineffectively because of the nonuniform distribution of the heat transfer across the volume of the exchanger. The optimal distribution of the transfer surface area for maximum heat exchanger effectiveness and constant total surface area is determined. It is found that a Dirac delta distribution of the transfer surface aligned along the diagonal of the crossflow exchanger gives the best performance; equal to that of a counterflow device. Design guidelines for optimal area allocation within crossflow heat exchangers are established. Compared to conventional designs, designs following these guidelines may lead to either a higher exchanger effectiveness for equal pressure drops and surface area, reduced pressure drops for equal exchanger effectiveness, or reduced weight and a near cubic form of the exchanger core for equal pressure drops and effectiveness.

  19. Optimal Shape Design of a Plane Diffuser in Turbulent Flow

    NASA Astrophysics Data System (ADS)

    Lim, Seokhyun; Choi, Haecheon

    2000-11-01

    Stratford (1959) experimentally designed an optimal shape of plane diffuser for maximum pressure recovery by having zero skin friction throughout the region of pressure rise. In the present study, we apply an algorithm of optimal shape design developed by Pironneau (1973, 1974) and Cabuk & Modi (1992) to a diffuser in turbulent flow, and show that maintaining zero skin friction in the pressure-rise region is an optimal condition for maximum pressure recovery at the diffuser exit. For turbulence model, we use the k-ɛ-v^2-f model by Durbin (1995) which is known to accurately predict flow with separation. Our results with this model agree well with the previous experimental and LES results for a diffuser shape tested by Obi et al. (1993). From this initial shape, an optimal diffuser shape for maximum pressure recovery is obtained through an iterative procedure. The optimal diffuser has indeed zero skin friction throughout the pressure-rise region, and thus there is no separation in the flow. For the optimal diffuser shape obtained, an LES is being conducted to investigate the turbulence characteristics near the zero-skin-friction wall. A preliminary result of LES will also be presented.

  20. Multidisciplinary Concurrent Design Optimization via the Internet

    NASA Technical Reports Server (NTRS)

    Woodard, Stanley E.; Kelkar, Atul G.; Koganti, Gopichand

    2001-01-01

    A methodology is presented which uses commercial design and analysis software and the Internet to perform concurrent multidisciplinary optimization. The methodology provides a means to develop multidisciplinary designs without requiring that all software be accessible from the same local network. The procedures are amenable to design and development teams whose members, expertise and respective software are not geographically located together. This methodology facilitates multidisciplinary teams working concurrently on a design problem of common interest. Partition of design software to different machines allows each constituent software to be used on the machine that provides the most economy and efficiency. The methodology is demonstrated on the concurrent design of a spacecraft structure and attitude control system. Results are compared to those derived from performing the design with an autonomous FORTRAN program.

  1. Efficient experimental design for uncertainty reduction in gene regulatory networks

    PubMed Central

    2015-01-01

    Background An accurate understanding of interactions among genes plays a major role in developing therapeutic intervention methods. Gene regulatory networks often contain a significant amount of uncertainty. The process of prioritizing biological experiments to reduce the uncertainty of gene regulatory networks is called experimental design. Under such a strategy, the experiments with high priority are suggested to be conducted first. Results The authors have already proposed an optimal experimental design method based upon the objective for modeling gene regulatory networks, such as deriving therapeutic interventions. The experimental design method utilizes the concept of mean objective cost of uncertainty (MOCU). MOCU quantifies the expected increase of cost resulting from uncertainty. The optimal experiment to be conducted first is the one which leads to the minimum expected remaining MOCU subsequent to the experiment. In the process, one must find the optimal intervention for every gene regulatory network compatible with the prior knowledge, which can be prohibitively expensive when the size of the network is large. In this paper, we propose a computationally efficient experimental design method. This method incorporates a network reduction scheme by introducing a novel cost function that takes into account the disruption in the ranking of potential experiments. We then estimate the approximate expected remaining MOCU at a lower computational cost using the reduced networks. Conclusions Simulation results based on synthetic and real gene regulatory networks show that the proposed approximate method has close performance to that of the optimal method but at lower computational cost. The proposed approximate method also outperforms the random selection policy significantly. A MATLAB software implementing the proposed experimental design method is available at http://gsp.tamu.edu/Publications/supplementary/roozbeh15a/. PMID:26423515

  2. Design Optimization of Structural Health Monitoring Systems

    SciTech Connect

    Flynn, Eric B.

    2014-03-06

    Sensor networks drive decisions. Approach: Design networks to minimize the expected total cost (in a statistical sense, i.e. Bayes Risk) associated with making wrong decisions and with installing maintaining and running the sensor network itself. Search for optimal solutions using Monte-Carlo-Sampling-Adapted Genetic Algorithm. Applications include structural health monitoring and surveillance.

  3. MDO can help resolve the designer's dilemma. [Multidisciplinary design optimization

    SciTech Connect

    Sobieszczanski-sobieski, Jaroslaw; Tulinius, J.R. Rockwell International Corp., El Segundo, CA )

    1991-09-01

    Multidisciplinary design optimization (MDO) is presented as a rapidly growing body of methods, algorithms, and techniques that will provide a quantum jump in the effectiveness and efficiency of the quantitative side of design, and will turn that side into an environment in which the qualitative side can thrive. MDO borrows from CAD/CAM for graphic visualization of geometrical and numerical data, data base technology, and in computer software and hardware. Expected benefits from this methodology are a rational, mathematically consistent approach to hypersonic aircraft designs, designs pushed closer to the optimum, and a design process either shortened or leaving time available for different concepts to be explored.

  4. MDO can help resolve the designer's dilemma. [multidisciplinary design optimization

    NASA Technical Reports Server (NTRS)

    Sobieszczanski-Sobieski, Jaroslaw; Tulinius, Jan R.

    1991-01-01

    Multidisciplinary design optimization (MDO) is presented as a rapidly growing body of methods, algorithms, and techniques that will provide a quantum jump in the effectiveness and efficiency of the quantitative side of design, and will turn that side into an environment in which the qualitative side can thrive. MDO borrows from CAD/CAM for graphic visualization of geometrical and numerical data, data base technology, and in computer software and hardware. Expected benefits from this methodology are a rational, mathematically consistent approach to hypersonic aircraft designs, designs pushed closer to the optimum, and a design process either shortened or leaving time available for different concepts to be explored.

  5. Drought Adaptation Mechanisms Should Guide Experimental Design.

    PubMed

    Gilbert, Matthew E; Medina, Viviana

    2016-08-01

    The mechanism, or hypothesis, of how a plant might be adapted to drought should strongly influence experimental design. For instance, an experiment testing for water conservation should be distinct from a damage-tolerance evaluation. We define here four new, general mechanisms for plant adaptation to drought such that experiments can be more easily designed based upon the definitions. A series of experimental methods are suggested together with appropriate physiological measurements related to the drought adaptation mechanisms. The suggestion is made that the experimental manipulation should match the rate, length, and severity of soil water deficit (SWD) necessary to test the hypothesized type of drought adaptation mechanism. PMID:27090148

  6. Design Oriented Structural Modeling for Airplane Conceptual Design Optimization

    NASA Technical Reports Server (NTRS)

    Livne, Eli

    1999-01-01

    The main goal for research conducted with the support of this grant was to develop design oriented structural optimization methods for the conceptual design of airplanes. Traditionally in conceptual design airframe weight is estimated based on statistical equations developed over years of fitting airplane weight data in data bases of similar existing air- planes. Utilization of such regression equations for the design of new airplanes can be justified only if the new air-planes use structural technology similar to the technology on the airplanes in those weight data bases. If any new structural technology is to be pursued or any new unconventional configurations designed the statistical weight equations cannot be used. In such cases any structural weight estimation must be based on rigorous "physics based" structural analysis and optimization of the airframes under consideration. Work under this grant progressed to explore airframe design-oriented structural optimization techniques along two lines of research: methods based on "fast" design oriented finite element technology and methods based on equivalent plate / equivalent shell models of airframes, in which the vehicle is modelled as an assembly of plate and shell components, each simulating a lifting surface or nacelle / fuselage pieces. Since response to changes in geometry are essential in conceptual design of airplanes, as well as the capability to optimize the shape itself, research supported by this grant sought to develop efficient techniques for parametrization of airplane shape and sensitivity analysis with respect to shape design variables. Towards the end of the grant period a prototype automated structural analysis code designed to work with the NASA Aircraft Synthesis conceptual design code ACS= was delivered to NASA Ames.

  7. Yakima Hatchery Experimental Design : Annual Progress Report.

    SciTech Connect

    Busack, Craig; Knudsen, Curtis; Marshall, Anne

    1991-08-01

    This progress report details the results and status of Washington Department of Fisheries' (WDF) pre-facility monitoring, research, and evaluation efforts, through May 1991, designed to support the development of an Experimental Design Plan (EDP) for the Yakima/Klickitat Fisheries Project (YKFP), previously termed the Yakima/Klickitat Production Project (YKPP or Y/KPP). This pre- facility work has been guided by planning efforts of various research and quality control teams of the project that are annually captured as revisions to the experimental design and pre-facility work plans. The current objective are as follows: to develop genetic monitoring and evaluation approach for the Y/KPP; to evaluate stock identification monitoring tools, approaches, and opportunities available to meet specific objectives of the experimental plan; and to evaluate adult and juvenile enumeration and sampling/collection capabilities in the Y/KPP necessary to measure experimental response variables.

  8. Optimization and Inverse Design of Pump Impeller

    NASA Astrophysics Data System (ADS)

    Miyauchi, S.; Zhu, B.; Luo, X.; Piao, B.; Matsumoto, H.; Sano, M.; Kassai, N.

    2012-11-01

    As for pump impellers, the meridional flow channel and blade-to-blade flow channel, which are relatively independent of each other but greatly affect performance, are designed in parallel. And the optimization design is used for the former and the inverse design is used for the latter. To verify this new design method, a mixed-flow impeller was made. Next, we use Tani's inverse design method for the blade loading of inverse design. It is useful enough to change a deceleration rate freely and greatly. And it can integrally express the rear blade loading of various methods by NACA, Zangeneh and Stratford. We controlled the deceleration rate by shape parameter m, and its value became almost same with Tani's recommended value of the laminar airfoil.

  9. Improving spacecraft design using a multidisciplinary design optimization methodology

    NASA Astrophysics Data System (ADS)

    Mosher, Todd Jon

    2000-10-01

    Spacecraft design has gone from maximizing performance under technology constraints to minimizing cost under performance constraints. This is characteristic of the "faster, better, cheaper" movement that has emerged within NASA. Currently spacecraft are "optimized" manually through a tool-assisted evaluation of a limited set of design alternatives. With this approach there is no guarantee that a systems-level focus will be taken and "feasibility" rather than "optimality" is commonly all that is achieved. To improve spacecraft design in the "faster, better, cheaper" era, a new approach using multidisciplinary design optimization (MDO) is proposed. Using MDO methods brings structure to conceptual spacecraft design by casting a spacecraft design problem into an optimization framework. Then, through the construction of a model that captures design and cost, this approach facilitates a quicker and more straightforward option synthesis. The final step is to automatically search the design space. As computer processor speed continues to increase, enumeration of all combinations, while not elegant, is one method that is straightforward to perform. As an alternative to enumeration, genetic algorithms are used and find solutions by reviewing fewer possible solutions with some limitations. Both methods increase the likelihood of finding an optimal design, or at least the most promising area of the design space. This spacecraft design methodology using MDO is demonstrated on three examples. A retrospective test for validation is performed using the Near Earth Asteroid Rendezvous (NEAR) spacecraft design. For the second example, the premise that aerobraking was needed to minimize mission cost and was mission enabling for the Mars Global Surveyor (MGS) mission is challenged. While one might expect no feasible design space for an MGS without aerobraking mission, a counterintuitive result is discovered. Several design options that don't use aerobraking are feasible and cost

  10. Application of heuristic optimization in aircraft design

    NASA Astrophysics Data System (ADS)

    Hu, Zhenning

    Genetic algorithms and the related heuristic optimization strategies are introduced and their applications in the aircraft design are developed. Generally speaking, genetic algorithms belong to non-deterministic direct search methods, which are most powerful in finding optimum or near-optimum solutions of a very complex system where a little priori knowledge is known. Therefore they have a wide application in aerospace systems. Two major aircraft optimal design projects are illustrated in this dissertation. The first is the application of material optimization of aligned fiber laminate composites in the presence of stress concentrations. After a large number of tests on laminates with different layers, genetic algorithms find an alignment pattern in a certain range for the Boeing Co. specified material. The second project is the application of piezoelectric actuator placement on a generic tail skins to reduce the 2nd mode vibration caused by buffet, which is part of a Boeing project to control the buffet effect on aircraft. In this project, genetic algorithms are closely involved with vibration analysis and finite element analysis. Actuator optimization strategies are first tested on the theoretical beam models to gain experience, and then the generic tail model is applied. Genetic algorithms achieve a great success in optimizing up to 888 actuator parameters on the tail skins.

  11. Reliability-based structural optimization: A proposed analytical-experimental study

    NASA Technical Reports Server (NTRS)

    Stroud, W. Jefferson; Nikolaidis, Efstratios

    1993-01-01

    An analytical and experimental study for assessing the potential of reliability-based structural optimization is proposed and described. In the study, competing designs obtained by deterministic and reliability-based optimization are compared. The experimental portion of the study is practical because the structure selected is a modular, actively and passively controlled truss that consists of many identical members, and because the competing designs are compared in terms of their dynamic performance and are not destroyed if failure occurs. The analytical portion of this study is illustrated on a 10-bar truss example. In the illustrative example, it is shown that reliability-based optimization can yield a design that is superior to an alternative design obtained by deterministic optimization. These analytical results provide motivation for the proposed study, which is underway.

  12. Optimal controller design for structural damage detection

    NASA Astrophysics Data System (ADS)

    Lew, Jiann-Shiun

    2005-03-01

    The virtual passive control technique has recently been applied to structural damage detection, where the virtual passive controller only uses the existing control devices, and no additional physical elements are attached to the tested structure. One important task is to design passive controllers that can enhance the sensitivity of the identified parameters, such as natural frequencies, to structural damage. This paper presents a novel study of an optimal controller design for structural damage detection. We apply not only passive controllers but also low-order and fixed-structure controllers, such as PID controllers. In the optimal control design, the performance of structural damage detection is based on the application of a neural network technique, which uses the pattern of the correlation between the natural frequency changes of the tested system and the damaged system.

  13. Optimizing the TRD design for ACCESS

    SciTech Connect

    Cherry, M. L.; Guzik, T. G.; Isbert, J.; Wefel, J. P.

    1999-01-22

    The present ACCESS design combines an ionization calorimeter with a transition radiation detector (TRD) to measure the cosmic ray composition and energy spectrum from H to Fe at energies above 1 TeV/nucleon to the 'knee' in the all particle spectrum. We are in the process of optimizing the TRD design to extend the range of the technique to as high an energy as possible given the constraints of the International Space Station mission and the need to coexist with the calorimeter. The current status of the design effort and preliminary results will be presented.

  14. Experimental Testing of Dynamically Optimized Photoelectron Beams

    SciTech Connect

    Rosenzweig, J. B.; Cook, A. M.; Dunning, M.; England, R. J.; Musumeci, P.; Bellaveglia, M.; Boscolo, M.; Catani, L.; Cianchi, A.; Di Pirro, G.; Ferrario, M.; Fillipetto, D.; Gatti, G.; Palumbo, L.; Vicario, C.; Serafini, L.; Jones, S.

    2006-11-27

    We discuss the design of and initial results from an experiment in space-charge dominated beam dynamics which explores a new regime of high-brightness electron beam generation at the SPARC photoinjector. The scheme under study employs the tendency of intense electron beams to rearrange to produce uniform density, giving a nearly ideal beam from the viewpoint of space charge-induced emittance. The experiments are aimed at testing the marriage of this idea with a related concept, emittance compensation. We show that this new regime of operating photoinjector may be the preferred method of obtaining highest brightness beams with lower energy spread. We discuss the design of the experiment, including developing of a novel time-dependent, aerogel-based imaging system. This system has been installed at SPARC, and first evidence for nearly uniformly filled ellipsoidal charge distributions recorded.

  15. Design criteria for optimal photosynthetic energy conversion

    NASA Astrophysics Data System (ADS)

    Fingerhut, Benjamin P.; Zinth, Wolfgang; de Vivie-Riedle, Regina

    2008-12-01

    Photochemical solar energy conversion is considered as an alternative of clean energy. For future light converting nano-machines photosynthetic reaction centers are used as prototypes optimized during evolution. We introduce a reaction scheme for global optimization and simulate the ultrafast charge separation in photochemical energy conversion. Multiple molecular charge carriers are involved in this process and are linked by Marcus-type electron transfer. In combination with evolutionary algorithms, we unravel the biological strategies for high quantum efficiency in photosynthetic reaction centers and extend these concepts to the design of artificial photochemical devices for energy conversion.

  16. Multidisciplinary Design Optimization on Conceptual Design of Aero-engine

    NASA Astrophysics Data System (ADS)

    Zhang, Xiao-bo; Wang, Zhan-xue; Zhou, Li; Liu, Zeng-wen

    2016-06-01

    In order to obtain better integrated performance of aero-engine during the conceptual design stage, multiple disciplines such as aerodynamics, structure, weight, and aircraft mission are required. Unfortunately, the couplings between these disciplines make it difficult to model or solve by conventional method. MDO (Multidisciplinary Design Optimization) methodology which can well deal with couplings of disciplines is considered to solve this coupled problem. Approximation method, optimization method, coordination method, and modeling method for MDO framework are deeply analyzed. For obtaining the more efficient MDO framework, an improved CSSO (Concurrent Subspace Optimization) strategy which is based on DOE (Design Of Experiment) and RSM (Response Surface Model) methods is proposed in this paper; and an improved DE (Differential Evolution) algorithm is recommended to solve the system-level and discipline-level optimization problems in MDO framework. The improved CSSO strategy and DE algorithm are evaluated by utilizing the numerical test problem. The result shows that the efficiency of improved methods proposed by this paper is significantly increased. The coupled problem of VCE (Variable Cycle Engine) conceptual design is solved by utilizing improved CSSO strategy, and the design parameter given by improved CSSO strategy is better than the original one. The integrated performance of VCE is significantly improved.

  17. Aircraft family design using enhanced collaborative optimization

    NASA Astrophysics Data System (ADS)

    Roth, Brian Douglas

    Significant progress has been made toward the development of multidisciplinary design optimization (MDO) methods that are well-suited to practical large-scale design problems. However, opportunities exist for further progress. This thesis describes the development of enhanced collaborative optimization (ECO), a new decomposition-based MDO method. To support the development effort, the thesis offers a detailed comparison of two existing MDO methods: collaborative optimization (CO) and analytical target cascading (ATC). This aids in clarifying their function and capabilities, and it provides inspiration for the development of ECO. The ECO method offers several significant contributions. First, it enhances communication between disciplinary design teams while retaining the low-order coupling between them. Second, it provides disciplinary design teams with more authority over the design process. Third, it resolves several troubling computational inefficiencies that are associated with CO. As a result, ECO provides significant computational savings (relative to CO) for the test cases and practical design problems described in this thesis. New aircraft development projects seldom focus on a single set of mission requirements. Rather, a family of aircraft is designed, with each family member tailored to a different set of requirements. This thesis illustrates the application of decomposition-based MDO methods to aircraft family design. This represents a new application area, since MDO methods have traditionally been applied to multidisciplinary problems. ECO offers aircraft family design the same benefits that it affords to multidisciplinary design problems. Namely, it simplifies analysis integration, it provides a means to manage problem complexity, and it enables concurrent design of all family members. In support of aircraft family design, this thesis introduces a new wing structural model with sufficient fidelity to capture the tradeoffs associated with component

  18. Optimal AFCS: particularities of real design

    NASA Astrophysics Data System (ADS)

    Platonov, A.; Zaitsev, Ie.; Chaciński, H.

    2015-09-01

    The paper discusses particularities of optimal adaptive communication systems (AFCS) design conditioned by the particularities of their architecture and way of functioning, as well as by the approach to their design. The main one is that AFCS employ the analog method of transmission (in the paper - amplitude modulation), and are intended for short-range transmission of signals from the analog sources. Another one is that AFCS design is carried out on the basis of strict results of concurrent analytical optimisation of the transmitting and receiving parts of the system. Below, general problems appearing during transition from the theoretical results to the real engineering design, as well as approach to their solution are discussed. Some concrete tasks of AFCS design are also considered.

  19. Phylogenetic information and experimental design in molecular systematics.

    PubMed Central

    Goldman, N

    1998-01-01

    Despite the widespread perception that evolutionary inference from molecular sequences is a statistical problem, there has been very little attention paid to questions of experimental design. Previous consideration of this topic has led to little more than an empirical folklore regarding the choice of suitable genes for analysis, and to dispute over the best choice of taxa for inclusion in data sets. I introduce what I believe are new methods that permit the quantification of phylogenetic information in a sequence alignment. The methods use likelihood calculations based on Markov-process models of nucleotide substitution allied with phylogenetic trees, and allow a general approach to optimal experimental design. Two examples are given, illustrating realistic problems in experimental design in molecular phylogenetics and suggesting more general conclusions about the choice of genomic regions, sequence lengths and taxa for evolutionary studies. PMID:9787470

  20. Robust Design Optimization via Failure Domain Bounding

    NASA Technical Reports Server (NTRS)

    Crespo, Luis G.; Kenny, Sean P.; Giesy, Daniel P.

    2007-01-01

    This paper extends and applies the strategies recently developed by the authors for handling constraints under uncertainty to robust design optimization. For the scope of this paper, robust optimization is a methodology aimed at problems for which some parameters are uncertain and are only known to belong to some uncertainty set. This set can be described by either a deterministic or a probabilistic model. In the methodology developed herein, optimization-based strategies are used to bound the constraint violation region using hyper-spheres and hyper-rectangles. By comparing the resulting bounding sets with any given uncertainty model, it can be determined whether the constraints are satisfied for all members of the uncertainty model (i.e., constraints are feasible) or not (i.e., constraints are infeasible). If constraints are infeasible and a probabilistic uncertainty model is available, upper bounds to the probability of constraint violation can be efficiently calculated. The tools developed enable approximating not only the set of designs that make the constraints feasible but also, when required, the set of designs for which the probability of constraint violation is below a prescribed admissible value. When constraint feasibility is possible, several design criteria can be used to shape the uncertainty model of performance metrics of interest. Worst-case, least-second-moment, and reliability-based design criteria are considered herein. Since the problem formulation is generic and the tools derived only require standard optimization algorithms for their implementation, these strategies are easily applicable to a broad range of engineering problems.

  1. Generalized mathematical models in design optimization

    NASA Technical Reports Server (NTRS)

    Papalambros, Panos Y.; Rao, J. R. Jagannatha

    1989-01-01

    The theory of optimality conditions of extremal problems can be extended to problems continuously deformed by an input vector. The connection between the sensitivity, well-posedness, stability and approximation of optimization problems is steadily emerging. The authors believe that the important realization here is that the underlying basis of all such work is still the study of point-to-set maps and of small perturbations, yet what has been identified previously as being just related to solution procedures is now being extended to study modeling itself in its own right. Many important studies related to the theoretical issues of parametric programming and large deformation in nonlinear programming have been reported in the last few years, and the challenge now seems to be in devising effective computational tools for solving these generalized design optimization models.

  2. Design Methods and Optimization for Morphing Aircraft

    NASA Technical Reports Server (NTRS)

    Crossley, William A.

    2005-01-01

    This report provides a summary of accomplishments made during this research effort. The major accomplishments are in three areas. The first is the use of a multiobjective optimization strategy to help identify potential morphing features that uses an existing aircraft sizing code to predict the weight, size and performance of several fixed-geometry aircraft that are Pareto-optimal based upon on two competing aircraft performance objectives. The second area has been titled morphing as an independent variable and formulates the sizing of a morphing aircraft as an optimization problem in which the amount of geometric morphing for various aircraft parameters are included as design variables. This second effort consumed most of the overall effort on the project. The third area involved a more detailed sizing study of a commercial transport aircraft that would incorporate a morphing wing to possibly enable transatlantic point-to-point passenger service.

  3. Direct optimization method for reentry trajectory design

    NASA Astrophysics Data System (ADS)

    Jallade, S.; Huber, P.; Potti, J.; Dutruel-Lecohier, G.

    The software package called `Reentry and Atmospheric Transfer Trajectory' (RATT) was developed under ESA contract for the design of atmospheric trajectories. It includes four software TOP (Trajectory OPtimization) programs, which optimize reentry and aeroassisted transfer trajectories. 6FD and 3FD (6 and 3 degrees of freedom Flight Dynamic) are devoted to the simulation of the trajectory. SCA (Sensitivity and Covariance Analysis) performs covariance analysis on a given trajectory with respect to different uncertainties and error sources. TOP provides the optimum guidance law of a three degree of freedom reentry of aeroassisted transfer (AAOT) trajectories. Deorbit and reorbit impulses (if necessary) can be taken into account in the optimization. A wide choice of cost function is available to the user such as the integrated heat flux, or the sum of the velocity impulses, or a linear combination of both of them for trajectory and vehicle design. The crossrange and the downrange can be maximized during reentry trajectory. Path constraints are available on the load factor, the heat flux and the dynamic pressure. Results on these proposed options are presented. TOPPHY is the part of the TOP software corresponding to the definition and the computation of the optimization problemphysics. TOPPHY can interface with several optimizes with dynamic solvers: TOPOP and TROPIC using direct collocation methods and PROMIS using direct multiple shooting method. TOPOP was developed in the frame of this contract, it uses Hermite polynomials for the collocation method and the NPSOL optimizer from the NAG library. Both TROPIC and PROMIS were developed by the DLR (Deutsche Forschungsanstalt fuer Luft und Raumfahrt) and use the SLSQP optimizer. For the dynamic equation resolution, TROPIC uses a collocation method with Splines and PROMIS uses a multiple shooting method with finite differences. The three different optimizers including dynamics were tested on the reentry trajectory of the

  4. Optimized design of LED plant lamp

    NASA Astrophysics Data System (ADS)

    Chen, Jian-sheng; Cai, Ruhai; Zhao, Yunyun; Zhao, Fuli; Yang, Bowen

    2014-12-01

    In order to fabricate the optimized LED plant lamp we demonstrated an optical spectral exploration. According to the mechanism of higher plant photosynthesis process and the spectral analysis we demonstrate an optical design of the LED plant lamp. Furthermore we built two kins of prototypes of the LED plant lamps which are suitable for the photosynthesis of higher green vegetables. Based on the simulation of the lamp box of the different alignment of the plants we carried out the growing experiment of green vegetable and obtain the optimized light illumination as well as the spectral profile. The results show that only blue and red light are efficient for the green leave vegetables. Our work is undoubtedly helpful for the LED plant lamping design and manufacture.

  5. Multiobjective optimization in integrated photonics design.

    PubMed

    Gagnon, Denis; Dumont, Joey; Dubé, Louis J

    2013-07-01

    We propose the use of the parallel tabu search algorithm (PTS) to solve combinatorial inverse design problems in integrated photonics. To assess the potential of this algorithm, we consider the problem of beam shaping using a two-dimensional arrangement of dielectric scatterers. The performance of PTS is compared to one of the most widely used optimization algorithms in photonics design, the genetic algorithm (GA). We find that PTS can produce comparable or better solutions than the GA, while requiring less computation time and fewer adjustable parameters. For the coherent beam shaping problem as a case study, we demonstrate how PTS can tackle multiobjective optimization problems and represent a robust and efficient alternative to GA. PMID:23811870

  6. Methodology on zoom system design and optimization

    NASA Astrophysics Data System (ADS)

    Ding, Quanxin; Liu, Hua

    2008-03-01

    For aim to establish effective methodology in research to design and evaluate on typical zoom sensor system, to satisfy the system requirements and achieve an advanced characteristics. Some methods about system analysis, especially task principle and key technique of core system, are analyzed deeply. Base on Gaussian photonics theory, zoom system differential equation, solves vector space distribution and integrated balance algorithm on global optimization system is studied. Dominate configuration of new idea system design and optimization, with which consecutive zoom and diffractive module equipped by great format photonics device, is established. The results of evaluated on a kind of typical zoom sensor system is presented, and achieves remarkable advantages on some criterions, such as Modulation Transfer Function (MTF), Spot Diagram (RMS) and Point Spread Function (PSF) etc., and in volume, weight, system efficiency and otherwise.

  7. Optimization Methodology for Unconventional Rocket Nozzle Design

    NASA Technical Reports Server (NTRS)

    Follett, W.

    1996-01-01

    Several current rocket engine concepts such as the bell-annular tripropellant engine, and the linear aerospike being proposed for the X-33, require unconventional three-dimensional rocket nozzles which must conform to rectangular or sector-shaped envelopes to meet integration constraints. These types of nozzles exist outside the current experience database, therefore, development of efficient design methods for these propulsion concepts is critical to the success of launch vehicle programs. Several approaches for optimizing rocket nozzles, including streamline tracing techniques, and the coupling of CFD analysis to optimization algorithms are described. The relative strengths and weaknesses of four classes of optimization algorithms are discussed: Gradient based methods, genetic algorithms, simplex methods, and surface response methods. Additionally, a streamline tracing technique, which provides a very computationally efficient means of defining a three-dimensional contour, is discussed. The performance of the various optimization methods on thrust optimization problems for tripropellant and aerospike concepts is assessed and recommendations are made for future development efforts.

  8. Slot design of optimized electromagnetic pump

    SciTech Connect

    Leboucher, L. . Institut de Mecanique); Villani, D. )

    1993-11-01

    Electromagnetic pumps are used for the transportation of liquid metals such as the cooling sodium of fast breeder nuclear reactors. The design of this induction machine is close to that of a tubular linear induction motor. A non uniform slot distribution is used to optimize electromagnetic pumps. This geometry is tested with a finite element code. The performances are compared with the regular slot distribution of Industrial prototypes.

  9. General purpose optimization software for engineering design

    NASA Technical Reports Server (NTRS)

    Vanderplaats, G. N.

    1990-01-01

    The author has developed several general purpose optimization programs over the past twenty years. The earlier programs were developed as research codes and served that purpose reasonably well. However, in taking the formal step from research to industrial application programs, several important lessons have been learned. Among these are the importance of clear documentation, immediate user support, and consistent maintenance. Most important has been the issue of providing software that gives a good, or at least acceptable, design at minimum computational cost. Here, the basic issues developing optimization software for industrial applications are outlined and issues of convergence rate, reliability, and relative minima are discussed. Considerable feedback has been received from users, and new software is being developed to respond to identified needs. The basic capabilities of this software are outlined. A major motivation for the development of commercial grade software is ease of use and flexibility, and these issues are discussed with reference to general multidisciplinary applications. It is concluded that design productivity can be significantly enhanced by the more widespread use of optimization as an everyday design tool.

  10. Optimal design of a tidal turbine

    NASA Astrophysics Data System (ADS)

    Kueny, J. L.; Lalande, T.; Herou, J. J.; Terme, L.

    2012-11-01

    An optimal design procedure has been applied to improve the design of an open-center tidal turbine. A specific software developed in C++ enables to generate the geometry adapted to the specific constraints imposed to this machine. Automatic scripts based on the AUTOGRID, IGG, FINE/TURBO and CFView software of the NUMECA CFD suite are used to evaluate all the candidate geometries. This package is coupled with the optimization software EASY, which is based on an evolutionary strategy completed by an artificial neural network. A new technique is proposed to guarantee the robustness of the mesh in the whole range of the design parameters. An important improvement of the initial geometry has been obtained. To limit the whole CPU time necessary for this optimization process, the geometry of the tidal turbine has been considered as axisymmetric, with a uniform upstream velocity. A more complete model (12 M nodes) has been built in order to analyze the effects related to the sea bed boundary layer, the proximity of the sea surface, the presence of an important triangular basement supporting the turbine and a possible incidence of the upstream velocity.

  11. Reliability Optimization Design for Contact Springs of AC Contactors Based on Adaptive Genetic Algorithm

    NASA Astrophysics Data System (ADS)

    Zhao, Sheng; Su, Xiuping; Wu, Ziran; Xu, Chengwen

    The paper illustrates the procedure of reliability optimization modeling for contact springs of AC contactors under nonlinear multi-constraint conditions. The adaptive genetic algorithm (AGA) is utilized to perform reliability optimization on the contact spring parameters of a type of AC contactor. A method that changes crossover and mutation rates at different times in the AGA can effectively avoid premature convergence, and experimental tests are performed after optimization. The experimental result shows that the mass of each optimized spring is reduced by 16.2%, while the reliability increases to 99.9% from 94.5%. The experimental result verifies the correctness and feasibility of this reliability optimization designing method.

  12. Optimization of the National Ignition Facility primary shield design

    SciTech Connect

    Annese, C.E.; Watkins, E.F.; Greenspan, E.; Miller, W.F.; Latkowski, J.; Lee, J.D.; Soran, P.; Tobin, M.L.

    1993-10-01

    Minimum cost design concepts of the primary shield for the National Ignition laser fusion experimental Facility (NIF) are searched with the help of the optimization code SWAN. The computational method developed for this search involves incorporating the time dependence of the delayed photon field within effective delayed photon production cross sections. This method enables one to address the time-dependent problem using relatively simple, time-independent transport calculations, thus significantly simplifying the design process. A novel approach was used for the identification of the optimal combination of constituents that will minimize the shield cost; it involves the generation, with SWAN, of effectiveness functions for replacing materials on an equal cost basis. The minimum cost shield design concept was found to consist of a mixture of polyethylene and low cost, low activation materials such as SiC, with boron added near the shield boundaries.

  13. Discrete optimization of isolator locations for vibration isolation systems: An analytical and experimental investigation

    SciTech Connect

    Ponslet, E.R.; Eldred, M.S.

    1996-05-17

    An analytical and experimental study is conducted to investigate the effect of isolator locations on the effectiveness of vibration isolation systems. The study uses isolators with fixed properties and evaluates potential improvements to the isolation system that can be achieved by optimizing isolator locations. Because the available locations for the isolators are discrete in this application, a Genetic Algorithm (GA) is used as the optimization method. The system is modeled in MATLAB{trademark} and coupled with the GA available in the DAKOTA optimization toolkit under development at Sandia National Laboratories. Design constraints dictated by hardware and experimental limitations are implemented through penalty function techniques. A series of GA runs reveal difficulties in the search on this heavily constrained, multimodal, discrete problem. However, the GA runs provide a variety of optimized designs with predicted performance from 30 to 70 times better than a baseline configuration. An alternate approach is also tested on this problem: it uses continuous optimization, followed by rounding of the solution to neighboring discrete configurations. Results show that this approach leads to either infeasible or poor designs. Finally, a number of optimized designs obtained from the GA searches are tested in the laboratory and compared to the baseline design. These experimental results show a 7 to 46 times improvement in vibration isolation from the baseline configuration.

  14. Simulation as an Aid to Experimental Design.

    ERIC Educational Resources Information Center

    Frazer, Jack W.; And Others

    1983-01-01

    Discusses simulation program to aid in the design of enzyme kinetic experimentation (includes sample runs). Concentration versus time profiles of any subset or all nine states of reactions can be displayed with/without simulated instrumental noise, allowing the user to estimate the practicality of any proposed experiment given known instrument…

  15. Optimal branching designs in respiratory systems

    NASA Astrophysics Data System (ADS)

    Park, Keunhwan; Kim, Wonjung; Kim, Ho-Young

    2015-11-01

    In nature, the size of the flow channels systematically decreases with multiple generations of branching, and a mother branch is ultimately divided into numerous terminal daughters. One important feature of branching designs is an increase in the total cross-sectional area along with generation, which provide more time and area for mass transfer at the terminal branches. However, the expansion of the total cross-sectional area can be costly due to the maintenance of redundant branches or the additional viscous resistance. Accordingly, we expect to find optimal designs in natural branching systems. Here we present two examples of branching designs in respiratory systems: fish gills and human lung airways. Fish gills consist of filaments with well-ordered lamellar structures. By developing a mathematical model of oxygen transfer rate as a function of the dimensions of fish gills, we demonstrate that the interlamellar distance has been optimized to maximize the oxygen transfer rate. Using the same framework, we examine the diameter reduction ratio in human lung airways, which branch by dichotomy with a systematic reduction of their diameters. Our mathematical model for oxygen transport in the airways enables us to unveil the design principle of human lung airways.

  16. Machine Learning Techniques in Optimal Design

    NASA Technical Reports Server (NTRS)

    Cerbone, Giuseppe

    1992-01-01

    Many important applications can be formalized as constrained optimization tasks. For example, we are studying the engineering domain of two-dimensional (2-D) structural design. In this task, the goal is to design a structure of minimum weight that bears a set of loads. A solution to a design problem in which there is a single load (L) and two stationary support points (S1 and S2) consists of four members, E1, E2, E3, and E4 that connect the load to the support points is discussed. In principle, optimal solutions to problems of this kind can be found by numerical optimization techniques. However, in practice [Vanderplaats, 1984] these methods are slow and they can produce different local solutions whose quality (ratio to the global optimum) varies with the choice of starting points. Hence, their applicability to real-world problems is severely restricted. To overcome these limitations, we propose to augment numerical optimization by first performing a symbolic compilation stage to produce: (a) objective functions that are faster to evaluate and that depend less on the choice of the starting point and (b) selection rules that associate problem instances to a set of recommended solutions. These goals are accomplished by successive specializations of the problem class and of the associated objective functions. In the end, this process reduces the problem to a collection of independent functions that are fast to evaluate, that can be differentiated symbolically, and that represent smaller regions of the overall search space. However, the specialization process can produce a large number of sub-problems. This is overcome by deriving inductively selection rules which associate problems to small sets of specialized independent sub-problems. Each set of candidate solutions is chosen to minimize a cost function which expresses the tradeoff between the quality of the solution that can be obtained from the sub-problem and the time it takes to produce it. The overall solution

  17. Topology optimization design of a space mirror

    NASA Astrophysics Data System (ADS)

    Liu, Jiazhen; Jiang, Bo

    2015-11-01

    As key components of the optical system of the space optical remote sensor, Space mirrors' surface accuracy had a direct impact that couldn't be ignored of the imaging quality of the remote sensor. In the future, large-diameter mirror would become an important trend in the development of space optical technology. However, a sharp increase in the mirror diameter would cause the deformation of the mirror and increase the thermal deformation caused by temperature variations. A reasonable lightweight structure designed to ensure the optical performance of the system to meet the requirements was required. As a new type of lightweight approach, topology optimization technology was an important direction of the current space optical remote sensing technology research. The lightweight design of rectangular mirror was studied. the variable density method of topology optimization was used. The mirror type precision of the mirror assemblies was obtained in different conditions. PV value was less than λ/10 and RMS value was less than λ/50(λ = 632.8nm). The results show that the entire The mirror assemblies can achieve a sufficiently high static rigidity, dynamic stiffness and thermal stability and has the capability of sufficient resistance to external environmental interference . Key words: topology optimization, space mirror, lightweight, space optical remote sensor

  18. Taguchi method of experimental design in materials education

    NASA Technical Reports Server (NTRS)

    Weiser, Martin W.

    1993-01-01

    Some of the advantages and disadvantages of the Taguchi Method of experimental design as applied to Materials Science will be discussed. This is a fractional factorial method that employs the minimum number of experimental trials for the information obtained. The analysis is also very simple to use and teach, which is quite advantageous in the classroom. In addition, the Taguchi loss function can be easily incorporated to emphasize that improvements in reproducibility are often at least as important as optimization of the response. The disadvantages of the Taguchi Method include the fact that factor interactions are normally not accounted for, there are zero degrees of freedom if all of the possible factors are used, and randomization is normally not used to prevent environmental biasing. In spite of these disadvantages it is felt that the Taguchi Method is extremely useful for both teaching experimental design and as a research tool, as will be shown with a number of brief examples.

  19. Design search and optimization in aerospace engineering.

    PubMed

    Keane, A J; Scanlan, J P

    2007-10-15

    In this paper, we take a design-led perspective on the use of computational tools in the aerospace sector. We briefly review the current state-of-the-art in design search and optimization (DSO) as applied to problems from aerospace engineering, focusing on those problems that make heavy use of computational fluid dynamics (CFD). This ranges over issues of representation, optimization problem formulation and computational modelling. We then follow this with a multi-objective, multi-disciplinary example of DSO applied to civil aircraft wing design, an area where this kind of approach is becoming essential for companies to maintain their competitive edge. Our example considers the structure and weight of a transonic civil transport wing, its aerodynamic performance at cruise speed and its manufacturing costs. The goals are low drag and cost while holding weight and structural performance at acceptable levels. The constraints and performance metrics are modelled by a linked series of analysis codes, the most expensive of which is a CFD analysis of the aerodynamics using an Euler code with coupled boundary layer model. Structural strength and weight are assessed using semi-empirical schemes based on typical airframe company practice. Costing is carried out using a newly developed generative approach based on a hierarchical decomposition of the key structural elements of a typical machined and bolted wing-box assembly. To carry out the DSO process in the face of multiple competing goals, a recently developed multi-objective probability of improvement formulation is invoked along with stochastic process response surface models (Krigs). This approach both mitigates the significant run times involved in CFD computation and also provides an elegant way of balancing competing goals while still allowing the deployment of the whole range of single objective optimizers commonly available to design teams. PMID:17519198

  20. Optimally designing games for behavioural research

    PubMed Central

    Rafferty, Anna N.; Zaharia, Matei; Griffiths, Thomas L.

    2014-01-01

    Computer games can be motivating and engaging experiences that facilitate learning, leading to their increasing use in education and behavioural experiments. For these applications, it is often important to make inferences about the knowledge and cognitive processes of players based on their behaviour. However, designing games that provide useful behavioural data are a difficult task that typically requires significant trial and error. We address this issue by creating a new formal framework that extends optimal experiment design, used in statistics, to apply to game design. In this framework, we use Markov decision processes to model players' actions within a game, and then make inferences about the parameters of a cognitive model from these actions. Using a variety of concept learning games, we show that in practice, this method can predict which games will result in better estimates of the parameters of interest. The best games require only half as many players to attain the same level of precision. PMID:25002821

  1. Multidisciplinary design optimization for sonic boom mitigation

    NASA Astrophysics Data System (ADS)

    Ozcer, Isik A.

    product design. The simulation tools are used to optimize three geometries for sonic boom mitigation. The first is a simple axisymmetric shape to be used as a generic nose component, the second is a delta wing with lift, and the third is a real aircraft with nose and wing optimization. The objectives are to minimize the pressure impulse or the peak pressure in the sonic boom signal, while keeping the drag penalty under feasible limits. The design parameters for the meridian profile of the nose shape are the lengths and the half-cone angles of the linear segments that make up the profile. The design parameters for the lifting wing are the dihedral angle, angle of attack, non-linear span-wise twist and camber distribution. The test-bed aircraft is the modified F-5E aircraft built by Northrop Grumman, designated the Shaped Sonic Boom Demonstrator. This aircraft is fitted with an optimized axisymmetric nose, and the wings are optimized to demonstrate optimization for sonic boom mitigation for a real aircraft. The final results predict 42% reduction in bow shock strength, 17% reduction in peak Deltap, 22% reduction in pressure impulse, 10% reduction in foot print size, 24% reduction in inviscid drag, and no loss in lift for the optimized aircraft. Optimization is carried out using response surface methodology, and the design matrices are determined using standard DoE techniques for quadratic response modeling.

  2. A FRAMEWORK TO DESIGN AND OPTIMIZE CHEMICAL FLOODING PROCESSES

    SciTech Connect

    Mojdeh Delshad; Gary A. Pope; Kamy Sepehrnoori

    2005-07-01

    The goal of this proposed research is to provide an efficient and user friendly simulation framework for screening and optimizing chemical/microbial enhanced oil recovery processes. The framework will include (1) a user friendly interface to identify the variables that have the most impact on oil recovery using the concept of experimental design and response surface maps, (2) UTCHEM reservoir simulator to perform the numerical simulations, and (3) an economic model that automatically imports the simulation production data to evaluate the profitability of a particular design. Such a reservoir simulation framework is not currently available to the oil industry. The objectives of Task 1 are to develop three primary modules representing reservoir, chemical, and well data. The modules will be interfaced with an already available experimental design model. The objective of the Task 2 is to incorporate UTCHEM reservoir simulator and the modules with the strategic variables and developing the response surface maps to identify the significant variables from each module. The objective of the Task 3 is to develop the economic model designed specifically for the chemical processes targeted in this proposal and interface the economic model with UTCHEM production output. Task 4 is on the validation of the framework and performing simulations of oil reservoirs to screen, design and optimize the chemical processes.

  3. A FRAMEWORK TO DESIGN AND OPTIMIZE CHEMICAL FLOODING PROCESSES

    SciTech Connect

    Mojdeh Delshad; Gary A. Pope; Kamy Sepehrnoori

    2004-11-01

    The goal of this proposed research is to provide an efficient and user friendly simulation framework for screening and optimizing chemical/microbial enhanced oil recovery processes. The framework will include (1) a user friendly interface to identify the variables that have the most impact on oil recovery using the concept of experimental design and response surface maps, (2) UTCHEM reservoir simulator to perform the numerical simulations, and (3) an economic model that automatically imports the simulation production data to evaluate the profitability of a particular design. Such a reservoir simulation framework is not currently available to the oil industry. The objectives of Task 1 are to develop three primary modules representing reservoir, chemical, and well data. The modules will be interfaced with an already available experimental design model. The objective of the Task 2 is to incorporate UTCHEM reservoir simulator and the modules with the strategic variables and developing the response surface maps to identify the significant variables from each module. The objective of the Task 3 is to develop the economic model designed specifically for the chemical processes targeted in this proposal and interface the economic model with UTCHEM production output. Task 4 is on the validation of the framework and performing simulations of oil reservoirs to screen, design and optimize the chemical processes.

  4. A Framework to Design and Optimize Chemical Flooding Processes

    SciTech Connect

    Mojdeh Delshad; Gary A. Pope; Kamy Sepehrnoori

    2006-08-31

    The goal of this proposed research is to provide an efficient and user friendly simulation framework for screening and optimizing chemical/microbial enhanced oil recovery processes. The framework will include (1) a user friendly interface to identify the variables that have the most impact on oil recovery using the concept of experimental design and response surface maps, (2) UTCHEM reservoir simulator to perform the numerical simulations, and (3) an economic model that automatically imports the simulation production data to evaluate the profitability of a particular design. Such a reservoir simulation framework is not currently available to the oil industry. The objectives of Task 1 are to develop three primary modules representing reservoir, chemical, and well data. The modules will be interfaced with an already available experimental design model. The objective of the Task 2 is to incorporate UTCHEM reservoir simulator and the modules with the strategic variables and developing the response surface maps to identify the significant variables from each module. The objective of the Task 3 is to develop the economic model designed specifically for the chemical processes targeted in this proposal and interface the economic model with UTCHEM production output. Task 4 is on the validation of the framework and performing simulations of oil reservoirs to screen, design and optimize the chemical processes.

  5. [Sun protection by optimally designed fabrics].

    PubMed

    Hoffmann, K; Hoffmann, A; Hanke, D; Böhringer, B; Schindling, G; Schön, U; Klotz, M L; Altmeyer, P

    1998-01-01

    A rising incidence worldwide of skin cancer has been observed for years. A high cumulative exposure to UV radiation is a major factor in the development of such neoplasms. Suitable protective measures are therefore becoming increasingly important. Textiles provide simple, effective and medically safe protection against UV radiation. At present, however, in Europe--in contrast to Australia--the UV protection factor (UPF) for summer textiles is not stated. It is a largely unknown fact that by far not all textiles offer sufficient protection. Our goal was to study the factors which determine the UV transmission of fabrics and, based on these findings, to design materials which provide enhanced protection. A spectrophotometric method was used to determine the UV transmission by fabrics made of cotton, silk, polyester and viscose. The UV protection factors of the fabrics were computed on the basis of the transmission data. The UV protection factor is dependent on the type of fibre, yarn and surface design, weight per square metre, moisture content, colour, finishing method and degree of wear. To optimize the UV protection via textiles, a viscose yarn with a low UV transmission was used. This yarn makes it possible to design light-weight summer fabrics with optimized UV protection. This development will make it possible to offer clothing with high UV protection on the European marked. This clothing will not be more expensive than normal products, so that effective prevention should be more available. PMID:9522187

  6. Design, optimization, and control of tensegrity structures

    NASA Astrophysics Data System (ADS)

    Masic, Milenko

    The contributions of this dissertation may be divided into four categories. The first category involves developing a systematic form-finding method for general and symmetric tensegrity structures. As an extension of the available results, different shape constraints are incorporated in the problem. Methods for treatment of these constraints are considered and proposed. A systematic formulation of the form-finding problem for symmetric tensegrity structures is introduced, and it uses the symmetry to reduce both the number of equations and the number of variables in the problem. The equilibrium analysis of modular tensegrities exploits their peculiar symmetry. The tensegrity similarity transformation completes the contributions in the area of enabling tools for tensegrity form-finding. The second group of contributions develops the methods for optimal mass-to-stiffness-ratio design of tensegrity structures. This technique represents the state-of-the-art for the static design of tensegrity structures. It is an extension of the results available for the topology optimization of truss structures. Besides guaranteeing that the final design satisfies the tensegrity paradigm, the problem constrains the structure from different modes of failure, which makes it very general. The open-loop control of the shape of modular tensegrities is the third contribution of the dissertation. This analytical result offers a closed form solution for the control of the reconfiguration of modular structures. Applications range from the deployment and stowing of large-scale space structures to the locomotion-inducing control for biologically inspired structures. The control algorithm is applicable regardless of the size of the structures, and it represents a very general result for a large class of tensegrities. Controlled deployments of large-scale tensegrity plates and tensegrity towers are shown as examples that demonstrate the full potential of this reconfiguration strategy. The last

  7. Inter occasion variability in individual optimal design.

    PubMed

    Kristoffersson, Anders N; Friberg, Lena E; Nyberg, Joakim

    2015-12-01

    Inter occasion variability (IOV) is of importance to consider in the development of a design where individual pharmacokinetic or pharmacodynamic parameters are of interest. IOV may adversely affect the precision of maximum a posteriori (MAP) estimated individual parameters, yet the influence of inclusion of IOV in optimal design for estimation of individual parameters has not been investigated. In this work two methods of including IOV in the maximum a posteriori Fisher information matrix (FIMMAP) are evaluated: (i) MAP occ-the IOV is included as a fixed effect deviation per occasion and individual, and (ii) POP occ-the IOV is included as an occasion random effect. Sparse sampling schedules were designed for two test models and compared to a scenario where IOV is ignored, either by omitting known IOV (Omit) or by mimicking a situation where unknown IOV has inflated the IIV (Inflate). Accounting for IOV in the FIMMAP markedly affected the designs compared to ignoring IOV and, as evaluated by stochastic simulation and estimation, resulted in superior precision in the individual parameters. In addition MAPocc and POP occ accurately predicted precision and shrinkage. For the investigated designs, the MAP occ method was on average slightly superior to POP occ and was less computationally intensive. PMID:26452548

  8. Sparse and optimal acquisition design for diffusion MRI and beyond

    PubMed Central

    Koay, Cheng Guan; Özarslan, Evren; Johnson, Kevin M.; Meyerand, M. Elizabeth

    2012-01-01

    strategy was found to be effective in finding the optimum configuration. It was found that the square design is the most robust (i.e., with stable condition numbers and A-optimal measures under varying experimental conditions) among many other possible designs of the same sample size. Under the same performance evaluation, the square design was found to be more robust than the widely used sampling schemes similar to that of 3D radial MRI and of diffusion spectrum imaging (DSI). Conclusions: A novel optimality criterion for sparse multiple-shell acquisition and quasimultiple-shell designs in diffusion MRI and an effective search strategy for finding the best configuration have been developed. The results are very promising, interesting, and practical for diffusion MRI acquisitions. PMID:22559620

  9. Involving students in experimental design: three approaches.

    PubMed

    McNeal, A P; Silverthorn, D U; Stratton, D B

    1998-12-01

    Many faculty want to involve students more actively in laboratories and in experimental design. However, just "turning them loose in the lab" is time-consuming and can be frustrating for both students and faculty. We describe three different ways of providing structures for labs that require students to design their own experiments but guide the choices. One approach emphasizes invertebrate preparations and classic techniques that students can learn fairly easily. Students must read relevant primary literature and learn each technique in one week, and then design and carry out their own experiments in the next week. Another approach provides a "design framework" for the experiments so that all students are using the same technique and the same statistical comparisons, whereas their experimental questions differ widely. The third approach involves assigning the questions or problems but challenging students to design good protocols to answer these questions. In each case, there is a mixture of structure and freedom that works for the level of the students, the resources available, and our particular aims. PMID:16161223

  10. Optimal air-breathing launch vehicle design

    NASA Technical Reports Server (NTRS)

    Hattis, P. D.

    1981-01-01

    A generalized two-point boundary problem methodology, similar to techniques used in deterministic optimal control studies, is applied to the design and flight analysis of a two-stage air-breathing launch vehicle. Simultaneous consideration is given to configuration and trajectory by treating geometry, dynamic discontinuities, and time-dependent flight variables all as controls to be optimized with respect to a single mathematical performance measure. While minimizing fuel consumption, inequality constraints are applied to dynamic pressure and specific force. The optimal system fuel consumption and staging Mach number are found to vary little with changes in the inequality constraints due to substantial geometry and trajectory adjustments. Staging, from an air-breathing first stage to a rocket-powered second stage, consistently occurs near Mach 3.5. The dynamic pressure bound has its most pronounced effects on vehicle geometry, particularly the air-breathing propulsion inlet area, and on the first-stage altitude profile. The specific force has its greatest influence on the second-stage thrust history.

  11. Global optimization methods for engineering design

    NASA Technical Reports Server (NTRS)

    Arora, Jasbir S.

    1990-01-01

    The problem is to find a global minimum for the Problem P. Necessary and sufficient conditions are available for local optimality. However, global solution can be assured only under the assumption of convexity of the problem. If the constraint set S is compact and the cost function is continuous on it, existence of a global minimum is guaranteed. However, in view of the fact that no global optimality conditions are available, a global solution can be found only by an exhaustive search to satisfy Inequality. The exhaustive search can be organized in such a way that the entire design space need not be searched for the solution. This way the computational burden is reduced somewhat. It is concluded that zooming algorithm for global optimizations appears to be a good alternative to stochastic methods. More testing is needed; a general, robust, and efficient local minimizer is required. IDESIGN was used in all numerical calculations which is based on a sequential quadratic programming algorithm, and since feasible set keeps on shrinking, a good algorithm to find an initial feasible point is required. Such algorithms need to be developed and evaluated.

  12. Multi-Disciplinary Design Optimization Using WAVE

    NASA Technical Reports Server (NTRS)

    Irwin, Keith

    2000-01-01

    develop an associative control structure (framework) in the UG WAVE environment enabling multi-disciplinary design of turbine propulsion systems. The capabilities of WAVE were evaluated to assess its use as a rapid optimization and productivity tool. This project also identified future WAVE product enhancements that will make the tool still more beneficial for product development.

  13. OPTIMIZATION OF DESIGN SPECIFICATIONS FOR LARGE DRY COOLING SYSTEMS

    EPA Science Inventory

    The report presents a methodology for optimizing design specifications of large, mechanical-draft, dry cooling systems. A multivariate, nonlinear, constrained optimization technique searches for the combination of design variables to determine the cooling system with the lowest a...

  14. Simulation as an aid to experimental design

    SciTech Connect

    Frazer, J.W.; Balaban, D.J.; Wang, J.L.

    1983-05-01

    A simulator of chemical reactions can aid the scientist in the design of experimentation. They are of great value when studying enzymatic kinetic reactions. One such simulator is a numerical ordinary differential equation solver which uses interactive graphics to provide the user with the capability to simulate an extremely wide range of enzyme reaction conditions for many types of single substrate reactions. The concentration vs. time profiles of any subset or all nine states of a complex reaction can be displayed with and without simulated instrumental noise. Thus the user can estimate the practicality of any proposed experimentation given known instrumental noise. The experimenter can readily determine which state provides the most information related to the proposed kinetic parameters and mechanism. A general discussion of the program including the nondimensionalization of the set of differential equations is included. Finally, several simulation examples are shown and the results discussed.

  15. Optimal design of a hybridization scheme with a fuel cell using genetic optimization

    NASA Astrophysics Data System (ADS)

    Rodriguez, Marco A.

    Fuel cell is one of the most dependable "green power" technologies, readily available for immediate application. It enables direct conversion of hydrogen and other gases into electric energy without any pollution of the environment. However, the efficient power generation is strictly stationary process that cannot operate under dynamic environment. Consequently, fuel cell becomes practical only within a specially designed hybridization scheme, capable of power storage and power management functions. The resultant technology could be utilized to its full potential only when both the fuel cell element and the entire hybridization scheme are optimally designed. The design optimization in engineering is among the most complex computational tasks due to its multidimensionality, nonlinearity, discontinuity and presence of constraints in the underlying optimization problem. this research aims at the optimal utilization of the fuel cell technology through the use of genetic optimization, and advance computing. This study implements genetic optimization in the definition of optimum hybridization rules for a PEM fuel cell/supercapacitor power system. PEM fuel cells exhibit high energy density but they are not intended for pulsating power draw applications. They work better in steady state operation and thus, are often hybridized. In a hybrid system, the fuel cell provides power during steady state operation while capacitors or batteries augment the power of the fuel cell during power surges. Capacitors and batteries can also be recharged when the motor is acting as a generator. Making analogies to driving cycles, three hybrid system operating modes are investigated: 'Flat' mode, 'Uphill' mode, and 'Downhill' mode. In the process of discovering the switching rules for these three modes, we also generate a model of a 30W PEM fuel cell. This study also proposes the optimum design of a 30W PEM fuel cell. The PEM fuel cell model and hybridization's switching rules are postulated

  16. Optimal Beam Combiner Design for Nulling Interferometers

    NASA Astrophysics Data System (ADS)

    Guyon, Olivier; Mennesson, Bertrand; Serabyn, Eugene; Martin, Stefan

    2013-08-01

    A scheme to optimally design a beam combiner is discussed for any predetermined fixed geometry nulling interferometer aimed at detection and characterization of exoplanets with multiple telescopes or a single telescope (aperture masking). We show that considerably higher order nulls can be achieved with 1D (one-dimensional) interferometer geometries than possible with 2D (two-dimensional) geometries with the same number of apertures. Any 1D interferometer with N apertures can achieve a 2(N - 1)-order null, while the order of the deepest null for a random 2D aperture geometry interferometer is the order of the Nth term in the Taylor expansion of ei(x2+y2) around x = 0, y = 0 (2nd order null for N = 2,3 4th order null for N = 4,5,6). We also show that an optimal beam combiner for nulling interferometry relies on only 0 or π phase shifts. Examples of nulling interferometer designs are shown to illustrate these findings.

  17. Optimal design of a touch trigger probe

    NASA Astrophysics Data System (ADS)

    Li, Rui-Jun; Xiang, Meng; Fan, Kuang-Chao; Zhou, Hao; Feng, Jian

    2015-02-01

    A tungsten stylus with a ruby ball tip was screwed into a floating plate, which was supported by four leaf springs. The displacement of the tip caused by the contact force in 3D could be transferred into the tilt or vertical displacement of a plane mirror mounted on the floating plate. A quadrant photo detector (QPD) based two dimensional angle sensor was used to detect the tilt or the vertical displacement of the plane mirror. The structural parameters of the probe are optimized for equal sensitivity and equal stiffness in a displacement range of +/-5 μm, and a restricted horizontal size of less than 40 mm. Simulation results indicated that the stiffness was less than 0.6 mN/μm and equal in 3D. Experimental results indicated that the probe could be used to achieve a resolution of 1 nm.

  18. CFD based draft tube hydraulic design optimization

    NASA Astrophysics Data System (ADS)

    McNabb, J.; Devals, C.; Kyriacou, S. A.; Murry, N.; Mullins, B. F.

    2014-03-01

    The draft tube design of a hydraulic turbine, particularly in low to medium head applications, plays an important role in determining the efficiency and power characteristics of the overall machine, since an important proportion of the available energy, being in kinetic form leaving the runner, needs to be recovered by the draft tube into static head. For large units, these efficiency and power characteristics can equate to large sums of money when considering the anticipated selling price of the energy produced over the machine's life-cycle. This same draft tube design is also a key factor in determining the overall civil costs of the powerhouse, primarily in excavation and concreting, which can amount to similar orders of magnitude as the price of the energy produced. Therefore, there is a need to find the optimum compromise between these two conflicting requirements. In this paper, an elaborate approach is described for dealing with this optimization problem. First, the draft tube's detailed geometry is defined as a function of a comprehensive set of design parameters (about 20 of which a subset is allowed to vary during the optimization process) and are then used in a non-uniform rational B-spline based geometric modeller to fully define the wetted surfaces geometry. Since the performance of the draft tube is largely governed by 3D viscous effects, such as boundary layer separation from the walls and swirling flow characteristics, which in turn governs the portion of the available kinetic energy which will be converted into pressure, a full 3D meshing and Navier-Stokes analysis is performed for each design. What makes this even more challenging is the fact that the inlet velocity distribution to the draft tube is governed by the runner at each of the various operating conditions that are of interest for the exploitation of the powerhouse. In order to determine these inlet conditions, a combined steady-state runner and an initial draft tube analysis, using a

  19. Design of an optimal snow observation network to estimate snowpack

    NASA Astrophysics Data System (ADS)

    Juan Collados Lara, Antonio; Pardo-Iguzquiza, Eulogio; Pulido-Velazquez, David

    2016-04-01

    Snow is an important water resource in many river basins that must be taken into account in hydrological modeling. Although the snow cover area may be nowadays estimated from satellite data, the snow pack thickness must be estimated from experimental data by using some interpolation procedure or hydrological models that approximates snow accumulation and fusion processes. The experimental data consist of hand probes and snow samples collected in a given number of locations that constitute the monitoring network. Assuming that there is an existing monitoring network, its optimization may imply the selection of an optimal network as a subset of the existing network (decrease of the existing network in the case that there are no funds for maintaining the full existing network) or to increase the existing network by one or more stations (optimal augmentation problem). In this work we propose a multicriterion approach for the optimal design of a snow network. These criteria include the estimation variance from a regression kriging approach for estimating thickness of the snowpack (using ground and satellite data), to minimize the total snow volume and accessibility criteria. We have also proposed a procedure to analyze the sensitivity of the results to the non-snow data deduced from the satellite information. We intent to minimize the uncertities in snowpack estimation. The methodology has been applied to estimation of the snow cover area and the design of the optimal snow observation network in Sierra Nevada mountain range in the Southern of Spain. Acknowledgments: This research has been partially supported by the GESINHIMPADAPT project (CGL2013-48424-C2-2-R) with Spanish MINECO funds. We would also like to thank ERHIN program and NASA DAAC for the data provided for this study.

  20. Optimal Design of Automotive Thermoelectric Air Conditioner (TEAC)

    NASA Astrophysics Data System (ADS)

    Attar, Alaa; Lee, HoSung; Weera, Sean

    2014-06-01

    The present work is an analytical study of the optimal design of an automotive thermoelectric air conditioner (TEAC) using a new optimal design method with dimensional analysis that has been recently developed by our research group. The optimal design gives not only the optimal current but also the optimal geometry (i.e., the number of thermocouples, the geometric factor, or the hot fluid parameters). The optimal design for the TEAC is carried out with two configurations: air-to-liquid and air-to-air heat exchangers.

  1. Maximum Entropy/Optimal Projection (MEOP) control design synthesis: Optimal quantification of the major design tradeoffs

    NASA Technical Reports Server (NTRS)

    Hyland, D. C.; Bernstein, D. S.

    1987-01-01

    The underlying philosophy and motivation of the optimal projection/maximum entropy (OP/ME) stochastic modeling and reduced control design methodology for high order systems with parameter uncertainties are discussed. The OP/ME design equations for reduced-order dynamic compensation including the effect of parameter uncertainties are reviewed. The application of the methodology to several Large Space Structures (LSS) problems of representative complexity is illustrated.

  2. The suitability of selected multidisciplinary design and optimization techniques to conceptual aerospace vehicle design

    NASA Technical Reports Server (NTRS)

    Olds, John R.

    1992-01-01

    Four methods for preliminary aerospace vehicle design are reviewed. The first three methods (classical optimization, system decomposition, and system sensitivity analysis (SSA)) employ numerical optimization techniques and numerical gradients to feed back changes in the design variables. The optimum solution is determined by stepping through a series of designs toward a final solution. Of these three, SSA is argued to be the most applicable to a large-scale highly coupled vehicle design where an accurate minimum of an objective function is required. With SSA, several tasks can be performed in parallel. The techniques of classical optimization and decomposition can be included in SSA, resulting in a very powerful design method. The Taguchi method is more of a 'smart' parametric design method that analyzes variable trends and interactions over designer specified ranges with a minimum of experimental analysis runs. Its advantages are its relative ease of use, ability to handle discrete variables, and ability to characterize the entire design space with a minimum of analysis runs.

  3. Optimizing Monitoring Designs under Alternative Objectives

    DOE PAGESBeta

    Gastelum, Jason A.; USA, Richland Washington; Porter, Ellen A.; USA, Richland Washington

    2014-12-31

    This paper describes an approach to identify monitoring designs that optimize detection of CO2 leakage from a carbon capture and sequestration (CCS) reservoir and compares the results generated under two alternative objective functions. The first objective function minimizes the expected time to first detection of CO2 leakage, the second more conservative objective function minimizes the maximum time to leakage detection across the set of realizations. The approach applies a simulated annealing algorithm that searches the solution space by iteratively mutating the incumbent monitoring design. The approach takes into account uncertainty by evaluating the performance of potential monitoring designs across amore » set of simulated leakage realizations. The approach relies on a flexible two-tiered signature to infer that CO2 leakage has occurred. This research is part of the National Risk Assessment Partnership, a U.S. Department of Energy (DOE) project tasked with conducting risk and uncertainty analysis in the areas of reservoir performance, natural leakage pathways, wellbore integrity, groundwater protection, monitoring, and systems level modeling.« less

  4. Optimal Ground Source Heat Pump System Design

    SciTech Connect

    Ozbek, Metin; Yavuzturk, Cy; Pinder, George

    2015-04-15

    Despite the facts that GSHPs first gained popularity as early as the 1940’s and they can achieve 30 to 60 percent in energy savings and carbon emission reductions relative to conventional HVAC systems, the use of geothermal energy in the U.S. has been less than 1 percent of the total energy consumption. The key barriers preventing this technically-mature technology from reaching its full commercial potential have been its high installation cost and limited consumer knowledge and trust in GSHP systems to deliver the technology in a cost-effective manner in the market place. Led by ENVIRON, with support from University Hartford and University of Vermont, the team developed and tested a software-based a decision making tool (‘OptGSHP’) for the least-cost design of ground-source heat pump (‘GSHP’) systems. OptGSHP combines state of the art optimization algorithms with GSHP-specific HVAC and groundwater flow and heat transport simulation. The particular strength of OptGSHP is in integrating heat transport due to groundwater flow into the design, which most of the GSHP designs do not get credit for and therefore are overdesigned.

  5. Optimizing Monitoring Designs under Alternative Objectives

    SciTech Connect

    Gastelum, Jason A.; USA, Richland Washington; Porter, Ellen A.; USA, Richland Washington

    2014-12-31

    This paper describes an approach to identify monitoring designs that optimize detection of CO2 leakage from a carbon capture and sequestration (CCS) reservoir and compares the results generated under two alternative objective functions. The first objective function minimizes the expected time to first detection of CO2 leakage, the second more conservative objective function minimizes the maximum time to leakage detection across the set of realizations. The approach applies a simulated annealing algorithm that searches the solution space by iteratively mutating the incumbent monitoring design. The approach takes into account uncertainty by evaluating the performance of potential monitoring designs across a set of simulated leakage realizations. The approach relies on a flexible two-tiered signature to infer that CO2 leakage has occurred. This research is part of the National Risk Assessment Partnership, a U.S. Department of Energy (DOE) project tasked with conducting risk and uncertainty analysis in the areas of reservoir performance, natural leakage pathways, wellbore integrity, groundwater protection, monitoring, and systems level modeling.

  6. Experimental reversion of the optimal quantum cloning and flipping processes

    SciTech Connect

    Sciarrino, Fabio; Secondi, Veronica; De Martini, Francesco

    2006-04-15

    The quantum cloner machine maps an unknown arbitrary input qubit into two optimal clones and one optimal flipped qubit. By combining linear and nonlinear optical methods we experimentally implement a scheme that, after the cloning transformation, restores the original input qubit in one of the output channels, by using local measurements, classical communication, and feedforward. This nonlocal method demonstrates how the information on the input qubit can be restored after the cloning process. The realization of the reversion process is expected to find useful applications in the field of modern multipartite quantum cryptography.

  7. Fuel optimal control of an experimental multi-mode system

    NASA Technical Reports Server (NTRS)

    Redmond, J.; Mayer, J. L.; Silverberg, L.

    1992-01-01

    In this paper, the dynamic characteristics associated with the fuel optimal control of a harmonic oscillator are utilized in the development of a near fuel optimal feedback control strategy for spacecraft vibration suppression. In this scheme, single level thrust actuators are governed by recursive computations of the standard deviations of displacement and velocity at the actuator's locations. The algorithm was tested on an experimental structure possessing a significant number of flexible body modes. The structure's response to both single and multiple mode excitation is presented.

  8. A statistical approach to optimizing concrete mixture design.

    PubMed

    Ahmad, Shamsad; Alghamdi, Saeid A

    2014-01-01

    A step-by-step statistical approach is proposed to obtain optimum proportioning of concrete mixtures using the data obtained through a statistically planned experimental program. The utility of the proposed approach for optimizing the design of concrete mixture is illustrated considering a typical case in which trial mixtures were considered according to a full factorial experiment design involving three factors and their three levels (3(3)). A total of 27 concrete mixtures with three replicates (81 specimens) were considered by varying the levels of key factors affecting compressive strength of concrete, namely, water/cementitious materials ratio (0.38, 0.43, and 0.48), cementitious materials content (350, 375, and 400 kg/m(3)), and fine/total aggregate ratio (0.35, 0.40, and 0.45). The experimental data were utilized to carry out analysis of variance (ANOVA) and to develop a polynomial regression model for compressive strength in terms of the three design factors considered in this study. The developed statistical model was used to show how optimization of concrete mixtures can be carried out with different possible options. PMID:24688405

  9. A Statistical Approach to Optimizing Concrete Mixture Design

    PubMed Central

    Alghamdi, Saeid A.

    2014-01-01

    A step-by-step statistical approach is proposed to obtain optimum proportioning of concrete mixtures using the data obtained through a statistically planned experimental program. The utility of the proposed approach for optimizing the design of concrete mixture is illustrated considering a typical case in which trial mixtures were considered according to a full factorial experiment design involving three factors and their three levels (33). A total of 27 concrete mixtures with three replicates (81 specimens) were considered by varying the levels of key factors affecting compressive strength of concrete, namely, water/cementitious materials ratio (0.38, 0.43, and 0.48), cementitious materials content (350, 375, and 400 kg/m3), and fine/total aggregate ratio (0.35, 0.40, and 0.45). The experimental data were utilized to carry out analysis of variance (ANOVA) and to develop a polynomial regression model for compressive strength in terms of the three design factors considered in this study. The developed statistical model was used to show how optimization of concrete mixtures can be carried out with different possible options. PMID:24688405

  10. Sampling design optimization for spatial functions

    USGS Publications Warehouse

    Olea, R.A.

    1984-01-01

    A new procedure is presented for minimizing the sampling requirements necessary to estimate a mappable spatial function at a specified level of accuracy. The technique is based on universal kriging, an estimation method within the theory of regionalized variables. Neither actual implementation of the sampling nor universal kriging estimations are necessary to make an optimal design. The average standard error and maximum standard error of estimation over the sampling domain are used as global indices of sampling efficiency. The procedure optimally selects those parameters controlling the magnitude of the indices, including the density and spatial pattern of the sample elements and the number of nearest sample elements used in the estimation. As an illustration, the network of observation wells used to monitor the water table in the Equus Beds of Kansas is analyzed and an improved sampling pattern suggested. This example demonstrates the practical utility of the procedure, which can be applied equally well to other spatial sampling problems, as the procedure is not limited by the nature of the spatial function. ?? 1984 Plenum Publishing Corporation.

  11. Space tourism optimized reusable spaceplane design

    NASA Astrophysics Data System (ADS)

    Penn, Jay P.; Lindley, Charles A.

    1997-01-01

    Market surveys suggest that a viable space tourism industry will require flight rates about two orders of magnitude higher than those required for conventional spacelift. Although enabling round-trip cost goals for a viable space tourism business are about $240 per pound ($529/kg), or $72,000 per passenger round-trip, goals should be about $50 per pound ($110/kg) or approximately $15,000 for a typical passenger and baggage. The lower price will probably open space tourism to the general population. Vehicle reliabilities must approach those of commercial aircraft as closely as possible. This paper addresses the development of spaceplanes optimized for the ultra-high flight rate and high reliability demands of the space tourism mission. It addresses the fundamental operability, reliability, and cost drivers needed to satisfy this mission need. Figures of merit similar to those used to evaluate the economic viability of conventional commercial aircraft are developed, including items such as payload/vehicle dry weight, turnaround time, propellant cost per passenger, and insurance and depreciation costs, which show that infrastructure can be developed for a viable space tourism industry. A reference spaceplane design optimized for space tourism is described. Subsystem allocations for reliability, operability, and costs are made and a route to developing such a capability is discussed. The vehicle's ability to also satisfy the traditional spacelift market is shown.

  12. Optimal design of robot accuracy compensators

    SciTech Connect

    Zhuang, H.; Roth, Z.S. . Robotics Center and Electrical Engineering Dept.); Hamano, Fumio . Dept. of Electrical Engineering)

    1993-12-01

    The problem of optimal design of robot accuracy compensators is addressed. Robot accuracy compensation requires that actual kinematic parameters of a robot be previously identified. Additive corrections of joint commands, including those at singular configurations, can be computed without solving the inverse kinematics problem for the actual robot. This is done by either the damped least-squares (DLS) algorithm or the linear quadratic regulator (LQR) algorithm, which is a recursive version of the DLS algorithm. The weight matrix in the performance index can be selected to achieve specific objectives, such as emphasizing end-effector's positioning accuracy over orientation accuracy or vice versa, or taking into account proximity to robot joint travel limits and singularity zones. The paper also compares the LQR and the DLS algorithms in terms of computational complexity, storage requirement, and programming convenience. Simulation results are provided to show the effectiveness of the algorithms.

  13. Three Program Architecture for Design Optimization

    NASA Technical Reports Server (NTRS)

    Miura, Hirokazu; Olson, Lawrence E. (Technical Monitor)

    1998-01-01

    In this presentation, I would like to review historical perspective on the program architecture used to build design optimization capabilities based on mathematical programming and other numerical search techniques. It is rather straightforward to classify the program architecture in three categories as shown above. However, the relative importance of each of the three approaches has not been static, instead dynamically changing as the capabilities of available computational resource increases. For example, we considered that the direct coupling architecture would never be used for practical problems, but availability of such computer systems as multi-processor. In this presentation, I would like to review the roles of three architecture from historical as well as current and future perspective. There may also be some possibility for emergence of hybrid architecture. I hope to provide some seeds for active discussion where we are heading to in the very dynamic environment for high speed computing and communication.

  14. Optimal design of a piezoelectric coupled beam for power harvesting

    NASA Astrophysics Data System (ADS)

    Wang, Quan; Wu, Nan

    2012-08-01

    An optimal design of a beam structure coupled with a piezoelectric patch is developed to achieve high efficiency of power transformation from a mechanical input to an electrical output power for the application of power harvesting. The power harvesting is realized by generating an electric voltage from the electromechanical coupling effect by the piezoelectric patch when the host beam is subjected to a dynamic loading. The electric power is then collected by a capacitor through an alternating current (AC)/direct current (DC) converter circuit (diode bridge circuit). To describe the power-harvesting process, a numerical model is developed to calculate the output voltage from the piezoelectric patch, the charge on the capacitor and the power-harvesting efficiency of the piezoelectric coupled vibrating beam. In addition, an experimental study is conducted to measure the generated voltage on the piezoelectric patch to verify the proposed numerical model. The effects of the excitation angular frequency and the patch size and location on the power-harvesting efficiency are discussed for the optimal design. The research provides a guideline for optimal designs of power-harvesting devices made of piezoelectric beam structures.

  15. Novel strategy for synthesis of magnetic dummy molecularly imprinted nanoparticles based on functionalized silica as an efficient sorbent for the determination of acrylamide in potato chips: Optimization by experimental design methodology.

    PubMed

    Arabi, Maryam; Ostovan, Abbas; Ghaedi, Mehrorang; Purkait, Mihir K

    2016-07-01

    This study discusses a novel and simple method for the preparation of magnetic dummy molecularly imprinted nanoparticles (MDMINPs). Firstly, Fe3O4 magnetic nanoparticles (MNPs) were synthesized as a magnetic component. Subsequently, MDMINPs were constructed via the sol-gel strategy using APTMS as the functional monomer. Urethane was considered as dummy template to avoid residual template and TEOS as the cross linker. The prepared MDMINPs were used for the pre-concentration of acrylamide from potato chips. Quantification was carried out by high performance liquid chromatography with UV detection (HPLC-UV). The impact of influential variables such as pH, amount of sorbent, sonication time and eluent volume were well investigated and optimized using a central composite design. The particles had excellent magnetic property and high selectivity to the targeted molecule. In optimized conditions, the recovery ranged from 94.0% to 98.0% with the detection limit of 0.35µgkg(-1). PMID:27154710

  16. Optimal screening designs for biomedical technology

    SciTech Connect

    Torney, D.C.; Bruno, W.J.; Knill, E.

    1997-10-01

    This is the final report of a three-year, Laboratory Directed Research and Development (LDRD) project at Los Alamos National Laboratory (LANL). Screening a large number of different types of molecules to isolate a few with desirable properties is essential in biomedical technology. For example, trying to find a particular gene in the Human genome could be akin to looking for a needle in a haystack. Fortunately, testing of mixtures, or pools, of molecules allows the desirable ones to be identified, using a number of experiments proportional only to the logarithm of the total number of experiments proportional only to the logarithm of the total number of types of molecules. We show how to capitalize upon this potential by using optimize pooling schemes, or designs. We propose efficient non-adaptive pooling designs, such as {open_quotes}random sets{close_quotes} designs and modified {open_quotes}row and column{close_quotes} designs. Our results have been applied in the pooling and unique-sequence screening of clone libraries used in the Human Genome Project and in the mapping of Human chromosome 16. This required the use of liquid-transferring robots and manifolds--for the largest clone libraries. Finally, we developed an efficient technique for finding the posterior probability each molecule has the desirable property, given the pool assay results. This technique works well, in practice, even if there are substantial rates of errors in the pool assay data. Both our methods and our results are relevant to a broad spectrum of research in modern biology.

  17. On the proper study design applicable to experimental balneology

    NASA Astrophysics Data System (ADS)

    Varga, Csaba

    2015-11-01

    The simple message of this paper is that it is the high time to reevaluate the strategies and optimize the efforts for investigation of thermal (spa) waters. Several articles trying to clear mode of action of medicinal waters have been published up to now. Almost all studies apply the unproven hypothesis, namely the inorganic ingredients are in close connection with healing effects of bathing. Change of paradigm would be highly necessary in this field taking into consideration the presence of several biologically active organic substances in these waters. A successful design for experimental mechanistic studies is approved.

  18. On the proper study design applicable to experimental balneology

    NASA Astrophysics Data System (ADS)

    Varga, Csaba

    2016-08-01

    The simple message of this paper is that it is the high time to reevaluate the strategies and optimize the efforts for investigation of thermal (spa) waters. Several articles trying to clear mode of action of medicinal waters have been published up to now. Almost all studies apply the unproven hypothesis, namely the inorganic ingredients are in close connection with healing effects of bathing. Change of paradigm would be highly necessary in this field taking into consideration the presence of several biologically active organic substances in these waters. A successful design for experimental mechanistic studies is approved.

  19. On the proper study design applicable to experimental balneology.

    PubMed

    Varga, Csaba

    2016-08-01

    The simple message of this paper is that it is the high time to reevaluate the strategies and optimize the efforts for investigation of thermal (spa) waters. Several articles trying to clear mode of action of medicinal waters have been published up to now. Almost all studies apply the unproven hypothesis, namely the inorganic ingredients are in close connection with healing effects of bathing. Change of paradigm would be highly necessary in this field taking into consideration the presence of several biologically active organic substances in these waters. A successful design for experimental mechanistic studies is approved. PMID:26597677

  20. Design and optimization of membrane-type acoustic metamaterials

    NASA Astrophysics Data System (ADS)

    Blevins, Matthew Grant

    One of the most common problems in noise control is the attenuation of low frequency noise. Typical solutions require barriers with high density and/or thickness. Membrane-type acoustic metamaterials are a novel type of engineered material capable of high low-frequency transmission loss despite their small thickness and light weight. These materials are ideally suited to applications with strict size and weight limitations such as aircraft, automobiles, and buildings. The transmission loss profile can be manipulated by changing the micro-level substructure, stacking multiple unit cells, or by creating multi-celled arrays. To date, analysis has focused primarily on experimental studies in plane-wave tubes and numerical modeling using finite element methods. These methods are inefficient when used for applications that require iterative changes to the structure of the material. To facilitate design and optimization of membrane-type acoustic metamaterials, computationally efficient dynamic models based on the impedance-mobility approach are proposed. Models of a single unit cell in a waveguide and in a baffle, a double layer of unit cells in a waveguide, and an array of unit cells in a baffle are studied. The accuracy of the models and the validity of assumptions used are verified using a finite element method. The remarkable computational efficiency of the impedance-mobility models compared to finite element methods enables implementation in design tools based on a graphical user interface and in optimization schemes. Genetic algorithms are used to optimize the unit cell design for a variety of noise reduction goals, including maximizing transmission loss for broadband, narrow-band, and tonal noise sources. The tools for design and optimization created in this work will enable rapid implementation of membrane-type acoustic metamaterials to solve real-world noise control problems.

  1. Chip Design Process Optimization Based on Design Quality Assessment

    NASA Astrophysics Data System (ADS)

    Häusler, Stefan; Blaschke, Jana; Sebeke, Christian; Rosenstiel, Wolfgang; Hahn, Axel

    2010-06-01

    Nowadays, the managing of product development projects is increasingly challenging. Especially the IC design of ASICs with both analog and digital components (mixed-signal design) is becoming more and more complex, while the time-to-market window narrows at the same time. Still, high quality standards must be fulfilled. Projects and their status are becoming less transparent due to this complexity. This makes the planning and execution of projects rather difficult. Therefore, there is a need for efficient project control. A main challenge is the objective evaluation of the current development status. Are all requirements successfully verified? Are all intermediate goals achieved? Companies often develop special solutions that are not reusable in other projects. This makes the quality measurement process itself less efficient and produces too much overhead. The method proposed in this paper is a contribution to solve these issues. It is applied at a German design house for analog mixed-signal IC design. This paper presents the results of a case study and introduces an optimized project scheduling on the basis of quality assessment results.

  2. Optimally designed fields for controlling molecular dynamics

    NASA Astrophysics Data System (ADS)

    Rabitz, Herschel

    1991-10-01

    This research concerns the development of molecular control theory techniques for designing optical fields capable of manipulating molecular dynamic phenomena. Although is has been long recognized that lasers should be capable of manipulating dynamic events, many frustrating years of intuitively driven laboratory studies only serve to illustrate the point that the task is complex and defies intuition. The principal new component in the present research is the recognition that this problem falls into the category of control theory and its inherent complexities require the use of modern control theory tools largely developed in the engineering disciplines. Thus, the research has initiated a transfer of the control theory concepts to the molecular scale. Although much contained effort will be needed to fully develop these concepts, the research in this grant set forth the basic components of the theory and carried out illustrative studies involving the design of optical fields capable of controlling rotational, vibrational and electronic degrees of freedom. Optimal control within the quantum mechanical molecular realm represents a frontier area with many possible ultimate applications. At this stage, the theoretical tools need to be joined with merging laboratory optical pulse shaping capabilities to illustrate the power of the concepts.

  3. Structural Optimization of a Force Balance Using a Computational Experiment Design

    NASA Technical Reports Server (NTRS)

    Parker, P. A.; DeLoach, R.

    2002-01-01

    This paper proposes a new approach to force balance structural optimization featuring a computational experiment design. Currently, this multi-dimensional design process requires the designer to perform a simplification by executing parameter studies on a small subset of design variables. This one-factor-at-a-time approach varies a single variable while holding all others at a constant level. Consequently, subtle interactions among the design variables, which can be exploited to achieve the design objectives, are undetected. The proposed method combines Modern Design of Experiments techniques to direct the exploration of the multi-dimensional design space, and a finite element analysis code to generate the experimental data. To efficiently search for an optimum combination of design variables and minimize the computational resources, a sequential design strategy was employed. Experimental results from the optimization of a non-traditional force balance measurement section are presented. An approach to overcome the unique problems associated with the simultaneous optimization of multiple response criteria is described. A quantitative single-point design procedure that reflects the designer's subjective impression of the relative importance of various design objectives, and a graphical multi-response optimization procedure that provides further insights into available tradeoffs among competing design objectives are illustrated. The proposed method enhances the intuition and experience of the designer by providing new perspectives on the relationships between the design variables and the competing design objectives providing a systematic foundation for advancements in structural design.

  4. Experimental Design to Evaluate Directed Adaptive Mutation in Mammalian Cells

    PubMed Central

    Chiaro, Christopher R; May, Tobias

    2014-01-01

    and have limited preliminary data from several pilot experiments. Cell growth and DNA sequence data indicate that we have identified a cell clone that exhibits several suitable characteristics, although further study is required to identify a more optimal cell clone. Conclusions The experimental approach is based on a quantum biological model of basis-dependent selection describing a novel mechanism of adaptive mutation. This project is currently inactive due to lack of funding. However, consistent with the objective of early reports, we describe a proposed study that has not produced publishable results, but is worthy of report because of the hypothesis, experimental design, and protocols. We outline the project’s rationale and experimental design, with its strengths and weaknesses, to stimulate discussion and analysis, and lay the foundation for future studies in this field. PMID:25491410

  5. Damage localization using experimental modal parameters and topology optimization

    NASA Astrophysics Data System (ADS)

    Niemann, Hanno; Morlier, Joseph; Shahdin, Amir; Gourinat, Yves

    2010-04-01

    This work focuses on the development of a damage detection and localization tool using the topology optimization feature of MSC.Nastran. This approach is based on the correlation of a local stiffness loss and the change in modal parameters due to damages in structures. The loss in stiffness is accounted by the topology optimization approach for updating undamaged numerical models towards similar models with embedded damages. Hereby, only a mass penalization and the changes in experimentally obtained modal parameters are used as objectives. The theoretical background for the implementation of this method is derived and programmed in a Nastran input file and the general feasibility of the approach is validated numerically, as well as experimentally by updating a model of an experimentally tested composite laminate specimen. The damages have been introduced to the specimen by controlled low energy impacts and high quality vibration tests have been conducted on the specimen for different levels of damage. These supervised experiments allow to test the numerical diagnosis tool by comparing the result with both NDT technics and results of previous works (concerning shifts in modal parameters due to damage). Good results have finally been achieved for the localization of the damages by the topology optimization.

  6. Optimization of Regression Models of Experimental Data Using Confirmation Points

    NASA Technical Reports Server (NTRS)

    Ulbrich, N.

    2010-01-01

    A new search metric is discussed that may be used to better assess the predictive capability of different math term combinations during the optimization of a regression model of experimental data. The new search metric can be determined for each tested math term combination if the given experimental data set is split into two subsets. The first subset consists of data points that are only used to determine the coefficients of the regression model. The second subset consists of confirmation points that are exclusively used to test the regression model. The new search metric value is assigned after comparing two values that describe the quality of the fit of each subset. The first value is the standard deviation of the PRESS residuals of the data points. The second value is the standard deviation of the response residuals of the confirmation points. The greater of the two values is used as the new search metric value. This choice guarantees that both standard deviations are always less or equal to the value that is used during the optimization. Experimental data from the calibration of a wind tunnel strain-gage balance is used to illustrate the application of the new search metric. The new search metric ultimately generates an optimized regression model that was already tested at regression model independent confirmation points before it is ever used to predict an unknown response from a set of regressors.

  7. Global and Local Optimization Algorithms for Optimal Signal Set Design

    PubMed Central

    Kearsley, Anthony J.

    2001-01-01

    The problem of choosing an optimal signal set for non-Gaussian detection was reduced to a smooth inequality constrained mini-max nonlinear programming problem by Gockenbach and Kearsley. Here we consider the application of several optimization algorithms, both global and local, to this problem. The most promising results are obtained when special-purpose sequential quadratic programming (SQP) algorithms are embedded into stochastic global algorithms.

  8. Design and optimization of a brachytherapy robot

    NASA Astrophysics Data System (ADS)

    Meltsner, Michael A.

    Trans-rectal ultrasound guided (TRUS) low dose rate (LDR) interstitial brachytherapy has become a popular procedure for the treatment of prostate cancer, the most common type of non-skin cancer among men. The current TRUS technique of LDR implantation may result in less than ideal coverage of the tumor with increased risk of negative response such as rectal toxicity and urinary retention. This technique is limited by the skill of the physician performing the implant, the accuracy of needle localization, and the inherent weaknesses of the procedure itself. The treatment may require 100 or more sources and 25 needles, compounding the inaccuracy of the needle localization procedure. A robot designed for prostate brachytherapy may increase the accuracy of needle placement while minimizing the effect of physician technique in the TRUS procedure. Furthermore, a robot may improve associated toxicities by utilizing angled insertions and freeing implantations from constraints applied by the 0.5 cm-spaced template used in the TRUS method. Within our group, Lin et al. have designed a new type of LDR source. The "directional" source is a seed designed to be partially shielded. Thus, a directional, or anisotropic, source does not emit radiation in all directions. The source can be oriented to irradiate cancerous tissues while sparing normal ones. This type of source necessitates a new, highly accurate method for localization in 6 degrees of freedom. A robot is the best way to accomplish this task accurately. The following presentation of work describes the invention and optimization of a new prostate brachytherapy robot that fulfills these goals. Furthermore, some research has been dedicated to the use of the robot to perform needle insertion tasks (brachytherapy, biopsy, RF ablation, etc.) in nearly any other soft tissue in the body. This can be accomplished with the robot combined with automatic, magnetic tracking.

  9. Space tourism optimized reusable spaceplane design

    SciTech Connect

    Penn, J.P.; Lindley, C.A.

    1997-01-01

    Market surveys suggest that a viable space tourism industry will require flight rates about two orders of magnitude higher than those required for conventional spacelift. Although enabling round-trip cost goals for a viable space tourism business are about {dollar_sign}240 per pound ({dollar_sign}529/kg), or {dollar_sign}72,000 per passenger round-trip, goals should be about {dollar_sign}50 per pound ({dollar_sign}110/kg) or approximately {dollar_sign}15,000 for a typical passenger and baggage. The lower price will probably open space tourism to the general population. Vehicle reliabilities must approach those of commercial aircraft as closely as possible. This paper addresses the development of spaceplanes optimized for the ultra-high flight rate and high reliability demands of the space tourism mission. It addresses the fundamental operability, reliability, and cost drivers needed to satisfy this mission need. Figures of merit similar to those used to evaluate the economic viability of conventional commercial aircraft are developed, including items such as payload/vehicle dry weight, turnaround time, propellant cost per passenger, and insurance and depreciation costs, which show that infrastructure can be developed for a viable space tourism industry. A reference spaceplane design optimized for space tourism is described. Subsystem allocations for reliability, operability, and costs are made and a route to developing such a capability is discussed. The vehicle{close_quote}s ability to also satisfy the traditional spacelift market is shown. {copyright} {ital 1997 American Institute of Physics.}

  10. Design time optimization for hardware watermarking protection of HDL designs.

    PubMed

    Castillo, E; Morales, D P; García, A; Parrilla, L; Todorovich, E; Meyer-Baese, U

    2015-01-01

    HDL-level design offers important advantages for the application of watermarking to IP cores, but its complexity also requires tools automating these watermarking algorithms. A new tool for signature distribution through combinational logic is proposed in this work. IPP@HDL, a previously proposed high-level watermarking technique, has been employed for evaluating the tool. IPP@HDL relies on spreading the bits of a digital signature at the HDL design level using combinational logic included within the original system. The development of this new tool for the signature distribution has not only extended and eased the applicability of this IPP technique, but it has also improved the signature hosting process itself. Three algorithms were studied in order to develop this automated tool. The selection of a cost function determines the best hosting solutions in terms of area and performance penalties on the IP core to protect. An 1D-DWT core and MD5 and SHA1 digital signatures were used in order to illustrate the benefits of the new tool and its optimization related to the extraction logic resources. Among the proposed algorithms, the alternative based on simulated annealing reduces the additional resources while maintaining an acceptable computation time and also saving designer effort and time. PMID:25861681

  11. Design Time Optimization for Hardware Watermarking Protection of HDL Designs

    PubMed Central

    Castillo, E.; Morales, D. P.; García, A.; Parrilla, L.; Todorovich, E.; Meyer-Baese, U.

    2015-01-01

    HDL-level design offers important advantages for the application of watermarking to IP cores, but its complexity also requires tools automating these watermarking algorithms. A new tool for signature distribution through combinational logic is proposed in this work. IPP@HDL, a previously proposed high-level watermarking technique, has been employed for evaluating the tool. IPP@HDL relies on spreading the bits of a digital signature at the HDL design level using combinational logic included within the original system. The development of this new tool for the signature distribution has not only extended and eased the applicability of this IPP technique, but it has also improved the signature hosting process itself. Three algorithms were studied in order to develop this automated tool. The selection of a cost function determines the best hosting solutions in terms of area and performance penalties on the IP core to protect. An 1D-DWT core and MD5 and SHA1 digital signatures were used in order to illustrate the benefits of the new tool and its optimization related to the extraction logic resources. Among the proposed algorithms, the alternative based on simulated annealing reduces the additional resources while maintaining an acceptable computation time and also saving designer effort and time. PMID:25861681

  12. Experimental Design for the LATOR Mission

    NASA Technical Reports Server (NTRS)

    Turyshev, Slava G.; Shao, Michael; Nordtvedt, Kenneth, Jr.

    2004-01-01

    This paper discusses experimental design for the Laser Astrometric Test Of Relativity (LATOR) mission. LATOR is designed to reach unprecedented accuracy of 1 part in 10(exp 8) in measuring the curvature of the solar gravitational field as given by the value of the key Eddington post-Newtonian parameter gamma. This mission will demonstrate the accuracy needed to measure effects of the next post-Newtonian order (near infinity G2) of light deflection resulting from gravity s intrinsic non-linearity. LATOR will provide the first precise measurement of the solar quadrupole moment parameter, J(sub 2), and will improve determination of a variety of relativistic effects including Lense-Thirring precession. The mission will benefit from the recent progress in the optical communication technologies the immediate and natural step above the standard radio-metric techniques. The key element of LATOR is a geometric redundancy provided by the laser ranging and long-baseline optical interferometry. We discuss the mission and optical designs, as well as the expected performance of this proposed mission. LATOR will lead to very robust advances in the tests of Fundamental physics: this mission could discover a violation or extension of general relativity, or reveal the presence of an additional long range interaction in the physical law. There are no analogs to the LATOR experiment; it is unique and is a natural culmination of solar system gravity experiments.

  13. Optimization of a sensitive method for the determination of nitro musk fragrances in waters by solid-phase microextraction and gas chromatography with micro electron capture detection using factorial experimental design.

    PubMed

    Polo, Maria; Garcia-Jares, Carmen; Llompart, Maria; Cela, Rafael

    2007-08-01

    A solid-phase microextraction method (SPME) followed by gas chromatography with micro electron capture detection for determining trace levels of nitro musk fragrances in residual waters was optimized. Four nitro musks, musk xylene, musk moskene, musk tibetene and musk ketone, were selected for the optimization of the method. Factors affecting the extraction process were studied using a multivariate approach. Two extraction modes (direct SPME and headspace SPME) were tried at different extraction temperatures using two fiber coatings [Carboxen-polydimethylsiloxane (CAR/PDMS) and polydimethylsiloxane-divinylbenzene (PDMS/DVB)] selected among five commercial tested fibers. Sample agitation and the salting-out effect were also factors studied. The main effects and interactions between the factors were studied for all the target compounds. An extraction temperature of 100 degrees C and sampling the headspace over the sample, using either CAR/PDMS or PDMS/DVB as fiber coatings, were found to be the experimental conditions that led to a more effective extraction. High sensitivity, with detection limits in the low nanogram per liter range, and good linearity and repeatability were achieved for all nitro musks. Since the method proposed performed well for real samples, it was applied to different water samples, including wastewater and sewage, in which some of the target compounds (musk xylene and musk ketone) were detected and quantified. PMID:17565486

  14. Optimization on Emergency Longitudinal Ventilation Design

    NASA Astrophysics Data System (ADS)

    Se, Camby M. K.; Yuen, Richard K. K.; Cheung, Sherman C. P.; Tu, Jiyuan

    2010-05-01

    Emergency ventilation design in longitudinally ventilated vehicular tunnels is vital to provide safe egress route for tunnel user under fire situations. In this study, the influences of the location of active fan groups on the upstream velocity are investigated using Computational Fluid Dynamics (CFD) techniques. The numeric model was firstly validated again the experimental data from Memorial Tunnel Fire Ventilation Test Program (MTFVTP). Based on the validated model, parametric studies were then preformed attempting to establish a semi-empirical correlation between the location of fan groups and the upstream velocity. In the presence of solid fire, it was found that the buoyant force by the fire source and inertial force by the fans interact with each other and resulted in a "leveling-off" characteristic when the inertial force is no longer dominating. Such interaction re-distributed the ventilation flow direction and sequentially reduces the magnitude of the upstream velocity. In other word, the industrial practice of activating furthest fan group may not be able to prevent the backlayering as a consequence of solid fires. Fans closer to the fire source are recommended to be activated for preventing the hazard of backlayering. Furthermore, through the parametric study, location of ventilation fans is found to have significant effect on the upstream velocity. Such finding suggests that other geometrical parameters could also impose adverse effects to the ventilation system. Existing empirical equation could be insufficient to cover all possible ventilation design scenarios.

  15. Experimental Optimization of a Free-to-Rotate Wing for Small UAS

    NASA Technical Reports Server (NTRS)

    Logan, Michael J.; DeLoach, Richard; Copeland, Tiwana; Vo, Steven

    2014-01-01

    This paper discusses an experimental investigation conducted to optimize a free-to-rotate wing for use on a small unmanned aircraft system (UAS). Although free-to-rotate wings have been used for decades on various small UAS and small manned aircraft, little is known about how to optimize these unusual wings for a specific application. The paper discusses some of the design rationale of the basic wing. In addition, three main parameters were selected for "optimization", wing camber, wing pivot location, and wing center of gravity (c.g.) location. A small apparatus was constructed to enable some simple experimental analysis of these parameters. A design-of-experiment series of tests were first conducted to discern which of the main optimization parameters were most likely to have the greatest impact on the outputs of interest, namely, some measure of "stability", some measure of the lift being generated at the neutral position, and how quickly the wing "recovers" from an upset. A second set of tests were conducted to develop a response-surface numerical representation of these outputs as functions of the three primary inputs. The response surface numerical representations are then used to develop an "optimum" within the trade space investigated. The results of the optimization are then tested experimentally to validate the predictions.

  16. Illumination system design with multi-step optimization

    NASA Astrophysics Data System (ADS)

    Magarill, Simon; Cassarly, William J.

    2015-08-01

    Automatic optimization algorithms can be used when designing illumination systems. For systems with many design variables, optimization using an adjustable set of variables at different steps of the process can provide different local minima. We present a few examples of implementing a multi-step optimization method. We have found that this approach can sometimes lead to more efficient solutions. In this paper we illustrate the effectiveness of using a commercially available optimization algorithm with a slightly modified procedure.

  17. Integrated topology and shape optimization in structural design

    NASA Technical Reports Server (NTRS)

    Bremicker, M.; Chirehdast, M.; Kikuchi, N.; Papalambros, P. Y.

    1990-01-01

    Structural optimization procedures usually start from a given design topology and vary its proportions or boundary shapes to achieve optimality under various constraints. Two different categories of structural optimization are distinguished in the literature, namely sizing and shape optimization. A major restriction in both cases is that the design topology is considered fixed and given. Questions concerning the general layout of a design (such as whether a truss or a solid structure should be used) as well as more detailed topology features (e.g., the number and connectivities of bars in a truss or the number of holes in a solid) have to be resolved by design experience before formulating the structural optimization model. Design quality of an optimized structure still depends strongly on engineering intuition. This article presents a novel approach for initiating formal structural optimization at an earlier stage, where the design topology is rigorously generated in addition to selecting shape and size dimensions. A three-phase design process is discussed: an optimal initial topology is created by a homogenization method as a gray level image, which is then transformed to a realizable design using computer vision techniques; this design is then parameterized and treated in detail by sizing and shape optimization. A fully automated process is described for trusses. Optimization of two dimensional solid structures is also discussed. Several application-oriented examples illustrate the usefulness of the proposed methodology.

  18. Optimizing Adhesive Design by Understanding Compliance.

    PubMed

    King, Daniel R; Crosby, Alfred J

    2015-12-23

    Adhesives have long been designed around a trade-off between adhesive strength and releasability. Geckos are of interest because they are the largest organisms which are able to climb utilizing adhesive toepads, yet can controllably release from surfaces and perform this action over and over again. Attempting to replicate the hierarchical, nanoscopic features which cover their toepads has been the primary focus of the adhesives field until recently. A new approach based on a scaling relation which states that reversible adhesive force capacity scales with (A/C)(1/2), where A is the area of contact and C is the compliance of the adhesive, has enabled the creation of high strength, reversible adhesives without requiring high aspect ratio, fibrillar features. Here we introduce an equation to calculate the compliance of adhesives, and utilize this equation to predict the shear adhesive force capacity of the adhesive based on the material components and geometric properties. Using this equation, we have investigated important geometric parameters which control force capacity and have shown that by controlling adhesive shape, adhesive force capacity can be increased by over 50% without varying pad size. Furthermore, we have demonstrated that compliance of the adhesive far from the interface still influences shear adhesive force capacity. Utilizing this equation will allow for the production of adhesives which are optimized for specific applications in commercial and industrial settings. PMID:26618537

  19. Multidisciplinary design optimization - An emerging new engineering discipline

    NASA Technical Reports Server (NTRS)

    Sobieszczanski-Sobieski, Jaroslaw

    1993-01-01

    A definition of the multidisciplinary design optimization (MDO) is introduced, and functionality and relationship of the MDO conceptual components are examined. The latter include design-oriented analysis, approximation concepts, mathematical system modeling, design space search, an optimization procedure, and a humane interface.

  20. Rapid convergence of airfoil design problems using progressive optimization

    NASA Astrophysics Data System (ADS)

    Dadone, A.; Grossman, B.

    An efficient formulation for the robust design optimization of compressible fluid flow problems is presented. The methodology has three essential ingredients: a highly accurate flow solver, robust and efficient design sensitivities from a discrete adjoint formulation based on a dissipative flow solver and progressive optimization, whereby a sequence of operations, containing a partially converged flow solution, followed by an adjoint solution followed by an optimization step is performed. Furthermore, the progressive optimization involves the use of progressively finer grids. The methodology is shown to be accurate, robust and highly efficient, with a converged design optimization produced in no more than the amount of computational work to perform from one to three flow analyses.

  1. A robust optimization methodology for preliminary aircraft design

    NASA Astrophysics Data System (ADS)

    Prigent, S.; Maréchal, P.; Rondepierre, A.; Druot, T.; Belleville, M.

    2016-05-01

    This article focuses on a robust optimization of an aircraft preliminary design under operational constraints. According to engineers' know-how, the aircraft preliminary design problem can be modelled as an uncertain optimization problem whose objective (the cost or the fuel consumption) is almost affine, and whose constraints are convex. It is shown that this uncertain optimization problem can be approximated in a conservative manner by an uncertain linear optimization program, which enables the use of the techniques of robust linear programming of Ben-Tal, El Ghaoui, and Nemirovski [Robust Optimization, Princeton University Press, 2009]. This methodology is then applied to two real cases of aircraft design and numerical results are presented.

  2. Optimal structural design of the Airborne Infrared Imager

    NASA Astrophysics Data System (ADS)

    Doyle, Keith B.; Cerrati, Vincent J.; Forman, Steven E.; Sultana, John A.

    1995-09-01

    The airborne infrared imager (AIRI) is a dual-band IR sensor designed to study air defense issues while wing mounted in a pod. The sensor consists of an optical bench attached to a two- axis inertially stabilized gimbal structure in elevation and azimuth. The gimbal assembly operates within an 18-inch diameter globe while meeting strict pointing and tracking requirements. Design conditions for the assembly include operational and nonoperational inertial, thermal, and dynamic loads. Primary design efforts centered on limiting the line-of- sight jitter of the optical system to 50 (mu) rad under the operating environment. An MSC/NASTRAN finite element model was developed for structural response predictions and correlated to experimental data. Design changes were aided by MSC/NASTRAN's optimization routine with the goal of maximizing the fundamental frequency of the gimbal assembly. The final structural design resultsed in a first natural frequency of 79 Hz using a titanium azimuthal gimbal, a stainless steel elevation gimbal, and an aluminum optical bench which met the design and performance requirements.

  3. Optimizing the design of geophysical experiments: Is it worthwhile?

    NASA Astrophysics Data System (ADS)

    Curtis, Andrew; Maurer, Hansruedi

    Determining the structure, composition, and state of the Earth's subsurface from measured data is the principal task of many geophysical experiments and surveys. Standard procedures involve the recording of appropriate data sets followed by the application of data analysis techniques to extract the desired information. While the importance of new tools for the analysis stage of an experiment is well recognized, much less attention seems to be paid to improving the data acquisition.A measure of the effort allocated to data analysis research relative to that devoted to data acquisition research is presented in Figure 1. Since 1955 there have been more than 10,000 publications on inversion methods alone, but in the same period only 100 papers on experimental design have appeared in journals. Considering that the acquisition component of an experiment defines what information will be contained in the data, and that no amount of data analysis can compensate for the lack of such information, we suggest that greater effort be made to improve survey planning techniques. Furthermore, given that logistical and financial constraints are often stringent and that relationships between geophysical data and model parameters describing the Earths subsurface are generally complicated, optimizing the design of an experiment may be quite challenging. Here we review experimental design procedures that optimize the benefit of a field survey, such that maximum information about the target structures is obtained at minimum cost. We also announce a new Web site and e-mail group set up as a forum for communication on survey design research and application.

  4. Design and optimization of bilayered tablet of Hydrochlorothiazide using the Quality-by-Design approach

    PubMed Central

    Dholariya, Yatin N; Bansod, Yogesh B; Vora, Rahul M; Mittal, Sandeep S; Shirsat, Ajinath Eknath; Bhingare, Chandrashekhar L

    2014-01-01

    Aim: The aim of the present study is to develop an optimize bilayered tablet using Hydrochlorothiazide (HCTZ) as a model drug candidate using quality by design (QbD) approach. Introduction and Method: The bilayered tablet gives biphasic drug release through loading dose; prepared using croscarmellose sodium a superdisintegrant and maintenance dose using several viscosity grades of hydrophilic polymers. The fundamental principle of QbD is to demonstrate understanding and control of pharmaceutical processes so as to deliver high quality pharmaceutical products with wide opportunities for continuous improvement. Risk assessment was carried out and subsequently 22 factorial designs in duplicate was selected to carry out design of experimentation (DOE) for evaluating the interactions and effects of the design factors on critical quality attribute. The design space was obtained by applying DOE and multivariate analysis, so as to ensure desired disintegration time (DT) and drug release is achieved. Bilayered tablet were evaluated for hardness, thickness, friability, drug content uniformity and in vitro drug dissolution. Result: Optimized formulation obtained from the design space exhibits DT of around 70 s, while DR T95% (time required to release 95% of the drug) was about 720 min. Kinetic studies of formulations revealed that erosion is the predominant mechanism for drug release. Conclusion: From the obtained results; it was concluded that independent variables have a significant effect over the dependent responses, which can be deduced from half normal plots, pareto charts and surface response graphs. The predicted values matched well with the experimental values and the result demonstrates the feasibility of the design model in the development and optimization of HCTZ bilayered tablet. PMID:25006554

  5. Optimal Design of a Center Support Quadruple Mass Gyroscope (CSQMG).

    PubMed

    Zhang, Tian; Zhou, Bin; Yin, Peng; Chen, Zhiyong; Zhang, Rong

    2016-01-01

    This paper reports a more complete description of the design process of the Center Support Quadruple Mass Gyroscope (CSQMG), a gyro expected to provide breakthrough performance for flat structures. The operation of the CSQMG is based on four lumped masses in a circumferential symmetric distribution, oscillating in anti-phase motion, and providing differential signal extraction. With its 4-fold symmetrical axes pattern, the CSQMG achieves a similar operation mode to Hemispherical Resonant Gyroscopes (HRGs). Compared to the conventional flat design, four Y-shaped coupling beams are used in this new pattern in order to adjust mode distribution and enhance the synchronization mechanism of operation modes. For the purpose of obtaining the optimal design of the CSQMG, a kind of applicative optimization flow is developed with a comprehensive derivation of the operation mode coordination, the pseudo mode inhibition, and the lumped mass twisting motion elimination. The experimental characterization of the CSQMG was performed at room temperature, and the center operation frequency is 6.8 kHz after tuning. Experiments show an Allan variance stability 0.12°/h (@100 s) and a white noise level about 0.72°/h/√Hz, which means that the CSQMG possesses great potential to achieve navigation grade performance. PMID:27136557

  6. Active cooling design for scramjet engines using optimization methods

    NASA Technical Reports Server (NTRS)

    Scotti, Stephen J.; Martin, Carl J.; Lucas, Stephen H.

    1988-01-01

    A methodology for using optimization in designing metallic cooling jackets for scramjet engines is presented. The optimal design minimizes the required coolant flow rate subject to temperature, mechanical-stress, and thermal-fatigue-life constraints on the cooling-jacket panels, and Mach-number and pressure constraints on the coolant exiting the panel. The analytical basis for the methodology is presented, and results for the optimal design of panels are shown to demonstrate its utility.

  7. Active cooling design for scramjet engines using optimization methods

    NASA Technical Reports Server (NTRS)

    Scotti, Stephen J.; Martin, Carl J.; Lucas, Stephen H.

    1988-01-01

    A methodology for using optimization in designing metallic cooling jackets for scramjet engines is presented. The optimal design minimizes the required coolant flow rate subject to temperature, mechanical-stress, and thermal-fatigue-life constraints on the cooling-jacket panels, and Mach-number and pressure contraints on the coolant exiting the panel. The analytical basis for the methodology is presented, and results for the optimal design of panels are shown to demonstrate its utility.

  8. Optimization and experimental validation of electrostatic adhesive geometry

    NASA Astrophysics Data System (ADS)

    Ruffatto, D.; Shah, J.; Spenko, M.

    This paper introduces a method to optimize the electrode geometry of electrostatic adhesives for robotic gripping, attachment, and manipulation applications. Electrostatic adhesion is achieved by applying a high voltage potential, on the order of kV, to a set of electrodes, which generates an electric field. The electric field polarizes the substrate material and creates an adhesion force. Previous attempts at creating electro-static adhesives have shown them to be effective, but researchers have made no effort to optimize the electrode configuration and geometry. We have shown that by optimizing the geometry of the electrode configuration, the electric field strength, and therefore the adhesion force, is enhanced. To accomplish this, Comsol Multiphysics was utilized to evaluate the average electric field generated by a given electrode geometry. Several electrode patterns were evaluated, including parallel conductors, concentric circles, Hilbert curves (a fractal geometry) and spirals. The arrangement of the electrodes in concentric circles with varying electrode widths proved to be the most effective. The most effective sizing was to use the smallest gap spacing allowable coupled with a variable electrode width. These results were experimentally validated on several different surfaces including drywall, wood, tile, glass, and steel. A new manufacturing process allowing for the fabrication of thin, conformal electro-static adhesive pads was utilized. By combining the optimized electrode geometry with the new fabrication process we are able to demonstrate a marked improvement of up to 500% in shear pressure when compared to previously published values.

  9. Optimal Cluster Sizes for Wireless Sensor Networks: An Experimental Analysis

    NASA Astrophysics Data System (ADS)

    Förster, Anna; Förster, Alexander; Murphy, Amy L.

    Node clustering and data aggregation are popular techniques to reduce energy consumption in large WSNs and a large body of literature has emerged describing various clustering protocols. Unfortunately, for practitioners wishing to exploit clustering in deployments, there is little help when trying to identify a protocol that meets their needs. This paper takes a step back from specific protocols to consider the fundamental question: what is the optimal cluster size in terms of the resulting communication generated to collect data. Our experimental analysis considers a wide range of parameters that characterize the WSN, and shows that in the most common cases, clusters in which all nodes can communicate in one hop to the cluster head are optimal.

  10. Electronic enclosure design using distributed particle swarm optimization

    NASA Astrophysics Data System (ADS)

    Scriven, Ian; Lu, Junwei; Lewis, Andrew

    2013-02-01

    This article proposes a method for designing electromagnetic compatibility shielding enclosures using a peer-to-peer based distributed optimization system based on a modified particle swarm optimization algorithm. This optimization system is used to obtain optimal solutions to a shielding enclosure design problem efficiently with respect to both electromagnetic shielding efficiency and thermal performance. During the optimization procedure it becomes evident that optimization algorithms and computational models must be properly matched in order to achieve efficient operation. The proposed system is designed to be tolerant of faults and resource heterogeneity, and as such would find use in environments where large-scale computing resources are not available, such as smaller engineering companies, where it would allow computer-aided design by optimization using existing resources with little to no financial outlay.

  11. Globally optimal trial design for local decision making.

    PubMed

    Eckermann, Simon; Willan, Andrew R

    2009-02-01

    Value of information methods allows decision makers to identify efficient trial design following a principle of maximizing the expected value to decision makers of information from potential trial designs relative to their expected cost. However, in health technology assessment (HTA) the restrictive assumption has been made that, prospectively, there is only expected value of sample information from research commissioned within jurisdiction. This paper extends the framework for optimal trial design and decision making within jurisdiction to allow for optimal trial design across jurisdictions. This is illustrated in identifying an optimal trial design for decision making across the US, the UK and Australia for early versus late external cephalic version for pregnant women presenting in the breech position. The expected net gain from locally optimal trial designs of US$0.72M is shown to increase to US$1.14M with a globally optimal trial design. In general, the proposed method of globally optimal trial design improves on optimal trial design within jurisdictions by: (i) reflecting the global value of non-rival information; (ii) allowing optimal allocation of trial sample across jurisdictions; (iii) avoiding market failure associated with free-rider effects, sub-optimal spreading of fixed costs and heterogeneity of trial information with multiple trials. PMID:18435429

  12. A new interval optimization method considering tolerance design

    NASA Astrophysics Data System (ADS)

    Jiang, C.; Xie, H. C.; Zhang, Z. G.; Han, X.

    2015-12-01

    This study considers the design variable uncertainty in the actual manufacturing process for a product or structure and proposes a new interval optimization method based on tolerance design, which can provide not only an optimal design but also the allowable maximal manufacturing errors that the design can bear. The design variables' manufacturing errors are depicted using the interval method, and an interval optimization model for the structure is constructed. A dimensionless design tolerance index is defined to describe the overall uncertainty of all design variables, and by combining the nominal objective function, a deterministic two-objective optimization model is built. The possibility degree of interval is used to represent the reliability of the constraints under uncertainty, through which the model is transformed to a deterministic optimization problem. Three numerical examples are investigated to verify the effectiveness of the present method.

  13. Optimal trajectories for flexible-link manipulator slewing using recursive quadratic programming: Experimental verification

    SciTech Connect

    Parker, G.G.; Eisler, G.R.; Feddema, J.T.

    1994-09-01

    Procedures for trajectory planning and control of flexible link robots are becoming increasingly important to satisfy performance requirements of hazardous waste removal efforts. It has been shown that utilizing link flexibility in designing open loop joint commands can result in improved performance as opposed to damping vibration throughout a trajectory. The efficient use of link compliance is exploited in this work. Specifically, experimental verification of minimum time, straight line tracking using a two-link planar flexible robot is presented. A numerical optimization process, using an experimentally verified modal model, is used for obtaining minimum time joint torque and angle histories. The optimal joint states are used as commands to the proportional-derivative servo actuated joints. These commands are precompensated for the nonnegligible joint servo actuator dynamics. Using the precompensated joint commands, the optimal joint angles are tracked with such fidelity that the tip tracking error is less than 2.5 cm.

  14. Entropy-Based Search Algorithm for Experimental Design

    NASA Astrophysics Data System (ADS)

    Malakar, N. K.; Knuth, K. H.

    2011-03-01

    The scientific method relies on the iterated processes of inference and inquiry. The inference phase consists of selecting the most probable models based on the available data; whereas the inquiry phase consists of using what is known about the models to select the most relevant experiment. Optimizing inquiry involves searching the parameterized space of experiments to select the experiment that promises, on average, to be maximally informative. In the case where it is important to learn about each of the model parameters, the relevance of an experiment is quantified by Shannon entropy of the distribution of experimental outcomes predicted by a probable set of models. If the set of potential experiments is described by many parameters, we must search this high-dimensional entropy space. Brute force search methods will be slow and computationally expensive. We present an entropy-based search algorithm, called nested entropy sampling, to select the most informative experiment for efficient experimental design. This algorithm is inspired by Skilling's nested sampling algorithm used in inference and borrows the concept of a rising threshold while a set of experiment samples are maintained. We demonstrate that this algorithm not only selects highly relevant experiments, but also is more efficient than brute force search. Such entropic search techniques promise to greatly benefit autonomous experimental design.

  15. A Modified Particle Swarm Optimization Technique for Finding Optimal Designs for Mixture Models.

    PubMed

    Wong, Weng Kee; Chen, Ray-Bing; Huang, Chien-Chih; Wang, Weichung

    2015-01-01

    Particle Swarm Optimization (PSO) is a meta-heuristic algorithm that has been shown to be successful in solving a wide variety of real and complicated optimization problems in engineering and computer science. This paper introduces a projection based PSO technique, named ProjPSO, to efficiently find different types of optimal designs, or nearly optimal designs, for mixture models with and without constraints on the components, and also for related models, like the log contrast models. We also compare the modified PSO performance with Fedorov's algorithm, a popular algorithm used to generate optimal designs, Cocktail algorithm, and the recent algorithm proposed by [1]. PMID:26091237

  16. A Modified Particle Swarm Optimization Technique for Finding Optimal Designs for Mixture Models

    PubMed Central

    Wong, Weng Kee; Chen, Ray-Bing; Huang, Chien-Chih; Wang, Weichung

    2015-01-01

    Particle Swarm Optimization (PSO) is a meta-heuristic algorithm that has been shown to be successful in solving a wide variety of real and complicated optimization problems in engineering and computer science. This paper introduces a projection based PSO technique, named ProjPSO, to efficiently find different types of optimal designs, or nearly optimal designs, for mixture models with and without constraints on the components, and also for related models, like the log contrast models. We also compare the modified PSO performance with Fedorov's algorithm, a popular algorithm used to generate optimal designs, Cocktail algorithm, and the recent algorithm proposed by [1]. PMID:26091237

  17. A multiple objective optimization approach to aircraft control systems design

    NASA Technical Reports Server (NTRS)

    Tabak, D.; Schy, A. A.; Johnson, K. G.; Giesy, D. P.

    1979-01-01

    The design of an aircraft lateral control system, subject to several performance criteria and constraints, is considered. While in the previous studies of the same model a single criterion optimization, with other performance requirements expressed as constraints, has been pursued, the current approach involves a multiple criteria optimization. In particular, a Pareto optimal solution is sought.

  18. INNOVATIVE METHODS FOR THE OPTIMIZATION OF GRAVITY STORM SEWER DESIGN

    EPA Science Inventory

    The purpose of this paper is to describe a new method for optimizing the design of urban storm sewer systems. Previous efforts to optimize gravity sewers have met with limited success because classical optimization methods require that the problem be well behaved, e.g. describ...

  19. Post-Optimality Analysis In Aerospace Vehicle Design

    NASA Technical Reports Server (NTRS)

    Braun, Robert D.; Kroo, Ilan M.; Gage, Peter J.

    1993-01-01

    This analysis pertains to the applicability of optimal sensitivity information to aerospace vehicle design. An optimal sensitivity (or post-optimality) analysis refers to computations performed once the initial optimization problem is solved. These computations may be used to characterize the design space about the present solution and infer changes in this solution as a result of constraint or parameter variations, without reoptimizing the entire system. The present analysis demonstrates that post-optimality information generated through first-order computations can be used to accurately predict the effect of constraint and parameter perturbations on the optimal solution. This assessment is based on the solution of an aircraft design problem in which the post-optimality estimates are shown to be within a few percent of the true solution over the practical range of constraint and parameter variations. Through solution of a reusable, single-stage-to-orbit, launch vehicle design problem, this optimal sensitivity information is also shown to improve the efficiency of the design process, For a hierarchically decomposed problem, this computational efficiency is realized by estimating the main-problem objective gradient through optimal sep&ivity calculations, By reducing the need for finite differentiation of a re-optimized subproblem, a significant decrease in the number of objective function evaluations required to reach the optimal solution is obtained.

  20. Two-stage microbial community experimental design.

    PubMed

    Tickle, Timothy L; Segata, Nicola; Waldron, Levi; Weingart, Uri; Huttenhower, Curtis

    2013-12-01

    Microbial community samples can be efficiently surveyed in high throughput by sequencing markers such as the 16S ribosomal RNA gene. Often, a collection of samples is then selected for subsequent metagenomic, metabolomic or other follow-up. Two-stage study design has long been used in ecology but has not yet been studied in-depth for high-throughput microbial community investigations. To avoid ad hoc sample selection, we developed and validated several purposive sample selection methods for two-stage studies (that is, biological criteria) targeting differing types of microbial communities. These methods select follow-up samples from large community surveys, with criteria including samples typical of the initially surveyed population, targeting specific microbial clades or rare species, maximizing diversity, representing extreme or deviant communities, or identifying communities distinct or discriminating among environment or host phenotypes. The accuracies of each sampling technique and their influences on the characteristics of the resulting selected microbial community were evaluated using both simulated and experimental data. Specifically, all criteria were able to identify samples whose properties were accurately retained in 318 paired 16S amplicon and whole-community metagenomic (follow-up) samples from the Human Microbiome Project. Some selection criteria resulted in follow-up samples that were strongly non-representative of the original survey population; diversity maximization particularly undersampled community configurations. Only selection of intentionally representative samples minimized differences in the selected sample set from the original microbial survey. An implementation is provided as the microPITA (Microbiomes: Picking Interesting Taxa for Analysis) software for two-stage study design of microbial communities. PMID:23949665

  1. Multidisciplinary design optimization: An emerging new engineering discipline

    NASA Technical Reports Server (NTRS)

    Sobieszczanski-Sobieski, Jaroslaw

    1993-01-01

    This paper defines the Multidisciplinary Design Optimization (MDO) as a new field of research endeavor and as an aid in the design of engineering systems. It examines the MDO conceptual components in relation to each other and defines their functions.

  2. A design optimization process for Space Station Freedom

    NASA Technical Reports Server (NTRS)

    Chamberlain, Robert G.; Fox, George; Duquette, William H.

    1990-01-01

    The Space Station Freedom Program is used to develop and implement a process for design optimization. Because the relative worth of arbitrary design concepts cannot be assessed directly, comparisons must be based on designs that provide the same performance from the point of view of station users; such designs can be compared in terms of life cycle cost. Since the technology required to produce a space station is widely dispersed, a decentralized optimization process is essential. A formulation of the optimization process is provided and the mathematical models designed to facilitate its implementation are described.

  3. A Matrix-Free Algorithm for Multidisciplinary Design Optimization

    NASA Astrophysics Data System (ADS)

    Lambe, Andrew Borean

    Multidisciplinary design optimization (MDO) is an approach to engineering design that exploits the coupling between components or knowledge disciplines in a complex system to improve the final product. In aircraft design, MDO methods can be used to simultaneously design the outer shape of the aircraft and the internal structure, taking into account the complex interaction between the aerodynamic forces and the structural flexibility. Efficient strategies are needed to solve such design optimization problems and guarantee convergence to an optimal design. This work begins with a comprehensive review of MDO problem formulations and solution algorithms. First, a fundamental MDO problem formulation is defined from which other formulations may be obtained through simple transformations. Using these fundamental problem formulations, decomposition methods from the literature are reviewed and classified. All MDO methods are presented in a unified mathematical notation to facilitate greater understanding. In addition, a novel set of diagrams, called extended design structure matrices, are used to simultaneously visualize both data communication and process flow between the many software components of each method. For aerostructural design optimization, modern decomposition-based MDO methods cannot efficiently handle the tight coupling between the aerodynamic and structural states. This fact motivates the exploration of methods that can reduce the computational cost. A particular structure in the direct and adjoint methods for gradient computation motivates the idea of a matrix-free optimization method. A simple matrix-free optimizer is developed based on the augmented Lagrangian algorithm. This new matrix-free optimizer is tested on two structural optimization problems and one aerostructural optimization problem. The results indicate that the matrix-free optimizer is able to efficiently solve structural and multidisciplinary design problems with thousands of variables and

  4. A Matrix-Free Algorithm for Multidisciplinary Design Optimization

    NASA Astrophysics Data System (ADS)

    Lambe, Andrew Borean

    Multidisciplinary design optimization (MDO) is an approach to engineering design that exploits the coupling between components or knowledge disciplines in a complex system to improve the final product. In aircraft design, MDO methods can be used to simultaneously design the outer shape of the aircraft and the internal structure, taking into account the complex interaction between the aerodynamic forces and the structural flexibility. Efficient strategies are needed to solve such design optimization problems and guarantee convergence to an optimal design. This work begins with a comprehensive review of MDO problem formulations and solution algorithms. First, a fundamental MDO problem formulation is defined from which other formulations may be obtained through simple transformations. Using these fundamental problem formulations, decomposition methods from the literature are reviewed and classified. All MDO methods are presented in a unified mathematical notation to facilitate greater understanding. In addition, a novel set of diagrams, called extended design structure matrices, are used to simultaneously visualize both data communication and process flow between the many software components of each method. For aerostructural design optimization, modern decomposition-based MDO methods cannot efficiently handle the tight coupling between the aerodynamic and structural states. This fact motivates the exploration of methods that can reduce the computational cost. A particular structure in the direct and adjoint methods for gradient computation. motivates the idea of a matrix-free optimization method. A simple matrix-free optimizer is developed based on the augmented Lagrangian algorithm. This new matrix-free optimizer is tested on two structural optimization problems and one aerostructural optimization problem. The results indicate that the matrix-free optimizer is able to efficiently solve structural and multidisciplinary design problems with thousands of variables and

  5. [Design and experimentation of marine optical buoy].

    PubMed

    Yang, Yue-Zhong; Sun, Zhao-Hua; Cao, Wen-Xi; Li, Cai; Zhao, Jun; Zhou, Wen; Lu, Gui-Xin; Ke, Tian-Cun; Guo, Chao-Ying

    2009-02-01

    Marine optical buoy is of important value in terms of calibration and validation of ocean color remote sensing, scientific observation, coastal environment monitoring, etc. A marine optical buoy system was designed which consists of a main and a slave buoy. The system can measure the distribution of irradiance and radiance over the sea surface, in the layer near sea surface and in the euphotic zone synchronously, during which some other parameters are also acquired such as spectral absorption and scattering coefficients of the water column, the velocity and direction of the wind, and so on. The buoy was positioned by GPS. The low-power integrated PC104 computer was used as the control core to collect data automatically. The data and commands were real-timely transmitted by CDMA/GPRS wireless networks or by the maritime satellite. The coastal marine experimentation demonstrated that the buoy has small pitch and roll rates in high sea state conditions and thus can meet the needs of underwater radiometric measurements, the data collection and remote transmission are reliable, and the auto-operated anti-biofouling devices can ensure that the optical sensors work effectively for a period of several months. PMID:19445253

  6. Maximizing the efficiency of a flexible propulsor using experimental optimization

    NASA Astrophysics Data System (ADS)

    Quinn, Daniel; Lauder, George; Smits, Alexander

    2014-11-01

    Experimental gradient-based optimization is used to maximize the propulsive efficiency of a heaving and pitching flexible panel. Optimum and near-optimum conditions are studied via direct force measurements and Particle Image Velocimetry (PIV). The net thrust and power are found to scale predictably with the frequency and amplitude of the leading edge, but the efficiency shows a complex multimodal response. Optimum pitch and heave motions are found to produce nearly twice the efficiencies of optimum heave-only motions. Efficiency is globally optimized when (1) the Strouhal number is within an optimal range that varies weakly with amplitude and boundary conditions; (2) the panel is actuated at a resonant frequency of the fluid-propulsor system; (3) heave amplitude is tuned such that trailing edge amplitude is maximized while flow along the body remains attached; and (4) the maximum pitch angle and phase lag are chosen so that the effective angle of attack is minimized. This work was supported by the Office of Naval Research under MURI Grant Number N00014-08-1-0642 (Program Director Dr. Bob Brizzolara), and the National Science Foundation under Grant DBI 1062052 (PI Lisa Fauci) and Grant EFRI-0938043 (PI George Lauder).

  7. Design Optimization for Anharmonic Linear Surface-Electrode Ion Trap

    NASA Astrophysics Data System (ADS)

    Liu, Wei; Chen, Shu-Ming; Chen, Ping-Xing; Wu, Wei

    2014-11-01

    An accurate and rapid method is proposed to optimize anharmonic linear surface-electrode ion trap design. Based on the method, we analyze the impact of the architectural parameters, including the width, number, and applied voltage of prerequisite active electrodes, on the number and spacing of trapped ions. Sets of optimal anharmonic trap design are given. Then the optimal designs are verified by using an ant colony optimization algorithm. The results show that the maximum ion position errors and maximum ion spacing errors are less than 1 μm up to 80. The mean of the maximum errors is nearly linear with respect to the number of trapped ions.

  8. Experimental designs and their recent advances in set-up, data interpretation, and analytical applications.

    PubMed

    Dejaegher, Bieke; Heyden, Yvan Vander

    2011-09-10

    In this review, the set-up and data interpretation of experimental designs (screening, response surface, and mixture designs) are discussed. Advanced set-ups considered are the application of D-optimal and supersaturated designs as screening designs. Advanced data interpretation approaches discussed are an adaptation of the algorithm of Dong and the estimation of factor effects from supersaturated design results. Finally, some analytical applications in separation science, on the one hand, and formulation-, product-, or process optimization, on the other, are discussed. PMID:21632194

  9. Optimal flexible sample size design with robust power.

    PubMed

    Zhang, Lanju; Cui, Lu; Yang, Bo

    2016-08-30

    It is well recognized that sample size determination is challenging because of the uncertainty on the treatment effect size. Several remedies are available in the literature. Group sequential designs start with a sample size based on a conservative (smaller) effect size and allow early stop at interim looks. Sample size re-estimation designs start with a sample size based on an optimistic (larger) effect size and allow sample size increase if the observed effect size is smaller than planned. Different opinions favoring one type over the other exist. We propose an optimal approach using an appropriate optimality criterion to select the best design among all the candidate designs. Our results show that (1) for the same type of designs, for example, group sequential designs, there is room for significant improvement through our optimization approach; (2) optimal promising zone designs appear to have no advantages over optimal group sequential designs; and (3) optimal designs with sample size re-estimation deliver the best adaptive performance. We conclude that to deal with the challenge of sample size determination due to effect size uncertainty, an optimal approach can help to select the best design that provides most robust power across the effect size range of interest. Copyright © 2016 John Wiley & Sons, Ltd. PMID:26999385

  10. Topology and boundary shape optimization as an integrated design tool

    NASA Technical Reports Server (NTRS)

    Bendsoe, Martin Philip; Rodrigues, Helder Carrico

    1990-01-01

    The optimal topology of a two dimensional linear elastic body can be computed by regarding the body as a domain of the plane with a high density of material. Such an optimal topology can then be used as the basis for a shape optimization method that computes the optimal form of the boundary curves of the body. This results in an efficient and reliable design tool, which can be implemented via common FEM mesh generator and CAD type input-output facilities.

  11. Control structure interaction/optimized design

    NASA Technical Reports Server (NTRS)

    Mclaren, Mark; Purvis, Chris

    1994-01-01

    The objective of this study is to apply the integrated design methodology to the mature GOES-1 spacecraft design, and to assess the possible advantages to be gained using this approach over the conventional sequential design approach used for the current design. In the process, the development of this technology into a tool that can be utilized for future near-term spacecraft designs is emphasized.

  12. A study of commuter airplane design optimization

    NASA Technical Reports Server (NTRS)

    Roskam, J.; Wyatt, R. D.; Griswold, D. A.; Hammer, J. L.

    1977-01-01

    Problems of commuter airplane configuration design were studied to affect a minimization of direct operating costs. Factors considered were the minimization of fuselage drag, methods of wing design, and the estimated drag of an airplane submerged in a propellor slipstream; all design criteria were studied under a set of fixed performance, mission, and stability constraints. Configuration design data were assembled for application by a computerized design methodology program similar to the NASA-Ames General Aviation Synthesis Program.

  13. Optimal fractional order PID design via Tabu Search based algorithm.

    PubMed

    Ateş, Abdullah; Yeroglu, Celaleddin

    2016-01-01

    This paper presents an optimization method based on the Tabu Search Algorithm (TSA) to design a Fractional-Order Proportional-Integral-Derivative (FOPID) controller. All parameter computations of the FOPID employ random initial conditions, using the proposed optimization method. Illustrative examples demonstrate the performance of the proposed FOPID controller design method. PMID:26652128

  14. Fatigue design of a cellular phone folder using regression model-based multi-objective optimization

    NASA Astrophysics Data System (ADS)

    Kim, Young Gyun; Lee, Jongsoo

    2016-08-01

    In a folding cellular phone, the folding device is repeatedly opened and closed by the user, which eventually results in fatigue damage, particularly to the front of the folder. Hence, it is important to improve the safety and endurance of the folder while also reducing its weight. This article presents an optimal design for the folder front that maximizes its fatigue endurance while minimizing its thickness. Design data for analysis and optimization were obtained experimentally using a test jig. Multi-objective optimization was carried out using a nonlinear regression model. Three regression methods were employed: back-propagation neural networks, logistic regression and support vector machines. The AdaBoost ensemble technique was also used to improve the approximation. Two-objective Pareto-optimal solutions were identified using the non-dominated sorting genetic algorithm (NSGA-II). Finally, a numerically optimized solution was validated against experimental product data, in terms of both fatigue endurance and thickness index.

  15. Design and Optimization of Composite Gyroscope Momentum Wheel Rings

    NASA Technical Reports Server (NTRS)

    Bednarcyk, Brett A.; Arnold, Steven M.

    2007-01-01

    Stress analysis and preliminary design/optimization procedures are presented for gyroscope momentum wheel rings composed of metallic, metal matrix composite, and polymer matrix composite materials. The design of these components involves simultaneously minimizing both true part volume and mass, while maximizing angular momentum. The stress analysis results are combined with an anisotropic failure criterion to formulate a new sizing procedure that provides considerable insight into the design of gyroscope momentum wheel ring components. Results compare the performance of two optimized metallic designs, an optimized SiC/Ti composite design, and an optimized graphite/epoxy composite design. The graphite/epoxy design appears to be far superior to the competitors considered unless a much greater premium is placed on volume efficiency compared to mass efficiency.

  16. Design optimization of an opposed piston brake caliper

    NASA Astrophysics Data System (ADS)

    Sergent, Nicolas; Tirovic, Marko; Voveris, Jeronimas

    2014-11-01

    Successful brake caliper designs must be light and stiff, preventing excessive deformation and extended brake pedal travel. These conflicting requirements are difficult to optimize owing to complex caliper geometry, loading and interaction of individual brake components (pads, disc and caliper). The article studies a fixed, four-pot (piston) caliper, and describes in detail the computer-based topology optimization methodology applied to obtain two optimized designs. At first sight, relatively different designs (named 'Z' and 'W') were obtained by minor changes to the designable volume and boundary conditions. However, on closer inspection, the same main bridge design features could be recognized. Both designs offered considerable reduction of caliper mass, by 19% and 28%, respectively. Further finite element analyses conducted on one of the optimized designs (Z caliper) showed which individual bridge features and their combinations are the most important in maintaining caliper stiffness.

  17. Design optimization of a magnetorheological brake in powered knee orthosis

    NASA Astrophysics Data System (ADS)

    Ma, Hao; Liao, Wei-Hsin

    2015-04-01

    Magneto-rheological (MR) fluids have been utilized in devices like orthoses and prostheses to generate controllable braking torque. In this paper, a flat shape rotary MR brake is designed for powered knee orthosis to provide adjustable resistance. Multiple disk structure with interior inner coil is adopted in the MR brake configuration. In order to increase the maximal magnetic flux, a novel internal structure design with smooth transition surface is proposed. Based on this design, a parameterized model of the MR brake is built for geometrical optimization. Multiple factors are considered in the optimization objective: braking torque, weight, and, particularly, average power consumption. The optimization is then performed with Finite Element Analysis (FEA), and the optimal design is obtained among the Pareto-optimal set considering the trade-offs in design objectives.

  18. Optimal shielding design for minimum materials cost or mass

    DOE PAGESBeta

    Woolley, Robert D.

    2015-12-02

    The mathematical underpinnings of cost optimal radiation shielding designs based on an extension of optimal control theory are presented, a heuristic algorithm to iteratively solve the resulting optimal design equations is suggested, and computational results for a simple test case are discussed. A typical radiation shielding design problem can have infinitely many solutions, all satisfying the problem's specified set of radiation attenuation requirements. Each such design has its own total materials cost. For a design to be optimal, no admissible change in its deployment of shielding materials can result in a lower cost. This applies in particular to very smallmore » changes, which can be restated using the calculus of variations as the Euler-Lagrange equations. Furthermore, the associated Hamiltonian function and application of Pontryagin's theorem lead to conditions for a shield to be optimal.« less

  19. Optimal shielding design for minimum materials cost or mass

    SciTech Connect

    Woolley, Robert D.

    2015-12-02

    The mathematical underpinnings of cost optimal radiation shielding designs based on an extension of optimal control theory are presented, a heuristic algorithm to iteratively solve the resulting optimal design equations is suggested, and computational results for a simple test case are discussed. A typical radiation shielding design problem can have infinitely many solutions, all satisfying the problem's specified set of radiation attenuation requirements. Each such design has its own total materials cost. For a design to be optimal, no admissible change in its deployment of shielding materials can result in a lower cost. This applies in particular to very small changes, which can be restated using the calculus of variations as the Euler-Lagrange equations. Furthermore, the associated Hamiltonian function and application of Pontryagin's theorem lead to conditions for a shield to be optimal.

  20. Optimizing spacecraft design - optimization engine development : progress and plans

    NASA Technical Reports Server (NTRS)

    Cornford, Steven L.; Feather, Martin S.; Dunphy, Julia R; Salcedo, Jose; Menzies, Tim

    2003-01-01

    At JPL and NASA, a process has been developed to perform life cycle risk management. This process requires users to identify: goals and objectives to be achieved (and their relative priorities), the various risks to achieving those goals and objectives, and options for risk mitigation (prevention, detection ahead of time, and alleviation). Risks are broadly defined to include the risk of failing to design a system with adequate performance, compatibility and robustness in addition to more traditional implementation and operational risks. The options for mitigating these different kinds of risks can include architectural and design choices, technology plans and technology back-up options, test-bed and simulation options, engineering models and hardware/software development techniques and other more traditional risk reduction techniques.

  1. Formulation of the multimission aircraft design optimization problem

    NASA Astrophysics Data System (ADS)

    Straussfogel, Dennis M.

    1998-12-01

    The conventional single-mission aircraft design optimization problem is reformulated to allow design and optimization for multiple missions. Defining the aircraft mission as a set of continuous scalar variables leads to the concepts of mission vector and mission space. The multimission aircraft design optimization problem becomes one of optimizing a design for several different points in the mission space, simultaneously. In the limit, a design can be optimized simultaneously for all points in the mission space. Mapping various points from the optimization design space into the mission space generates actual and theoretical optimum performance surfaces. The multimission optimum configuration is defined as the single configuration that minimizes the difference between the actual performance and the theoretical optimum performance. The multimission aircraft design optimization method can be applied over several distinct mission by summing the differences between actual and theoretical optimum performance, or over all possible missions by integrating the difference between the actual and theoretical optimum performance surfaces. The concepts associated with the mission vector, mission space, and multimission optimum configuration are presented in mathematical form. The objective function for the multimission optimization problem is expressed both as a summation over discrete mission points and as an integral over the entire mission space. Mathematical expressions for objective functions based on both single and multiple objectives are developed and presented. A weighting function, emphasizing certain parts of the mission space over others, is also discussed. The multimission aircraft design optimization method is applied to an elementary wing-design optimization problem for an executive jet. The optimization problem is solved numerically for single and multiple objectives, with and without a functional constraint. The effect of the different objective functions and

  2. Trajectory optimization software for planetary mission design

    NASA Technical Reports Server (NTRS)

    D'Amario, Louis A.

    1989-01-01

    The development history and characteristics of the interactive trajectory-optimization programs MOSES (D'Amario et al., 1981) and PLATO (D'Amario et al., 1982) are briefly reviewed, with an emphasis on their application to the Galileo mission. The requirements imposed by a mission involving flybys of several planetary satellites or planets are discussed; the formulation of the parameter-optimization problem is outlined; and particular attention is given to the use of multiconic methods to model the gravitational attraction of Jupiter in MOSES. Diagrams and tables of numerical data are included.

  3. Optimal Design of RF Energy Harvesting Device Using Genetic Algorithm

    NASA Astrophysics Data System (ADS)

    Mori, T.; Sato, Y.; Adriano, R.; Igarashi, H.

    2015-11-01

    This paper presents optimal design of an RF energy harvesting device using genetic algorithm (GA). In the present RF harvester, a planar spiral antenna (PSA) is loaded with matching and rectifying circuits. On the first stage of the optimal design, the shape parameters of PSA are optimized using . Then, the equivalent circuit of the optimized PSA is derived for optimization of the circuits. Finally, the parameters of RF energy harvesting circuit are optimized to maximize the output power using GA. It is shown that the present optimization increases the output power by a factor of five. The manufactured energy harvester starts working when the input electric field is greater than 0.5 V/m.

  4. Three-dimensional subsonic diffuser design optimization and analysis

    NASA Astrophysics Data System (ADS)

    Zhang, Wei-Li

    A novel methodology is developed to integrate state-of-the-art CFD analysis, the Non-uniform Rational B-Spline technique (NURBS) and optimization theory to reduce total pressure distortion and sustain or improve total pressure recovery within a curved three dimensional subsonic diffuser. Diffusing S-shaped ducts are representative of curved subsonic diffusers and are characterized by the S-shaped curvature of the duct's centerline and their increasing cross-sectional area. For aircraft inlet applications the measure of duct aerodynamic performance is the ability to decelerate the flow to the desired velocity while maintaining high total pressure recovery and flow near-uniformity. Reduced total pressure recovery lowers propulsion efficiency, whereas nonuniform flow conditions at the engine face lower engine stall and surge limits. Three degrees of freedom are employed as the number of independent design variables. The change of the surface shape is assumed to be Gaussian. The design variables are the location of the flow separation, the width and height of the Gaussian change. The General Aerodynamic Simulation Program (GASP) with the Baldwin-Lomax turbulence model is employed for the flow field prediction and proved to give good agreement with the experimental results for the baseline diffuser geometry. With the automatic change of the design variables, the configuration of the diffuser surface shape is able to be changed while keeping the entrance and exit of the diffuser unchanged in order to meet the specification of the engine and inlet. A trade study was performed which analyzed more than 10 configurations of the modified diffuser. Surface static pressure, surface flow visualization, and exit plane total pressure and transverse velocity data were acquired. The aerodynamic performance of each configuration was assessed by calculating total pressure recovery and spatial distortion elements. The automated design optimization is performed with a gradient

  5. Optimal design of plate-fin heat exchangers by particle swarm optimization

    NASA Astrophysics Data System (ADS)

    Yousefi, M.; Darus, A. N.

    2011-12-01

    This study explores the application of Particle Swarm Optimization (PSO) for optimization of a cross-flow plate fin heat exchanger. Minimization total annual cost is the target of optimization. Seven design parameters, namely, heat exchanger length at hot and cold sides, fin height, fin frequency, fin thickness, fin-strip length and number of hot side layers are selected as optimization variables. A case study from the literature proves the effectiveness of the proposed algorithm in case of achieving more accurate results.

  6. Web Based Learning Support for Experimental Design in Molecular Biology.

    ERIC Educational Resources Information Center

    Wilmsen, Tinri; Bisseling, Ton; Hartog, Rob

    An important learning goal of a molecular biology curriculum is a certain proficiency level in experimental design. Currently students are confronted with experimental approaches in textbooks, in lectures and in the laboratory. However, most students do not reach a satisfactory level of competence in the design of experimental approaches. This…

  7. Execution of Multidisciplinary Design Optimization Approaches on Common Test Problems

    NASA Technical Reports Server (NTRS)

    Balling, R. J.; Wilkinson, C. A.

    1997-01-01

    A class of synthetic problems for testing multidisciplinary design optimization (MDO) approaches is presented. These test problems are easy to reproduce because all functions are given as closed-form mathematical expressions. They are constructed in such a way that the optimal value of all variables and the objective is unity. The test problems involve three disciplines and allow the user to specify the number of design variables, state variables, coupling functions, design constraints, controlling design constraints, and the strength of coupling. Several MDO approaches were executed on two sample synthetic test problems. These approaches included single-level optimization approaches, collaborative optimization approaches, and concurrent subspace optimization approaches. Execution results are presented, and the robustness and efficiency of these approaches an evaluated for these sample problems.

  8. Optimal experiment design for quantum state tomography: Fair, precise, and minimal tomography

    SciTech Connect

    Nunn, J.; Smith, B. J.; Puentes, G.; Walmsley, I. A.; Lundeen, J. S.

    2010-04-15

    Given an experimental setup and a fixed number of measurements, how should one take data to optimally reconstruct the state of a quantum system? The problem of optimal experiment design (OED) for quantum state tomography was first broached by Kosut et al.[R. Kosut, I. Walmsley, and H. Rabitz, e-print arXiv:quant-ph/0411093 (2004)]. Here we provide efficient numerical algorithms for finding the optimal design, and analytic results for the case of 'minimal tomography'. We also introduce the average OED, which is independent of the state to be reconstructed, and the optimal design for tomography (ODT), which minimizes tomographic bias. Monte Carlo simulations confirm the utility of our results for qubits. Finally, we adapt our approach to deal with constrained techniques such as maximum-likelihood estimation. We find that these are less amenable to optimization than cruder reconstruction methods, such as linear inversion.

  9. Optimal Design of Aortic Leaflet Prosthesis

    NASA Technical Reports Server (NTRS)

    Ghista, Dhanjoo N.; Reul, Helmut; Ray, Gautam; Chandran, K. B.

    1978-01-01

    The design criteria for an optimum prosthetic-aortic leaflet valve are a smooth washout in the valve cusps, minimal leaflet stress, minimal transmembrane pressure for the valve to open, an adequate lifetime (for a given blood-compatible leaflet material's fatigue data). A rigorous design analysis is presented to obtain the prosthetic tri-leaflet aortic valve leaflet's optimum design parameters. Four alternative optimum leaflet geometries are obtained to satisfy the criteria of a smooth washout and minimal leaflet stress. The leaflet thicknesses of these four optimum designs are determined by satisfying the two remaining design criteria for minimal transmembrane opening pressure and adequate fatigue lifetime, which are formulated in terms of the elastic and fatigue properties of the selected leaflet material - Avcothane-51 (of the Avco-Everett Co. of Massachusetts). Prosthetic valves are fabricated on the basis of the optimum analysis and the resulting detailed engineering drawings of the designs are also presented in the paper.

  10. Optimal cost basis for seismic design

    SciTech Connect

    Hadjian, A.H.

    1993-09-01

    The paper summarizes a methodology for establishing seismic design levels based on a cost-benefit assessment. The methodology requires the development of costs and benefits for varying design levels of vibratory ground motion and surface fault displacements. The optimum design level for a given structure is that which gives the minimum total direct and earthquake consequence costs. The example used is the surface facilities of the proposed high-level waste repository at Yucca Mountain, Nevada.

  11. Factorial Design to Optimize Biosurfactant Production by Yarrowia lipolytica

    PubMed Central

    Fontes, Gizele Cardoso; Fonseca Amaral, Priscilla Filomena; Nele, Marcio; Zarur Coelho, Maria Alice

    2010-01-01

    In order to improve biosurfactant production by Yarrowia lipolytica IMUFRJ 50682, a factorial design was carried out. A 24 full factorial design was used to investigate the effects of nitrogen sources (urea, ammonium sulfate, yeast extract, and peptone) on maximum variation of surface tension (ΔST) and emulsification index (EI). The best results (67.7% of EI and 20.9 mN m−1 of ΔST) were obtained in a medium composed of 10 g 1−1 of ammonium sulfate and 0.5 g 1−1 of yeast extract. Then, the effects of carbon sources (glycerol, hexadecane, olive oil, and glucose) were evaluated. The most favorable medium for biosurfactant production was composed of both glucose (4% w/v) and glycerol (2% w/v), which provided an EI of 81.3% and a ΔST of 19.5 mN m−1. The experimental design optimization enhanced ΔEI by 110.7% and ΔST by 108.1% in relation to the standard process. PMID:20368788

  12. Design of nanoporous materials with optimal sorption capacity

    NASA Astrophysics Data System (ADS)

    Zhang, Xuan; Urita, Koki; Moriguchi, Isamu; Tartakovsky, Daniel M.

    2015-06-01

    Modern technological advances have enabled one to manufacture nanoporous materials with a prescribed pore structure. This raises a possibility of using controllable pore-scale parameters (e.g., pore size and connectivity) to design materials with desired macroscopic properties (e.g., diffusion coefficient and adsorption capacity). By relating these two scales, the homogenization theory (or other upscaling techniques) provides a means of guiding the experimental design. To demonstrate this approach, we consider a class of nanoporous materials whose pore space consists of nanotunnels interconnected by nanotube bridges. Such hierarchical nanoporous carbons with mesopores and micropores have shown high specific electric double layer capacitances and high rate capability in an organic electrolyte. We express the anisotropic diffusion coefficient and adsorption coefficient of such materials in terms of the tunnels' properties (pore radius and inter-pore throat width) and their connectivity (spacing between the adjacent tunnels and nanotube-bridge density). Our analysis is applicable for solutes that undergo a non-equilibrium Langmuir adsorption reaction on the surfaces of fluid-filled pores, but other homogeneous and heterogeneous reactions can be handled in a similar fashion. The presented results can be used to guide the design of nanoporous materials with optimal permeability and sorption capacity.

  13. Origami Optimization: Role of Symmetry in Accelerating Design

    NASA Astrophysics Data System (ADS)

    Buskohl, Philip; Fuchi, Kazuko; Bazzan, Giorgio; Durstock, Michael; Reich, Gregory; Joo, James; Vaia, Richard

    Origami structures morph between 2D and 3D conformations along predetermined fold lines that efficiently program the form, function and mobility of the structure. Design optimization tools have recently been developed to predict optimal fold patterns with mechanics-based metrics, such as the maximal energy storage, auxetic response and actuation. Origami actuator design problems possess inherent symmetries associated with the grid, mechanical boundary conditions and the objective function, which are often exploited to reduce the design space and computational cost of optimization. However, enforcing symmetry eliminates the prediction of potentially better performing asymmetric designs, which are more likely to exist given the discrete nature of fold line optimization. To better understand this effect, actuator design problems with different combinations of rotation and reflection symmetries were optimized while varying the number of folds allowed in the final design. In each case, the optimal origami patterns transitioned between symmetric and asymmetric solutions depended on the number of folds available for the design, with fewer symmetries present with more fold lines allowed. This study investigates the interplay of symmetry and discrete vs continuous optimization in origami actuators and provides insight into how the symmetries of the reference grid regulate the performance landscape. This work was supported by the Air Force Office of Scientific Research.

  14. Lessons Learned During Solutions of Multidisciplinary Design Optimization Problems

    NASA Technical Reports Server (NTRS)

    Patnaik, Suna N.; Coroneos, Rula M.; Hopkins, Dale A.; Lavelle, Thomas M.

    2000-01-01

    Optimization research at NASA Glenn Research Center has addressed the design of structures, aircraft and airbreathing propulsion engines. During solution of the multidisciplinary problems several issues were encountered. This paper lists four issues and discusses the strategies adapted for their resolution: (1) The optimization process can lead to an inefficient local solution. This deficiency was encountered during design of an engine component. The limitation was overcome through an augmentation of animation into optimization. (2) Optimum solutions obtained were infeasible for aircraft and air-breathing propulsion engine problems. Alleviation of this deficiency required a cascading of multiple algorithms. (3) Profile optimization of a beam produced an irregular shape. Engineering intuition restored the regular shape for the beam. (4) The solution obtained for a cylindrical shell by a subproblem strategy converged to a design that can be difficult to manufacture. Resolution of this issue remains a challenge. The issues and resolutions are illustrated through six problems: (1) design of an engine component, (2) synthesis of a subsonic aircraft, (3) operation optimization of a supersonic engine, (4) design of a wave-rotor-topping device, (5) profile optimization of a cantilever beam, and (6) design of a cvlindrical shell. The combined effort of designers and researchers can bring the optimization method from academia to industry.

  15. Experimental study on active vibration control using genetic algorithm-based system identification and optimized positive position feedback

    NASA Astrophysics Data System (ADS)

    Orszulik, Ryan R.; Shan, Jinjun

    2012-12-01

    A genetic algorithm is implemented to identify the transfer function of an experimental system consisting of a flexible manipulator with a collocated piezoelectric sensor/actuator pair. A multi-mode positive position feedback controller is then designed based upon the identified transfer function. To this end, the same iteratively implemented genetic algorithm is used to optimize all controller parameters by minimization of the closed loop H∞-norm. The designed controller is then applied for vibration suppression on the experimental system.

  16. Optimal design of spatial distribution networks

    NASA Astrophysics Data System (ADS)

    Gastner, Michael T.; Newman, M. E. J.

    2006-07-01

    We consider the problem of constructing facilities such as hospitals, airports, or malls in a country with a nonuniform population density, such that the average distance from a person’s home to the nearest facility is minimized. We review some previous approximate treatments of this problem that indicate that the optimal distribution of facilities should have a density that increases with population density, but does so slower than linearly, as the two-thirds power. We confirm this result numerically for the particular case of the United States with recent population data using two independent methods, one a straightforward regression analysis, the other based on density-dependent map projections. We also consider strategies for linking the facilities to form a spatial network, such as a network of flights between airports, so that the combined cost of maintenance of and travel on the network is minimized. We show specific examples of such optimal networks for the case of the United States.

  17. Total energy control system autopilot design with constrained parameter optimization

    NASA Technical Reports Server (NTRS)

    Ly, Uy-Loi; Voth, Christopher

    1990-01-01

    A description is given of the application of a multivariable control design method (SANDY) based on constrained parameter optimization to the design of a multiloop aircraft flight control system. Specifically, the design method is applied to the direct synthesis of a multiloop AFCS inner-loop feedback control system based on total energy control system (TECS) principles. The design procedure offers a structured approach for the determination of a set of stabilizing controller design gains that meet design specifications in closed-loop stability, command tracking performance, disturbance rejection, and limits on control activities. The approach can be extended to a broader class of multiloop flight control systems. Direct tradeoffs between many real design goals are rendered systematic by proper formulation of the design objectives and constraints. Satisfactory designs are usually obtained in few iterations. Performance characteristics of the optimized TECS design have been improved, particularly in the areas of closed-loop damping and control activity in the presence of turbulence.

  18. Preserving omnidirectionality in optimized asymmetric transformation optics designs

    NASA Astrophysics Data System (ADS)

    Popa, Bogdan-Ioan; Cummer, Steven A.

    2016-04-01

    Optimization techniques are efficient methods to simplify the design of transformation optics devices. Asymmetric devices obtained through these methods typically lose omnidirectionality and perform well only when illuminated from a finite set of directions. We present here an optimization approach that results in simplified, omnidirectional designs of fairly large bandwidths. The method leverages a class of coordinate transformations that result in transformation media whose material parameters follow identical isocontours. We show that discretizing these media in a finite number of layers that follow the common isocontours and optimizing the material parameters inside each layer is an effective way to significantly simplify the design complexity while preserving most of the original performance including omnidirectionality.

  19. Field Quality Optimization in a Common Coil Magnet Design

    SciTech Connect

    Gupta, Ramesh; Ramberger, Suitbert

    1999-09-01

    This paper presents the results of initial field quality optimization of body and end harmonics in a 'common coil magnet design'. It is shown that a good field quality, as required in accelerator magnets, can be obtained by distributing conductor blocks in such a way that they simulate an elliptical coil geometry. This strategy assures that the amount of conductor used in this block design is similar to that is used in a conventional cosine theta design. An optimized yoke that keeps all harmonics small over the entire range of operation using a single power supply is also presented. The field harmonics are primarily optimized with the computer program ROXIE.

  20. Demonstrating the benefits of template-based design-technology co-optimization

    NASA Astrophysics Data System (ADS)

    Liebmann, Lars; Hibbeler, Jason; Hieter, Nathaniel; Pileggi, Larry; Jhaveri, Tejas; Moe, Matthew; Rovner, Vyacheslav

    2010-03-01

    The concept of template-based design-technology co-optimization as a means of curbing escalating design complexity and increasing technology qualification risk is described. Data is presented highlighting the design efficacy of this proposal in terms of power, performance, and area benefits, quantifying the specific contributions of complex logic gates in this design optimization. Experimental results from 32nm technology node bulk CMOS wafers are presented to quantify the variability and design-margin reductions as well as yield and manufacturability improvements achievable with the proposed template-based design-technology co-optimization technique. The paper closes with data showing the predictable composability of individual templates, demonstrating a fundamental requirement of this proposal.

  1. Optimal design of geodesically stiffened composite cylindrical shells

    NASA Technical Reports Server (NTRS)

    Gendron, G.; Guerdal, Z.

    1992-01-01

    An optimization system based on the finite element code Computations Structural Mechanics (CSM) Testbed and the optimization program, Automated Design Synthesis (ADS), is described. The optimization system can be used to obtain minimum-weight designs of composite stiffened structures. Ply thickness, ply orientations, and stiffener heights can be used as design variables. Buckling, displacement, and material failure constraints can be imposed on the design. The system is used to conduct a design study of geodesically stiffened shells. For comparison purposes, optimal designs of unstiffened shells and shells stiffened by rings and stingers are also obtained. Trends in the design of geodesically stiffened shells are identified. An approach to include local stress concentrations during the design optimization process is then presented. The method is based on a global/local analysis technique. It employs spline interpolation functions to determine displacements and rotations from a global model which are used as 'boundary conditions' for the local model. The organization of the strategy in the context of an optimization process is described. The method is validated with an example.

  2. New approaches to the design optimization of hydrofoils

    NASA Astrophysics Data System (ADS)

    Beyhaghi, Pooriya; Meneghello, Gianluca; Bewley, Thomas

    2015-11-01

    Two simulation-based approaches are developed to optimize the design of hydrofoils for foiling catamarans, with the objective of maximizing efficiency (lift/drag). In the first, a simple hydrofoil model based on the vortex-lattice method is coupled with a hybrid global and local optimization algorithm that combines our Delaunay-based optimization algorithm with a Generalized Pattern Search. This optimization procedure is compared with the classical Newton-based optimization method. The accuracy of the vortex-lattice simulation of the optimized design is compared with a more accurate and computationally expensive LES-based simulation. In the second approach, the (expensive) LES model of the flow is used directly during the optimization. A modified Delaunay-based optimization algorithm is used to maximize the efficiency of the optimization, which measures a finite-time averaged approximation of the infinite-time averaged value of an ergodic and stationary process. Since the optimization algorithm takes into account the uncertainty of the finite-time averaged approximation of the infinite-time averaged statistic of interest, the total computational time of the optimization algorithm is significantly reduced. Results from the two different approaches are compared.

  3. A study of commuter airplane design optimization

    NASA Technical Reports Server (NTRS)

    Keppel, B. V.; Eysink, H.; Hammer, J.; Hawley, K.; Meredith, P.; Roskam, J.

    1978-01-01

    The usability of the general aviation synthesis program (GASP) was enhanced by the development of separate computer subroutines which can be added as a package to this assembly of computerized design methods or used as a separate subroutine program to compute the dynamic longitudinal, lateral-directional stability characteristics for a given airplane. Currently available analysis methods were evaluated to ascertain those most appropriate for the design functions which the GASP computerized design program performs. Methods for providing proper constraint and/or analysis functions for GASP were developed as well as the appropriate subroutines.

  4. Optimization applications in aircraft engine design and test

    NASA Technical Reports Server (NTRS)

    Pratt, T. K.

    1984-01-01

    Starting with the NASA-sponsored STAEBL program, optimization methods based primarily upon the versatile program COPES/CONMIN were introduced over the past few years to a broad spectrum of engineering problems in structural optimization, engine design, engine test, and more recently, manufacturing processes. By automating design and testing processes, many repetitive and costly trade-off studies have been replaced by optimization procedures. Rather than taking engineers and designers out of the loop, optimization has, in fact, put them more in control by providing sophisticated search techniques. The ultimate decision whether to accept or reject an optimal feasible design still rests with the analyst. Feedback obtained from this decision process has been invaluable since it can be incorporated into the optimization procedure to make it more intelligent. On several occasions, optimization procedures have produced novel designs, such as the nonsymmetric placement of rotor case stiffener rings, not anticipated by engineering designers. In another case, a particularly difficult resonance contraint could not be satisfied using hand iterations for a compressor blade, when the STAEBL program was applied to the problem, a feasible solution was obtained in just two iterations.

  5. Optimizing Balanced Incomplete Block Designs for Educational Assessments

    ERIC Educational Resources Information Center

    van der Linden, Wim J.; Veldkamp, Bernard P.; Carlson, James E.

    2004-01-01

    A popular design in large-scale educational assessments as well as any other type of survey is the balanced incomplete block design. The design is based on an item pool split into a set of blocks of items that are assigned to sets of "assessment booklets." This article shows how the problem of calculating an optimal balanced incomplete block…

  6. Optimization and analysis of a CFJ-airfoil using adaptive meta-model based design optimization

    NASA Astrophysics Data System (ADS)

    Whitlock, Michael D.

    Although strong potential for Co-Flow Jet (CFJ) flow separation control system has been demonstrated in existing literature, there has been little effort applied towards the optimization of the design for a given application. The high dimensional design space makes any optimization computationally intensive. This work presents the optimization of a CFJ airfoil as applied to a low Reynolds Number regimen using meta-model based design optimization (MBDO). The approach consists of computational fluid dynamics (CFD) analysis coupled with a surrogate model derived using Kriging. A genetic algorithm (GA) is then used to perform optimization on the efficient surrogate model. MBDO was shown to be an effective and efficient approach to solving the CFJ design problem. The final solution set was found to decrease drag by 100% while increasing lift by 42%. When validated, the final solution was found to be within one standard deviation of the CFD model it was representing.

  7. EXPERIMENTAL DESIGN: STATISTICAL CONSIDERATIONS AND ANALYSIS

    Technology Transfer Automated Retrieval System (TEKTRAN)

    In this book chapter, information on how field experiments in invertebrate pathology are designed and the data collected, analyzed, and interpreted is presented. The practical and statistical issues that need to be considered and the rationale and assumptions behind different designs or procedures ...

  8. Experimental qualification of a code for optimizing gamma irradiation facilities

    NASA Astrophysics Data System (ADS)

    Mosse, D. C.; Leizier, J. J. M.; Keraron, Y.; Lallemant, T. F.; Perdriau, P. D. M.

    Dose computation codes are a prerequisite for the design of gamma irradiation facilities. Code quality is a basic factor in the achievement of sound economic and technical performance by the facility. This paper covers the validation of a code by reference dosimetry experiments. Developed by the "Société Générale pour les Techniques Nouvelles" (SGN), a supplier of irradiation facilities and member of the CEA Group, the code is currently used by that company. (ERHART, KERARON, 1986) Experimental data were obtained under conditions representative of those prevailing in the gamma irradiation of foodstuffs. Irradiation was performed in POSEIDON, a Cobalt 60 cell of ORIS-I. Several Cobalt 60 rods of known activity are arranged in a planar array typical of industrial irradiation facilities. Pallet density is uniform, ranging from 0 (air) to 0.6. Reference dosimetry measurements were performed by the "Laboratoire de Métrologie des Rayonnements Ionisants" (LMRI) of the "Bureau National de Métrologie" (BNM). The procedure is based on the positioning of more than 300 ESR/alanine dosemeters throughout the various target volumes used. The reference quantity was the absorbed dose in water. The code was validated by a comparison of experimental and computed data. It has proved to be an effective tool for the design of facilities meeting the specific requirements applicable to foodstuff irradiation, which are frequently found difficult to meet.

  9. Managing Model Data Introduced Uncertainties in Simulator Predictions for Generation IV Systems via Optimum Experimental Design

    SciTech Connect

    Turinsky, Paul J; Abdel-Khalik, Hany S; Stover, Tracy E

    2011-03-31

    An optimization technique has been developed to select optimized experimental design specifications to produce data specifically designed to be assimilated to optimize a given reactor concept. Data from the optimized experiment is assimilated to generate posteriori uncertainties on the reactor concept’s core attributes from which the design responses are computed. The reactor concept is then optimized with the new data to realize cost savings by reducing margin. The optimization problem iterates until an optimal experiment is found to maximize the savings. A new generation of innovative nuclear reactor designs, in particular fast neutron spectrum recycle reactors, are being considered for the application of closing the nuclear fuel cycle in the future. Safe and economical design of these reactors will require uncertainty reduction in basic nuclear data which are input to the reactor design. These data uncertainty propagate to design responses which in turn require the reactor designer to incorporate additional safety margin into the design, which often increases the cost of the reactor. Therefore basic nuclear data needs to be improved and this is accomplished through experimentation. Considering the high cost of nuclear experiments, it is desired to have an optimized experiment which will provide the data needed for uncertainty reduction such that a reactor design concept can meet its target accuracies or to allow savings to be realized by reducing the margin required due to uncertainty propagated from basic nuclear data. However, this optimization is coupled to the reactor design itself because with improved data the reactor concept can be re-optimized itself. It is thus desired to find the experiment that gives the best optimized reactor design. Methods are first established to model both the reactor concept and the experiment and to efficiently propagate the basic nuclear data uncertainty through these models to outputs. The representativity of the experiment

  10. Optimized planar micro-optic concentrator design

    NASA Astrophysics Data System (ADS)

    Pan, Jui-Wen; Su, Yu-Chung; Lee, Sheng-Yi

    2016-06-01

    The structural parameters of a planar micro-optic concentrator are optimized. First, the direct-loss is minimized by altering the relationship between the f-number of the lenslet, the angle of the micro-structure and the ray paths in the planar micro-optic concentrator. Second, the size of the micro-structure is made equal to the mini-blur size of the lenslet in order to reduce the non-direct loss. Last, the f-number and the entrance pupil diameter of the lenslet are determined by the relationships among the f-number, the entrance pupil diameter, the optical efficiency, the acceptance angle and the thickness of the planar micro-optic concentrator from the optical simulation results. For an optimized planar micro-optic concentrator with a 300× concentration, the f-number of the lenslet, the EPD of the lenslet, the angle of the micro-structure and the thickness of the planar micro-optic concentrator are 2.6, 1.49 mm, 120 degrees and 5.97 mm, respectively. For micro-structures 28.95 μm, 51.24 μm and 88.29 μm in size, the half acceptance angles of the planar micro-optic concentrator are 0.115 degrees, 0.275 degrees and 0.55 degrees, respectively, and the optical efficiencies are 81.23%, 71.92% and 50.02%, respectively.

  11. Integrated structure/control law design by multilevel optimization

    NASA Technical Reports Server (NTRS)

    Gilbert, Michael G.; Schmidt, David K.

    1989-01-01

    A new approach to integrated structure/control law design based on multilevel optimization is presented. This new approach is applicable to aircraft and spacecraft and allows for the independent design of the structure and control law. Integration of the designs is achieved through use of an upper level coordination problem formulation within the multilevel optimization framework. The method requires the use of structure and control law design sensitivity information. A general multilevel structure/control law design problem formulation is given, and the use of Linear Quadratic Gaussian (LQG) control law design and design sensitivity methods within the formulation is illustrated. Results of three simple integrated structure/control law design examples are presented. These results show the capability of structure and control law design tradeoffs to improve controlled system performance within the multilevel approach.

  12. Conceptual design report, CEBAF basic experimental equipment

    SciTech Connect

    1990-04-13

    The Continuous Electron Beam Accelerator Facility (CEBAF) will be dedicated to basic research in Nuclear Physics using electrons and photons as projectiles. The accelerator configuration allows three nearly continuous beams to be delivered simultaneously in three experimental halls, which will be equipped with complementary sets of instruments: Hall A--two high resolution magnetic spectrometers; Hall B--a large acceptance magnetic spectrometer; Hall C--a high-momentum, moderate resolution, magnetic spectrometer and a variety of more dedicated instruments. This report contains a short description of the initial complement of experimental equipment to be installed in each of the three halls.

  13. Post-optimality analysis in aerospace vehicle design

    NASA Technical Reports Server (NTRS)

    Braun, Robert D.; Kroo, Ilan M.; Gage, Peter J.

    1993-01-01

    This analysis pertains to the applicability of optimal sensitivity information to aerospace vehicle design. The present analysis demonstrates that post-optimality information generated through first-order computations can be used to accurately predict file effect of constraint and parameter perturbations on the optimal solution. This assessment is based on the solution of an aircraft design problem in which the post-optimality estimates are shown to be within a few percent of the true solution over the practical range of constraint and parameter variations. Through solution of a reusable, single-stage-to-orbit, launch vehicle design problem, this optimal sensitivity information is also shown to improve the efficiency of the design process. For a hierarchically decomposed problem, this computational efficiency is realizable by estimating the main-problem objective gradient through optimal sensitivity calculations. By reducing the need for finite differentiation of a re-optimized subproblem, a significant decrease in the number of objective function evaluations required to reach the optimal solution is obtained.

  14. Precision of Sensitivity in the Design Optimization of Indeterminate Structures

    NASA Technical Reports Server (NTRS)

    Patnaik, Surya N.; Pai, Shantaram S.; Hopkins, Dale A.

    2006-01-01

    Design sensitivity is central to most optimization methods. The analytical sensitivity expression for an indeterminate structural design optimization problem can be factored into a simple determinate term and a complicated indeterminate component. Sensitivity can be approximated by retaining only the determinate term and setting the indeterminate factor to zero. The optimum solution is reached with the approximate sensitivity. The central processing unit (CPU) time to solution is substantially reduced. The benefit that accrues from using the approximate sensitivity is quantified by solving a set of problems in a controlled environment. Each problem is solved twice: first using the closed-form sensitivity expression, then using the approximation. The problem solutions use the CometBoards testbed as the optimization tool with the integrated force method as the analyzer. The modification that may be required, to use the stiffener method as the analysis tool in optimization, is discussed. The design optimization problem of an indeterminate structure contains many dependent constraints because of the implicit relationship between stresses, as well as the relationship between the stresses and displacements. The design optimization process can become problematic because the implicit relationship reduces the rank of the sensitivity matrix. The proposed approximation restores the full rank and enhances the robustness of the design optimization method.

  15. Optimality criteria design and stress constraint processing

    NASA Technical Reports Server (NTRS)

    Levy, R.

    1982-01-01

    Methods for pre-screening stress constraints into either primary or side-constraint categories are reviewed; a projection method, which is developed from prior cycle stress resultant history, is introduced as an additional screening parameter. Stress resultant projections are also employed to modify the traditional stress-ratio, side-constraint boundary. A special application of structural modification reanalysis is applied to the critical stress constraints to provide feasible designs that are preferable to those obtained by conventional scaling. Sample problem executions show relatively short run times and fewer design cycle iterations to achieve low structural weights; those attained are comparable to the minimum values developed elsewhere.

  16. Experimental Stream Facility: Design and Research

    EPA Science Inventory

    The Experimental Stream Facility (ESF) is a valuable research tool for the U.S. Environmental Protection Agency’s (EPA) Office of Research and Development’s (ORD) laboratories in Cincinnati, Ohio. This brochure describes the ESF, which is one of only a handful of research facilit...

  17. Minimum weight design of structures via optimality criteria

    NASA Technical Reports Server (NTRS)

    Kiusalaas, J.

    1972-01-01

    The state of the art of automated structural design through the use of optimality criteria, with emphasis on aerospace applications is reviewed. Constraints on stresses, displacements, and buckling strengths under static loading, as well as lower bound limits on natural frequencies and flutter speeds are presented. It is presumed that the reader is experienced in finite element methods of analysis, but is not familiar with optimal design techniques.

  18. Optimal design of plasmonic waveguide using multiobjective genetic algorithm

    NASA Astrophysics Data System (ADS)

    Jung, Jaehoon

    2016-01-01

    An approach for multiobjective optimal design of a plasmonic waveguide is presented. We use a multiobjective extension of a genetic algorithm to find the Pareto-optimal geometries. The design variables are the geometrical parameters of the waveguide. The objective functions are chosen as the figure of merit defined as the ratio between the propagation distance and effective mode size and the normalized coupling length between adjacent waveguides at the telecom wavelength of 1550 nm.

  19. Optimal design for epidemiological studies subject to designed missingness.

    PubMed

    Morara, Michele; Ryan, Louise; Houseman, Andres; Strauss, Warren

    2007-12-01

    In large epidemiological studies, budgetary or logistical constraints will typically preclude study investigators from measuring all exposures, covariates and outcomes of interest on all study subjects. We develop a flexible theoretical framework that incorporates a number of familiar designs such as case control and cohort studies, as well as multistage sampling designs. Our framework also allows for designed missingness and includes the option for outcome dependent designs. Our formulation is based on maximum likelihood and generalizes well known results for inference with missing data to the multistage setting. A variety of techniques are applied to streamline the computation of the Hessian matrix for these designs, facilitating the development of an efficient software tool to implement a wide variety of designs. PMID:18080755

  20. Application of optimization techniques to vehicle design: A review

    NASA Technical Reports Server (NTRS)

    Prasad, B.; Magee, C. L.

    1984-01-01

    The work that has been done in the last decade or so in the application of optimization techniques to vehicle design is discussed. Much of the work reviewed deals with the design of body or suspension (chassis) components for reduced weight. Also reviewed are studies dealing with system optimization problems for improved functional performance, such as ride or handling. In reviewing the work on the use of optimization techniques, one notes the transition from the rare mention of the methods in the 70's to an increased effort in the early 80's. Efficient and convenient optimization and analysis tools still need to be developed so that they can be regularly applied in the early design stage of the vehicle development cycle to be most effective. Based on the reported applications, an attempt is made to assess the potential for automotive application of optimization techniques. The major issue involved remains the creation of quantifiable means of analysis to be used in vehicle design. The conventional process of vehicle design still contains much experience-based input because it has not yet proven possible to quantify all important constraints. This restraint on the part of the analysis will continue to be a major limiting factor in application of optimization to vehicle design.

  1. Integrated design optimization research and development in an industrial environment

    NASA Astrophysics Data System (ADS)

    Kumar, V.; German, Marjorie D.; Lee, S.-J.

    1989-04-01

    An overview is given of a design optimization project that is in progress at the GE Research and Development Center for the past few years. The objective of this project is to develop a methodology and a software system for design automation and optimization of structural/mechanical components and systems. The effort focuses on research and development issues and also on optimization applications that can be related to real-life industrial design problems. The overall technical approach is based on integration of numerical optimization techniques, finite element methods, CAE and software engineering, and artificial intelligence/expert systems (AI/ES) concepts. The role of each of these engineering technologies in the development of a unified design methodology is illustrated. A software system DESIGN-OPT has been developed for both size and shape optimization of structural components subjected to static as well as dynamic loadings. By integrating this software with an automatic mesh generator, a geometric modeler and an attribute specification computer code, a software module SHAPE-OPT has been developed for shape optimization. Details of these software packages together with their applications to some 2- and 3-dimensional design problems are described.

  2. Integrated design optimization research and development in an industrial environment

    NASA Technical Reports Server (NTRS)

    Kumar, V.; German, Marjorie D.; Lee, S.-J.

    1989-01-01

    An overview is given of a design optimization project that is in progress at the GE Research and Development Center for the past few years. The objective of this project is to develop a methodology and a software system for design automation and optimization of structural/mechanical components and systems. The effort focuses on research and development issues and also on optimization applications that can be related to real-life industrial design problems. The overall technical approach is based on integration of numerical optimization techniques, finite element methods, CAE and software engineering, and artificial intelligence/expert systems (AI/ES) concepts. The role of each of these engineering technologies in the development of a unified design methodology is illustrated. A software system DESIGN-OPT has been developed for both size and shape optimization of structural components subjected to static as well as dynamic loadings. By integrating this software with an automatic mesh generator, a geometric modeler and an attribute specification computer code, a software module SHAPE-OPT has been developed for shape optimization. Details of these software packages together with their applications to some 2- and 3-dimensional design problems are described.

  3. A new method of optimal design for a two-dimensional diffuser by using dynamic programming

    NASA Technical Reports Server (NTRS)

    Gu, Chuangang; Zhang, Moujin; Chen, XI; Miao, Yongmiao

    1991-01-01

    A new method for predicting the optimal velocity distribution on the wall of a two dimensional diffuser is presented. The method uses dynamic programming to solve the optimal control problem with inequality constraints of state variables. The physical model of optimization is designed to prevent the separation of the boundary layer while approaching the maximum pressure ratio in a diffuser of a specified length. The computational results are in fair agreement with the experimental ones. Optimal velocity distribution on a diffuser wall is said to occur when the flow decelerates quickly at first and then smoothly, while the flow is near separation, but always protected from it. The optimal velocity distribution can be used to design the contour of the diffuser.

  4. Design of ophthalmic lens by using optimized aspheric surface coefficients

    NASA Astrophysics Data System (ADS)

    Chang, Ming-Wen; Sun, Wen-Shing; Tien, Chuen-Lin

    1998-09-01

    Coddington's equations can be used to eliminate the oblique astigmatic error in the design of ophthalmic lens of spherical or other conicoidal surfaces. But it is difficult to get satisfactory result in the designing of the nonconic aspheric ophthalmic lens. In this paper we present an efficient approach based on optimization of aspheric coefficients, which enables the design program to obtain the minimum aberrations. Many higher order coefficients of aspheric surfaces can easily result in inflection point, which increases the difficulty in manufacturing. We solved the problem by taking it as one of the optimization constraints. The design of nonconic aspheric ophthalmic lens could also make the spectacle lenses well thinner in thickness and well flatter in shape than the design of spherical ophthalmic lens and other conicoidal ophthalmic lens. Damped least square methods are used in our design. Aspherical myopia ophthalmic lenses, aspherical hypermetropic lenses and cataract lenses were designed. Comparisons of design examples' results are given.

  5. Rapid Modeling, Assembly and Simulation in Design Optimization

    NASA Technical Reports Server (NTRS)

    Housner, Jerry

    1997-01-01

    A new capability for design is reviewed. This capability provides for rapid assembly of detail finite element models early in the design process where costs are most effectively impacted. This creates an engineering environment which enables comprehensive analysis and design optimization early in the design process. Graphical interactive computing makes it possible for the engineer to interact with the design while performing comprehensive design studies. This rapid assembly capability is enabled by the use of Interface Technology, to couple independently created models which can be archived and made accessible to the designer. Results are presented to demonstrate the capability.

  6. Intelligent Space Tube Optimization for speeding ground water remedial design.

    PubMed

    Kalwij, Ineke M; Peralta, Richard C

    2008-01-01

    An innovative Intelligent Space Tube Optimization (ISTO) two-stage approach facilitates solving complex nonlinear flow and contaminant transport management problems. It reduces computational effort of designing optimal ground water remediation systems and strategies for an assumed set of wells. ISTO's stage 1 defines an adaptive mobile space tube that lengthens toward the optimal solution. The space tube has overlapping multidimensional subspaces. Stage 1 generates several strategies within the space tube, trains neural surrogate simulators (NSS) using the limited space tube data, and optimizes using an advanced genetic algorithm (AGA) with NSS. Stage 1 speeds evaluating assumed well locations and combinations. For a large complex plume of solvents and explosives, ISTO stage 1 reaches within 10% of the optimal solution 25% faster than an efficient AGA coupled with comprehensive tabu search (AGCT) does by itself. ISTO input parameters include space tube radius and number of strategies used to train NSS per cycle. Larger radii can speed convergence to optimality for optimizations that achieve it but might increase the number of optimizations reaching it. ISTO stage 2 automatically refines the NSS-AGA stage 1 optimal strategy using heuristic optimization (we used AGCT), without using NSS surrogates. Stage 2 explores the entire solution space. ISTO is applicable for many heuristic optimization settings in which the numerical simulator is computationally intensive, and one would like to reduce that burden. PMID:18754799

  7. Finite element based electric motor design optimization

    NASA Technical Reports Server (NTRS)

    Campbell, C. Warren

    1993-01-01

    The purpose of this effort was to develop a finite element code for the analysis and design of permanent magnet electric motors. These motors would drive electromechanical actuators in advanced rocket engines. The actuators would control fuel valves and thrust vector control systems. Refurbishing the hydraulic systems of the Space Shuttle after each flight is costly and time consuming. Electromechanical actuators could replace hydraulics, improve system reliability, and reduce down time.

  8. Variable-complexity optimization applied to airfoil design

    NASA Astrophysics Data System (ADS)

    Thokala, Praveen; Martins, Joaquim R. R. A.

    2007-04-01

    Variable-complexity methods are applied to aerodynamic shape design problems with the objective of reducing the total computational cost of the optimization process. Two main strategies are employed: the use of different levels of fidelity in the analysis models (variable fidelity) and the use of different sets of design variables (variable parameterization). Variable-fidelity methods with three different types of corrections are implemented and applied to a set of two-dimensional airfoil optimization problems that use computational fluid dynamics for the analysis. Variable parameterization is also used to solve the same problems. Both strategies are shown to reduce the computational cost of the optimization.

  9. Optimal aeroelastic design of an oblique wing structure

    NASA Technical Reports Server (NTRS)

    Gwin, L. B.

    1974-01-01

    A procedure is presented for determining the optimal cover panel thickness of a wing structure to meet specified strength and static aeroelastic divergence requirements for minimum weight. Efficient reanalysis techniques using discrete structural and aerodynamic methods are used in conjunction with redesign algorithms driven by optimality criteria. The optimality conditions for the divergence constraint are established, and expressions are obtained for derivatives of the dynamic pressure at divergence with respect to design variables. The procedure is applied to an oblique wing aircraft where strength and stiffness are critical design considerations for sizing the cover thickness of the wing structure.

  10. Optimal brushless DC motor design using genetic algorithms

    NASA Astrophysics Data System (ADS)

    Rahideh, A.; Korakianitis, T.; Ruiz, P.; Keeble, T.; Rothman, M. T.

    2010-11-01

    This paper presents a method for the optimal design of a slotless permanent magnet brushless DC (BLDC) motor with surface mounted magnets using a genetic algorithm. Characteristics of the motor are expressed as functions of motor geometries. The objective function is a combination of losses, volume and cost to be minimized simultaneously. Electrical and mechanical requirements (i.e. voltage, torque and speed) and other limitations (e.g. upper and lower limits of the motor geometries) are cast into constraints of the optimization problem. One sample case is used to illustrate the design and optimization technique.

  11. Application of clustering global optimization to thin film design problems.

    PubMed

    Lemarchand, Fabien

    2014-03-10

    Refinement techniques usually calculate an optimized local solution, which is strongly dependent on the initial formula used for the thin film design. In the present study, a clustering global optimization method is used which can iteratively change this initial formula, thereby progressing further than in the case of local optimization techniques. A wide panel of local solutions is found using this procedure, resulting in a large range of optical thicknesses. The efficiency of this technique is illustrated by two thin film design problems, in particular an infrared antireflection coating, and a solar-selective absorber coating. PMID:24663856

  12. Optimal design of Purcell's three-link swimmer.

    PubMed

    Giraldi, Laetitia; Martinon, Pierre; Zoppello, Marta

    2015-02-01

    In this paper we address the question of the optimal design for the Purcell three-link swimmer. More precisely, we investigate the best link length ratio which maximizes its displacement. The dynamics of the swimmer is expressed as an ordinary differential equation, using the resistive force theory. Among a set of optimal strategies of deformation (strokes), we provide an asymptotic estimate of the displacement for small deformations, from which we derive the optimal link ratio. Numerical simulations are in good agreement with this theoretical estimate and also cover larger amplitudes of deformation. Compared with the classical design of the Purcell swimmer, we observe a gain in displacement of roughly 60%. PMID:25768602

  13. Towards Robust Designs Via Multiple-Objective Optimization Methods

    NASA Technical Reports Server (NTRS)

    Man Mohan, Rai

    2006-01-01

    Fabricating and operating complex systems involves dealing with uncertainty in the relevant variables. In the case of aircraft, flow conditions are subject to change during operation. Efficiency and engine noise may be different from the expected values because of manufacturing tolerances and normal wear and tear. Engine components may have a shorter life than expected because of manufacturing tolerances. In spite of the important effect of operating- and manufacturing-uncertainty on the performance and expected life of the component or system, traditional aerodynamic shape optimization has focused on obtaining the best design given a set of deterministic flow conditions. Clearly it is important to both maintain near-optimal performance levels at off-design operating conditions, and, ensure that performance does not degrade appreciably when the component shape differs from the optimal shape due to manufacturing tolerances and normal wear and tear. These requirements naturally lead to the idea of robust optimal design wherein the concept of robustness to various perturbations is built into the design optimization procedure. The basic ideas involved in robust optimal design will be included in this lecture. The imposition of the additional requirement of robustness results in a multiple-objective optimization problem requiring appropriate solution procedures. Typically the costs associated with multiple-objective optimization are substantial. Therefore efficient multiple-objective optimization procedures are crucial to the rapid deployment of the principles of robust design in industry. Hence the companion set of lecture notes (Single- and Multiple-Objective Optimization with Differential Evolution and Neural Networks ) deals with methodology for solving multiple-objective Optimization problems efficiently, reliably and with little user intervention. Applications of the methodologies presented in the companion lecture to robust design will be included here. The

  14. Optimal experimental dynamical decoupling of both longitudinal and transverse relaxations

    NASA Astrophysics Data System (ADS)

    Zhen, Xing-Long; Zhang, Fei-Hao; Feng, Guanru; Li, Hang; Long, Gui-Lu

    2016-02-01

    Both longitudinal and transverse relaxations exist in the practical environment. Their simultaneous eliminations are extremely demanding in real applications. Previous experimental work has focused mainly on the suppression of transverse relaxation. In this paper we investigate the performance of three important dynamical decoupling schemes—quadratic dynamical decoupling, periodic dynamical decoupling, and concatenated dynamical decoupling—in an environment with hybrid errors. We propose a method to engineer arbitrary environment by modulating the control field. The technique developed here is universal and can be applied to other quantum information processing systems. Three-dimensional filter functions technique is utilized to analyze the fidelity decay of a one-qubit state protected by dynamical decoupling sequences. This enables us to quantitatively compare the performance of different dynamical decoupling sequences and demonstrate the superiority of quadratic dynamical decoupling in experiments for the first time. Our work reveals that quadratic dynamical decoupling is optimal conditioned on the appropriate noise properties. The difference of constructing dynamical decoupling sequences with various Pauli pulses is also investigated.

  15. Formulation for Simultaneous Aerodynamic Analysis and Design Optimization

    NASA Technical Reports Server (NTRS)

    Hou, G. W.; Taylor, A. C., III; Mani, S. V.; Newman, P. A.

    1993-01-01

    An efficient approach for simultaneous aerodynamic analysis and design optimization is presented. This approach does not require the performance of many flow analyses at each design optimization step, which can be an expensive procedure. Thus, this approach brings us one step closer to meeting the challenge of incorporating computational fluid dynamic codes into gradient-based optimization techniques for aerodynamic design. An adjoint-variable method is introduced to nullify the effect of the increased number of design variables in the problem formulation. The method has been successfully tested on one-dimensional nozzle flow problems, including a sample problem with a normal shock. Implementations of the above algorithm are also presented that incorporate Newton iterations to secure a high-quality flow solution at the end of the design process. Implementations with iterative flow solvers are possible and will be required for large, multidimensional flow problems.

  16. Study on aerodynamic design optimization of turbomachinery blades

    NASA Astrophysics Data System (ADS)

    Chen, Naixing; Zhang, Hongwu; Huang, Weiguang; Xu, Yanji

    2005-12-01

    This paper describes the study on aerodynamics design optimization of turbomachinery blading developed by the authors at the Institute of Engineering Thermophysics, Chinese Academy of Sciences, during the recent few years. The present paper describes the aspects mainly on how to use a rapid approach of profiling a 3D blading and of grid generation for computation, a fast and accurate viscous computation method and an appropriate optimization methodology including a blade parameterization algorithm to optimize turbomachinery blading aerodynamically. Any blade configuration can be expressed by three curves, they are the camber lines, the thickness distributions and the radial stacking line, and then the blade geometry can be easily parameterized by a number of parameters with three polynomials. A gradient-based parameterization analytical method and a response surface method were applied herein for blade optimization. It was found that the optimization process provides reliable design for turbomachinery with reasonable computing time.

  17. Optimization of the GRAPE Polarimeter Design

    NASA Astrophysics Data System (ADS)

    McConnell, Mark

    The Gamma Ray Polarimeter Experiment (GRAPE) is designed to investigate one of the most exotic phenomena in the universe - gamma-ray bursts (GRB). There has been intense observational and theoretical research in recent years, but research in this area has been largely focused on studies of time histories, spectra, and spatial distributions. Theoretical models show that a more complete understanding of the inner structure of GRBs, including the geometry and physical processes close to the central engine, requires the exploitation of gamma-ray polarimetry. Over the past several years, we have developed the GRAPE instrument to measure the polarization of gamma-rays from GRBs over the energy range of 50 to 500 keV. The GRAPE design is a modular one in which several independent modules are required to achieve sufficient sensitivity. A single module fits on the front end of a 2-inch square flat-panel multi-anode photomultiplier tube (MAPMT). The first operational balloon flight took in place in September of 2011 from Ft. Sumner, NM. The purpose of the 2011 flight was to validate the science capability of GRAPE by measuring the Crab polarization with a collimated array of 16 modules. The limited success of that flight led to a second validation flight (also from Ft. Sumner) in the fall of 2014, with significantly improved shielding and a larger array of modules. That flight proved too short to make a full observation of the Crab. Although we did not succeed in measuring the polarization of the Crab with a high degree of confidence, we feel that we are nonetheless prepared to move forward with our program. Our next goal is to fly GRAPE on a long duration balloon (LDB) platform to collect data on a significant sample of GRBs. Our experience with the first two balloon flights, coupled with further design efforts focused on orbital payloads, has led to an improved polarimeter concept that represents a natural evolution of the current design. It is this new concept that we are

  18. New approaches to optimization in aerospace conceptual design

    NASA Technical Reports Server (NTRS)

    Gage, Peter J.

    1995-01-01

    Aerospace design can be viewed as an optimization process, but conceptual studies are rarely performed using formal search algorithms. Three issues that restrict the success of automatic search are identified in this work. New approaches are introduced to address the integration of analyses and optimizers, to avoid the need for accurate gradient information and a smooth search space (required for calculus-based optimization), and to remove the restrictions imposed by fixed complexity problem formulations. (1) Optimization should be performed in a flexible environment. A quasi-procedural architecture is used to conveniently link analysis modules and automatically coordinate their execution. It efficiently controls a large-scale design tasks. (2) Genetic algorithms provide a search method for discontinuous or noisy domains. The utility of genetic optimization is demonstrated here, but parameter encodings and constraint-handling schemes must be carefully chosen to avoid premature convergence to suboptimal designs. The relationship between genetic and calculus-based methods is explored. (3) A variable-complexity genetic algorithm is created to permit flexible parameterization, so that the level of description can change during optimization. This new optimizer automatically discovers novel designs in structural and aerodynamic tasks.

  19. Optimal design of a subsurface redox barrier

    SciTech Connect

    Chilakapati, A.

    1999-06-01

    Harmful contaminants such as chromium (Cr{sup +6}), and TCE can be removed from groundwater by reactions with reduced subsurface sediments. Establishing an in situ Fe(II) barrier through the reduction of soil-bound Fe(III) to Fe(II) by injecting a sodium dithionite (Na{sub 2}S{sub 2}O{sub 4}) solution is studied. Critical to this problem is the possible formation and expansion of a zone around the injection, where all the soil-bound Fe(III) is reduced to Fe(II). Different reaction models apply inside and outside of this zone so that a determination of this moving boundary is a fundamental part of the solution. The complete analytic solution to this problem was used to develop optimal process parameters, such as injection rate and operational time, that maximize the radius of the Fe(III)-reduced zone when a given mass of sodium dithionite is injected at a well. When a large reduction [>63% of initially present Fe(III)] is desired, the results indicate that it is better to use a low flow rate to form a Fe(III)-free zone around the injection. The opposite is true for smaller reductions (<63%), so that a faster injection rate that avoids the formation of the Fe(III)-free zone yields a larger reduction zone.

  20. Efficient optimal design of smooth optical freeform surfaces using ray targeting

    NASA Astrophysics Data System (ADS)

    Wu, Rengmao; Wang, Huihui; Liu, Peng; Zhang, Yaqin; Zheng, Zhenrong; Li, Haifeng; Liu, Xu

    2013-07-01

    An optimization design method is proposed for generating smooth freeform reflective and refractive surfaces. In this method, two optimization steps are employed for ray targeting. The first step aims to ensure the shape of the target illumination, and the second step is employed to further improve the irradiance uniformity. These two steps can provide significant savings of time because the time consuming Monte Carlo raytracing is not used during the optimization process. Both smooth freeform reflective surfaces and smooth freeform refractive surfaces can be designed, and the target illumination could be achieved just by controlling the positions of several hundred predefined rays on the target plane with these two steps. The simulation results and the experimental tests show that this optimization design method is robust and efficient.

  1. The Implications of "Contamination" for Experimental Design in Education

    ERIC Educational Resources Information Center

    Rhoads, Christopher H.

    2011-01-01

    Experimental designs that randomly assign entire clusters of individuals (e.g., schools and classrooms) to treatments are frequently advocated as a way of guarding against contamination of the estimated average causal effect of treatment. However, in the absence of contamination, experimental designs that randomly assign intact clusters to…

  2. Optimizing experimental procedures for quantitative evaluation of crop plant performance in high throughput phenotyping systems

    PubMed Central

    Junker, Astrid; Muraya, Moses M.; Weigelt-Fischer, Kathleen; Arana-Ceballos, Fernando; Klukas, Christian; Melchinger, Albrecht E.; Meyer, Rhonda C.; Riewe, David; Altmann, Thomas

    2015-01-01

    Detailed and standardized protocols for plant cultivation in environmentally controlled conditions are an essential prerequisite to conduct reproducible experiments with precisely defined treatments. Setting up appropriate and well defined experimental procedures is thus crucial for the generation of solid evidence and indispensable for successful plant research. Non-invasive and high throughput (HT) phenotyping technologies offer the opportunity to monitor and quantify performance dynamics of several hundreds of plants at a time. Compared to small scale plant cultivations, HT systems have much higher demands, from a conceptual and a logistic point of view, on experimental design, as well as the actual plant cultivation conditions, and the image analysis and statistical methods for data evaluation. Furthermore, cultivation conditions need to be designed that elicit plant performance characteristics corresponding to those under natural conditions. This manuscript describes critical steps in the optimization of procedures for HT plant phenotyping systems. Starting with the model plant Arabidopsis, HT-compatible methods were tested, and optimized with regard to growth substrate, soil coverage, watering regime, experimental design (considering environmental inhomogeneities) in automated plant cultivation and imaging systems. As revealed by metabolite profiling, plant movement did not affect the plants' physiological status. Based on these results, procedures for maize HT cultivation and monitoring were established. Variation of maize vegetative growth in the HT phenotyping system did match well with that observed in the field. The presented results outline important issues to be considered in the design of HT phenotyping experiments for model and crop plants. It thereby provides guidelines for the setup of HT experimental procedures, which are required for the generation of reliable and reproducible data of phenotypic variation for a broad range of applications. PMID

  3. Optimal structural design via optimality criteria as a nonsmooth mechanics problem

    NASA Astrophysics Data System (ADS)

    Tzaferopoulos, M. Ap.; Stravroulakis, G. E.

    1995-06-01

    In the theory of plastic structural design via optimality criteria (due to W. Prager), the optimal design problem is transformed to a nonlinear elastic structural analysis problem with appropriate stress-strain laws, which generally include complete vertical branches. In this context, the concept of structural universe (in the sense of G. Rozvany) permits the treatment of complicated optimal layout problems. Recent progress in the field of nonsmooth mechanics makes the solution of structural analysis problems with this kind of 'complete' law possible. Elements from the two fields are combined in this paper for the solution of optimal design and layout problems for structures. The optimal layout of plane trusses with various specific cost functions is studied here as a representative problem. The use of convex, continuous and piecewise linear specific cost functions for the structural members leads to problems of linear variational inequalities or equivalently piecewise linear, convex but nonsmooth optimization problems, which are solved by means of an iterative algorithm based on sequential linear programming techniques. Numerical examples illustrate the theory and its applicability to practical engineering structures. Following a parametric investigation of an optimal bridge design, certain aspects of the optimal truss layout problem are discussed, which can be extended to other types of structural systems as well.

  4. First wall/blanket/shield design and optimization system

    SciTech Connect

    Gohar, Y.; Baker, C.; Attaya, H.; Cha, Y.; Majumdar, S.; Scandora, T.

    1988-02-01

    First wall/blanket/shield design and optimization system (BSDOS) has been developed to provide a state-of-the-art design tool for fast accurate analysis. In addition, it has been designed to perform several other functions: (1) allowing comparison and evaluation studies for different concepts using the same data bases and ground rules, (2) permitting the use of any figure of merit in the evaluation studies, (3) optimizing the first wall/blanket/shield design parameters for any figure of merit under several design constraints, (4) permitting the use of different reactor parameters in the evaluation and optimization analyses, (5) allowing the use of improved eingineering data bases to study the impact on the design performance for planning future research and development, and (6) evaluating the effect of the data base uncertainties on the design performance. BSDOS is the first design and optimization system to couple the highly interacting neutronics, heat transfer, thermal hydraulics, stress analysis, radioactivity and decay-heat analyses, tritium balance, and capital cost. A brief description of the main features of BSDOS is given in this paper. Also, results from using BSDOS to perform design analysis for several reactor components are presented. 17 refs., 1 fig., 2 tabs.

  5. Scalar and Multivariate Approaches for Optimal Network Design in Antarctica

    NASA Astrophysics Data System (ADS)

    Hryniw, Natalia

    Observations are crucial for weather and climate, not only for daily forecasts and logistical purposes, for but maintaining representative records and for tuning atmospheric models. Here scalar theory for optimal network design is expanded in a multivariate framework, to allow for optimal station siting for full field optimization. Ensemble sensitivity theory is expanded to produce the covariance trace approach, which optimizes for the trace of the covariance matrix. Relative entropy is also used for multivariate optimization as an information theory approach for finding optimal locations. Antarctic surface temperature data is used as a testbed for these methods. Both methods produce different results which are tied to the fundamental physical parameters of the Antarctic temperature field.

  6. Improved method for transonic airfoil design-by-optimization

    NASA Technical Reports Server (NTRS)

    Kennelly, R. A., Jr.

    1983-01-01

    An improved method for use of optimization techniques in transonic airfoil design is demonstrated. FLO6QNM incorporates a modified quasi-Newton optimization package, and is shown to be more reliable and efficient than the method developed previously at NASA-Ames, which used the COPES/CONMIN optimization problem. The design codes are compared on a series of test cases with known solutions, and the effects of problem scaling, proximity of initial point to solution, and objective function precision are studied. In contrast to the older method, well-converged solutions are shown to be attainable in the context of engineering design using computational fluid dynamics tools, a new result. The improvements are due to better performance by the optimization routine and to the use of problem-adaptive finite difference step sizes for gradient evaluation.

  7. Improved method for transonic airfoil design-by-optimization

    NASA Technical Reports Server (NTRS)

    Kennelly, R. A., Jr.

    1983-01-01

    An improved method for use of optimization techniques in transonic airfoil design is demonstrated. FLO6QNM incorporates a modified quasi-Newton optimization package, and is shown to be more reliable and efficient than the method developed previously at NASA-Ames, which used the COPES/CONMIN optimization program. The design codes are compared on a series of test cases with known solutions, and the effects of problem scaling, proximity of initial point to solution, and objective function precision are studied. In contrast to the older method, well-converged solutions are shown to be attainable in the context of engineering design using computational fluid dynamics tools, a new result. The improvements are due to better performance by the optimization routine and to the use of problem-adaptive finite difference step sizes for gradient evaluation.

  8. Fuel Injector Design Optimization for an Annular Scramjet Geometry

    NASA Astrophysics Data System (ADS)

    Steffen, Christopher J., Jr.

    2003-01-01

    A four-parameter, three-level, central composite experiment design has been used to optimize the configuration of an annular scramjet injector geometry using computational fluid dynamics. The computational fluid dynamic solutions played the role of computer experiments, and response surface methodology was used to capture the simulation results for mixing efficiency and total pressure recovery within the scramjet flowpath. An optimization procedure, based upon the response surface results of mixing efficiency, was used to compare the optimal design configuration against the target efficiency value of 92.5%. The results of three different optimization procedures are presented and all point to the need to look outside the current design space for different injector geometries that can meet or exceed the stated mixing efficiency target.

  9. Tabu search method with random moves for globally optimal design

    NASA Astrophysics Data System (ADS)

    Hu, Nanfang

    1992-09-01

    Optimum engineering design problems are usually formulated as non-convex optimization problems of continuous variables. Because of the absence of convexity structure, they can have multiple minima, and global optimization becomes difficult. Traditional methods of optimization, such as penalty methods, can often be trapped at a local optimum. The tabu search method with random moves to solve approximately these problems is introduced. Its reliability and efficiency are examined with the help of standard test functions. By the analysis of the implementations, it is seen that this method is easy to use, and no derivative information is necessary. It outperforms the random search method and composite genetic algorithm. In particular, it is applied to minimum weight design examples of a three-bar truss, coil springs, a Z-section and a channel section. For the channel section, the optimal design using the tabu search method with random moves saved 26.14 percent over the weight of the SUMT method.

  10. On Optimal Input Design and Model Selection for Communication Channels

    SciTech Connect

    Li, Yanyan; Djouadi, Seddik M; Olama, Mohammed M

    2013-01-01

    In this paper, the optimal model (structure) selection and input design which minimize the worst case identification error for communication systems are provided. The problem is formulated using metric complexity theory in a Hilbert space setting. It is pointed out that model selection and input design can be handled independently. Kolmogorov n-width is used to characterize the representation error introduced by model selection, while Gel fand and Time n-widths are used to represent the inherent error introduced by input design. After the model is selected, an optimal input which minimizes the worst case identification error is shown to exist. In particular, it is proven that the optimal model for reducing the representation error is a Finite Impulse Response (FIR) model, and the optimal input is an impulse at the start of the observation interval. FIR models are widely popular in communication systems, such as, in Orthogonal Frequency Division Multiplexing (OFDM) systems.

  11. Designing and optimizing a healthcare kiosk for the community.

    PubMed

    Lyu, Yongqiang; Vincent, Christopher James; Chen, Yu; Shi, Yuanchun; Tang, Yida; Wang, Wenyao; Liu, Wei; Zhang, Shuangshuang; Fang, Ke; Ding, Ji

    2015-03-01

    Investigating new ways to deliver care, such as the use of self-service kiosks to collect and monitor signs of wellness, supports healthcare efficiency and inclusivity. Self-service kiosks offer this potential, but there is a need for solutions to meet acceptable standards, e.g. provision of accurate measurements. This study investigates the design and optimization of a prototype healthcare kiosk to collect vital signs measures. The design problem was decomposed, formalized, focused and used to generate multiple solutions. Systematic implementation and evaluation allowed for the optimization of measurement accuracy, first for individuals and then for a population. The optimized solution was tested independently to check the suitability of the methods, and quality of the solution. The process resulted in a reduction of measurement noise and an optimal fit, in terms of the positioning of measurement devices. This guaranteed the accuracy of the solution and provides a general methodology for similar design problems. PMID:25479985

  12. Computer Language For Optimization Of Design

    NASA Technical Reports Server (NTRS)

    Scotti, Stephen J.; Lucas, Stephen H.

    1991-01-01

    SOL is computer language geared to solution of design problems. Includes mathematical modeling and logical capabilities of computer language like FORTRAN; also includes additional power of nonlinear mathematical programming methods at language level. SOL compiler takes SOL-language statements and generates equivalent FORTRAN code and system calls. Provides syntactic and semantic checking for recovery from errors and provides detailed reports containing cross-references to show where each variable used. Implemented on VAX/VMS computer systems. Requires VAX FORTRAN compiler to produce executable program.

  13. An optimal trajectory design for debris deorbiting

    NASA Astrophysics Data System (ADS)

    Ouyang, Gaoxiang; Dong, Xin; Li, Xin; Zhang, Yang

    2016-01-01

    The problem of deorbiting debris is studied in this paper. As a feasible measure, a disposable satellite would be launched, attach to debris, and deorbit the space debris using a technology named electrodynamic tether (EDT). In order to deorbit multiple debris as many as possible, a suboptimal but feasible and efficient trajectory set has been designed to allow a deorbiter satellite tour the LEO small bodies per one mission. Finally a simulation given by this paper showed that a 600 kg satellite is capable of deorbiting 6 debris objects in about 230 days.

  14. Role of Design Standards in Wind Plant Optimization (Presentation)

    SciTech Connect

    Veers, P.; Churchfield, M.; Lee, S.; Moon, J.; Larsen, G.

    2013-10-01

    When a turbine is optimized, it is done within the design constraints established by the objective criteria in the international design standards used to certify a design. Since these criteria are multifaceted, it is a challenging task to conduct the optimization, but it can be done. The optimization is facilitated by the fact that a standard turbine model is subjected to standard inflow conditions that are well characterized in the standard. Examples of applying these conditions to rotor optimization are examined. In other cases, an innovation may provide substantial improvement in one area, but be challenged to impact all of the myriad design load cases. When a turbine is placed in a wind plant, the challenge is magnified. Typical design practice optimizes the turbine for stand-alone operation, and then runs a check on the actual site conditions, including wakes from all nearby turbines. Thus, each turbine in a plant has unique inflow conditions. The possibility of creating objective and consistent inflow conditions for turbines within a plant, for used in optimization of the turbine and the plant, are examined with examples taken from LES simulation.

  15. Geometry Modeling and Grid Generation for Design and Optimization

    NASA Technical Reports Server (NTRS)

    Samareh, Jamshid A.

    1998-01-01

    Geometry modeling and grid generation (GMGG) have played and will continue to play an important role in computational aerosciences. During the past two decades, tremendous progress has occurred in GMGG; however, GMGG is still the biggest bottleneck to routine applications for complicated Computational Fluid Dynamics (CFD) and Computational Structures Mechanics (CSM) models for analysis, design, and optimization. We are still far from incorporating GMGG tools in a design and optimization environment for complicated configurations. It is still a challenging task to parameterize an existing model in today's Computer-Aided Design (CAD) systems, and the models created are not always good enough for automatic grid generation tools. Designers may believe their models are complete and accurate, but unseen imperfections (e.g., gaps, unwanted wiggles, free edges, slivers, and transition cracks) often cause problems in gridding for CSM and CFD. Despite many advances in grid generation, the process is still the most labor-intensive and time-consuming part of the computational aerosciences for analysis, design, and optimization. In an ideal design environment, a design engineer would use a parametric model to evaluate alternative designs effortlessly and optimize an existing design for a new set of design objectives and constraints. For this ideal environment to be realized, the GMGG tools must have the following characteristics: (1) be automated, (2) provide consistent geometry across all disciplines, (3) be parametric, and (4) provide sensitivity derivatives. This paper will review the status of GMGG for analysis, design, and optimization processes, and it will focus on some emerging ideas that will advance the GMGG toward the ideal design environment.

  16. Irradiation Design for an Experimental Murine Model

    SciTech Connect

    Ballesteros-Zebadua, P.; Moreno-Jimenez, S.; Suarez-Campos, J. E.; Celis, M. A.; Larraga-Gutierrez, J. M.; Garcia-Garduno, O. A.; Rubio-Osornio, M. C.; Custodio-Ramirez, V.; Paz, C.

    2010-12-07

    In radiotherapy and stereotactic radiosurgery, small animal experimental models are frequently used, since there are still a lot of unsolved questions about the biological and biochemical effects of ionizing radiation. This work presents a method for small-animal brain radiotherapy compatible with a dedicated 6MV Linac. This rodent model is focused on the research of the inflammatory effects produced by ionizing radiation in the brain. In this work comparisons between Pencil Beam and Monte Carlo techniques, were used in order to evaluate accuracy of the calculated dose using a commercial planning system. Challenges in this murine model are discussed.

  17. Superconducting Fault Current Limiter optimized design

    NASA Astrophysics Data System (ADS)

    Tixador, Pascal; Badel, Arnaud

    2015-11-01

    The SuperConducting Fault Current Limiter (SCFCL) appears as one of the most promising SC applications for the electrical grids. Despite its advantages and many successful field experiences the market of SCFCL has difficulties to take off even if the first orders for permanent operation in grids are taken. The analytical design of resistive SCFCL will be discussed with the objective to reduce the quantity of SC conductor (length and section) to be more cost-effective. For that the SC conductor must have a high resistivity in normal state. It can be achieved by using high resistivity alloy for shunt, such as Hastelloy®. One of the most severe constraint is that the SCFCL should operate safely for any faults, especially those with low prospective short-circuit currents. This constraint requires to properly design the thickness of the SC tape in order to limit the hot spot temperature. An operation at 65 K appears as very interesting since it decreases the SC cost at least by a factor 2 with a simple LN2 cryogenics. Taking into account the cost reduction in a near future, the SC conductor cost could be rather low, half a dollar per kV A.

  18. Design sensitivity analysis and optimization tool (DSO) for sizing design applications

    NASA Technical Reports Server (NTRS)

    Chang, Kuang-Hua; Choi, Kyung K.; Perng, Jyh-Hwa

    1992-01-01

    The DSO tool, a structural design software system that provides the designer with a graphics-based menu-driven design environment to perform easy design optimization for general applications, is presented. Three design stages, preprocessing, design sensitivity analysis, and postprocessing, are implemented in the DSO to allow the designer to carry out the design process systematically. A framework, including data base, user interface, foundation class, and remote module, has been designed and implemented to facilitate software development for the DSO. A number of dedicated commercial software/packages have been integrated in the DSO to support the design procedures. Instead of parameterizing an FEM, design parameters are defined on a geometric model associated with physical quantities, and the continuum design sensitivity analysis theory is implemented to compute design sensitivity coefficients using postprocessing data from the analysis codes. A tracked vehicle road wheel is given as a sizing design application to demonstrate the DSO's easy and convenient design optimization process.

  19. Multiobjective hyper heuristic scheme for system design and optimization

    NASA Astrophysics Data System (ADS)

    Rafique, Amer Farhan

    2012-11-01

    As system design is becoming more and more multifaceted, integrated, and complex, the traditional single objective optimization trends of optimal design are becoming less and less efficient and effective. Single objective optimization methods present a unique optimal solution whereas multiobjective methods present pareto front. The foremost intent is to predict a reasonable distributed pareto-optimal solution set independent of the problem instance through multiobjective scheme. Other objective of application of intended approach is to improve the worthiness of outputs of the complex engineering system design process at the conceptual design phase. The process is automated in order to provide the system designer with the leverage of the possibility of studying and analyzing a large multiple of possible solutions in a short time. This article presents Multiobjective Hyper Heuristic Optimization Scheme based on low level meta-heuristics developed for the application in engineering system design. Herein, we present a stochastic function to manage meta-heuristics (low-level) to augment surety of global optimum solution. Generic Algorithm, Simulated Annealing and Swarm Intelligence are used as low-level meta-heuristics in this study. Performance of the proposed scheme is investigated through a comprehensive empirical analysis yielding acceptable results. One of the primary motives for performing multiobjective optimization is that the current engineering systems require simultaneous optimization of conflicting and multiple. Random decision making makes the implementation of this scheme attractive and easy. Injecting feasible solutions significantly alters the search direction and also adds diversity of population resulting in accomplishment of pre-defined goals set in the proposed scheme.

  20. Performance Trend of Different Algorithms for Structural Design Optimization

    NASA Technical Reports Server (NTRS)

    Patnaik, Surya N.; Coroneos, Rula M.; Guptill, James D.; Hopkins, Dale A.

    1996-01-01

    Nonlinear programming algorithms play an important role in structural design optimization. Fortunately, several algorithms with computer codes are available. At NASA Lewis Research Center, a project was initiated to assess performance of different optimizers through the development of a computer code CometBoards. This paper summarizes the conclusions of that research. CometBoards was employed to solve sets of small, medium and large structural problems, using different optimizers on a Cray-YMP8E/8128 computer. The reliability and efficiency of the optimizers were determined from the performance of these problems. For small problems, the performance of most of the optimizers could be considered adequate. For large problems however, three optimizers (two sequential quadratic programming routines, DNCONG of IMSL and SQP of IDESIGN, along with the sequential unconstrained minimizations technique SUMT) outperformed others. At optimum, most optimizers captured an identical number of active displacement and frequency constraints but the number of active stress constraints differed among the optimizers. This discrepancy can be attributed to singularity conditions in the optimization and the alleviation of this discrepancy can improve the efficiency of optimizers.