Science.gov

Sample records for driveway mean-variance optimization

  1. Portfolio optimization with mean-variance model

    NASA Astrophysics Data System (ADS)

    Hoe, Lam Weng; Siew, Lam Weng

    2016-06-01

    Investors wish to achieve the target rate of return at the minimum level of risk in their investment. Portfolio optimization is an investment strategy that can be used to minimize the portfolio risk and can achieve the target rate of return. The mean-variance model has been proposed in portfolio optimization. The mean-variance model is an optimization model that aims to minimize the portfolio risk which is the portfolio variance. The objective of this study is to construct the optimal portfolio using the mean-variance model. The data of this study consists of weekly returns of 20 component stocks of FTSE Bursa Malaysia Kuala Lumpur Composite Index (FBMKLCI). The results of this study show that the portfolio composition of the stocks is different. Moreover, investors can get the return at minimum level of risk with the constructed optimal mean-variance portfolio.

  2. Multiperiod Mean-Variance Portfolio Optimization via Market Cloning

    SciTech Connect

    Ankirchner, Stefan; Dermoune, Azzouz

    2011-08-15

    The problem of finding the mean variance optimal portfolio in a multiperiod model can not be solved directly by means of dynamic programming. In order to find a solution we therefore first introduce independent market clones having the same distributional properties as the original market, and we replace the portfolio mean and variance by their empirical counterparts. We then use dynamic programming to derive portfolios maximizing a weighted sum of the empirical mean and variance. By letting the number of market clones converge to infinity we are able to solve the original mean variance problem.

  3. PET image reconstruction: mean, variance, and optimal minimax criterion

    NASA Astrophysics Data System (ADS)

    Liu, Huafeng; Gao, Fei; Guo, Min; Xue, Liying; Nie, Jing; Shi, Pengcheng

    2015-04-01

    Given the noise nature of positron emission tomography (PET) measurements, it is critical to know the image quality and reliability as well as expected radioactivity map (mean image) for both qualitative interpretation and quantitative analysis. While existing efforts have often been devoted to providing only the reconstructed mean image, we present a unified framework for joint estimation of the mean and corresponding variance of the radioactivity map based on an efficient optimal min-max criterion. The proposed framework formulates the PET image reconstruction problem to be a transformation from system uncertainties to estimation errors, where the minimax criterion is adopted to minimize the estimation errors with possibly maximized system uncertainties. The estimation errors, in the form of a covariance matrix, express the measurement uncertainties in a complete way. The framework is then optimized by ∞-norm optimization and solved with the corresponding H∞ filter. Unlike conventional statistical reconstruction algorithms, that rely on the statistical modeling methods of the measurement data or noise, the proposed joint estimation stands from the point of view of signal energies and can handle from imperfect statistical assumptions to even no a priori statistical assumptions. The performance and accuracy of reconstructed mean and variance images are validated using Monte Carlo simulations. Experiments on phantom scans with a small animal PET scanner and real patient scans are also conducted for assessment of clinical potential.

  4. Swarm based mean-variance mapping optimization (MVMOS) for solving economic dispatch

    NASA Astrophysics Data System (ADS)

    Khoa, T. H.; Vasant, P. M.; Singh, M. S. Balbir; Dieu, V. N.

    2014-10-01

    The economic dispatch (ED) is an essential optimization task in the power generation system. It is defined as the process of allocating the real power output of generation units to meet required load demand so as their total operating cost is minimized while satisfying all physical and operational constraints. This paper introduces a novel optimization which named as Swarm based Mean-variance mapping optimization (MVMOS). The technique is the extension of the original single particle mean-variance mapping optimization (MVMO). Its features make it potentially attractive algorithm for solving optimization problems. The proposed method is implemented for three test power systems, including 3, 13 and 20 thermal generation units with quadratic cost function and the obtained results are compared with many other methods available in the literature. Test results have indicated that the proposed method can efficiently implement for solving economic dispatch.

  5. Firefly Algorithm for Cardinality Constrained Mean-Variance Portfolio Optimization Problem with Entropy Diversity Constraint

    PubMed Central

    2014-01-01

    Portfolio optimization (selection) problem is an important and hard optimization problem that, with the addition of necessary realistic constraints, becomes computationally intractable. Nature-inspired metaheuristics are appropriate for solving such problems; however, literature review shows that there are very few applications of nature-inspired metaheuristics to portfolio optimization problem. This is especially true for swarm intelligence algorithms which represent the newer branch of nature-inspired algorithms. No application of any swarm intelligence metaheuristics to cardinality constrained mean-variance (CCMV) portfolio problem with entropy constraint was found in the literature. This paper introduces modified firefly algorithm (FA) for the CCMV portfolio model with entropy constraint. Firefly algorithm is one of the latest, very successful swarm intelligence algorithm; however, it exhibits some deficiencies when applied to constrained problems. To overcome lack of exploration power during early iterations, we modified the algorithm and tested it on standard portfolio benchmark data sets used in the literature. Our proposed modified firefly algorithm proved to be better than other state-of-the-art algorithms, while introduction of entropy diversity constraint further improved results. PMID:24991645

  6. Mean-variance portfolio analysis data for optimizing community-based photovoltaic investment.

    PubMed

    Shakouri, Mahmoud; Lee, Hyun Woo

    2016-03-01

    The amount of electricity generated by Photovoltaic (PV) systems is affected by factors such as shading, building orientation and roof slope. To increase electricity generation and reduce volatility in generation of PV systems, a portfolio of PV systems can be made which takes advantages of the potential synergy among neighboring buildings. This paper contains data supporting the research article entitled: PACPIM: new decision-support model of optimized portfolio analysis for community-based photovoltaic investment [1]. We present a set of data relating to physical properties of 24 houses in Oregon, USA, along with simulated hourly electricity data for the installed PV systems. The developed Matlab code to construct optimized portfolios is also provided in . The application of these files can be generalized to variety of communities interested in investing on PV systems. PMID:26937458

  7. Conversations across Meaning Variance

    ERIC Educational Resources Information Center

    Cordero, Alberto

    2013-01-01

    Progressive interpretations of scientific theories have long been denounced as naive, because of the inescapability of meaning variance. The charge reportedly applies to recent realist moves that focus on theory-parts rather than whole theories. This paper considers the question of what "theory-parts" of epistemic significance (if any) relevantly…

  8. Driveway identification signing and marking

    SciTech Connect

    Not Available

    1986-09-01

    A primary purpose of a public highway is to provide access to commercial establishments. Driveways to fast-food restaurants, banks, retail stores, and office buildings line many road sections. A bewildering array of reflectorized, unreflectorized, lighted, and unlighted high- and low-mounted signs with varying messages direct motorists in and out of many of the driveways, while other access points have no signs. In a very few cases, the Manual of Uniform Traffic Control Devices-type signing for one-way roads has been installed for the one-way-type circulation design at some of these establishments. Should MUTCD-type one-way signs be recommended practice to control and provide a more uniform identification of these driveways when necessary. Are there signs not presently in the MUTCD that should be added to cope with this situation.

  9. Mean-Variance Hedging on Uncertain Time Horizon in a Market with a Jump

    SciTech Connect

    Kharroubi, Idris; Lim, Thomas; Ngoupeyou, Armand

    2013-12-15

    In this work, we study the problem of mean-variance hedging with a random horizon T∧τ, where T is a deterministic constant and τ is a jump time of the underlying asset price process. We first formulate this problem as a stochastic control problem and relate it to a system of BSDEs with a jump. We then provide a verification theorem which gives the optimal strategy for the mean-variance hedging using the solution of the previous system of BSDEs. Finally, we prove that this system of BSDEs admits a solution via a decomposition approach coming from filtration enlargement theory.

  10. The cost of geothermal energy in the western US region:a portfolio-based approach a mean-variance portfolio optimization of the regions' generating mix to 2013.

    SciTech Connect

    Beurskens, Luuk (ECN-Energy Research Centre of the Netherland); Jansen, Jaap C. (ECN-Energy Research Centre of the Netherlands); Awerbuch, Shimon Ph.D. (.University of Sussex, Brighton, UK); Drennen, Thomas E.

    2005-09-01

    Energy planning represents an investment-decision problem. Investors commonly evaluate such problems using portfolio theory to manage risk and maximize portfolio performance under a variety of unpredictable economic outcomes. Energy planners need to similarly abandon their reliance on traditional, ''least-cost'' stand-alone technology cost estimates and instead evaluate conventional and renewable energy sources on the basis of their portfolio cost--their cost contribution relative to their risk contribution to a mix of generating assets. This report describes essential portfolio-theory ideas and discusses their application in the Western US region. The memo illustrates how electricity-generating mixes can benefit from additional shares of geothermal and other renewables. Compared to fossil-dominated mixes, efficient portfolios reduce generating cost while including greater renewables shares in the mix. This enhances energy security. Though counter-intuitive, the idea that adding more costly geothermal can actually reduce portfolio-generating cost is consistent with basic finance theory. An important implication is that in dynamic and uncertain environments, the relative value of generating technologies must be determined not by evaluating alternative resources, but by evaluating alternative resource portfolios. The optimal results for the Western US Region indicate that compared to the EIA target mixes, there exist generating mixes with larger geothermal shares at equal-or-lower expected cost and risk.

  11. Continuous-Time Mean-Variance Portfolio Selection with Random Horizon

    SciTech Connect

    Yu, Zhiyong

    2013-12-15

    This paper examines the continuous-time mean-variance optimal portfolio selection problem with random market parameters and random time horizon. Treating this problem as a linearly constrained stochastic linear-quadratic optimal control problem, I explicitly derive the efficient portfolios and efficient frontier in closed forms based on the solutions of two backward stochastic differential equations. Some related issues such as a minimum variance portfolio and a mutual fund theorem are also addressed. All the results are markedly different from those in the problem with deterministic exit time. A key part of my analysis involves proving the global solvability of a stochastic Riccati equation, which is interesting in its own right.

  12. 9 CFR 313.1 - Livestock pens, driveways and ramps.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 9 Animals and Animal Products 2 2010-01-01 2010-01-01 false Livestock pens, driveways and ramps... INSPECTION AND CERTIFICATION HUMANE SLAUGHTER OF LIVESTOCK § 313.1 Livestock pens, driveways and ramps. (a) Livestock pens, driveways and ramps shall be maintained in good repair. They shall be free from sharp...

  13. 43 CFR 3815.7 - Mining claims subject to stock driveway withdrawals.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 43 Public Lands: Interior 2 2012-10-01 2012-10-01 false Mining claims subject to stock driveway... SUBJECT TO LOCATION Mineral Locations in Stock Driveway Withdrawals § 3815.7 Mining claims subject to stock driveway withdrawals. Mining claims on lands within stock driveway withdrawals, located prior...

  14. 43 CFR 3815.7 - Mining claims subject to stock driveway withdrawals.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 43 Public Lands: Interior 2 2013-10-01 2013-10-01 false Mining claims subject to stock driveway... SUBJECT TO LOCATION Mineral Locations in Stock Driveway Withdrawals § 3815.7 Mining claims subject to stock driveway withdrawals. Mining claims on lands within stock driveway withdrawals, located prior...

  15. 43 CFR 3815.7 - Mining claims subject to stock driveway withdrawals.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 43 Public Lands: Interior 2 2011-10-01 2011-10-01 false Mining claims subject to stock driveway... SUBJECT TO LOCATION Mineral Locations in Stock Driveway Withdrawals § 3815.7 Mining claims subject to stock driveway withdrawals. Mining claims on lands within stock driveway withdrawals, located prior...

  16. Photocopy of original blackandwhite silver gelatin print, TWELFTH STREET DRIVEWAY ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    Photocopy of original black-and-white silver gelatin print, TWELFTH STREET DRIVEWAY ENTRANCE, August 31, 1929, photographer Commercial Photo Company - Internal Revenue Service Headquarters Building, 1111 Constitution Avenue Northwest, Washington, District of Columbia, DC

  17. FACILITY 89. FRONT OBLIQUE TAKEN FROM DRIVEWAY. VIEW FACING NORTHEAST. ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    FACILITY 89. FRONT OBLIQUE TAKEN FROM DRIVEWAY. VIEW FACING NORTHEAST. - U.S. Naval Base, Pearl Harbor, Naval Housing Area Makalapa, Junior Officers' Quarters Type K, Makin Place, & Halawa, Makalapa, & Midway Drives, Pearl City, Honolulu County, HI

  18. 7. View of south court and driveway toward main entrance; ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    7. View of south court and driveway toward main entrance; and parts of north and south wings of main building; facing east. - Mission Motel, South Court, 9235 MacArthur Boulevard, Oakland, Alameda County, CA

  19. 7. ELEVATION OF STREET (NORTH) FACADE FROM DRIVEWAY OF LOWELL'S ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    7. ELEVATION OF STREET (NORTH) FACADE FROM DRIVEWAY OF LOWELL'S FORMER RESIDENCE. NOTE BUILDERS VERTICALLY ALIGNED STEM OF BOATS WITH CORNER OF HOUSE BEHIND CAMERA POSITION. - Lowell's Boat Shop, 459 Main Street, Amesbury, Essex County, MA

  20. 2. View from the mansion formal entrance driveway toward the ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    2. View from the mansion formal entrance driveway toward the big meadow at the Billings Farm & Museum. The driveway is flanked by granite gateposts surmounted by wrought iron urn lamps. The view includes a manicured hemlock hedge (Tsuga canadensis) retained by a stone wall at left, and white birch (Betula species) under-planted with ferns at center. - Marsh-Billings-Rockefeller National Historical Park, 54 Elm Street, Woodstock, Windsor County, VT

  1. Risk-Averse Multi-Armed Bandit Problems Under Mean-Variance Measure

    NASA Astrophysics Data System (ADS)

    Vakili, Sattar; Zhao, Qing

    2016-09-01

    The multi-armed bandit problems have been studied mainly under the measure of expected total reward accrued over a horizon of length $T$. In this paper, we address the issue of risk in multi-armed bandit problems and develop parallel results under the measure of mean-variance, a commonly adopted risk measure in economics and mathematical finance. We show that the model-specific regret and the model-independent regret in terms of the mean-variance of the reward process are lower bounded by $\\Omega(\\log T)$ and $\\Omega(T^{2/3})$, respectively. We then show that variations of the UCB policy and the DSEE policy developed for the classic risk-neutral MAB achieve these lower bounds.

  2. Robust Programming Problems Based on the Mean-Variance Model Including Uncertainty Factors

    NASA Astrophysics Data System (ADS)

    Hasuike, Takashi; Ishii, Hiroaki

    2009-01-01

    This paper considers robust programming problems based on the mean-variance model including uncertainty sets and fuzzy factors. Since these problems are not well-defined problems due to fuzzy factors, it is hard to solve them directly. Therefore, introducing chance constraints, fuzzy goals and possibility measures, the proposed models are transformed into the deterministic equivalent problems. Furthermore, in order to solve these equivalent problems efficiently, the solution method is constructed introducing the mean-absolute deviation and doing the equivalent transformations.

  3. R package MVR for Joint Adaptive Mean-Variance Regularization and Variance Stabilization

    PubMed Central

    Dazard, Jean-Eudes; Xu, Hua; Rao, J. Sunil

    2015-01-01

    We present an implementation in the R language for statistical computing of our recent non-parametric joint adaptive mean-variance regularization and variance stabilization procedure. The method is specifically suited for handling difficult problems posed by high-dimensional multivariate datasets (p ≫ n paradigm), such as in ‘omics’-type data, among which are that the variance is often a function of the mean, variable-specific estimators of variances are not reliable, and tests statistics have low powers due to a lack of degrees of freedom. The implementation offers a complete set of features including: (i) normalization and/or variance stabilization function, (ii) computation of mean-variance-regularized t and F statistics, (iii) generation of diverse diagnostic plots, (iv) synthetic and real ‘omics’ test datasets, (v) computationally efficient implementation, using C interfacing, and an option for parallel computing, (vi) manual and documentation on how to setup a cluster. To make each feature as user-friendly as possible, only one subroutine per functionality is to be handled by the end-user. It is available as an R package, called MVR (‘Mean-Variance Regularization’), downloadable from the CRAN. PMID:26819572

  4. Measuring kinetics of complex single ion channel data using mean-variance histograms.

    PubMed Central

    Patlak, J B

    1993-01-01

    The measurement of single ion channel kinetics is difficult when those channels exhibit subconductance events. When the kinetics are fast, and when the current magnitudes are small, as is the case for Na+, Ca2+, and some K+ channels, these difficulties can lead to serious errors in the estimation of channel kinetics. I present here a method, based on the construction and analysis of mean-variance histograms, that can overcome these problems. A mean-variance histogram is constructed by calculating the mean current and the current variance within a brief "window" (a set of N consecutive data samples) superimposed on the digitized raw channel data. Systematic movement of this window over the data produces large numbers of mean-variance pairs which can be assembled into a two-dimensional histogram. Defined current levels (open, closed, or sublevel) appear in such plots as low variance regions. The total number of events in such low variance regions is estimated by curve fitting and plotted as a function of window width. This function decreases with the same time constants as the original dwell time probability distribution for each of the regions. The method can therefore be used: 1) to present a qualitative summary of the single channel data from which the signal-to-noise ratio, open channel noise, steadiness of the baseline, and number of conductance levels can be quickly determined; 2) to quantify the dwell time distribution in each of the levels exhibited. In this paper I present the analysis of a Na+ channel recording that had a number of complexities. The signal-to-noise ratio was only about 8 for the main open state, open channel noise, and fast flickers to other states were present, as were a substantial number of subconductance states. "Standard" half-amplitude threshold analysis of these data produce open and closed time histograms that were well fitted by the sum of two exponentials, but with apparently erroneous time constants, whereas the mean-variance

  5. 2. GENERAL VIEW: MAIN DRIVEWAY: CORD CABIN IS TO THE ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    2. GENERAL VIEW: MAIN DRIVEWAY: CORD CABIN IS TO THE RIGHT OF KIOSK THE FAGEOL CABIN IS IN THE BACKGROUND. - Camp Richardson Resort, Cord Cabin, U.S. Highway 89, 3 miles west of State Highway 50 & 89, South Lake Tahoe, El Dorado County, CA

  6. 9 CFR 313.1 - Livestock pens, driveways and ramps.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 9 Animals and Animal Products 2 2013-01-01 2013-01-01 false Livestock pens, driveways and ramps. 313.1 Section 313.1 Animals and Animal Products FOOD SAFETY AND INSPECTION SERVICE, DEPARTMENT OF AGRICULTURE AGENCY ORGANIZATION AND TERMINOLOGY; MANDATORY MEAT AND POULTRY PRODUCTS INSPECTION AND...

  7. 9 CFR 313.1 - Livestock pens, driveways and ramps.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 9 Animals and Animal Products 2 2014-01-01 2014-01-01 false Livestock pens, driveways and ramps. 313.1 Section 313.1 Animals and Animal Products FOOD SAFETY AND INSPECTION SERVICE, DEPARTMENT OF AGRICULTURE AGENCY ORGANIZATION AND TERMINOLOGY; MANDATORY MEAT AND POULTRY PRODUCTS INSPECTION AND...

  8. 9 CFR 313.1 - Livestock pens, driveways and ramps.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 9 Animals and Animal Products 2 2011-01-01 2011-01-01 false Livestock pens, driveways and ramps. 313.1 Section 313.1 Animals and Animal Products FOOD SAFETY AND INSPECTION SERVICE, DEPARTMENT OF AGRICULTURE AGENCY ORGANIZATION AND TERMINOLOGY; MANDATORY MEAT AND POULTRY PRODUCTS INSPECTION AND...

  9. 9 CFR 313.1 - Livestock pens, driveways and ramps.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 9 Animals and Animal Products 2 2012-01-01 2012-01-01 false Livestock pens, driveways and ramps. 313.1 Section 313.1 Animals and Animal Products FOOD SAFETY AND INSPECTION SERVICE, DEPARTMENT OF AGRICULTURE AGENCY ORGANIZATION AND TERMINOLOGY; MANDATORY MEAT AND POULTRY PRODUCTS INSPECTION AND VOLUNTARY INSPECTION AND CERTIFICATION...

  10. Full-Depth Asphalt Pavements for Parking Lots and Driveways.

    ERIC Educational Resources Information Center

    Asphalt Inst., College Park, MD.

    The latest information for designing full-depth asphalt pavements for parking lots and driveways is covered in relationship to the continued increase in vehicle registration. It is based on The Asphalt Institute's Thickness Design Manual, Series No. 1 (MS-1), Seventh Edition, which covers all aspects of asphalt pavement thickness design in detail,…

  11. THE KINETH HOUSE FROM THE CENTRAL DRIVEWAY, LOOKING NORTHEAST. The ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    THE KINETH HOUSE FROM THE CENTRAL DRIVEWAY, LOOKING NORTHEAST. The two-story vernacular farm house, with its early addition to the east, faces Highway 20. Its yard is lined with split rail fence. - Kineth Farm, House, 19162 STATE ROUTE 20, Coupeville, Island County, WA

  12. Nondestructive probing of means, variances, and correlations of ultracold-atomic-system densities via qubit impurities

    NASA Astrophysics Data System (ADS)

    Elliott, T. J.; Johnson, T. H.

    2016-04-01

    We show how impurity atoms can measure moments of ultracold-atomic-gas densities, using the example of bosons in a one-dimensional lattice. This builds on a body of work regarding the probing of systems by measuring the dephasing of an immersed qubit. We show that this dephasing is captured by a function resembling characteristic functions of probability theory, of which the derivatives at short times reveal moments of the system operator to which the qubit couples. For a qubit formed by an impurity atom, in a system of ultracold atoms, this operator can be the density of the system at the location of the impurity, and thus means, variances, and correlations of the atomic densities are accessible.

  13. Quantifying Systemic Risk by Solutions of the Mean-Variance Risk Model

    PubMed Central

    Morgenstern, Ingo

    2016-01-01

    The world is still recovering from the financial crisis peaking in September 2008. The triggering event was the bankruptcy of Lehman Brothers. To detect such turmoils, one can investigate the time-dependent behaviour of correlations between assets or indices. These cross-correlations have been connected to the systemic risks within markets by several studies in the aftermath of this crisis. We study 37 different US indices which cover almost all aspects of the US economy and show that monitoring an average investor’s behaviour can be used to quantify times of increased risk. In this paper the overall investing strategy is approximated by the ground-states of the mean-variance model along the efficient frontier bound to real world constraints. Changes in the behaviour of the average investor is utlilized as a early warning sign. PMID:27351482

  14. Quantifying Systemic Risk by Solutions of the Mean-Variance Risk Model.

    PubMed

    Jurczyk, Jan; Eckrot, Alexander; Morgenstern, Ingo

    2016-01-01

    The world is still recovering from the financial crisis peaking in September 2008. The triggering event was the bankruptcy of Lehman Brothers. To detect such turmoils, one can investigate the time-dependent behaviour of correlations between assets or indices. These cross-correlations have been connected to the systemic risks within markets by several studies in the aftermath of this crisis. We study 37 different US indices which cover almost all aspects of the US economy and show that monitoring an average investor's behaviour can be used to quantify times of increased risk. In this paper the overall investing strategy is approximated by the ground-states of the mean-variance model along the efficient frontier bound to real world constraints. Changes in the behaviour of the average investor is utlilized as a early warning sign. PMID:27351482

  15. Research on regularized mean-variance portfolio selection strategy with modified Roy safety-first principle.

    PubMed

    Atta Mills, Ebenezer Fiifi Emire; Yan, Dawen; Yu, Bo; Wei, Xinyuan

    2016-01-01

    We propose a consolidated risk measure based on variance and the safety-first principle in a mean-risk portfolio optimization framework. The safety-first principle to financial portfolio selection strategy is modified and improved. Our proposed models are subjected to norm regularization to seek near-optimal stable and sparse portfolios. We compare the cumulative wealth of our preferred proposed model to a benchmark, S&P 500 index for the same period. Our proposed portfolio strategies have better out-of-sample performance than the selected alternative portfolio rules in literature and control the downside risk of the portfolio returns. PMID:27386363

  16. Self-Averaging Property of Minimal Investment Risk of Mean-Variance Model

    PubMed Central

    Shinzato, Takashi

    2015-01-01

    In portfolio optimization problems, the minimum expected investment risk is not always smaller than the expected minimal investment risk. That is, using a well-known approach from operations research, it is possible to derive a strategy that minimizes the expected investment risk, but this strategy does not always result in the best rate of return on assets. Prior to making investment decisions, it is important to an investor to know the potential minimal investment risk (or the expected minimal investment risk) and to determine the strategy that will maximize the return on assets. We use the self-averaging property to analyze the potential minimal investment risk and the concentrated investment level for the strategy that gives the best rate of return. We compare the results from our method with the results obtained by the operations research approach and with those obtained by a numerical simulation using the optimal portfolio. The results of our method and the numerical simulation are in agreement, but they differ from that of the operations research approach. PMID:26225761

  17. 9 CFR 355.15 - Inedible material operating and storage rooms; outer premises, docks, driveways, etc.; fly...

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... storage rooms; outer premises, docks, driveways, etc.; fly-breeding material; nuisances. 355.15 Section....15 Inedible material operating and storage rooms; outer premises, docks, driveways, etc.; fly... any material in which flies may breed, or the maintenance of any nuisance on the premises shall not...

  18. 9 CFR 355.15 - Inedible material operating and storage rooms; outer premises, docks, driveways, etc.; fly...

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... storage rooms; outer premises, docks, driveways, etc.; fly-breeding material; nuisances. 355.15 Section....15 Inedible material operating and storage rooms; outer premises, docks, driveways, etc.; fly... any material in which flies may breed, or the maintenance of any nuisance on the premises shall not...

  19. DRAWING R100131, COMPANY OFFICERS' AREA, BUILDING LOCATIONS, DRIVEWAYS, AND SIDEWALKS, ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    DRAWING R-1001-31, COMPANY OFFICERS' AREA, BUILDING LOCATIONS, DRIVEWAYS, AND SIDEWALKS, LAS LOMAS AND BUENA VISTA DRIVES. Ink on linen, signed by H.B. Nurse. Date has been erased, but probably June 15, 1933. Also marked "PWC 104288." - Hamilton Field, East of Nave Drive, Novato, Marin County, CA

  20. DRAWING R100132, FIELD OFFICERS' AREA, BUILDING LOCATIONS, DRIVEWAYS, AND SIDEWALKS, ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    DRAWING R-1001-32, FIELD OFFICERS' AREA, BUILDING LOCATIONS, DRIVEWAYS, AND SIDEWALKS, SOUTH CIRCLE, CASA GRANDE REAL, AND SEQUOIA DRIVES. Ink on linen, signed by H.B. Nurse. Date has been erased, but probably June 15, 1933. Also marked "PWC 104289." - Hamilton Field, East of Nave Drive, Novato, Marin County, CA

  1. Risk modelling in portfolio optimization

    NASA Astrophysics Data System (ADS)

    Lam, W. H.; Jaaman, Saiful Hafizah Hj.; Isa, Zaidi

    2013-09-01

    Risk management is very important in portfolio optimization. The mean-variance model has been used in portfolio optimization to minimize the investment risk. The objective of the mean-variance model is to minimize the portfolio risk and achieve the target rate of return. Variance is used as risk measure in the mean-variance model. The purpose of this study is to compare the portfolio composition as well as performance between the optimal portfolio of mean-variance model and equally weighted portfolio. Equally weighted portfolio means the proportions that are invested in each asset are equal. The results show that the portfolio composition of the mean-variance optimal portfolio and equally weighted portfolio are different. Besides that, the mean-variance optimal portfolio gives better performance because it gives higher performance ratio than the equally weighted portfolio.

  2. 9 CFR 309.7 - Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 9 Animals and Animal Products 2 2010-01-01 2010-01-01 false Livestock affected with anthrax... INSPECTION § 309.7 Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways. (a) Any livestock found on ante-mortem inspection to be affected with anthrax shall be...

  3. 9 CFR 309.7 - Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 9 Animals and Animal Products 2 2012-01-01 2012-01-01 false Livestock affected with anthrax... INSPECTION § 309.7 Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways. (a) Any livestock found on ante-mortem inspection to be affected with anthrax shall be...

  4. 9 CFR 309.7 - Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 9 Animals and Animal Products 2 2011-01-01 2011-01-01 false Livestock affected with anthrax... INSPECTION § 309.7 Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways. (a) Any livestock found on ante-mortem inspection to be affected with anthrax shall be...

  5. 9 CFR 309.7 - Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 9 Animals and Animal Products 2 2014-01-01 2014-01-01 false Livestock affected with anthrax... INSPECTION § 309.7 Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways. (a) Any livestock found on ante-mortem inspection to be affected with anthrax shall be...

  6. 9 CFR 309.7 - Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 9 Animals and Animal Products 2 2013-01-01 2013-01-01 false Livestock affected with anthrax... INSPECTION § 309.7 Livestock affected with anthrax; cleaning and disinfection of infected livestock pens and driveways. (a) Any livestock found on ante-mortem inspection to be affected with anthrax shall be...

  7. 9 CFR 355.15 - Inedible material operating and storage rooms; outer premises, docks, driveways, etc.; fly...

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... storage rooms; outer premises, docks, driveways, etc.; fly-breeding material; nuisances. 355.15 Section...-breeding material; nuisances. All operating and storage rooms and departments of inspected plants used for... any material in which flies may breed, or the maintenance of any nuisance on the premises shall not...

  8. 9 CFR 355.15 - Inedible material operating and storage rooms; outer premises, docks, driveways, etc.; fly...

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... storage rooms; outer premises, docks, driveways, etc.; fly-breeding material; nuisances. 355.15 Section...-breeding material; nuisances. All operating and storage rooms and departments of inspected plants used for... any material in which flies may breed, or the maintenance of any nuisance on the premises shall not...

  9. 9 CFR 355.15 - Inedible material operating and storage rooms; outer premises, docks, driveways, etc.; fly...

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... storage rooms; outer premises, docks, driveways, etc.; fly-breeding material; nuisances. 355.15 Section...-breeding material; nuisances. All operating and storage rooms and departments of inspected plants used for... any material in which flies may breed, or the maintenance of any nuisance on the premises shall not...

  10. Portfolio optimization with skewness and kurtosis

    NASA Astrophysics Data System (ADS)

    Lam, Weng Hoe; Jaaman, Saiful Hafizah Hj.; Isa, Zaidi

    2013-04-01

    Mean and variance of return distributions are two important parameters of the mean-variance model in portfolio optimization. However, the mean-variance model will become inadequate if the returns of assets are not normally distributed. Therefore, higher moments such as skewness and kurtosis cannot be ignored. Risk averse investors prefer portfolios with high skewness and low kurtosis so that the probability of getting negative rates of return will be reduced. The objective of this study is to compare the portfolio compositions as well as performances between the mean-variance model and mean-variance-skewness-kurtosis model by using the polynomial goal programming approach. The results show that the incorporation of skewness and kurtosis will change the optimal portfolio compositions. The mean-variance-skewness-kurtosis model outperforms the mean-variance model because the mean-variance-skewness-kurtosis model takes skewness and kurtosis into consideration. Therefore, the mean-variance-skewness-kurtosis model is more appropriate for the investors of Malaysia in portfolio optimization.

  11. Portfolio optimization using median-variance approach

    NASA Astrophysics Data System (ADS)

    Wan Mohd, Wan Rosanisah; Mohamad, Daud; Mohamed, Zulkifli

    2013-04-01

    Optimization models have been applied in many decision-making problems particularly in portfolio selection. Since the introduction of Markowitz's theory of portfolio selection, various approaches based on mathematical programming have been introduced such as mean-variance, mean-absolute deviation, mean-variance-skewness and conditional value-at-risk (CVaR) mainly to maximize return and minimize risk. However most of the approaches assume that the distribution of data is normal and this is not generally true. As an alternative, in this paper, we employ the median-variance approach to improve the portfolio optimization. This approach has successfully catered both types of normal and non-normal distribution of data. With this actual representation, we analyze and compare the rate of return and risk between the mean-variance and the median-variance based portfolio which consist of 30 stocks from Bursa Malaysia. The results in this study show that the median-variance approach is capable to produce a lower risk for each return earning as compared to the mean-variance approach.

  12. Algorithms for optimizing CT fluence control

    NASA Astrophysics Data System (ADS)

    Hsieh, Scott S.; Pelc, Norbert J.

    2014-03-01

    The ability to customize the incident x-ray fluence in CT via beam-shaping filters or mA modulation is known to improve image quality and/or reduce radiation dose. Previous work has shown that complete control of x-ray fluence (ray-by-ray fluence modulation) would further improve dose efficiency. While complete control of fluence is not currently possible, emerging concepts such as dynamic attenuators and inverse-geometry CT allow nearly complete control to be realized. Optimally using ray-by-ray fluence modulation requires solving a very high-dimensional optimization problem. Most optimization techniques fail or only provide approximate solutions. We present efficient algorithms for minimizing mean or peak variance given a fixed dose limit. The reductions in variance can easily be translated to reduction in dose, if the original variance met image quality requirements. For mean variance, a closed form solution is derived. The peak variance problem is recast as iterated, weighted mean variance minimization, and at each iteration it is possible to bound the distance to the optimal solution. We apply our algorithms in simulations of scans of the thorax and abdomen. Peak variance reductions of 45% and 65% are demonstrated in the abdomen and thorax, respectively, compared to a bowtie filter alone. Mean variance shows smaller gains (about 15%).

  13. Large deviations and portfolio optimization

    NASA Astrophysics Data System (ADS)

    Sornette, Didier

    Risk control and optimal diversification constitute a major focus in the finance and insurance industries as well as, more or less consciously, in our everyday life. We present a discussion of the characterization of risks and of the optimization of portfolios that starts from a simple illustrative model and ends by a general functional integral formulation. A major item is that risk, usually thought of as one-dimensional in the conventional mean-variance approach, has to be addressed by the full distribution of losses. Furthermore, the time-horizon of the investment is shown to play a major role. We show the importance of accounting for large fluctuations and use the theory of Cramér for large deviations in this context. We first treat a simple model with a single risky asset that exemplifies the distinction between the average return and the typical return and the role of large deviations in multiplicative processes, and the different optimal strategies for the investors depending on their size. We then analyze the case of assets whose price variations are distributed according to exponential laws, a situation that is found to describe daily price variations reasonably well. Several portfolio optimization strategies are presented that aim at controlling large risks. We end by extending the standard mean-variance portfolio optimization theory, first within the quasi-Gaussian approximation and then using a general formulation for non-Gaussian correlated assets in terms of the formalism of functional integrals developed in the field theory of critical phenomena.

  14. Static vs stochastic optimization: A case study of FTSE Bursa Malaysia sectorial indices

    SciTech Connect

    Mamat, Nur Jumaadzan Zaleha; Jaaman, Saiful Hafizah; Ahmad, Rokiah Rozita

    2014-06-19

    Traditional portfolio optimization methods in the likes of Markowitz' mean-variance model and semi-variance model utilize static expected return and volatility risk from historical data to generate an optimal portfolio. The optimal portfolio may not truly be optimal in reality due to the fact that maximum and minimum values from the data may largely influence the expected return and volatility risk values. This paper considers distributions of assets' return and volatility risk to determine a more realistic optimized portfolio. For illustration purposes, the sectorial indices data in FTSE Bursa Malaysia is employed. The results show that stochastic optimization provides more stable information ratio.

  15. Static vs stochastic optimization: A case study of FTSE Bursa Malaysia sectorial indices

    NASA Astrophysics Data System (ADS)

    Mamat, Nur Jumaadzan Zaleha; Jaaman, Saiful Hafizah; Ahmad, Rokiah@Rozita

    2014-06-01

    Traditional portfolio optimization methods in the likes of Markowitz' mean-variance model and semi-variance model utilize static expected return and volatility risk from historical data to generate an optimal portfolio. The optimal portfolio may not truly be optimal in reality due to the fact that maximum and minimum values from the data may largely influence the expected return and volatility risk values. This paper considers distributions of assets' return and volatility risk to determine a more realistic optimized portfolio. For illustration purposes, the sectorial indices data in FTSE Bursa Malaysia is employed. The results show that stochastic optimization provides more stable information ratio.

  16. Inverse Optimization: A New Perspective on the Black-Litterman Model.

    PubMed

    Bertsimas, Dimitris; Gupta, Vishal; Paschalidis, Ioannis Ch

    2012-12-11

    The Black-Litterman (BL) model is a widely used asset allocation model in the financial industry. In this paper, we provide a new perspective. The key insight is to replace the statistical framework in the original approach with ideas from inverse optimization. This insight allows us to significantly expand the scope and applicability of the BL model. We provide a richer formulation that, unlike the original model, is flexible enough to incorporate investor information on volatility and market dynamics. Equally importantly, our approach allows us to move beyond the traditional mean-variance paradigm of the original model and construct "BL"-type estimators for more general notions of risk such as coherent risk measures. Computationally, we introduce and study two new "BL"-type estimators and their corresponding portfolios: a Mean Variance Inverse Optimization (MV-IO) portfolio and a Robust Mean Variance Inverse Optimization (RMV-IO) portfolio. These two approaches are motivated by ideas from arbitrage pricing theory and volatility uncertainty. Using numerical simulation and historical backtesting, we show that both methods often demonstrate a better risk-reward tradeoff than their BL counterparts and are more robust to incorrect investor views. PMID:25382873

  17. Inverse Optimization: A New Perspective on the Black-Litterman Model

    PubMed Central

    Bertsimas, Dimitris; Gupta, Vishal; Paschalidis, Ioannis Ch.

    2014-01-01

    The Black-Litterman (BL) model is a widely used asset allocation model in the financial industry. In this paper, we provide a new perspective. The key insight is to replace the statistical framework in the original approach with ideas from inverse optimization. This insight allows us to significantly expand the scope and applicability of the BL model. We provide a richer formulation that, unlike the original model, is flexible enough to incorporate investor information on volatility and market dynamics. Equally importantly, our approach allows us to move beyond the traditional mean-variance paradigm of the original model and construct “BL”-type estimators for more general notions of risk such as coherent risk measures. Computationally, we introduce and study two new “BL”-type estimators and their corresponding portfolios: a Mean Variance Inverse Optimization (MV-IO) portfolio and a Robust Mean Variance Inverse Optimization (RMV-IO) portfolio. These two approaches are motivated by ideas from arbitrage pricing theory and volatility uncertainty. Using numerical simulation and historical backtesting, we show that both methods often demonstrate a better risk-reward tradeoff than their BL counterparts and are more robust to incorrect investor views. PMID:25382873

  18. Turnover, account value and diversification of real traders: evidence of collective portfolio optimizing behavior

    NASA Astrophysics Data System (ADS)

    Morton de Lachapelle, David; Challet, Damien

    2010-07-01

    Despite the availability of very detailed data on financial markets, agent-based modeling is hindered by the lack of information about real trader behavior. This makes it impossible to validate agent-based models, which are thus reverse-engineering attempts. This work is a contribution towards building a set of stylized facts about the traders themselves. Using the client database of Swissquote Bank SA, the largest online Swiss broker, we find empirical relationships between turnover, account values and the number of assets in which a trader is invested. A theory based on simple mean-variance portfolio optimization that crucially includes variable transaction costs is able to reproduce faithfully the observed behaviors. We finally argue that our results bring to light the collective ability of a population to construct a mean-variance portfolio that takes into account the structure of transaction costs.

  19. Robust Portfolio Optimization Using Pseudodistances

    PubMed Central

    2015-01-01

    The presence of outliers in financial asset returns is a frequently occurring phenomenon which may lead to unreliable mean-variance optimized portfolios. This fact is due to the unbounded influence that outliers can have on the mean returns and covariance estimators that are inputs in the optimization procedure. In this paper we present robust estimators of mean and covariance matrix obtained by minimizing an empirical version of a pseudodistance between the assumed model and the true model underlying the data. We prove and discuss theoretical properties of these estimators, such as affine equivariance, B-robustness, asymptotic normality and asymptotic relative efficiency. These estimators can be easily used in place of the classical estimators, thereby providing robust optimized portfolios. A Monte Carlo simulation study and applications to real data show the advantages of the proposed approach. We study both in-sample and out-of-sample performance of the proposed robust portfolios comparing them with some other portfolios known in literature. PMID:26468948

  20. Dynamics of mean-variance-skewness of cumulative crop yield impact temporal yield variance

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Production risk associated with cropping systems influences farmers’ decisions to adopt a new management practice or a production system. Cumulative yield (CY), temporal yield variance (TYV) and coefficient of variation (CV) were used to assess the risk associated with adopting combinations of new m...

  1. Life history traits and exploitation affect the spatial mean-variance relationship in fish abundance.

    PubMed

    Kuo, Ting-chun; Mandal, Sandip; Yamauchi, Atsushi; Hsieh, Chih-hao

    2016-05-01

    Fishing is expected to alter the spatial heterogeneity of fishes. As an effective index to quantify spatial heterogeneity, the exponent b in Taylor's power law (V = aMb) measures how spatial variance (V) varies with changes in mean abundance (M) of a population, with larger b indicating higher spatial aggregation potential (i.e., more heterogeneity). Theory predicts b is related with life history traits, but empirical evidence is lacking. Using 50-yr spatiotemporal data from the California Current Ecosystem, we examined fishing and life history effects on Taylor's exponent by comparing spatial distributions of exploited and unexploited fishes living in the same environment. We found that unexploited species with smaller size and generation time exhibit larger b, supporting theoretical prediction. In contrast, this relationship in exploited species is much weaker, as the exponents of large exploited species were higher than unexploited species with similar traits. Our results suggest that fishing may increase spatial aggregation potential of a species, likely through degrading their size/age structure. Results of moving-window cross-correlation analyses on b vs. age structure indices (mean age and age evenness) for some exploited species corroborate our findings. Furthermore, through linking our findings to other fundamental ecological patterns (occupancy-abundance and size-abundance relationships), we provide theoretical arguments for the usefulness of monitoring the exponent b for management purposes. We propose that age/size-truncated species might have lower recovery rate in spatial occupancy, and the spatial variance-mass relationship of a species might be non-linear. Our findings provide theoretical basis explaining why fishery management strategy should be concerned with changes to the age and spatial structure of exploited fishes. PMID:27349101

  2. Optimal Solar PV Arrays Integration for Distributed Generation

    SciTech Connect

    Omitaomu, Olufemi A; Li, Xueping

    2012-01-01

    Solar photovoltaic (PV) systems hold great potential for distributed energy generation by installing PV panels on rooftops of residential and commercial buildings. Yet challenges arise along with the variability and non-dispatchability of the PV systems that affect the stability of the grid and the economics of the PV system. This paper investigates the integration of PV arrays for distributed generation applications by identifying a combination of buildings that will maximize solar energy output and minimize system variability. Particularly, we propose mean-variance optimization models to choose suitable rooftops for PV integration based on Markowitz mean-variance portfolio selection model. We further introduce quantity and cardinality constraints to result in a mixed integer quadratic programming problem. Case studies based on real data are presented. An efficient frontier is obtained for sample data that allows decision makers to choose a desired solar energy generation level with a comfortable variability tolerance level. Sensitivity analysis is conducted to show the tradeoffs between solar PV energy generation potential and variability.

  3. Optimal trading strategies—a time series approach

    NASA Astrophysics Data System (ADS)

    Bebbington, Peter A.; Kühn, Reimer

    2016-05-01

    Motivated by recent advances in the spectral theory of auto-covariance matrices, we are led to revisit a reformulation of Markowitz’ mean-variance portfolio optimization approach in the time domain. In its simplest incarnation it applies to a single traded asset and allows an optimal trading strategy to be found which—for a given return—is minimally exposed to market price fluctuations. The model is initially investigated for a range of synthetic price processes, taken to be either second order stationary, or to exhibit second order stationary increments. Attention is paid to consequences of estimating auto-covariance matrices from small finite samples, and auto-covariance matrix cleaning strategies to mitigate against these are investigated. Finally we apply our framework to real world data.

  4. Optimal portfolio strategy with cross-correlation matrix composed by DCCA coefficients: Evidence from the Chinese stock market

    NASA Astrophysics Data System (ADS)

    Sun, Xuelian; Liu, Zixian

    2016-02-01

    In this paper, a new estimator of correlation matrix is proposed, which is composed of the detrended cross-correlation coefficients (DCCA coefficients), to improve portfolio optimization. In contrast to Pearson's correlation coefficients (PCC), DCCA coefficients acquired by the detrended cross-correlation analysis (DCCA) method can describe the nonlinear correlation between assets, and can be decomposed in different time scales. These properties of DCCA make it possible to improve the investment effect and more valuable to investigate the scale behaviors of portfolios. The minimum variance portfolio (MVP) model and the Mean-Variance (MV) model are used to evaluate the effectiveness of this improvement. Stability analysis shows the effect of two kinds of correlation matrices on the estimation error of portfolio weights. The observed scale behaviors are significant to risk management and could be used to optimize the portfolio selection.

  5. Dynamic Optimization

    NASA Technical Reports Server (NTRS)

    Laird, Philip

    1992-01-01

    We distinguish static and dynamic optimization of programs: whereas static optimization modifies a program before runtime and is based only on its syntactical structure, dynamic optimization is based on the statistical properties of the input source and examples of program execution. Explanation-based generalization is a commonly used dynamic optimization method, but its effectiveness as a speedup-learning method is limited, in part because it fails to separate the learning process from the program transformation process. This paper describes a dynamic optimization technique called a learn-optimize cycle that first uses a learning element to uncover predictable patterns in the program execution and then uses an optimization algorithm to map these patterns into beneficial transformations. The technique has been used successfully for dynamic optimization of pure Prolog.

  6. Information Extraction of High Resolution Remote Sensing Images Based on the Calculation of Optimal Segmentation Parameters

    PubMed Central

    Zhu, Hongchun; Cai, Lijie; Liu, Haiying; Huang, Wei

    2016-01-01

    Multi-scale image segmentation and the selection of optimal segmentation parameters are the key processes in the object-oriented information extraction of high-resolution remote sensing images. The accuracy of remote sensing special subject information depends on this extraction. On the basis of WorldView-2 high-resolution data, the optimal segmentation parameters methodof object-oriented image segmentation and high-resolution image information extraction, the following processes were conducted in this study. Firstly, the best combination of the bands and weights was determined for the information extraction of high-resolution remote sensing image. An improved weighted mean-variance method was proposed andused to calculatethe optimal segmentation scale. Thereafter, the best shape factor parameter and compact factor parameters were computed with the use of the control variables and the combination of the heterogeneity and homogeneity indexes. Different types of image segmentation parameters were obtained according to the surface features. The high-resolution remote sensing images were multi-scale segmented with the optimal segmentation parameters. Ahierarchical network structure was established by setting the information extraction rules to achieve object-oriented information extraction. This study presents an effective and practical method that can explain expert input judgment by reproducible quantitative measurements. Furthermore the results of this procedure may be incorporated into a classification scheme. PMID:27362762

  7. Information Extraction of High Resolution Remote Sensing Images Based on the Calculation of Optimal Segmentation Parameters.

    PubMed

    Zhu, Hongchun; Cai, Lijie; Liu, Haiying; Huang, Wei

    2016-01-01

    Multi-scale image segmentation and the selection of optimal segmentation parameters are the key processes in the object-oriented information extraction of high-resolution remote sensing images. The accuracy of remote sensing special subject information depends on this extraction. On the basis of WorldView-2 high-resolution data, the optimal segmentation parameters methodof object-oriented image segmentation and high-resolution image information extraction, the following processes were conducted in this study. Firstly, the best combination of the bands and weights was determined for the information extraction of high-resolution remote sensing image. An improved weighted mean-variance method was proposed andused to calculatethe optimal segmentation scale. Thereafter, the best shape factor parameter and compact factor parameters were computed with the use of the control variables and the combination of the heterogeneity and homogeneity indexes. Different types of image segmentation parameters were obtained according to the surface features. The high-resolution remote sensing images were multi-scale segmented with the optimal segmentation parameters. Ahierarchical network structure was established by setting the information extraction rules to achieve object-oriented information extraction. This study presents an effective and practical method that can explain expert input judgment by reproducible quantitative measurements. Furthermore the results of this procedure may be incorporated into a classification scheme. PMID:27362762

  8. Dispositional Optimism

    PubMed Central

    Carver, Charles S.; Scheier, Michael F.

    2014-01-01

    Optimism is a cognitive construct (expectancies regarding future outcomes) that also relates to motivation: optimistic people exert effort, whereas pessimistic people disengage from effort. Study of optimism began largely in health contexts, finding positive associations between optimism and markers of better psychological and physical health. Physical health effects likely occur through differences in both health-promoting behaviors and physiological concomitants of coping. Recently, the scientific study of optimism has extended to the realm of social relations: new evidence indicates that optimists have better social connections, partly because they work harder at them. In this review, we examine the myriad ways this trait can benefit an individual, and our current understanding of the biological basis of optimism. PMID:24630971

  9. Distributed Optimization

    NASA Technical Reports Server (NTRS)

    Macready, William; Wolpert, David

    2005-01-01

    We demonstrate a new framework for analyzing and controlling distributed systems, by solving constrained optimization problems with an algorithm based on that framework. The framework is ar. information-theoretic extension of conventional full-rationality game theory to allow bounded rational agents. The associated optimization algorithm is a game in which agents control the variables of the optimization problem. They do this by jointly minimizing a Lagrangian of (the probability distribution of) their joint state. The updating of the Lagrange parameters in that Lagrangian is a form of automated annealing, one that focuses the multi-agent system on the optimal pure strategy. We present computer experiments for the k-sat constraint satisfaction problem and for unconstrained minimization of NK functions.

  10. Prospective Optimization

    PubMed Central

    Sejnowski, Terrence J.; Poizner, Howard; Lynch, Gary; Gepshtein, Sergei; Greenspan, Ralph J.

    2014-01-01

    Human performance approaches that of an ideal observer and optimal actor in some perceptual and motor tasks. These optimal abilities depend on the capacity of the cerebral cortex to store an immense amount of information and to flexibly make rapid decisions. However, behavior only approaches these limits after a long period of learning while the cerebral cortex interacts with the basal ganglia, an ancient part of the vertebrate brain that is responsible for learning sequences of actions directed toward achieving goals. Progress has been made in understanding the algorithms used by the brain during reinforcement learning, which is an online approximation of dynamic programming. Humans also make plans that depend on past experience by simulating different scenarios, which is called prospective optimization. The same brain structures in the cortex and basal ganglia that are active online during optimal behavior are also active offline during prospective optimization. The emergence of general principles and algorithms for goal-directed behavior has consequences for the development of autonomous devices in engineering applications. PMID:25328167

  11. Optimal Fluoridation

    PubMed Central

    Lee, John R.

    1975-01-01

    Optimal fluoridation has been defined as that fluoride exposure which confers maximal cariostasis with minimal toxicity and its values have been previously determined to be 0.5 to 1 mg per day for infants and 1 to 1.5 mg per day for an average child. Total fluoride ingestion and urine excretion were studied in Marin County, California, children in 1973 before municipal water fluoridation. Results showed fluoride exposure to be higher than anticipated and fulfilled previously accepted criteria for optimal fluoridation. Present and future water fluoridation plans need to be reevaluated in light of total environmental fluoride exposure. PMID:1130041

  12. Gear optimization

    NASA Technical Reports Server (NTRS)

    Vanderplaats, G. N.; Chen, Xiang; Zhang, Ning-Tian

    1988-01-01

    The use of formal numerical optimization methods for the design of gears is investigated. To achieve this, computer codes were developed for the analysis of spur gears and spiral bevel gears. These codes calculate the life, dynamic load, bending strength, surface durability, gear weight and size, and various geometric parameters. It is necessary to calculate all such important responses because they all represent competing requirements in the design process. The codes developed here were written in subroutine form and coupled to the COPES/ADS general purpose optimization program. This code allows the user to define the optimization problem at the time of program execution. Typical design variables include face width, number of teeth and diametral pitch. The user is free to choose any calculated response as the design objective to minimize or maximize and may impose lower and upper bounds on any calculated responses. Typical examples include life maximization with limits on dynamic load, stress, weight, etc. or minimization of weight subject to limits on life, dynamic load, etc. The research codes were written in modular form for easy expansion and so that they could be combined to create a multiple reduction optimization capability in future.

  13. Neutralizer optimization

    NASA Technical Reports Server (NTRS)

    Patterson, Michael J.; Mohajeri, Kayhan

    1991-01-01

    The preliminary results of a test program to optimize a neutralizer design for 30 cm xenon ion thrusters are discussed. The impact of neutralizer geometry, neutralizer axial location, and local magnetic fields on neutralizer performance is discussed. The effect of neutralizer performance on overall thruster performance is quantified, for thruster operation in the 0.5-3.2 kW power range. Additionally, these data are compared to data published for other north-south stationkeeping (NSSK) and primary propulsion xenon ion thruster neutralizers.

  14. [SIAM conference on optimization

    SciTech Connect

    Not Available

    1992-05-10

    Abstracts are presented of 63 papers on the following topics: large-scale optimization, interior-point methods, algorithms for optimization, problems in control, network optimization methods, and parallel algorithms for optimization problems.

  15. Optimal refrigerator

    NASA Astrophysics Data System (ADS)

    Allahverdyan, Armen E.; Hovhannisyan, Karen; Mahler, Guenter

    2010-05-01

    We study a refrigerator model which consists of two n -level systems interacting via a pulsed external field. Each system couples to its own thermal bath at temperatures Th and Tc , respectively (θ≡Tc/Th<1) . The refrigerator functions in two steps: thermally isolated interaction between the systems driven by the external field and isothermal relaxation back to equilibrium. There is a complementarity between the power of heat transfer from the cold bath and the efficiency: the latter nullifies when the former is maximized and vice versa. A reasonable compromise is achieved by optimizing the product of the heat-power and efficiency over the Hamiltonian of the two systems. The efficiency is then found to be bounded from below by ζCA=(1)/(1-θ)-1 (an analog of the Curzon-Ahlborn efficiency), besides being bound from above by the Carnot efficiency ζC=(1)/(1-θ)-1 . The lower bound is reached in the equilibrium limit θ→1 . The Carnot bound is reached (for a finite power and a finite amount of heat transferred per cycle) for lnn≫1 . If the above maximization is constrained by assuming homogeneous energy spectra for both systems, the efficiency is bounded from above by ζCA and converges to it for n≫1 .

  16. Optimal refrigerator.

    PubMed

    Allahverdyan, Armen E; Hovhannisyan, Karen; Mahler, Guenter

    2010-05-01

    We study a refrigerator model which consists of two n -level systems interacting via a pulsed external field. Each system couples to its own thermal bath at temperatures T h and T c, respectively (θ ≡ T c/T h < 1). The refrigerator functions in two steps: thermally isolated interaction between the systems driven by the external field and isothermal relaxation back to equilibrium. There is a complementarity between the power of heat transfer from the cold bath and the efficiency: the latter nullifies when the former is maximized and vice versa. A reasonable compromise is achieved by optimizing the product of the heat-power and efficiency over the Hamiltonian of the two systems. The efficiency is then found to be bounded from below by [formula: see text] (an analog of the Curzon-Ahlborn efficiency), besides being bound from above by the Carnot efficiency [formula: see text]. The lower bound is reached in the equilibrium limit θ → 1. The Carnot bound is reached (for a finite power and a finite amount of heat transferred per cycle) for ln n > 1. If the above maximization is constrained by assuming homogeneous energy spectra for both systems, the efficiency is bounded from above by ζ CA and converges to it for n > 1. PMID:20866207

  17. RECOVERY ACT - Robust Optimization for Connectivity and Flows in Dynamic Complex Networks

    SciTech Connect

    Balasundaram, Balabhaskar; Butenko, Sergiy; Boginski, Vladimir; Uryasev, Stan

    2013-12-25

    to capture uncertainty and risk using appropriate probabilistic, statistical and optimization concepts. The main difficulty arising in addressing these issues is the dramatic increase in the computational complexity of the resulting optimization problems. This project studied novel models and methodologies for risk-averse network optimization- specifically, network design, network flows and cluster detection problems under uncertainty. The approach taken was to incorporate a quantitative risk measure known as conditional value-at-risk that is widely used in financial applications. This approach presents a viable alternate modeling and optimization framework to chance-constrained optimization and mean-variance optimization, one that also facilitates the detection of risk-averse solutions.

  18. Particle Swarm Optimization Toolbox

    NASA Technical Reports Server (NTRS)

    Grant, Michael J.

    2010-01-01

    The Particle Swarm Optimization Toolbox is a library of evolutionary optimization tools developed in the MATLAB environment. The algorithms contained in the library include a genetic algorithm (GA), a single-objective particle swarm optimizer (SOPSO), and a multi-objective particle swarm optimizer (MOPSO). Development focused on both the SOPSO and MOPSO. A GA was included mainly for comparison purposes, and the particle swarm optimizers appeared to perform better for a wide variety of optimization problems. All algorithms are capable of performing unconstrained and constrained optimization. The particle swarm optimizers are capable of performing single and multi-objective optimization. The SOPSO and MOPSO algorithms are based on swarming theory and bird-flocking patterns to search the trade space for the optimal solution or optimal trade in competing objectives. The MOPSO generates Pareto fronts for objectives that are in competition. A GA, based on Darwin evolutionary theory, is also included in the library. The GA consists of individuals that form a population in the design space. The population mates to form offspring at new locations in the design space. These offspring contain traits from both of the parents. The algorithm is based on this combination of traits from parents to hopefully provide an improved solution than either of the original parents. As the algorithm progresses, individuals that hold these optimal traits will emerge as the optimal solutions. Due to the generic design of all optimization algorithms, each algorithm interfaces with a user-supplied objective function. This function serves as a "black-box" to the optimizers in which the only purpose of this function is to evaluate solutions provided by the optimizers. Hence, the user-supplied function can be numerical simulations, analytical functions, etc., since the specific detail of this function is of no concern to the optimizer. These algorithms were originally developed to support entry

  19. Aristos Optimization Package

    Energy Science and Technology Software Center (ESTSC)

    2007-03-01

    Aristos is a Trilinos package for nonlinear continuous optimization, based on full-space sequential quadratic programming (SQP) methods. Aristos is specifically designed for the solution of large-scale constrained optimization problems in which the linearized constraint equations require iterative (i.e. inexact) linear solver techniques. Aristos' unique feature is an efficient handling of inexactness in linear system solves. Aristos currently supports the solution of equality-constrained convex and nonconvex optimization problems. It has been used successfully in the areamore » of PDE-constrained optimization, for the solution of nonlinear optimal control, optimal design, and inverse problems.« less

  20. Multidisciplinary Optimization for Aerospace Using Genetic Optimization

    NASA Technical Reports Server (NTRS)

    Pak, Chan-gi; Hahn, Edward E.; Herrera, Claudia Y.

    2007-01-01

    In support of the ARMD guidelines NASA's Dryden Flight Research Center is developing a multidisciplinary design and optimization tool This tool will leverage existing tools and practices, and allow the easy integration and adoption of new state-of-the-art software. Optimization has made its way into many mainstream applications. For example NASTRAN(TradeMark) has its solution sequence 200 for Design Optimization, and MATLAB(TradeMark) has an Optimization Tool box. Other packages, such as ZAERO(TradeMark) aeroelastic panel code and the CFL3D(TradeMark) Navier-Stokes solver have no built in optimizer. The goal of the tool development is to generate a central executive capable of using disparate software packages ina cross platform network environment so as to quickly perform optimization and design tasks in a cohesive streamlined manner. A provided figure (Figure 1) shows a typical set of tools and their relation to the central executive. Optimization can take place within each individual too, or in a loop between the executive and the tool, or both.

  1. Optimization of aerospace structures

    NASA Technical Reports Server (NTRS)

    Keith, Theo G., Jr.; Patnaik, Surya N.

    1994-01-01

    Research carried out is grouped under two topics: (1) Design Optimization, and (2) Integrated Force Method of Analysis. Design Optimization Research Topics are singularity alleviation enhances structural optimization methods, computer based design capability extended through substructure synthesis, and optimality criteria provides optimum design for a select class of structural problems. Integrated Force Method of Analysis Research Topics are boundary compatibility formulation improves stress analysis of shell structures. Brief descriptions of the four topics are appended.

  2. Cyclone performance and optimization

    SciTech Connect

    Leith, D.

    1990-12-15

    An empirical model for predicting pressure drop across a cyclone, developed by Dirgo (1988), is presented. The model was developed through a statistical analysis of pressure drop data for 98 cyclone designs. This model is used with the efficiency model of Iozia and Leith (1990) to develop an optimization curve which predicts the minimum pressure drop on the dimension ratios of the optimized cyclone for a given aerodynamic cut diameter, d{sub 50}. The effect of variation in cyclone height, cyclone diameter, and flow on the optimization is determined. The optimization results are used to develop a design procedure for optimized cyclones. 33 refs., 10 figs., 4 tabs.

  3. Hope, optimism and delusion

    PubMed Central

    McGuire-Snieckus, Rebecca

    2014-01-01

    Optimism is generally accepted by psychiatrists, psychologists and other caring professionals as a feature of mental health. Interventions typically rely on cognitive-behavioural tools to encourage individuals to ‘stop negative thought cycles’ and to ‘challenge unhelpful thoughts’. However, evidence suggests that most individuals have persistent biases of optimism and that excessive optimism is not conducive to mental health. How helpful is it to facilitate optimism in individuals who are likely to exhibit biases of optimism already? By locating the cause of distress at the individual level and ‘unhelpful’ cognitions, does this minimise wider systemic social and economic influences on mental health? PMID:25237497

  4. Integrated controls design optimization

    DOEpatents

    Lou, Xinsheng; Neuschaefer, Carl H.

    2015-09-01

    A control system (207) for optimizing a chemical looping process of a power plant includes an optimizer (420), an income algorithm (230) and a cost algorithm (225) and a chemical looping process models. The process models are used to predict the process outputs from process input variables. Some of the process in puts and output variables are related to the income of the plant; and some others are related to the cost of the plant operations. The income algorithm (230) provides an income input to the optimizer (420) based on a plurality of input parameters (215) of the power plant. The cost algorithm (225) provides a cost input to the optimizer (420) based on a plurality of output parameters (220) of the power plant. The optimizer (420) determines an optimized operating parameter solution based on at least one of the income input and the cost input, and supplies the optimized operating parameter solution to the power plant.

  5. Particle Swarm Optimization

    NASA Technical Reports Server (NTRS)

    Venter, Gerhard; Sobieszczanski-Sobieski Jaroslaw

    2002-01-01

    The purpose of this paper is to show how the search algorithm known as particle swarm optimization performs. Here, particle swarm optimization is applied to structural design problems, but the method has a much wider range of possible applications. The paper's new contributions are improvements to the particle swarm optimization algorithm and conclusions and recommendations as to the utility of the algorithm, Results of numerical experiments for both continuous and discrete applications are presented in the paper. The results indicate that the particle swarm optimization algorithm does locate the constrained minimum design in continuous applications with very good precision, albeit at a much higher computational cost than that of a typical gradient based optimizer. However, the true potential of particle swarm optimization is primarily in applications with discrete and/or discontinuous functions and variables. Additionally, particle swarm optimization has the potential of efficient computation with very large numbers of concurrently operating processors.

  6. Supercomputer optimizations for stochastic optimal control applications

    NASA Technical Reports Server (NTRS)

    Chung, Siu-Leung; Hanson, Floyd B.; Xu, Huihuang

    1991-01-01

    Supercomputer optimizations for a computational method of solving stochastic, multibody, dynamic programming problems are presented. The computational method is valid for a general class of optimal control problems that are nonlinear, multibody dynamical systems, perturbed by general Markov noise in continuous time, i.e., nonsmooth Gaussian as well as jump Poisson random white noise. Optimization techniques for vector multiprocessors or vectorizing supercomputers include advanced data structures, loop restructuring, loop collapsing, blocking, and compiler directives. These advanced computing techniques and superconducting hardware help alleviate Bellman's curse of dimensionality in dynamic programming computations, by permitting the solution of large multibody problems. Possible applications include lumped flight dynamics models for uncertain environments, such as large scale and background random aerospace fluctuations.

  7. Search-based optimization

    NASA Technical Reports Server (NTRS)

    Wheeler, Ward C.

    2003-01-01

    The problem of determining the minimum cost hypothetical ancestral sequences for a given cladogram is known to be NP-complete (Wang and Jiang, 1994). Traditionally, point estimations of hypothetical ancestral sequences have been used to gain heuristic, upper bounds on cladogram cost. These include procedures with such diverse approaches as non-additive optimization of multiple sequence alignment, direct optimization (Wheeler, 1996), and fixed-state character optimization (Wheeler, 1999). A method is proposed here which, by extending fixed-state character optimization, replaces the estimation process with a search. This form of optimization examines a diversity of potential state solutions for cost-efficient hypothetical ancestral sequences and can result in greatly more parsimonious cladograms. Additionally, such an approach can be applied to other NP-complete phylogenetic optimization problems such as genomic break-point analysis. c2003 The Willi Hennig Society. Published by Elsevier Science (USA). All rights reserved.

  8. Search-based optimization.

    PubMed

    Wheeler, Ward C

    2003-08-01

    The problem of determining the minimum cost hypothetical ancestral sequences for a given cladogram is known to be NP-complete (Wang and Jiang, 1994). Traditionally, point estimations of hypothetical ancestral sequences have been used to gain heuristic, upper bounds on cladogram cost. These include procedures with such diverse approaches as non-additive optimization of multiple sequence alignment, direct optimization (Wheeler, 1996), and fixed-state character optimization (Wheeler, 1999). A method is proposed here which, by extending fixed-state character optimization, replaces the estimation process with a search. This form of optimization examines a diversity of potential state solutions for cost-efficient hypothetical ancestral sequences and can result in greatly more parsimonious cladograms. Additionally, such an approach can be applied to other NP-complete phylogenetic optimization problems such as genomic break-point analysis. PMID:14531408

  9. Energy optimization system

    DOEpatents

    Zhou, Zhi; de Bedout, Juan Manuel; Kern, John Michael; Biyik, Emrah; Chandra, Ramu Sharat

    2013-01-22

    A system for optimizing customer utility usage in a utility network of customer sites, each having one or more utility devices, where customer site is communicated between each of the customer sites and an optimization server having software for optimizing customer utility usage over one or more networks, including private and public networks. A customer site model for each of the customer sites is generated based upon the customer site information, and the customer utility usage is optimized based upon the customer site information and the customer site model. The optimization server can be hosted by an external source or within the customer site. In addition, the optimization processing can be partitioned between the customer site and an external source.

  10. Homotopy optimization methods for global optimization.

    SciTech Connect

    Dunlavy, Daniel M.; O'Leary, Dianne P.

    2005-12-01

    We define a new method for global optimization, the Homotopy Optimization Method (HOM). This method differs from previous homotopy and continuation methods in that its aim is to find a minimizer for each of a set of values of the homotopy parameter, rather than to follow a path of minimizers. We define a second method, called HOPE, by allowing HOM to follow an ensemble of points obtained by perturbation of previous ones. We relate this new method to standard methods such as simulated annealing and show under what circumstances it is superior. We present results of extensive numerical experiments demonstrating performance of HOM and HOPE.

  11. Structural optimization using optimality criteria methods

    NASA Technical Reports Server (NTRS)

    Khot, N. S.; Berke, L.

    1984-01-01

    Optimality criteria methods take advantage of some concepts as those of statically determinate or indeterminate structures, and certain variational principles of structural dynamics, to develop efficient algorithms for the sizing of structures that are subjected to stiffness-related constraints. Some of the methods and iterative strategies developed over the last decade for calculations of the Lagrange multipliers in stressand displacement-limited problems, as well as for satisfying the appropriate optimality criterion, are discussed. The application of these methods are illustrated by solving problems with stress and displacement constraints.

  12. Conceptual design optimization study

    NASA Technical Reports Server (NTRS)

    Hollowell, S. J.; Beeman, E. R., II; Hiyama, R. M.

    1990-01-01

    The feasibility of applying multilevel functional decomposition and optimization techniques to conceptual design of advanced fighter aircraft was investigated. Applying the functional decomposition techniques to the conceptual design phase appears to be feasible. The initial implementation of the modified design process will optimize wing design variables. A hybrid approach, combining functional decomposition techniques for generation of aerodynamic and mass properties linear sensitivity derivatives with existing techniques for sizing mission performance and optimization, is proposed.

  13. Smoothers for Optimization Problems

    NASA Technical Reports Server (NTRS)

    Arian, Eyal; Ta'asan, Shlomo

    1996-01-01

    We present a multigrid one-shot algorithm, and a smoothing analysis, for the numerical solution of optimal control problems which are governed by an elliptic PDE. The analysis provides a simple tool to determine a smoothing minimization process which is essential for multigrid application. Numerical results include optimal control of boundary data using different discretization schemes and an optimal shape design problem in 2D with Dirichlet boundary conditions.

  14. Control and optimization system

    DOEpatents

    Xinsheng, Lou

    2013-02-12

    A system for optimizing a power plant includes a chemical loop having an input for receiving an input parameter (270) and an output for outputting an output parameter (280), a control system operably connected to the chemical loop and having a multiple controller part (230) comprising a model-free controller. The control system receives the output parameter (280), optimizes the input parameter (270) based on the received output parameter (280), and outputs an optimized input parameter (270) to the input of the chemical loop to control a process of the chemical loop in an optimized manner.

  15. Erosion and Optimal Transport

    NASA Astrophysics Data System (ADS)

    Birnir, Bjorn; Rowlett, Julie

    2010-05-01

    We show that the land-surface equation of Birnir, Smith and Merchant, describing erosion of transport limited surfaces have unique weak solutions. The theory of optimal transport is then used to show that these equations constitute an optimal transport of the sediment by the water flow.

  16. Data communication network optimization

    SciTech Connect

    Hessel, S.R.

    1987-01-01

    Research that will ultimately improve the quality of the decisions made in the design of packet-data communication networks is described. Three topics are specifically addressed: an unusually general method of optimizing networks is developed; a network model is developed that facilitates the optimization and it itself valuable; and the statistical nature of the data-communication traffic on networks is examined to support the model. The model permits the analysis of delays in heterogeneous networks of complex topology. The focus of this dissertation is on the optimization of networks. The optimization is built upon the model. The algorithm varies capacity and routing to minimize network cost. The capacity-assignment algorithm developed is an iterative algorithm that allows arbitrarily complex queuing models to be used. It is because of this algorithm that the optimization can be applied to a wide range of heterogeneous network. The routing algorithms is basically the flow deviation algorithm of Gerla.

  17. Algorithms for bilevel optimization

    NASA Technical Reports Server (NTRS)

    Alexandrov, Natalia; Dennis, J. E., Jr.

    1994-01-01

    General multilevel nonlinear optimization problems arise in design of complex systems and can be used as a means of regularization for multi-criteria optimization problems. Here, for clarity in displaying our ideas, we restrict ourselves to general bi-level optimization problems, and we present two solution approaches. Both approaches use a trust-region globalization strategy, and they can be easily extended to handle the general multilevel problem. We make no convexity assumptions, but we do assume that the problem has a nondegenerate feasible set. We consider necessary optimality conditions for the bi-level problem formulations and discuss results that can be extended to obtain multilevel optimization formulations with constraints at each level.

  18. Flight plan optimization

    NASA Astrophysics Data System (ADS)

    Dharmaseelan, Anoop; Adistambha, Keyne D.

    2015-05-01

    Fuel cost accounts for 40 percent of the operating cost of an airline. Fuel cost can be minimized by planning a flight on optimized routes. The routes can be optimized by searching best connections based on the cost function defined by the airline. The most common algorithm that used to optimize route search is Dijkstra's. Dijkstra's algorithm produces a static result and the time taken for the search is relatively long. This paper experiments a new algorithm to optimize route search which combines the principle of simulated annealing and genetic algorithm. The experimental results of route search, presented are shown to be computationally fast and accurate compared with timings from generic algorithm. The new algorithm is optimal for random routing feature that is highly sought by many regional operators.

  19. Optimal Limited Contingency Planning

    NASA Technical Reports Server (NTRS)

    Meuleau, Nicolas; Smith, David E.

    2003-01-01

    For a given problem, the optimal Markov policy over a finite horizon is a conditional plan containing a potentially large number of branches. However, there are applications where it is desirable to strictly limit the number of decision points and branches in a plan. This raises the question of how one goes about finding optimal plans containing only a limited number of branches. In this paper, we present an any-time algorithm for optimal k-contingency planning. It is the first optimal algorithm for limited contingency planning that is not an explicit enumeration of possible contingent plans. By modelling the problem as a partially observable Markov decision process, it implements the Bellman optimality principle and prunes the solution space. We present experimental results of applying this algorithm to some simple test cases.

  20. Rapid Optimization Library

    Energy Science and Technology Software Center (ESTSC)

    2014-05-13

    ROL provides interfaces to and implementations of algorithms for gradient-based unconstrained and constrained optimization. ROL can be used to optimize the response of any client simulation code that evaluates scalar-valued response functions. If the client code can provide gradient information for the response function, ROL will take advantage of it, resulting in faster runtimes. ROL's interfaces are matrix-free, in other words ROL only uses evaluations of scalar-valued and vector-valued functions. ROL can be used tomore » solve optimal design problems and inverse problems based on a variety of simulation software.« less

  1. Optimal control computer programs

    NASA Technical Reports Server (NTRS)

    Kuo, F.

    1992-01-01

    The solution of the optimal control problem, even with low order dynamical systems, can usually strain the analytical ability of most engineers. The understanding of this subject matter, therefore, would be greatly enhanced if a software package existed that could simulate simple generic problems. Surprisingly, despite a great abundance of commercially available control software, few, if any, address the part of optimal control in its most generic form. The purpose of this paper is, therefore, to present a simple computer program that will perform simulations of optimal control problems that arise from the first necessary condition and the Pontryagin's maximum principle.

  2. Thermophotovoltaic Array Optimization

    SciTech Connect

    SBurger; E Brown; K Rahner; L Danielson; J Openlander; J Vell; D Siganporia

    2004-07-29

    A systematic approach to thermophotovoltaic (TPV) array design and fabrication was used to optimize the performance of a 192-cell TPV array. The systematic approach began with cell selection criteria that ranked cells and then matched cell characteristics to maximize power output. Following cell selection, optimization continued with an array packaging design and fabrication techniques that introduced negligible electrical interconnect resistance and minimal parasitic losses while maintaining original cell electrical performance. This paper describes the cell selection and packaging aspects of array optimization as applied to fabrication of a 192-cell array.

  3. Optimization of evaporative cooling

    NASA Astrophysics Data System (ADS)

    Sackett, C. A.; Bradley, C. C.; Hulet, R. G.

    1997-05-01

    Recent experiments have used forced evaporative cooling to produce Bose-Einstein condensation in dilute gases. The evaporative cooling process can be optimized to provide the maximum phase-space density with a specified number of atoms remaining. We show that this global optimization is approximately achieved by locally optimizing the cooling efficiency at each instant. We discuss how this method can be implemented, and present the results for our 7Li trap. The predicted behavior of the gas is found to agree well with experiment.

  4. Rapid Optimization Library

    SciTech Connect

    Denis Rldzal, Drew Kouri

    2014-05-13

    ROL provides interfaces to and implementations of algorithms for gradient-based unconstrained and constrained optimization. ROL can be used to optimize the response of any client simulation code that evaluates scalar-valued response functions. If the client code can provide gradient information for the response function, ROL will take advantage of it, resulting in faster runtimes. ROL's interfaces are matrix-free, in other words ROL only uses evaluations of scalar-valued and vector-valued functions. ROL can be used to solve optimal design problems and inverse problems based on a variety of simulation software.

  5. Topology optimized microbioreactors.

    PubMed

    Schäpper, Daniel; Lencastre Fernandes, Rita; Lantz, Anna Eliasson; Okkels, Fridolin; Bruus, Henrik; Gernaey, Krist V

    2011-04-01

    This article presents the fusion of two hitherto unrelated fields--microbioreactors and topology optimization. The basis for this study is a rectangular microbioreactor with homogeneously distributed immobilized brewers yeast cells (Saccharomyces cerevisiae) that produce a recombinant protein. Topology optimization is then used to change the spatial distribution of cells in the reactor in order to optimize for maximal product flow out of the reactor. This distribution accounts for potentially negative effects of, for example, by-product inhibition. We show that the theoretical improvement in productivity is at least fivefold compared with the homogeneous reactor. The improvements obtained by applying topology optimization are largest where either nutrition is scarce or inhibition effects are pronounced. PMID:21404253

  6. Nuclear Energy Density Optimization

    SciTech Connect

    Kortelainen, Erno M; Lesinski, Thomas; More, J.; Nazarewicz, W.; Sarich, J.; Schunck, N.; Stoitsov, M. V.; Wild, S.

    2010-01-01

    We carry out state-of-the-art optimization of a nuclear energy density of Skyrme type in the framework of the Hartree-Fock-Bogoliubov (HFB) theory. The particle-hole and particle-particle channels are optimized simultaneously, and the experimental data set includes both spherical and deformed nuclei. The new model-based, derivative-free optimization algorithm used in this work has been found to be significantly better than standard optimization methods in terms of reliability, speed, accuracy, and precision. The resulting parameter set UNEDFpre results in good agreement with experimental masses, radii, and deformations and seems to be free of finite-size instabilities. An estimate of the reliability of the obtained parameterization is given, based on standard statistical methods. We discuss new physics insights offered by the advanced covariance analysis.

  7. TOOLKIT FOR ADVANCED OPTIMIZATION

    Energy Science and Technology Software Center (ESTSC)

    2000-10-13

    The TAO project focuses on the development of software for large scale optimization problems. TAO uses an object-oriented design to create a flexible toolkit with strong emphasis on the reuse of external tools where appropriate. Our design enables bi-directional connection to lower level linear algebra support (for example, parallel sparse matrix data structures) as well as higher level application frameworks. The Toolkist for Advanced Optimization (TAO) is aimed at teh solution of large-scale optimization problemsmore » on high-performance architectures. Our main goals are portability, performance, scalable parallelism, and an interface independent of the architecture. TAO is suitable for both single-processor and massively-parallel architectures. The current version of TAO has algorithms for unconstrained and bound-constrained optimization.« less

  8. General shape optimization capability

    NASA Technical Reports Server (NTRS)

    Chargin, Mladen K.; Raasch, Ingo; Bruns, Rudolf; Deuermeyer, Dawson

    1991-01-01

    A method is described for calculating shape sensitivities, within MSC/NASTRAN, in a simple manner without resort to external programs. The method uses natural design variables to define the shape changes in a given structure. Once the shape sensitivities are obtained, the shape optimization process is carried out in a manner similar to property optimization processes. The capability of this method is illustrated by two examples: the shape optimization of a cantilever beam with holes, loaded by a point load at the free end (with the shape of the holes and the thickness of the beam selected as the design variables), and the shape optimization of a connecting rod subjected to several different loading and boundary conditions.

  9. Library for Nonlinear Optimization

    Energy Science and Technology Software Center (ESTSC)

    2001-10-09

    OPT++ is a C++ object-oriented library for nonlinear optimization. This incorporates an improved implementation of an existing capability and two new algorithmic capabilities based on existing journal articles and freely available software.

  10. RF Gun Optimization Study

    SciTech Connect

    A. S. Hofler; P. Evtushenko; M. Krasilnikov

    2007-08-01

    Injector gun design is an iterative process where the designer optimizes a few nonlinearly interdependent beam parameters to achieve the required beam quality for a particle accelerator. Few tools exist to automate the optimization process and thoroughly explore the parameter space. The challenging beam requirements of new accelerator applications such as light sources and electron cooling devices drive the development of RF and SRF photo injectors. RF and SRF gun design is further complicated because the bunches are space charge dominated and require additional emittance compensation. A genetic algorithm has been successfully used to optimize DC photo injector designs for Cornell* and Jefferson Lab**, and we propose studying how the genetic algorithm techniques can be applied to the design of RF and SRF gun injectors. In this paper, we report on the initial phase of the study where we model and optimize gun designs that have been benchmarked with beam measurements and simulation.

  11. Modeling using optimization routines

    NASA Technical Reports Server (NTRS)

    Thomas, Theodore

    1995-01-01

    Modeling using mathematical optimization dynamics is a design tool used in magnetic suspension system development. MATLAB (software) is used to calculate minimum cost and other desired constraints. The parameters to be measured are programmed into mathematical equations. MATLAB will calculate answers for each set of inputs; inputs cover the boundary limits of the design. A Magnetic Suspension System using Electromagnets Mounted in a Plannar Array is a design system that makes use of optimization modeling.

  12. Propeller design by optimization

    NASA Technical Reports Server (NTRS)

    Rizk, M. H.; Jou, W.-H.

    1986-01-01

    The feasibility of designing propellers by an optimization procedure is investigated. A scheme, which solves the full potential flow equation about a propeller by line relaxation, is modified so that the iterative solutions of the flow equation and the design parameters are updated simultaneously. Some technical problems in using optimization for designing propellers with maximum efficiency are identified. Approaches for overcoming these problems are presented.

  13. Depth Optimization Study

    DOE Data Explorer

    Kawase, Mitsuhiro

    2009-11-22

    The zipped file contains a directory of data and routines used in the NNMREC turbine depth optimization study (Kawase et al., 2011), and calculation results thereof. For further info, please contact Mitsuhiro Kawase at kawase@uw.edu. Reference: Mitsuhiro Kawase, Patricia Beba, and Brian Fabien (2011), Finding an Optimal Placement Depth for a Tidal In-Stream Conversion Device in an Energetic, Baroclinic Tidal Channel, NNMREC Technical Report.

  14. Cyclone performance and optimization

    SciTech Connect

    Leith, D.

    1990-09-15

    The objectives of this project are: to characterize the gas flow pattern within cyclones, to revise the theory for cyclone performance on the basis of these findings, and to design and test cyclones whose dimensions have been optimized using revised performance theory. This work is important because its successful completion will aid in the technology for combustion of coal in pressurized, fluidized beds. This quarter, an empirical model for predicting pressure drop across a cyclone was developed through a statistical analysis of pressure drop data for 98 cyclone designs. The model is shown to perform better than the pressure drop models of First (1950), Alexander (1949), Barth (1956), Stairmand (1949), and Shepherd-Lapple (1940). This model is used with the efficiency model of Iozia and Leith (1990) to develop an optimization curve which predicts the minimum pressure drop and the dimension rations of the optimized cyclone for a given aerodynamic cut diameter, d{sub 50}. The effect of variation in cyclone height, cyclone diameter, and flow on the optimization curve is determined. The optimization results are used to develop a design procedure for optimized cyclones. 37 refs., 10 figs., 4 tabs.

  15. Software for global optimization

    SciTech Connect

    Mockus, L.

    1994-12-31

    The interactive graphical software that implements numeric methods and other techniques to solve global optimization problems is presented. The Bayesian approach to the optimization is the underlying idea of numeric methods used. Software is designed to solve deterministic and stochastic problems of different complexity and with many variables. It includes global and local optimization methods for differentiable and nondifferentiable functions. Implemented numerical techniques for global optimization vary from simple Monte-Carlo simulation to Bayesian methods by J. Mockus and extrapolation theory based methods by Zilinskas. Local optimization techniques includes simplex method of Nelder and Mead method of nonlinear programming by Shitkowski, and method of stochastic approximation with Bayesian step size control by J. Mockus. Software is interactive, it allows user to start and stop chosen method of global or local optimization, define and change its parameters and examine the solution process. Out-put from solution process is both numerical and graphical. Currently available graphical features are the projection of the objective function on a chosen plane and convergence plot. Both these features let the user easily observe solution process and interactively modify it. More features can be added in a standard way. It is up to the user how many graphical and numerical output features activate or deactivate at any given time. Software is implemented in C++ using X Windows as graphical platform.

  16. Thermoacoustic Refrigerator's Stack Optimization

    NASA Astrophysics Data System (ADS)

    El-Fawal, Mawahib Hassan; Mohd-Ghazali, Normah; Yaacob, Mohd. Shafik; Darus, Amer Nordin

    2010-06-01

    The standing wave thermoacoustic refrigerator, which uses sound generation to transfer heat, was developed rapidly during the past four decades. It was regarded as a new, promising and environmentally benign alternative to conventional compression vapor refrigerators, although it was not competitive regarding the coefficient of performance (COP) yet. Thus the aim of this paper is to enhance thermoacoustic refrigerator's stack performance through optimization. A computational optimization procedure of thermoacoustic stack design was fully developed. The procedure was designed to achieve optimal coefficient of performance based on most of the design and operating parameters. Cooling load and acoustic power governing equations were set assuming the linear thermoacoustic theory. Lagrange multipliers method was used as an optimization technique tool to solve the governing equations. Numerical analyses results of the developed design procedure are presented. The results showed that the stack design parameters are the most significant parameters for the optimal overall performance. The coefficient of performance obtained increases by about 48.8% from the published experimental optimization methods. The results are in good agreement with past established studies.

  17. Optimization using Extremal Dynamics

    NASA Astrophysics Data System (ADS)

    Boettcher, Stefan

    2001-03-01

    We explore a new heuristic for finding high-quality solutions to NP-hard optimization problems which we have recently introduced [see ``Nature's Way of Optimizing," Artificial Intelligence 119, 275-286 (2000) and cond-mat/0010337]. The method, called extremal optimization, is inspired by self-organized criticality, a concept introduced to describe emergent complexity in physical systems. Extremal optimization successively replaces extremely undesirable elements of a single sub-optimal solution with new, random ones. Large fluctuations ensue that efficiently explore many local optima. With only one adjustable parameter, its performance has proved competitive with more elaborate methods, especially near phase transitions which are believed to contain the hardest instances. In particular, extremal optimization is superior to simulated annealing in the partitioning of sparse graphs, it finds the overlap of all ground-states at the phase transition of the 3-coloring problem, and it provides independent confirmation for the ground-state energy of spin glasses, previously obtained with elaborate genetic algorithms.

  18. Training a quantum optimizer

    NASA Astrophysics Data System (ADS)

    Wecker, Dave; Hastings, Matthew B.; Troyer, Matthias

    2016-08-01

    We study a variant of the quantum approximate optimization algorithm [E. Farhi, J. Goldstone, and S. Gutmann, arXiv:1411.4028] with a slightly different parametrization and a different objective: rather than looking for a state which approximately solves an optimization problem, our goal is to find a quantum algorithm that, given an instance of the maximum 2-satisfiability problem (MAX-2-SAT), will produce a state with high overlap with the optimal state. Using a machine learning approach, we chose a "training set" of instances and optimized the parameters to produce a large overlap for the training set. We then tested these optimized parameters on a larger instance set. As a training set, we used a subset of the hard instances studied by Crosson, Farhi, C. Y.-Y. Lin, H.-H. Lin, and P. Shor (CFLLS) (arXiv:1401.7320). When tested, on the full set, the parameters that we find produce a significantly larger overlap than the optimized annealing times of CFLLS. Testing on other random instances from 20 to 28 bits continues to show improvement over annealing, with the improvement being most notable on the hardest instances. Further tests on instances of MAX-3-SAT also showed improvement on the hardest instances. This algorithm may be a possible application for near-term quantum computers with limited coherence times.

  19. Shape optimization and CAD

    NASA Technical Reports Server (NTRS)

    Rasmussen, John

    1990-01-01

    Structural optimization has attracted the attention since the days of Galileo. Olhoff and Taylor have produced an excellent overview of the classical research within this field. However, the interest in structural optimization has increased greatly during the last decade due to the advent of reliable general numerical analysis methods and the computer power necessary to use them efficiently. This has created the possibility of developing general numerical systems for shape optimization. Several authors, eg., Esping; Braibant & Fleury; Bennet & Botkin; Botkin, Yang, and Bennet; and Stanton have published practical and successful applications of general optimization systems. Ding and Homlein have produced extensive overviews of available systems. Furthermore, a number of commercial optimization systems based on well-established finite element codes have been introduced. Systems like ANSYS, IDEAS, OASIS, and NISAOPT are widely known examples. In parallel to this development, the technology of computer aided design (CAD) has gained a large influence on the design process of mechanical engineering. The CAD technology has already lived through a rapid development driven by the drastically growing capabilities of digital computers. However, the systems of today are still considered as being only the first generation of a long row of computer integrated manufacturing (CIM) systems. These systems to come will offer an integrated environment for design, analysis, and fabrication of products of almost any character. Thus, the CAD system could be regarded as simply a database for geometrical information equipped with a number of tools with the purpose of helping the user in the design process. Among these tools are facilities for structural analysis and optimization as well as present standard CAD features like drawing, modeling, and visualization tools. The state of the art of structural optimization is that a large amount of mathematical and mechanical techniques are

  20. Optimization of Heat Exchangers

    SciTech Connect

    Ivan Catton

    2010-10-01

    The objective of this research is to develop tools to design and optimize heat exchangers (HE) and compact heat exchangers (CHE) for intermediate loop heat transport systems found in the very high temperature reator (VHTR) and other Generation IV designs by addressing heat transfer surface augmentation and conjugate modeling. To optimize heat exchanger, a fast running model must be created that will allow for multiple designs to be compared quickly. To model a heat exchanger, volume averaging theory, VAT, is used. VAT allows for the conservation of mass, momentum and energy to be solved for point by point in a 3 dimensional computer model of a heat exchanger. The end product of this project is a computer code that can predict an optimal configuration for a heat exchanger given only a few constraints (input fluids, size, cost, etc.). As VAT computer code can be used to model characteristics )pumping power, temperatures, and cost) of heat exchangers more quickly than traditional CFD or experiment, optimization of every geometric parameter simultaneously can be made. Using design of experiment, DOE and genetric algorithms, GE, to optimize the results of the computer code will improve heat exchanger disign.

  1. Generic Optimization Program

    Energy Science and Technology Software Center (ESTSC)

    1998-07-01

    GenOpt is a generic optimization program for nonlinear, constrained optimization. For evaluating the objective function, any simulation program that communicates over text files can be coupled to GenOpt without code modification. No analytic properties of the objective function are used by GenOpt. ptimization algorithms and numerical methods can be implemented in a library and shared among users. Gencpt offers an interlace between the optimization algorithm and its kernel to make the implementation of new algorithmsmore » fast and easy. Different algorithms of constrained and unconstrained minimization can be added to a library. Algorithms for approximation derivatives and performing line-search will be implemented. The objective function is evaluated as a black-box function by an external simulation program. The kernel of GenOpt deals with the data I/O, result sotrage and report, interlace to the external simulation program, and error handling. An abstract optimization class offers methods to interface the GenOpt kernel and the optimization algorithm library.« less

  2. Reverse Osmosis Optimization

    SciTech Connect

    McMordie Stoughton, Kate; Duan, Xiaoli; Wendel, Emily M.

    2013-08-26

    This technology evaluation was prepared by Pacific Northwest National Laboratory on behalf of the U.S. Department of Energy’s Federal Energy Management Program (FEMP). ¬The technology evaluation assesses techniques for optimizing reverse osmosis (RO) systems to increase RO system performance and water efficiency. This evaluation provides a general description of RO systems, the influence of RO systems on water use, and key areas where RO systems can be optimized to reduce water and energy consumption. The evaluation is intended to help facility managers at Federal sites understand the basic concepts of the RO process and system optimization options, enabling them to make informed decisions during the system design process for either new projects or recommissioning of existing equipment. This evaluation is focused on commercial-sized RO systems generally treating more than 80 gallons per hour.¬

  3. Reverse Osmosis Optimization

    SciTech Connect

    2013-08-01

    This technology evaluation was prepared by Pacific Northwest National Laboratory on behalf of the U.S. Department of Energy’s Federal Energy Management Program (FEMP). The technology evaluation assesses techniques for optimizing reverse osmosis (RO) systems to increase RO system performance and water efficiency. This evaluation provides a general description of RO systems, the influence of RO systems on water use, and key areas where RO systems can be optimized to reduce water and energy consumption. The evaluation is intended to help facility managers at Federal sites understand the basic concepts of the RO process and system optimization options, enabling them to make informed decisions during the system design process for either new projects or recommissioning of existing equipment. This evaluation is focused on commercial-sized RO systems generally treating more than 80 gallons per hour.

  4. Timing optimization control

    SciTech Connect

    Johnson, E.A.; Leung, C.; Schira, J.J.

    1983-03-01

    A closed loop timing optimization control for an internal combustion engine closed about the instantaneous rotational velocity of the engine's crankshaft is disclosed herein. The optimization control computes from the instantaneous rotational velocity of the engine's crankshaft, a signal indicative of the angle at which the crankshaft has a maximum rotational velocity for the torque impulses imparted to the engine's crankshaft by the burning of an air/fuel mixture in each of the engine's combustion chambers and generates a timing correction signal for each of the engine's combustion chambers. The timing correction signals, applied to the engine timing control, modifies the time at which the ignition signal, injection signals or both are generated such that the rotational velocity of the engine's crankshaft has a maximum value at a predetermined angle for each torque impulse generated optimizing the conversion of the combustion energy to rotational torque.

  5. Optimally combined confidence limits

    NASA Astrophysics Data System (ADS)

    Janot, P.; Le Diberder, F.

    1998-02-01

    An analytical and optimal procedure to combine statistically independent sets of confidence levels on a quantity is presented. This procedure does not impose any constraint on the methods followed by each analysis to derive its own limit. It incorporates the a priori statistical power of each of the analyses to be combined, in order to optimize the overall sensitivity. It can, in particular, be used to combine the mass limits obtained by several analyses searching for the Higgs boson in different decay channels, with different selection efficiencies, mass resolution and expected background. It can also be used to combine the mass limits obtained by several experiments (e.g. ALEPH, DELPHI, L3 and OPAL, at LEP 2) independently of the method followed by each of these experiments to derive their own limit. A method to derive the limit set by one analysis is also presented, along with an unbiased prescription to optimize the expected mass limit in the no-signal-hypothesis.

  6. Fuzzy logic controller optimization

    DOEpatents

    Sepe, Jr., Raymond B; Miller, John Michael

    2004-03-23

    A method is provided for optimizing a rotating induction machine system fuzzy logic controller. The fuzzy logic controller has at least one input and at least one output. Each input accepts a machine system operating parameter. Each output produces at least one machine system control parameter. The fuzzy logic controller generates each output based on at least one input and on fuzzy logic decision parameters. Optimization begins by obtaining a set of data relating each control parameter to at least one operating parameter for each machine operating region. A model is constructed for each machine operating region based on the machine operating region data obtained. The fuzzy logic controller is simulated with at least one created model in a feedback loop from a fuzzy logic output to a fuzzy logic input. Fuzzy logic decision parameters are optimized based on the simulation.

  7. Optimal symmetric flight studies

    NASA Technical Reports Server (NTRS)

    Weston, A. R.; Menon, P. K. A.; Bilimoria, K. D.; Cliff, E. M.; Kelley, H. J.

    1985-01-01

    Several topics in optimal symmetric flight of airbreathing vehicles are examined. In one study, an approximation scheme designed for onboard real-time energy management of climb-dash is developed and calculations for a high-performance aircraft presented. In another, a vehicle model intermediate in complexity between energy and point-mass models is explored and some quirks in optimal flight characteristics peculiar to the model uncovered. In yet another study, energy-modelling procedures are re-examined with a view to stretching the range of validity of zeroth-order approximation by special choice of state variables. In a final study, time-fuel tradeoffs in cruise-dash are examined for the consequences of nonconvexities appearing in the classical steady cruise-dash model. Two appendices provide retrospective looks at two early publications on energy modelling and related optimal control theory.

  8. Optimization of Combinatorial Mutagenesis

    NASA Astrophysics Data System (ADS)

    Parker, Andrew S.; Griswold, Karl E.; Bailey-Kellogg, Chris

    Protein engineering by combinatorial site-directed mutagenesis evaluates a portion of the sequence space near a target protein, seeking variants with improved properties (stability, activity, immunogenicity, etc.). In order to improve the hit-rate of beneficial variants in such mutagenesis libraries, we develop methods to select optimal positions and corresponding sets of the mutations that will be used, in all combinations, in constructing a library for experimental evaluation. Our approach, OCoM (Optimization of Combinatorial Mutagenesis), encompasses both degenerate oligonucleotides and specified point mutations, and can be directed accordingly by requirements of experimental cost and library size. It evaluates the quality of the resulting library by one- and two-body sequence potentials, averaged over the variants. To ensure that it is not simply recapitulating extant sequences, it balances the quality of a library with an explicit evaluation of the novelty of its members. We show that, despite dealing with a combinatorial set of variants, in our approach the resulting library optimization problem is actually isomorphic to single-variant optimization. By the same token, this means that the two-body sequence potential results in an NP-hard optimization problem. We present an efficient dynamic programming algorithm for the one-body case and a practically-efficient integer programming approach for the general two-body case. We demonstrate the effectiveness of our approach in designing libraries for three different case study proteins targeted by previous combinatorial libraries - a green fluorescent protein, a cytochrome P450, and a beta lactamase. We found that OCoM worked quite efficiently in practice, requiring only 1 hour even for the massive design problem of selecting 18 mutations to generate 107 variants of a 443-residue P450. We demonstrate the general ability of OCoM in enabling the protein engineer to explore and evaluate trade-offs between quality and

  9. Terascale Optimal PDE Simulations

    SciTech Connect

    David Keyes

    2009-07-28

    The Terascale Optimal PDE Solvers (TOPS) Integrated Software Infrastructure Center (ISIC) was created to develop and implement algorithms and support scientific investigations performed by DOE-sponsored researchers. These simulations often involve the solution of partial differential equations (PDEs) on terascale computers. The TOPS Center researched, developed and deployed an integrated toolkit of open-source, optimal complexity solvers for the nonlinear partial differential equations that arise in many DOE application areas, including fusion, accelerator design, global climate change and reactive chemistry. The algorithms created as part of this project were also designed to reduce current computational bottlenecks by orders of magnitude on terascale computers, enabling scientific simulation on a scale heretofore impossible.

  10. Optimal Quantum Phase Estimation

    SciTech Connect

    Dorner, U.; Smith, B. J.; Lundeen, J. S.; Walmsley, I. A.; Demkowicz-Dobrzanski, R.; Banaszek, K.; Wasilewski, W.

    2009-01-30

    By using a systematic optimization approach, we determine quantum states of light with definite photon number leading to the best possible precision in optical two-mode interferometry. Our treatment takes into account the experimentally relevant situation of photon losses. Our results thus reveal the benchmark for precision in optical interferometry. Although this boundary is generally worse than the Heisenberg limit, we show that the obtained precision beats the standard quantum limit, thus leading to a significant improvement compared to classical interferometers. We furthermore discuss alternative states and strategies to the optimized states which are easier to generate at the cost of only slightly lower precision.

  11. Neural wiring optimization.

    PubMed

    Cherniak, Christopher

    2012-01-01

    Combinatorial network optimization theory concerns minimization of connection costs among interconnected components in systems such as electronic circuits. As an organization principle, similar wiring minimization can be observed at various levels of nervous systems, invertebrate and vertebrate, including primate, from placement of the entire brain in the body down to the subcellular level of neuron arbor geometry. In some cases, the minimization appears either perfect, or as good as can be detected with current methods. One question such best-of-all-possible-brains results raise is, what is the map of such optimization, does it have a distinct neural domain? PMID:22230636

  12. Shape optimization for DSA

    NASA Astrophysics Data System (ADS)

    Ouaknin, Gaddiel; Laachi, Nabil; Delaney, Kris; Fredrickson, Glenn; Gibou, Frederic

    2016-03-01

    Directed self-assembly using block copolymers for positioning vertical interconnect access in integrated circuits relies on the proper shape of a confined domain in which polymers will self-assemble into the targeted design. Finding that shape, i.e., solving the inverse problem, is currently mainly based on trial and error approaches. We introduce a level-set based algorithm that makes use of a shape optimization strategy coupled with self-consistent field theory to solve the inverse problem in an automated way. It is shown that optimal shapes are found for different targeted topologies with accurate placement and distances between the different components.

  13. Distributed Optimization System

    DOEpatents

    Hurtado, John E.; Dohrmann, Clark R.; Robinett, III, Rush D.

    2004-11-30

    A search system and method for controlling multiple agents to optimize an objective using distributed sensing and cooperative control. The search agent can be one or more physical agents, such as a robot, and can be software agents for searching cyberspace. The objective can be: chemical sources, temperature sources, radiation sources, light sources, evaders, trespassers, explosive sources, time dependent sources, time independent sources, function surfaces, maximization points, minimization points, and optimal control of a system such as a communication system, an economy, a crane, and a multi-processor computer.

  14. Optimization in Ecology

    ERIC Educational Resources Information Center

    Cody, Martin L.

    1974-01-01

    Discusses the optimality of natural selection, ways of testing for optimum solutions to problems of time - or energy-allocation in nature, optimum patterns in spatial distribution and diet breadth, and how best to travel over a feeding area so that food intake is maximized. (JR)

  15. Toward Optimal Transport Networks

    NASA Technical Reports Server (NTRS)

    Alexandrov, Natalia; Kincaid, Rex K.; Vargo, Erik P.

    2008-01-01

    Strictly evolutionary approaches to improving the air transport system a highly complex network of interacting systems no longer suffice in the face of demand that is projected to double or triple in the near future. Thus evolutionary approaches should be augmented with active design methods. The ability to actively design, optimize and control a system presupposes the existence of predictive modeling and reasonably well-defined functional dependences between the controllable variables of the system and objective and constraint functions for optimization. Following recent advances in the studies of the effects of network topology structure on dynamics, we investigate the performance of dynamic processes on transport networks as a function of the first nontrivial eigenvalue of the network's Laplacian, which, in turn, is a function of the network s connectivity and modularity. The last two characteristics can be controlled and tuned via optimization. We consider design optimization problem formulations. We have developed a flexible simulation of network topology coupled with flows on the network for use as a platform for computational experiments.

  16. Optimizing Computer Technology Integration

    ERIC Educational Resources Information Center

    Dillon-Marable, Elizabeth; Valentine, Thomas

    2006-01-01

    The purpose of this study was to better understand what optimal computer technology integration looks like in adult basic skills education (ABSE). One question guided the research: How is computer technology integration best conceptualized and measured? The study used the Delphi method to map the construct of computer technology integration and…

  17. Optimization of digital designs

    NASA Technical Reports Server (NTRS)

    Whitaker, Sterling R. (Inventor); Miles, Lowell H. (Inventor)

    2009-01-01

    An application specific integrated circuit is optimized by translating a first representation of its digital design to a second representation. The second representation includes multiple syntactic expressions that admit a representation of a higher-order function of base Boolean values. The syntactic expressions are manipulated to form a third representation of the digital design.

  18. Fourier Series Optimization Opportunity

    ERIC Educational Resources Information Center

    Winkel, Brian

    2008-01-01

    This note discusses the introduction of Fourier series as an immediate application of optimization of a function of more than one variable. Specifically, it is shown how the study of Fourier series can be motivated to enrich a multivariable calculus class. This is done through discovery learning and use of technology wherein students build the…

  19. Optimal ciliary beating patterns

    NASA Astrophysics Data System (ADS)

    Vilfan, Andrej; Osterman, Natan

    2011-11-01

    We introduce a measure for energetic efficiency of single or collective biological cilia. We define the efficiency of a single cilium as Q2 / P , where Q is the volume flow rate of the pumped fluid and P is the dissipated power. For ciliary arrays, we define it as (ρQ) 2 / (ρP) , with ρ denoting the surface density of cilia. We then numerically determine the optimal beating patterns according to this criterion. For a single cilium optimization leads to curly, somewhat counterintuitive patterns. But when looking at a densely ciliated surface, the optimal patterns become remarkably similar to what is observed in microorganisms like Paramecium. The optimal beating pattern then consists of a fast effective stroke and a slow sweeping recovery stroke. Metachronal waves lead to a significantly higher efficiency than synchronous beating. Efficiency also increases with an increasing density of cilia up to the point where crowding becomes a problem. We finally relate the pumping efficiency of cilia to the swimming efficiency of a spherical microorganism and show that the experimentally estimated efficiency of Paramecium is surprisingly close to the theoretically possible optimum.

  20. Optimization of Systran System.

    ERIC Educational Resources Information Center

    Toma, Peter P.; And Others

    This report describes an optimization phase of the SYSTRAN (System Translation) machine translation technique. The most distinctive characteristic of SYSTRAN is the absence of pre-editing; the program reads tapes containing raw and unedited Russian texts, carries out dictionary and table lookups, performs all syntactic analysis procedures, and…

  1. Optimal Ski Jump

    ERIC Educational Resources Information Center

    Rebilas, Krzysztof

    2013-01-01

    Consider a skier who goes down a takeoff ramp, attains a speed "V", and jumps, attempting to land as far as possible down the hill below (Fig. 1). At the moment of takeoff the angle between the skier's velocity and the horizontal is [alpha]. What is the optimal angle [alpha] that makes the jump the longest possible for the fixed magnitude of the…

  2. Optimized multisectioned acoustic liners

    NASA Technical Reports Server (NTRS)

    Baumeister, K. J.

    1979-01-01

    A critical examination is presented of the use of optimized axially segmented acoustic liners to increase the attenuation of a liner. New calculations show that segmenting is most efficient at high frequencies with relatively long duct lengths where the attenuation is low for both uniform and segmented liners. Statistical considerations indicate little advantage in using optimized liners with more than two segments while the bandwidth of an optimized two-segment liner is shown to be nearly equal to that of a uniform liner. Multielement liner calculations show a large degradation in performance due to changes in assumed input modal structure. Finally, in order to substantiate previous and future analytical results, in-house (finite difference) and contractor (mode matching) programs are used to generate theoretical attenuations for a number of liner configurations for liners in a rectangular duct with no mean flow. Overall, the use of optimized multisectioned liners (sometimes called phased liners) fails to offer sufficient advantage over a uniform liner to warrant their use except in low frequency single mode application.

  3. OPTIMIZING EXPOSURE MEASUREMENT TECHNIQUES

    EPA Science Inventory

    The research reported in this task description addresses one of a series of interrelated NERL tasks with the common goal of optimizing the predictive power of low cost, reliable exposure measurements for the planned Interagency National Children's Study (NCS). Specifically, we w...

  4. Optimization in Cardiovascular Modeling

    NASA Astrophysics Data System (ADS)

    Marsden, Alison L.

    2014-01-01

    Fluid mechanics plays a key role in the development, progression, and treatment of cardiovascular disease. Advances in imaging methods and patient-specific modeling now reveal increasingly detailed information about blood flow patterns in health and disease. Building on these tools, there is now an opportunity to couple blood flow simulation with optimization algorithms to improve the design of surgeries and devices, incorporating more information about the flow physics in the design process to augment current medical knowledge. In doing so, a major challenge is the need for efficient optimization tools that are appropriate for unsteady fluid mechanics problems, particularly for the optimization of complex patient-specific models in the presence of uncertainty. This article reviews the state of the art in optimization tools for virtual surgery, device design, and model parameter identification in cardiovascular flow and mechanobiology applications. In particular, it reviews trade-offs between traditional gradient-based methods and derivative-free approaches, as well as the need to incorporate uncertainties. Key future challenges are outlined, which extend to the incorporation of biological response and the customization of surgeries and devices for individual patients.

  5. Is Optimism Real?

    ERIC Educational Resources Information Center

    Simmons, Joseph P.; Massey, Cade

    2012-01-01

    Is optimism real, or are optimistic forecasts just cheap talk? To help answer this question, we investigated whether optimistic predictions persist in the face of large incentives to be accurate. We asked National Football League football fans to predict the winner of a single game. Roughly half (the partisans) predicted a game involving their…

  6. Optimizing Conferencing Freeware

    ERIC Educational Resources Information Center

    Baggaley, Jon; Klaas, Jim; Wark, Norine; Depow, Jim

    2005-01-01

    The increasing range of options provided by two popular conferencing freeware products, "Yahoo Messenger" and "MSN Messenger," are discussed. Each tool contains features designed primarily for entertainment purposes, which can be customized for use in online education. This report provides suggestions for optimizing the educational potential of…

  7. Online Search Optimization.

    ERIC Educational Resources Information Center

    Homan, Michael; Worley, Penny

    This course syllabus describes methods for optimizing online searching, using as an example searching on the National Library of Medicine (NLM) online system. Four major activities considered are the online interview, query analysis and search planning, online interaction, and post-search analysis. Within the context of these activities, concepts…

  8. Optimal GENCO bidding strategy

    NASA Astrophysics Data System (ADS)

    Gao, Feng

    Electricity industries worldwide are undergoing a period of profound upheaval. The conventional vertically integrated mechanism is being replaced by a competitive market environment. Generation companies have incentives to apply novel technologies to lower production costs, for example: Combined Cycle units. Economic dispatch with Combined Cycle units becomes a non-convex optimization problem, which is difficult if not impossible to solve by conventional methods. Several techniques are proposed here: Mixed Integer Linear Programming, a hybrid method, as well as Evolutionary Algorithms. Evolutionary Algorithms share a common mechanism, stochastic searching per generation. The stochastic property makes evolutionary algorithms robust and adaptive enough to solve a non-convex optimization problem. This research implements GA, EP, and PS algorithms for economic dispatch with Combined Cycle units, and makes a comparison with classical Mixed Integer Linear Programming. The electricity market equilibrium model not only helps Independent System Operator/Regulator analyze market performance and market power, but also provides Market Participants the ability to build optimal bidding strategies based on Microeconomics analysis. Supply Function Equilibrium (SFE) is attractive compared to traditional models. This research identifies a proper SFE model, which can be applied to a multiple period situation. The equilibrium condition using discrete time optimal control is then developed for fuel resource constraints. Finally, the research discusses the issues of multiple equilibria and mixed strategies, which are caused by the transmission network. Additionally, an advantage of the proposed model for merchant transmission planning is discussed. A market simulator is a valuable training and evaluation tool to assist sellers, buyers, and regulators to understand market performance and make better decisions. A traditional optimization model may not be enough to consider the distributed

  9. An optimal structural design algorithm using optimality criteria

    NASA Technical Reports Server (NTRS)

    Taylor, J. E.; Rossow, M. P.

    1976-01-01

    An algorithm for optimal design is given which incorporates several of the desirable features of both mathematical programming and optimality criteria, while avoiding some of the undesirable features. The algorithm proceeds by approaching the optimal solution through the solutions of an associated set of constrained optimal design problems. The solutions of the constrained problems are recognized at each stage through the application of optimality criteria based on energy concepts. Two examples are described in which the optimal member size and layout of a truss is predicted, given the joint locations and loads.

  10. Optimization of combinatorial mutagenesis.

    PubMed

    Parker, Andrew S; Griswold, Karl E; Bailey-Kellogg, Chris

    2011-11-01

    Protein engineering by combinatorial site-directed mutagenesis evaluates a portion of the sequence space near a target protein, seeking variants with improved properties (e.g., stability, activity, immunogenicity). In order to improve the hit-rate of beneficial variants in such mutagenesis libraries, we develop methods to select optimal positions and corresponding sets of the mutations that will be used, in all combinations, in constructing a library for experimental evaluation. Our approach, OCoM (Optimization of Combinatorial Mutagenesis), encompasses both degenerate oligonucleotides and specified point mutations, and can be directed accordingly by requirements of experimental cost and library size. It evaluates the quality of the resulting library by one- and two-body sequence potentials, averaged over the variants. To ensure that it is not simply recapitulating extant sequences, it balances the quality of a library with an explicit evaluation of the novelty of its members. We show that, despite dealing with a combinatorial set of variants, in our approach the resulting library optimization problem is actually isomorphic to single-variant optimization. By the same token, this means that the two-body sequence potential results in an NP-hard optimization problem. We present an efficient dynamic programming algorithm for the one-body case and a practically-efficient integer programming approach for the general two-body case. We demonstrate the effectiveness of our approach in designing libraries for three different case study proteins targeted by previous combinatorial libraries--a green fluorescent protein, a cytochrome P450, and a beta lactamase. We found that OCoM worked quite efficiently in practice, requiring only 1 hour even for the massive design problem of selecting 18 mutations to generate 10⁷ variants of a 443-residue P450. We demonstrate the general ability of OCoM in enabling the protein engineer to explore and evaluate trade-offs between quality and

  11. Optimizing Thomson's jumping ring

    NASA Astrophysics Data System (ADS)

    Tjossem, Paul J. H.; Brost, Elizabeth C.

    2011-04-01

    The height to which rings will jump in a Thomson jumping ring apparatus is the central question posed by this popular lecture demonstration. We develop a simple time-averaged inductive-phase-lag model for the dependence of the jump height on the ring material, its mass, and temperature and apply it to measurements of the jump height for a set of rings made by slicing copper and aluminum alloy pipe into varying lengths. The data confirm a peak jump height that grows, narrows, and shifts to smaller optimal mass when the rings are cooled to 77 K. The model explains the ratio of the cooled/warm jump heights for a given ring, the reduction in optimal mass as the ring is cooled, and the shape of the mass resonance. The ring that jumps the highest is found to have a characteristic resistance equal to the inductive reactance of the set of rings.

  12. Combinatorial optimization games

    SciTech Connect

    Deng, X.; Ibaraki, Toshihide; Nagamochi, Hiroshi

    1997-06-01

    We introduce a general integer programming formulation for a class of combinatorial optimization games, which immediately allows us to improve the algorithmic result for finding amputations in the core (an important solution concept in cooperative game theory) of the network flow game on simple networks by Kalai and Zemel. An interesting result is a general theorem that the core for this class of games is nonempty if and only if a related linear program has an integer optimal solution. We study the properties for this mathematical condition to hold for several interesting problems, and apply them to resolve algorithmic and complexity issues for their cores along the line as put forward in: decide whether the core is empty; if the core is empty, find an imputation in the core; given an imputation x, test whether x is in the core. We also explore the properties of totally balanced games in this succinct formulation of cooperative games.

  13. Heliostat cost optimization study

    NASA Astrophysics Data System (ADS)

    von Reeken, Finn; Weinrebe, Gerhard; Keck, Thomas; Balz, Markus

    2016-05-01

    This paper presents a methodology for a heliostat cost optimization study. First different variants of small, medium sized and large heliostats are designed. Then the respective costs, tracking and optical quality are determined. For the calculation of optical quality a structural model of the heliostat is programmed and analyzed using finite element software. The costs are determined based on inquiries and from experience with similar structures. Eventually the levelised electricity costs for a reference power tower plant are calculated. Before each annual simulation run the heliostat field is optimized. Calculated LCOEs are then used to identify the most suitable option(s). Finally, the conclusions and findings of this extensive cost study are used to define the concept of a new cost-efficient heliostat called `Stellio'.

  14. Cyclone performance and optimization

    SciTech Connect

    Leith, D.

    1990-06-15

    The objectives of this project are: to characterize the gas flow pattern within cyclones, to revise the theory for cyclone performance on the basis of these findings, and to design and test cyclones whose dimensions have been optimized using revised performance theory. This work is important because its successful completion will aid in the technology for combustion of coal in pressurized, fluidized beds. During the past quarter, we have nearly completed modeling work that employs the flow field measurements made during the past six months. In addition, we have begun final work using the results of this project to develop improved design methods for cyclones. This work involves optimization using the Iozia-Leith efficiency model and the Dirgo pressure drop model. This work will be completed this summer. 9 figs.

  15. Optimism in Active Learning

    PubMed Central

    Collet, Timothé; Pietquin, Olivier

    2015-01-01

    Active learning is the problem of interactively constructing the training set used in classification in order to reduce its size. It would ideally successively add the instance-label pair that decreases the classification error most. However, the effect of the addition of a pair is not known in advance. It can still be estimated with the pairs already in the training set. The online minimization of the classification error involves a tradeoff between exploration and exploitation. This is a common problem in machine learning for which multiarmed bandit, using the approach of Optimism int the Face of Uncertainty, has proven very efficient these last years. This paper introduces three algorithms for the active learning problem in classification using Optimism in the Face of Uncertainty. Experiments lead on built-in problems and real world datasets demonstrate that they compare positively to state-of-the-art methods. PMID:26681934

  16. Optimal Electric Utility Expansion

    Energy Science and Technology Software Center (ESTSC)

    1989-10-10

    SAGE-WASP is designed to find the optimal generation expansion policy for an electrical utility system. New units can be automatically selected from a user-supplied list of expansion candidates which can include hydroelectric and pumped storage projects. The existing system is modeled. The calculational procedure takes into account user restrictions to limit generation configurations to an area of economic interest. The optimization program reports whether the restrictions acted as a constraint on the solution. All expansionmore » configurations considered are required to pass a user supplied reliability criterion. The discount rate and escalation rate are treated separately for each expansion candidate and for each fuel type. All expenditures are separated into local and foreign accounts, and a weighting factor can be applied to foreign expenditures.« less

  17. Optimal firing rate estimation

    NASA Technical Reports Server (NTRS)

    Paulin, M. G.; Hoffman, L. F.

    2001-01-01

    We define a measure for evaluating the quality of a predictive model of the behavior of a spiking neuron. This measure, information gain per spike (Is), indicates how much more information is provided by the model than if the prediction were made by specifying the neuron's average firing rate over the same time period. We apply a maximum Is criterion to optimize the performance of Gaussian smoothing filters for estimating neural firing rates. With data from bullfrog vestibular semicircular canal neurons and data from simulated integrate-and-fire neurons, the optimal bandwidth for firing rate estimation is typically similar to the average firing rate. Precise timing and average rate models are limiting cases that perform poorly. We estimate that bullfrog semicircular canal sensory neurons transmit in the order of 1 bit of stimulus-related information per spike.

  18. NEMO Oceanic Model Optimization

    NASA Astrophysics Data System (ADS)

    Epicoco, I.; Mocavero, S.; Murli, A.; Aloisio, G.

    2012-04-01

    NEMO is an oceanic model used by the climate community for stand-alone or coupled experiments. Its parallel implementation, based on MPI, limits the exploitation of the emerging computational infrastructures at peta and exascale, due to the weight of communications. As case study we considered the MFS configuration developed at INGV with a resolution of 1/16° tailored on the Mediterranenan Basin. The work is focused on the analysis of the code on the MareNostrum cluster and on the optimization of critical routines. The first performance analysis of the model aimed at establishing how much the computational performance are influenced by the GPFS file system or the local disks and wich is the best domain decomposition. The results highlight that the exploitation of local disks can reduce the wall clock time up to 40% and that the best performance is achieved with a 2D decomposition when the local domain has a square shape. A deeper performance analysis highlights the obc_rad, dyn_spg and tra_adv routines are the most time consuming routines. The obc_rad implements the evaluation of the open boundaries and it has been the first routine to be optimized. The communication pattern implemented in obc_rad routine has been redesigned. Before the introduction of the optimizations all processes were involved in the communication, but only the processes on the boundaries have the actual data to be exchanged and only the data on the boundaries must be exchanged. Moreover the data along the vertical levels are "packed" and sent with only one MPI_send invocation. The overall efficiency increases compared with the original version, as well as the parallel speed-up. The execution time was reduced of about 33.81%. The second phase of optimization involved the SOR solver routine, implementing the Red-Black Successive-Over-Relaxation method. The high frequency of exchanging data among processes represent the most part of the overall communication time. The number of communication is

  19. Optimize acid gas removal

    SciTech Connect

    Nicholas, D.M.; Wilkins, J.T.

    1983-09-01

    Innovative design of physical solvent plants for acid gas removal can materially reduce both installation and operating costs. A review of the design considerations for one physical solvent process (Selexol) points to numerous arrangements for potential improvement. These are evaluated for a specific case in four combinations that identify an optimum for the case in question but, more importantly, illustrate the mechanism for use for such optimization elsewhere.

  20. Optimal Prenatal Care

    PubMed Central

    Reynolds, J. L.

    1982-01-01

    Optimal prenatal care begins before conception, when health habits can be reviewed. The most important task of the initial prenatal assessment is establishing dates. Ongoing assessments should emphasize measurement of symphisis to fundus height, maternal nutrition and screening, especially for urinary tract infection and gestational diabetes. Prenatal care is an excellent opportunity for patient education and involvement of the family. Good prenatal care is today's best health investment. PMID:21286515

  1. Singularity in structural optimization

    NASA Technical Reports Server (NTRS)

    Patnaik, S. N.; Guptill, J. D.; Berke, L.

    1993-01-01

    The conditions under which global and local singularities may arise in structural optimization are examined. Examples of these singularities are presented, and a framework is given within which the singularities can be recognized. It is shown, in particular, that singularities can be identified through the analysis of stress-displacement relations together with compatibility conditions or the displacement-stress relations derived by the integrated force method of structural analysis. Methods of eliminating the effects of singularities are suggested and illustrated numerically.

  2. Trajectory Optimization: OTIS 4

    NASA Technical Reports Server (NTRS)

    Riehl, John P.; Sjauw, Waldy K.; Falck, Robert D.; Paris, Stephen W.

    2010-01-01

    The latest release of the Optimal Trajectories by Implicit Simulation (OTIS4) allows users to simulate and optimize aerospace vehicle trajectories. With OTIS4, one can seamlessly generate optimal trajectories and parametric vehicle designs simultaneously. New features also allow OTIS4 to solve non-aerospace continuous time optimal control problems. The inputs and outputs of OTIS4 have been updated extensively from previous versions. Inputs now make use of objectoriented constructs, including one called a metastring. Metastrings use a greatly improved calculator and common nomenclature to reduce the user s workload. They allow for more flexibility in specifying vehicle physical models, boundary conditions, and path constraints. The OTIS4 calculator supports common mathematical functions, Boolean operations, and conditional statements. This allows users to define their own variables for use as outputs, constraints, or objective functions. The user-defined outputs can directly interface with other programs, such as spreadsheets, plotting packages, and visualization programs. Internally, OTIS4 has more explicit and implicit integration procedures, including high-order collocation methods, the pseudo-spectral method, and several variations of multiple shooting. Users may switch easily between the various methods. Several unique numerical techniques such as automated variable scaling and implicit integration grid refinement, support the integration methods. OTIS4 is also significantly more user friendly than previous versions. The installation process is nearly identical on various platforms, including Microsoft Windows, Apple OS X, and Linux operating systems. Cross-platform scripts also help make the execution of OTIS and post-processing of data easier. OTIS4 is supplied free by NASA and is subject to ITAR (International Traffic in Arms Regulations) restrictions. Users must have a Fortran compiler, and a Python interpreter is highly recommended.

  3. HOMER® Micropower Optimization Model

    SciTech Connect

    Lilienthal, P.

    2005-01-01

    NREL has developed the HOMER micropower optimization model. The model can analyze all of the available small power technologies individually and in hybrid configurations to identify least-cost solutions to energy requirements. This capability is valuable to a diverse set of energy professionals and applications. NREL has actively supported its growing user base and developed training programs around the model. These activities are helping to grow the global market for solar technologies.

  4. Optimal Centroid Position Estimation

    SciTech Connect

    Candy, J V; McClay, W A; Awwal, A S; Ferguson, S W

    2004-07-23

    The alignment of high energy laser beams for potential fusion experiments demand high precision and accuracy by the underlying positioning algorithms. This paper discusses the feasibility of employing online optimal position estimators in the form of model-based processors to achieve the desired results. Here we discuss the modeling, development, implementation and processing of model-based processors applied to both simulated and actual beam line data.

  5. Structural optimization of framed structures using generalized optimality criteria

    NASA Technical Reports Server (NTRS)

    Kolonay, R. M.; Venkayya, Vipperla B.; Tischler, V. A.; Canfield, R. A.

    1989-01-01

    The application of a generalized optimality criteria to framed structures is presented. The optimality conditions, Lagrangian multipliers, resizing algorithm, and scaling procedures are all represented as a function of the objective and constraint functions along with their respective gradients. The optimization of two plane frames under multiple loading conditions subject to stress, displacement, generalized stiffness, and side constraints is presented. These results are compared to those found by optimizing the frames using a nonlinear mathematical programming technique.

  6. Flood Bypass Capacity Optimization

    NASA Astrophysics Data System (ADS)

    Siclari, A.; Hui, R.; Lund, J. R.

    2015-12-01

    Large river flows can damage adjacent flood-prone areas, by exceeding river channel and levee capacities. Particularly large floods are difficult to contain in leveed river banks alone. Flood bypasses often can efficiently reduce flood risks, where excess river flow is diverted over a weir to bypasses, that incur much less damage and cost. Additional benefits of bypasses include ecosystem protection, agriculture, groundwater recharge and recreation. Constructing or expanding an existing bypass costs in land purchase easements, and levee setbacks. Accounting for such benefits and costs, this study develops a simple mathematical model for optimizing flood bypass capacity using benefit-cost and risk analysis. Application to the Yolo Bypass, an existing bypass along the Sacramento River in California, estimates optimal capacity that economically reduces flood damage and increases various benefits, especially for agriculture. Land availability is likely to limit bypass expansion. Compensation for landowners could relax such limitations. Other economic values could affect the optimal results, which are shown by sensitivity analysis on major parameters. By including land geography into the model, location of promising capacity expansions can be identified.

  7. Optimization of neutron source

    SciTech Connect

    Hooper, E.B.

    1993-11-09

    I consider here the optimization of the two component neutron source, allowing beam species and energy to vary. A simple model is developed, based on the earlier publications, that permits the optimum to be obtained simply. The two component plasma, with one species of hot ion (D{sup +} or T{sup +}) and the complementary species of cold ion, is easy to analyze in the case of a spatially uniform cold plasma, as to good approximation the total number of hot ions is important but not their spatial distribution. Consequently, the optimization can ignore spatial effects. The problem of a plasma with both types of hot ions and cold ions is rather more difficult, as the neutron production by hot-hot interactions is sensitive to their spatial distributions. Consequently, consideration of this problem will be delayed to a future memorandum. The basic model is that used in the published articles on the two-component, beam-plasma mirror source. I integrate the Fokker-Planck equation analytically, obtaining good agreement with previous numerical results. This simplifies the optimization, by providing a functional form for the neutron production. The primary result is expressed in terms of the power efficiency: watts of neutrons/watts of primary power. The latter includes the positive ion neutralization efficiency. At 150 keV, the present model obtains an efficiency of 0.66%, compared with 0.53% of the earlier calculation.

  8. Scanner matching optimization

    NASA Astrophysics Data System (ADS)

    Kupers, Michiel; Klingbeil, Patrick; Tschischgale, Joerg; Buhl, Stefan; Hempel, Fritjof

    2009-03-01

    Cost of ownership of scanners for the manufacturing of front end layers is becoming increasingly expensive. The ability to quickly switch the production of a layer to another scanner in case it is down is important. This paper presents a method to match the scanner grids in the most optimal manner so that use of front end scanners in effect becomes interchangeable. A breakdown of the various components of overlay is given and we discuss methods to optimize the matching strategy in the fab. A concern here is how to separate the scanner and process induced effects. We look at the relative contributions of intrafield and interfield errors caused by the scanner and the process. Experimental results of a method to control the scanner grid are presented and discussed. We compare the overlay results before and after optimizing the scanner grids and show that the matching penalty is reduced by 20%. We conclude with some thoughts on the need to correct the remaining matching errors.

  9. Very Large Scale Optimization

    NASA Technical Reports Server (NTRS)

    Vanderplaats, Garrett; Townsend, James C. (Technical Monitor)

    2002-01-01

    The purpose of this research under the NASA Small Business Innovative Research program was to develop algorithms and associated software to solve very large nonlinear, constrained optimization tasks. Key issues included efficiency, reliability, memory, and gradient calculation requirements. This report describes the general optimization problem, ten candidate methods, and detailed evaluations of four candidates. The algorithm chosen for final development is a modern recreation of a 1960s external penalty function method that uses very limited computer memory and computational time. Although of lower efficiency, the new method can solve problems orders of magnitude larger than current methods. The resulting BIGDOT software has been demonstrated on problems with 50,000 variables and about 50,000 active constraints. For unconstrained optimization, it has solved a problem in excess of 135,000 variables. The method includes a technique for solving discrete variable problems that finds a "good" design, although a theoretical optimum cannot be guaranteed. It is very scalable in that the number of function and gradient evaluations does not change significantly with increased problem size. Test cases are provided to demonstrate the efficiency and reliability of the methods and software.

  10. Structural optimization of large structural systems by optimality criteria methods

    NASA Technical Reports Server (NTRS)

    Berke, Laszlo

    1992-01-01

    The fundamental concepts of the optimality criteria method of structural optimization are presented. The effect of the separability properties of the objective and constraint functions on the optimality criteria expressions is emphasized. The single constraint case is treated first, followed by the multiple constraint case with a more complex evaluation of the Lagrange multipliers. Examples illustrate the efficiency of the method.

  11. Infrared Drying Parameter Optimization

    NASA Astrophysics Data System (ADS)

    Jackson, Matthew R.

    In recent years, much research has been done to explore direct printing methods, such as screen and inkjet printing, as alternatives to the traditional lithographic process. The primary motivation is reduction of the material costs associated with producing common electronic devices. Much of this research has focused on developing inkjet or screen paste formulations that can be printed on a variety of substrates, and which have similar conductivity performance to the materials currently used in the manufacturing of circuit boards and other electronic devices. Very little research has been done to develop a process that would use direct printing methods to manufacture electronic devices in high volumes. This study focuses on developing and optimizing a drying process for conductive copper ink in a high volume manufacturing setting. Using an infrared (IR) dryer, it was determined that conductive copper prints could be dried in seconds or minutes as opposed to tens of minutes or hours that it would take with other drying devices, such as a vacuum oven. In addition, this study also identifies significant parameters that can affect the conductivity of IR dried prints. Using designed experiments and statistical analysis; the dryer parameters were optimized to produce the best conductivity performance for a specific ink formulation and substrate combination. It was determined that for an ethylene glycol, butanol, 1-methoxy 2- propanol ink formulation printed on Kapton, the optimal drying parameters consisted of a dryer height of 4 inches, a temperature setting between 190 - 200°C, and a dry time of 50-65 seconds depending on the printed film thickness as determined by the number of print passes. It is important to note that these parameters are optimized specifically for the ink formulation and substrate used in this study. There is still much research that needs to be done into optimizing the IR dryer for different ink substrate combinations, as well as developing a

  12. Ames Optimized TCA Configuration

    NASA Technical Reports Server (NTRS)

    Cliff, Susan E.; Reuther, James J.; Hicks, Raymond M.

    1999-01-01

    Configuration design at Ames was carried out with the SYN87-SB (single block) Euler code using a 193 x 49 x 65 C-H grid. The Euler solver is coupled to the constrained (NPSOL) and the unconstrained (QNMDIF) optimization packages. Since the single block grid is able to model only wing-body configurations, the nacelle/diverter effects were included in the optimization process by SYN87's option to superimpose the nacelle/diverter interference pressures on the wing. These interference pressures were calculated using the AIRPLANE code. AIRPLANE is an Euler solver that uses a unstructured tetrahedral mesh and is capable of computations about arbitrary complete configurations. In addition, the buoyancy effects of the nacelle/diverters were also included in the design process by imposing the pressure field obtained during the design process onto the triangulated surfaces of the nacelle/diverter mesh generated by AIRPLANE. The interference pressures and nacelle buoyancy effects are added to the final forces after each flow field calculation. Full details of the (recently enhanced) ghost nacelle capability are given in a related talk. The pseudo nacelle corrections were greatly improved during this design cycle. During the Ref H and Cycle 1 design activities, the nacelles were only translated and pitched. In the cycle 2 design effort the nacelles can translate vertically, and pitch to accommodate the changes in the lower surface geometry. The diverter heights (between their leading and trailing edges) were modified during design as the shape of the lower wing changed, with the drag of the diverter changing accordingly. Both adjoint and finite difference gradients were used during optimization. The adjoint-based gradients were found to give good direction in the design space for configurations near the starting point, but as the design approached a minimum, the finite difference gradients were found to be more accurate. Use of finite difference gradients was limited by the

  13. Turbine Performance Optimization Task Status

    NASA Technical Reports Server (NTRS)

    Griffin, Lisa W.; Turner, James E. (Technical Monitor)

    2001-01-01

    Capability to optimize for turbine performance and accurately predict unsteady loads will allow for increased reliability, Isp, and thrust-to-weight. The development of a fast, accurate aerodynamic design, analysis, and optimization system is required.

  14. Multiobjective optimization of temporal processes.

    PubMed

    Song, Zhe; Kusiak, Andrew

    2010-06-01

    This paper presents a dynamic predictive-optimization framework of a nonlinear temporal process. Data-mining (DM) and evolutionary strategy algorithms are integrated in the framework for solving the optimization model. DM algorithms learn dynamic equations from the process data. An evolutionary strategy algorithm is then applied to solve the optimization problem guided by the knowledge extracted by the DM algorithm. The concept presented in this paper is illustrated with the data from a power plant, where the goal is to maximize the boiler efficiency and minimize the limestone consumption. This multiobjective optimization problem can be either transformed into a single-objective optimization problem through preference aggregation approaches or into a Pareto-optimal optimization problem. The computational results have shown the effectiveness of the proposed optimization framework. PMID:19900853

  15. Combinatorial optimization in foundry practice

    NASA Astrophysics Data System (ADS)

    Antamoshkin, A. N.; Masich, I. S.

    2016-04-01

    The multicriteria mathematical model of foundry production capacity planning is suggested in the paper. The model is produced in terms of pseudo-Boolean optimization theory. Different search optimization methods were used to solve the obtained problem.

  16. Taking Stock of Unrealistic Optimism

    PubMed Central

    Shepperd, James A.; Klein, William M. P.; Waters, Erika A.; Weinstein, Neil D.

    2015-01-01

    Researchers have used terms such as unrealistic optimism and optimistic bias to refer to concepts that are similar but not synonymous. Drawing from three decades of research, we critically discuss how researchers define unrealistic optimism and we identify four types that reflect different measurement approaches: unrealistic absolute optimism at the individual and group level and unrealistic comparative optimism at the individual and group level. In addition, we discuss methodological criticisms leveled against research on unrealistic optimism and note that the criticisms are primarily relevant to only one type—the group form of unrealistic comparative optimism. We further clarify how the criticisms are not nearly as problematic even for unrealistic comparative optimism as they might seem. Finally, we note boundary conditions on the different types of unrealistic optimism and reflect on five broad questions that deserve further attention. PMID:26045714

  17. A Primer on Unrealistic Optimism

    PubMed Central

    Shepperd, James A.; Waters, Erika; Weinstein, Neil D.; Klein, William M. P.

    2014-01-01

    People display unrealistic optimism in their predictions for countless events, believing that their personal future outcomes will be more desirable than can possibly be true. We summarize the vast literature on unrealistic optimism by focusing on four broad questions: What is unrealistic optimism; when does it occur; why does it occur; and what are its consequences. PMID:26089606

  18. Optimization of stall regulated rotors

    SciTech Connect

    Fuglsang, P.L.; Madsen, H.A.

    1995-09-01

    The present work deals with the optimization of stall regulated rotors for wind turbines. Two different optimization methods are presented. The first method is a single design point optimization procedure, whereas the second is a multi pointed optimization technique which is founded on a general optimization algorithm. The use of an optimization algorithm offers the possibility to treat complex optimization problems concerning the entire rotor geometry. The two methods are compared through design of a 20 kW rotor showing good agreement. By use of the optimization algorithm, different aspects of modern wind turbine design layout are investigated. The improvement of the annual energy production by optimizing the airfoil characteristics in addition to the blade chord and twist has been found marginal compared to a case where a standard NACA 634x airfoil family is used. The optimal ratio of swept area to rated power is found depending strongly on the value of the specified maximum loads. Optimization of rotors to specific wind regimes has not been found favorable. In general, the results show that the optimization algorithm is an useful aid to the design.

  19. Optimal Test Construction. Research Report.

    ERIC Educational Resources Information Center

    Veldkamp, Bernard P.

    This paper discusses optimal test construction, which deals with the selection of items from a pool to construct a test that performs optimally with respect to the objective of the test and simultaneously meets all test specifications. Optimal test construction problems can be formulated as mathematical decision models. Algorithms and heuristics…

  20. Metacognitive Control and Optimal Learning

    ERIC Educational Resources Information Center

    Son, Lisa K.; Sethi, Rajiv

    2006-01-01

    The notion of optimality is often invoked informally in the literature on metacognitive control. We provide a precise formulation of the optimization problem and show that optimal time allocation strategies depend critically on certain characteristics of the learning environment, such as the extent of time pressure, and the nature of the uptake…

  1. Optimism and Well-Being

    ERIC Educational Resources Information Center

    Reivich, Karen

    2010-01-01

    Dictionary definitions of optimism encompass two related concepts. The first of these is a hopeful disposition or a conviction that good will ultimately prevail. The second, broader conception of optimism refers to the belief, or the inclination to believe, that the world is the best of all possible worlds. In psychological research, optimism has…

  2. Multicriteria VMAT optimization

    SciTech Connect

    Craft, David; McQuaid, Dualta; Wala, Jeremiah; Chen, Wei; Salari, Ehsan; Bortfeld, Thomas

    2012-02-15

    Purpose: To make the planning of volumetric modulated arc therapy (VMAT) faster and to explore the tradeoffs between planning objectives and delivery efficiency. Methods: A convex multicriteria dose optimization problem is solved for an angular grid of 180 equi-spaced beams. This allows the planner to navigate the ideal dose distribution Pareto surface and select a plan of desired target coverage versus organ at risk sparing. The selected plan is then made VMAT deliverable by a fluence map merging and sequencing algorithm, which combines neighboring fluence maps based on a similarity score and then delivers the merged maps together, simplifying delivery. Successive merges are made as long as the dose distribution quality is maintained. The complete algorithm is called VMERGE. Results: VMERGE is applied to three cases: a prostate, a pancreas, and a brain. In each case, the selected Pareto-optimal plan is matched almost exactly with the VMAT merging routine, resulting in a high quality plan delivered with a single arc in less than 5 min on average. Conclusions: VMERGE offers significant improvements over existing VMAT algorithms. The first is the multicriteria planning aspect, which greatly speeds up planning time and allows the user to select the plan, which represents the most desirable compromise between target coverage and organ at risk sparing. The second is the user-chosen epsilon-optimality guarantee of the final VMAT plan. Finally, the user can explore the tradeoff between delivery time and plan quality, which is a fundamental aspect of VMAT that cannot be easily investigated with current commercial planning systems.

  3. Optimal heat transport

    NASA Astrophysics Data System (ADS)

    Souza, Andre; Doering, Charles R.

    2015-11-01

    The transport of heat by buoyancy driven flows, i.e., thermal convection plays a central role in many natural phenomena and an understanding of how to control its mechanisms is relevant to many engineering applications. In this talk we will consider a variational formulation of optimal heat transport in simple geometries. Numerical results, limits on heat transport, and a comparison to Rayleigh-Bénard convection will be presented. Research supported by NSF Awards PHY-1205219, PHY-1338407, PHY-1443836, PHY-1533555 and DMS-1515161.

  4. Nonconvex optimization and jamming

    NASA Astrophysics Data System (ADS)

    Kallus, Yoav

    Recent work on the jamming transition of particles with short-range interactions has drawn connections with models based on minimization problems with linear inequality constraints and a concave objective. These properties reduce the continuous optimization problem to a discrete search among the corners of the feasible polytope. I will discuss results from simulations of models with and without quenched disorder, exhibiting critical power laws, scaling collapse, and protocol dependence. These models are also well-suited for study using tools of algebraic topology, which I will discuss briefly. Supported by an Omidyar Fellowship at the Santa Fe Institute.

  5. Constructing optimal entanglement witnesses

    SciTech Connect

    Chruscinski, Dariusz; Pytel, Justyna; Sarbicki, Gniewomir

    2009-12-15

    We provide a class of indecomposable entanglement witnesses. In 4x4 case, it reproduces the well-known Breuer-Hall witness. We prove that these witnesses are optimal and atomic, i.e., they are able to detect the 'weakest' quantum entanglement encoded into states with positive partial transposition. Equivalently, we provide a construction of indecomposable atomic maps in the algebra of 2kx2k complex matrices. It is shown that their structural physical approximations give rise to entanglement breaking channels. This result supports recent conjecture by Korbicz et al. [Phys. Rev. A 78, 062105 (2008)].

  6. Optimized hydrogen piston engines

    SciTech Connect

    Smith, J.R.

    1994-05-10

    Hydrogen piston engines can be simultaneously optimized for improved thermal efficiency and for extremely low emissions. Using these engines in constant-speed, constant-load systems such as series hybrid-electric automobiles or home cogeneration systems can result in significantly improved energy efficiency. For the same electrical energy produced, the emissions from such engines can be comparable to those from natural gas-fired steam power plants. These hydrogen-fueled high-efficiency, low-emission (HELE) engines are a mechanical equivalent of hydrogen fuel cells. HELE engines could facilitate the transition to a hydrogen fuel cell economy using near-term technology.

  7. Optimization of radiation protection

    SciTech Connect

    Lochard, J.

    1981-07-01

    The practical and theoretical problems raised by the optimization of radiological protection merit a review of decision-making methods, their relevance, and the way in which they are used in order to better determine what role they should play in the decision-making process. Following a brief summary of the theoretical background of the cost-benefit analysis, we examine the methodological choices implicit in the model presented in the International Commission on Radiological Protection Publication No. 26 and, particularly, the consequences of the theory that the level of radiation protection, the benefits, and the production costs of an activity can be treated separately.

  8. Design Optimization Toolkit: Users' Manual

    SciTech Connect

    Aguilo Valentin, Miguel Alejandro

    2014-07-01

    The Design Optimization Toolkit (DOTk) is a stand-alone C++ software package intended to solve complex design optimization problems. DOTk software package provides a range of solution methods that are suited for gradient/nongradient-based optimization, large scale constrained optimization, and topology optimization. DOTk was design to have a flexible user interface to allow easy access to DOTk solution methods from external engineering software packages. This inherent flexibility makes DOTk barely intrusive to other engineering software packages. As part of this inherent flexibility, DOTk software package provides an easy-to-use MATLAB interface that enables users to call DOTk solution methods directly from the MATLAB command window.

  9. An improved cockroach swarm optimization.

    PubMed

    Obagbuwa, I C; Adewumi, A O

    2014-01-01

    Hunger component is introduced to the existing cockroach swarm optimization (CSO) algorithm to improve its searching ability and population diversity. The original CSO was modelled with three components: chase-swarming, dispersion, and ruthless; additional hunger component which is modelled using partial differential equation (PDE) method is included in this paper. An improved cockroach swarm optimization (ICSO) is proposed in this paper. The performance of the proposed algorithm is tested on well known benchmarks and compared with the existing CSO, modified cockroach swarm optimization (MCSO), roach infestation optimization RIO, and hungry roach infestation optimization (HRIO). The comparison results show clearly that the proposed algorithm outperforms the existing algorithms. PMID:24959611

  10. Synthesizing optimal waste blends

    SciTech Connect

    Narayan, V.; Diwekar, W.M.; Hoza, M.

    1996-10-01

    Vitrification of tank wastes to form glass is a technique that will be used for the disposal of high-level waste at Hanford. Process and storage economics show that minimizing the total number of glass logs produced is the key to keeping cost as low as possible. The amount of glass produced can be reduced by blending of the wastes. The optimal way to combine the tanks to minimize the vole of glass can be determined from a discrete blend calculation. However, this problem results in a combinatorial explosion as the number of tanks increases. Moreover, the property constraints make this problem highly nonconvex where many algorithms get trapped in local minima. In this paper the authors examine the use of different combinatorial optimization approaches to solve this problem. A two-stage approach using a combination of simulated annealing and nonlinear programming (NLP) is developed. The results of different methods such as the heuristics approach based on human knowledge and judgment, the mixed integer nonlinear programming (MINLP) approach with GAMS, and branch and bound with lower bound derived from the structure of the given blending problem are compared with this coupled simulated annealing and NLP approach.

  11. Optimal Image Subtraction

    NASA Astrophysics Data System (ADS)

    Gal-Yam, Avishay; Zackay, Barak; Ofek, Eran O.

    2016-01-01

    Transient detection and flux measurement via image subtraction are fundamental to time domain astronomy. Starting from first principles, we develop the optimal linear statistic for transient detection and flux measurement and any other image-difference hypothesis testing. We derive a simple closed-form statistic that: (1) Is mathematically proven to be the optimal subtraction statistic in the limit of background-dominated noise within the family of linear solutions, that contains all previously suggested methods; (2) Does not leave subtraction or deconvolution artifacts, even in the vicinity of bright stars; (3) Is an order of magnitude faster to compute than popular methods; (4) Allowsautomatic transient detection down to the theoretical sensitivity limit by providing a reliable, well-defined detection significance; (5) Is symmetric to the interchange of the new and reference images; (6) Is numerically stable; and (7) Is trivial to implement. We demonstratethat the correct way to prepare a reference image is the proper image co-addition presented in Zackay & Ofek 2015. Finally, we show a proper image subtraction statistic, that, along with its point spread functions, is a sufficient statistic for any decision or measurement on the difference image. This allows accurate filtration of image artifacts such as cosmic rays and hot pixels. We demonstrate this method on simulated data as well as on observations from the Palomar Transient Factory.

  12. Optimization Methods in Sherpa

    NASA Astrophysics Data System (ADS)

    Siemiginowska, Aneta; Nguyen, Dan T.; Doe, Stephen M.; Refsdal, Brian L.

    2009-09-01

    Forward fitting is a standard technique used to model X-ray data. A statistic, usually assumed weighted chi^2 or Poisson likelihood (e.g. Cash), is minimized in the fitting process to obtain a set of the best model parameters. Astronomical models often have complex forms with many parameters that can be correlated (e.g. an absorbed power law). Minimization is not trivial in such setting, as the statistical parameter space becomes multimodal and finding the global minimum is hard. Standard minimization algorithms can be found in many libraries of scientific functions, but they are usually focused on specific functions. However, Sherpa designed as general fitting and modeling application requires very robust optimization methods that can be applied to variety of astronomical data (X-ray spectra, images, timing, optical data etc.). We developed several optimization algorithms in Sherpa targeting a wide range of minimization problems. Two local minimization methods were built: Levenberg-Marquardt algorithm was obtained from MINPACK subroutine LMDIF and modified to achieve the required robustness; and Nelder-Mead simplex method has been implemented in-house based on variations of the algorithm described in the literature. A global search Monte-Carlo method has been implemented following a differential evolution algorithm presented by Storn and Price (1997). We will present the methods in Sherpa and discuss their usage cases. We will focus on the application to Chandra data showing both 1D and 2D examples. This work is supported by NASA contract NAS8-03060 (CXC).

  13. E85 Optimized Engine

    SciTech Connect

    Bower, Stanley

    2011-12-31

    A 5.0L V8 twin-turbocharged direct injection engine was designed, built, and tested for the purpose of assessing the fuel economy and performance in the F-Series pickup of the Dual Fuel engine concept and of an E85 optimized FFV engine. Additionally, production 3.5L gasoline turbocharged direct injection (GTDI) EcoBoost engines were converted to Dual Fuel capability and used to evaluate the cold start emissions and fuel system robustness of the Dual Fuel engine concept. Project objectives were: to develop a roadmap to demonstrate a minimized fuel economy penalty for an F-Series FFV truck with a highly boosted, high compression ratio spark ignition engine optimized to run with ethanol fuel blends up to E85; to reduce FTP 75 energy consumption by 15% - 20% compared to an equally powered vehicle with a current production gasoline engine; and to meet ULEV emissions, with a stretch target of ULEV II / Tier II Bin 4. All project objectives were met or exceeded.

  14. Optimal Synchronizability of Bearings

    NASA Astrophysics Data System (ADS)

    Araújo, N. A. M.; Seybold, H.; Baram, R. M.; Herrmann, H. J.; Andrade, J. S., Jr.

    2013-02-01

    Bearings are mechanical dissipative systems that, when perturbed, relax toward a synchronized (bearing) state. Here we find that bearings can be perceived as physical realizations of complex networks of oscillators with asymmetrically weighted couplings. Accordingly, these networks can exhibit optimal synchronization properties through fine-tuning of the local interaction strength as a function of node degree [Motter, Zhou, and Kurths, Phys. Rev. E 71, 016116 (2005)PLEEE81539-3755]. We show that, in analogy, the synchronizability of bearings can be maximized by counterbalancing the number of contacts and the inertia of their constituting rotor disks through the mass-radius relation, m˜rα, with an optimal exponent α=α× which converges to unity for a large number of rotors. Under this condition, and regardless of the presence of a long-tailed distribution of disk radii composing the mechanical system, the average participation per disk is maximized and the energy dissipation rate is homogeneously distributed among elementary rotors.

  15. Code Optimization Techniques

    SciTech Connect

    MAGEE,GLEN I.

    2000-08-03

    Computers transfer data in a number of different ways. Whether through a serial port, a parallel port, over a modem, over an ethernet cable, or internally from a hard disk to memory, some data will be lost. To compensate for that loss, numerous error detection and correction algorithms have been developed. One of the most common error correction codes is the Reed-Solomon code, which is a special subset of BCH (Bose-Chaudhuri-Hocquenghem) linear cyclic block codes. In the AURA project, an unmanned aircraft sends the data it collects back to earth so it can be analyzed during flight and possible flight modifications made. To counter possible data corruption during transmission, the data is encoded using a multi-block Reed-Solomon implementation with a possibly shortened final block. In order to maximize the amount of data transmitted, it was necessary to reduce the computation time of a Reed-Solomon encoding to three percent of the processor's time. To achieve such a reduction, many code optimization techniques were employed. This paper outlines the steps taken to reduce the processing time of a Reed-Solomon encoding and the insight into modern optimization techniques gained from the experience.

  16. Diffusion with optimal resetting

    NASA Astrophysics Data System (ADS)

    Evans, Martin R.; Majumdar, Satya N.

    2011-10-01

    We consider the mean time to absorption by an absorbing target of a diffusive particle with the addition of a process whereby the particle is reset to its initial position with rate r. We consider several generalizations of the model of Evans and Majumdar (2011 Phys. Rev. Lett.106 160601): (i) a space-dependent resetting rate r(x); (ii) resetting to a random position z drawn from a resetting distribution { P}(z); and (iii) a spatial distribution for the absorbing target PT(x). As an example of (i) we show that the introduction of a non-resetting window around the initial position can reduce the mean time to absorption provided that the initial position is sufficiently far from the target. We address the problem of optimal resetting, that is, minimizing the mean time to absorption for a given target distribution. For an exponentially decaying target distribution centred at the origin we show that a transition in the optimal resetting distribution occurs as the target distribution narrows.

  17. Optimization of inclusive fitness.

    PubMed

    Grafen, Alan

    2006-02-01

    The first fully explicit argument is given that broadly supports a widespread belief among whole-organism biologists that natural selection tends to lead to organisms acting as if maximizing their inclusive fitness. The use of optimization programs permits a clear statement of what this belief should be understood to mean, in contradistinction to the common mathematical presumption that it should be formalized as some kind of Lyapunov or even potential function. The argument reveals new details and uncovers latent assumptions. A very general genetic architecture is allowed, and there is arbitrary uncertainty. However, frequency dependence of fitnesses is not permitted. The logic of inclusive fitness immediately draws together various kinds of intra-genomic conflict, and the concept of 'p-family' is introduced. Inclusive fitness is thus incorporated into the formal Darwinism project, which aims to link the mathematics of motion (difference and differential equations) used to describe gene frequency trajectories with the mathematics of optimization used to describe purpose and design. Important questions remain to be answered in the fundamental theory of inclusive fitness. PMID:16046225

  18. Parametric Design Optimization By Integrating CAD Systems And Optimization Tools

    NASA Astrophysics Data System (ADS)

    Rehan, M.; Olabi, A. G.

    2009-11-01

    Designing a cost effective product in minimum time is a complex process. In order to achieve this goal the requirement of optimum designs are becoming more important. One of the time consuming factor in the design optimization cycle is the modifications of Computer Aided Design (CAD) model after optimization. In conventional design optimization techniques the design engineer has to update the CAD model after receiving optimum design from optimization tools. It is worthwhile using parametric design optimization process to minimize the optimization cycle time. This paper presents a comprehensive study to integrate the optimization parameters between CAD system and optimization tools which were driven from a single user environment. Finally, design optimization of a Compressed Natural Gas (CNG) cylinder was implemented as case study. In this case study the optimization tools were fully integrated with CAD system, therefore, all the deliverables including; part design, drawings and assembly can be automatically updated after achieving the optimum geometry having minimum volume and satisfying all imposed constraints.

  19. Particle swarm optimization for complex nonlinear optimization problems

    NASA Astrophysics Data System (ADS)

    Alexandridis, Alex; Famelis, Ioannis Th.; Tsitouras, Charalambos

    2016-06-01

    This work presents the application of a technique belonging to evolutionary computation, namely particle swarm optimization (PSO), to complex nonlinear optimization problems. To be more specific, a PSO optimizer is setup and applied to the derivation of Runge-Kutta pairs for the numerical solution of initial value problems. The effect of critical PSO operational parameters on the performance of the proposed scheme is thoroughly investigated.

  20. Ultimate open pit stochastic optimization

    NASA Astrophysics Data System (ADS)

    Marcotte, Denis; Caron, Josiane

    2013-02-01

    Classical open pit optimization (maximum closure problem) is made on block estimates, without directly considering the block grades uncertainty. We propose an alternative approach of stochastic optimization. The stochastic optimization is taken as the optimal pit computed on the block expected profits, rather than expected grades, computed from a series of conditional simulations. The stochastic optimization generates, by construction, larger ore and waste tonnages than the classical optimization. Contrary to the classical approach, the stochastic optimization is conditionally unbiased for the realized profit given the predicted profit. A series of simulated deposits with different variograms are used to compare the stochastic approach, the classical approach and the simulated approach that maximizes expected profit among simulated designs. Profits obtained with the stochastic optimization are generally larger than the classical or simulated pit. The main factor controlling the relative gain of stochastic optimization compared to classical approach and simulated pit is shown to be the information level as measured by the boreholes spacing/range ratio. The relative gains of the stochastic approach over the classical approach increase with the treatment costs but decrease with mining costs. The relative gains of the stochastic approach over the simulated pit approach increase both with the treatment and mining costs. At early stages of an open pit project, when uncertainty is large, the stochastic optimization approach appears preferable to the classical approach or the simulated pit approach for fair comparison of the values of alternative projects and for the initial design and planning of the open pit.

  1. Desalination Plant Optimization

    Energy Science and Technology Software Center (ESTSC)

    1992-10-01

    MSF21 and VTE21 perform design and costing calculations for multistage flash evaporator (MSF) and multieffect vertical tube evaporator (VTE) desalination plants. An optimization capability is available, if desired. The MSF plant consists of a recovery section, reject section, brine heater, and associated buildings and equipment. Operating costs and direct and indirect capital costs for plant, buildings, site, and intakes are calculated. Computations are based on the first and last stages of each section and amore » typical middle recovery stage. As a result, the program runs rapidly but does not give stage by stage parameters. The VTE plant consists of vertical tube effects, multistage flash preheater, condenser, and brine heater and associated buildings and equipment. Design computations are done for each vertical tube effect, but preheater computations are based on the first and last stages and a typical middle stage.« less

  2. [Stress and optimal ageing].

    PubMed

    Gogol, Manfred

    2015-08-01

    Stress is a stimulus or incident which has an exogenic or endogenic influence on an organism and leads to a biological and/or psychological adaptation from the organism by adaptation. Stressors can be differentiated by the temporal impact (e.g. acute, chronic or acute on chronic), strength and quality. The consequences of stress exposure and adaptation can be measured at the cellular level and as (sub) clinical manifestations, where this process can be biologically seen as a continuum. Over the course of life there is an accumulation of stress incidents resulting in a diminution of the capability for adaptation and repair mechanisms. By means of various interventions it is possible to improve the individual capability for adaptation but it is not currently definitively possible to disentangle alterations due to ageing and the development of diseases. As a consequence the term "healthy ageing" should be replaced by the concept of "optimal ageing". PMID:26208575

  3. DENSE MEDIA CYCLONE OPTIMIZATION

    SciTech Connect

    Gerald H. Luttrell

    2002-04-11

    The test data obtained from the Baseline Assessment that compares the performance of the density traces to that of different sizes of coal particles is now complete. The experimental results show that the tracer data can indeed be used to accurately predict HMC performance. The following conclusions were drawn: (i) the tracer curve is slightly sharper than curve for coarsest size fraction of coal (probably due to the greater resolution of the tracer technique), (ii) the Ep increases with decreasing coal particle size, and (iii) the Ep values are not excessively large for the well-maintained HMC circuits. The major problems discovered were associated with improper apex-to-vortex finder ratios and particle hang-up due to media segregation. Only one plant yielded test data that were typical of a fully optimized level of performance.

  4. Cyclone performance and optimization

    SciTech Connect

    Leith, D.

    1989-06-15

    The objectives of this project are: to characterize the gas flow pattern within cyclones, to revise the theory for cyclone performance on the basis of these findings, and to design and test cyclones whose dimensions have been optimized using revised performance theory. This work is important because its successful completion will aid in the technology for combustion of coal in pressurized, fluidized beds. We have now received all the equipment necessary for the flow visualization studies described over the last two progress reports. We have begun more detailed studies of the gas flow pattern within cyclones as detailed below. Third, we have begun studies of the effect of particle concentration on cyclone performance. This work is critical to application of our results to commercial operations. 1 fig.

  5. DENSE MEDIA CYCLONE OPTIMIZATION

    SciTech Connect

    Gerald H. Luttrell

    2002-01-14

    During the past quarter, float-sink analyses were completed for four of seven circuits evaluated in this project. According to the commercial laboratory, the analyses for the remaining three sites will be finished by mid February 2002. In addition, it was necessary to repeat several of the float-sink tests to resolve problems identified during the analysis of the experimental data. In terms of accomplishments, a website is being prepared to distribute project findings and software to the public. This site will include (i) an operators manual for HMC operation and maintenance (already available in hard copy), (ii) an expert system software package for evaluating and optimizing HMC performance (in development), and (iii) a spreadsheet-based process model for plant designers (in development). Several technology transfer activities were also carried out including the publication of project results in proceedings and the training of plant operations via workshops.

  6. Optimizing process vacuum condensers

    SciTech Connect

    Lines, J.R.; Tice, D.W.

    1997-09-01

    Vacuum condensers play a critical role in supporting vacuum processing operations. Although they may appear similar to atmospheric units, vacuum condensers have their own special designs, considerations and installation needs. By adding vacuum condensers, precondensers and intercondensers, system cost efficiency can be optimized. Vacuum-condensing systems permit reclamation of high-value product by use of a precondenser, or reduce operating costs with intercondensers. A precondenser placed between the vacuum vessel and ejector system will recover valuable process vapors and reduce vapor load to an ejector system--minimizing the system`s capital and operating costs. Similarly, an intercondenser positioned between ejector stages can condense motive steam and process vapors and reduce vapor load to downstream ejectors as well as lower capital and operating costs. The paper describes vacuum condenser systems, types of vacuum condensers, shellside condensing, tubeside condensing, noncondensable gases, precondenser pressure drop, system interdependency, equipment installation, and equipment layout.

  7. Optimization of electrolytic cells

    SciTech Connect

    Alkire, R.; Soon, S-A.; Sradther, M.

    1985-05-01

    A methodology was developed for optimizing electrolytic cells described by a potential field distribution along with material, voltage, and economic balance equations. In the present study, the cell consisted of two flow-through porous electrodes separated by a membrane. The model consisted of two nonlinear differential equations, 1 variables, eight equality constraints, and five inequality constraints. The optimum solutions were obtained for simple economic objectives with use of a successive quadratic programming method. The sensitivity of the optimum to operating variables and design constraints was found with the use of Lagrange multipliers. The method may be applied to an electrolytic cell which can be modeled by a combination of differential, algebraic, and polynomial (curve-fit) equations.

  8. Optimal traps in graphene

    NASA Astrophysics Data System (ADS)

    Downing, C. A.; Pearce, A. R.; Churchill, R. J.; Portnoi, M. E.

    2015-10-01

    We transform the two-dimensional Dirac-Weyl equation, which governs the charge carriers in graphene, into a nonlinear first-order differential equation for scattering phase shift, using the so-called variable-phase method. This allows us to utilize the Levinson theorem, relating scattering phase shifts of a slow particle to its bound states, to find zero-energy bound states created electrostatically in realistic structures. These confined states are formed at critical potential strengths, which leads us to posit the use of "optimal traps" to combat the chiral tunneling found in graphene: this could be explored experimentally with an artificial network of point charges held above the graphene layer. We also discuss scattering on these states and find that the s states create a dominant peak in the scattering cross section as the energy tends towards the Dirac point energy, suggesting a dominant contribution to the resistivity.

  9. Optimally interacting minds

    PubMed Central

    Bahrami, Bahador; Olsen, Karsten; Latham, Peter E.; Roepstorff, Andreas; Rees, Geraint; Frith, Chris D.

    2012-01-01

    In everyday life, many believe that ‘two heads are better than one’. Indeed, our ability to solve problems together appears to be fundamental to the current dominance, and future survival, of the human species. But are two heads really better than one? We addressed this question in the context of a collective low-level perceptual decision-making task. For two observers of nearly equal sensitivity, two heads were definitely better than one, provided that they were given the opportunity to communicate freely, even in the absence of any feedback about decision outcomes. But for observers with very different sensitivities, two heads were worse than the better one. These seemingly discrepant patterns of group behaviour can be explained by a model in which two heads are Bayes optimal under the assumption that individuals accurately communicate their level of confidence on every trial. PMID:20798320

  10. Optimality in Data Assimilation

    NASA Astrophysics Data System (ADS)

    Nearing, Grey; Yatheendradas, Soni

    2016-04-01

    It costs a lot more to develop and launch an earth-observing satellite than it does to build a data assimilation system. As such, we propose that it is important to understand the efficiency of our assimilation algorithms at extracting information from remote sensing retrievals. To address this, we propose that it is necessary to adopt completely general definition of "optimality" that explicitly acknowledges all differences between the parametric constraints of our assimilation algorithm (e.g., Gaussianity, partial linearity, Markovian updates) and the true nature of the environmetnal system and observing system. In fact, it is not only possible, but incredibly straightforward, to measure the optimality (in this more general sense) of any data assimilation algorithm as applied to any intended model or natural system. We measure the information content of remote sensing data conditional on the fact that we are already running a model and then measure the actual information extracted by data assimilation. The ratio of the two is an efficiency metric, and optimality is defined as occurring when the data assimilation algorithm is perfectly efficient at extracting information from the retrievals. We measure the information content of the remote sensing data in a way that, unlike triple collocation, does not rely on any a priori presumed relationship (e.g., linear) between the retrieval and the ground truth, however, like triple-collocation, is insensitive to the spatial mismatch between point-based measurements and grid-scale retrievals. This theory and method is therefore suitable for use with both dense and sparse validation networks. Additionally, the method we propose is *constructive* in the sense that it provides guidance on how to improve data assimilation systems. All data assimilation strategies can be reduced to approximations of Bayes' law, and we measure the fractions of total information loss that are due to individual assumptions or approximations in the

  11. Optimized nanoporous materials.

    SciTech Connect

    Braun, Paul V.; Langham, Mary Elizabeth; Jacobs, Benjamin W.; Ong, Markus D.; Narayan, Roger J.; Pierson, Bonnie E.; Gittard, Shaun D.; Robinson, David B.; Ham, Sung-Kyoung; Chae, Weon-Sik; Gough, Dara V.; Wu, Chung-An Max; Ha, Cindy M.; Tran, Kim L.

    2009-09-01

    Nanoporous materials have maximum practical surface areas for electrical charge storage; every point in an electrode is within a few atoms of an interface at which charge can be stored. Metal-electrolyte interfaces make best use of surface area in porous materials. However, ion transport through long, narrow pores is slow. We seek to understand and optimize the tradeoff between capacity and transport. Modeling and measurements of nanoporous gold electrodes has allowed us to determine design principles, including the fact that these materials can deplete salt from the electrolyte, increasing resistance. We have developed fabrication techniques to demonstrate architectures inspired by these principles that may overcome identified obstacles. A key concept is that electrodes should be as close together as possible; this is likely to involve an interpenetrating pore structure. However, this may prove extremely challenging to fabricate at the finest scales; a hierarchically porous structure can be a worthy compromise.

  12. OPTIMAL NETWORK TOPOLOGY DESIGN

    NASA Technical Reports Server (NTRS)

    Yuen, J. H.

    1994-01-01

    This program was developed as part of a research study on the topology design and performance analysis for the Space Station Information System (SSIS) network. It uses an efficient algorithm to generate candidate network designs (consisting of subsets of the set of all network components) in increasing order of their total costs, and checks each design to see if it forms an acceptable network. This technique gives the true cost-optimal network, and is particularly useful when the network has many constraints and not too many components. It is intended that this new design technique consider all important performance measures explicitly and take into account the constraints due to various technical feasibilities. In the current program, technical constraints are taken care of by the user properly forming the starting set of candidate components (e.g. nonfeasible links are not included). As subsets are generated, they are tested to see if they form an acceptable network by checking that all requirements are satisfied. Thus the first acceptable subset encountered gives the cost-optimal topology satisfying all given constraints. The user must sort the set of "feasible" link elements in increasing order of their costs. The program prompts the user for the following information for each link: 1) cost, 2) connectivity (number of stations connected by the link), and 3) the stations connected by that link. Unless instructed to stop, the program generates all possible acceptable networks in increasing order of their total costs. The program is written only to generate topologies that are simply connected. Tests on reliability, delay, and other performance measures are discussed in the documentation, but have not been incorporated into the program. This program is written in PASCAL for interactive execution and has been implemented on an IBM PC series computer operating under PC DOS. The disk contains source code only. This program was developed in 1985.

  13. Optimal management strategies in variable environments: Stochastic optimal control methods

    USGS Publications Warehouse

    Williams, B.K.

    1985-01-01

    Dynamic optimization was used to investigate the optimal defoliation of salt desert shrubs in north-western Utah. Management was formulated in the context of optimal stochastic control theory, with objective functions composed of discounted or time-averaged biomass yields. Climatic variability and community patterns of salt desert shrublands make the application of stochastic optimal control both feasible and necessary. A primary production model was used to simulate shrub responses and harvest yields under a variety of climatic regimes and defoliation patterns. The simulation results then were used in an optimization model to determine optimal defoliation strategies. The latter model encodes an algorithm for finite state, finite action, infinite discrete time horizon Markov decision processes. Three questions were addressed: (i) What effect do changes in weather patterns have on optimal management strategies? (ii) What effect does the discounting of future returns have? (iii) How do the optimal strategies perform relative to certain fixed defoliation strategies? An analysis was performed for the three shrub species, winterfat (Ceratoides lanata), shadscale (Atriplex confertifolia) and big sagebrush (Artemisia tridentata). In general, the results indicate substantial differences among species in optimal control strategies, which are associated with differences in physiological and morphological characteristics. Optimal policies for big sagebrush varied less with variation in climate, reserve levels and discount rates than did either shadscale or winterfat. This was attributed primarily to the overwintering of photosynthetically active tissue and to metabolic activity early in the growing season. Optimal defoliation of shadscale and winterfat generally was more responsive to differences in plant vigor and climate, reflecting the sensitivity of these species to utilization and replenishment of carbohydrate reserves. Similarities could be seen in the influence of both

  14. GAPS IN SUPPORT VECTOR OPTIMIZATION

    SciTech Connect

    STEINWART, INGO; HUSH, DON; SCOVEL, CLINT; LIST, NICOLAS

    2007-01-29

    We show that the stopping criteria used in many support vector machine (SVM) algorithms working on the dual can be interpreted as primal optimality bounds which in turn are known to be important for the statistical analysis of SVMs. To this end we revisit the duality theory underlying the derivation of the dual and show that in many interesting cases primal optimality bounds are the same as known dual optimality bounds.

  15. Recent developments in multilevel optimization

    NASA Technical Reports Server (NTRS)

    Vanderplaats, Garret N.; Kim, D.-S.

    1989-01-01

    Recent developments in multilevel optimization are briefly reviewed. The general nature of the multilevel design task, the use of approximations to develop and solve the analysis design task, the structure of the formal multidiscipline optimization problem, a simple cantilevered beam which demonstrates the concepts of multilevel design and the basic mathematical details of the optimization task and the system level are among the topics discussed.

  16. Structural optimization by multilevel decomposition

    NASA Technical Reports Server (NTRS)

    Sobieszczanski-Sobieski, J.; James, B.; Dovi, A.

    1983-01-01

    A method is described for decomposing an optimization problem into a set of subproblems and a coordination problem which preserves coupling between the subproblems. The method is introduced as a special case of multilevel, multidisciplinary system optimization and its algorithm is fully described for two level optimization for structures assembled of finite elements of arbitrary type. Numerical results are given for an example of a framework to show that the decomposition method converges and yields results comparable to those obtained without decomposition. It is pointed out that optimization by decomposition should reduce the design time by allowing groups of engineers, using different computers to work concurrently on the same large problem.

  17. Stochastic Optimization of Complex Systems

    SciTech Connect

    Birge, John R.

    2014-03-20

    This project focused on methodologies for the solution of stochastic optimization problems based on relaxation and penalty methods, Monte Carlo simulation, parallel processing, and inverse optimization. The main results of the project were the development of a convergent method for the solution of models that include expectation constraints as in equilibrium models, improvement of Monte Carlo convergence through the use of a new method of sample batch optimization, the development of new parallel processing methods for stochastic unit commitment models, and the development of improved methods in combination with parallel processing for incorporating automatic differentiation methods into optimization.

  18. Structural Optimization in automotive design

    NASA Technical Reports Server (NTRS)

    Bennett, J. A.; Botkin, M. E.

    1984-01-01

    Although mathematical structural optimization has been an active research area for twenty years, there has been relatively little penetration into the design process. Experience indicates that often this is due to the traditional layout-analysis design process. In many cases, optimization efforts have been outgrowths of analysis groups which are themselves appendages to the traditional design process. As a result, optimization is often introduced into the design process too late to have a significant effect because many potential design variables have already been fixed. A series of examples are given to indicate how structural optimization has been effectively integrated into the design process.

  19. RNA based evolutionary optimization

    NASA Astrophysics Data System (ADS)

    Schuster, Peter

    1993-12-01

    . Evolutionary optimization of two-letter sequences in thus more difficult than optimization in the world of natural RNA sequences with four bases. This fact might explain the usage of four bases in the genetic language of nature. Finally we study the mapping from RNA sequences into secondary structures and explore the topology of RNA shape space. We find that ‘neutral paths’ connecting neighbouring sequences with identical structures go very frequently through entire sequence space. Sequences folding into common structures are found everywhere in sequence space. Hence, evolution can migrate to almost every part of sequence space without ‘hill climbing’ and only small fractions of the entire number of sequences have to be searched in order to find suitable structures.

  20. Stability versus Optimality

    NASA Astrophysics Data System (ADS)

    Inanloo, B.

    2011-12-01

    The Caspian Sea is considered to be the largest inland body of water in the world, which located between the Caucasus Mountains and Central Asia. The Caspian Sea has been a source of the most contentious international conflicts between five littoral states now borders the sea: Azerbaijan, Iran, Kazakhstan, Russia, and Turkmenistan. The conflict over the legal status of this international body of water as an aftermath of the breakup of the Soviet Union in 1991. Since then the parties have been negotiating without coming up with any agreement neither on the ownerships of waters, nor the oil and natural gas beneath them. The number of involved stakeholders, the unusual characteristics of the Caspian Sea in considering it as a lake or a sea, and a large number of external parties are interested in the valuable resources of the Sea has made this conflict complex and unique. This paper intends to apply methods to find the best allocation schemes considering acceptability and stability of selected solution to share the Caspian Sea and its resources fairly and efficiently. Although, there are several allocation methods in solving such allocation problems, however, most of those seek a socially optimal solution that can satisfy majority of criteria or decision makers, while, in practice, especially in multi-nation problems, such solution may not be necessarily a stable solution and to be acceptable to all parties. Hence, there is need to apply a method that considers stability and acceptability of solutions to find a solution with high chance to be agreed upon that. Application of some distance-based methods in studying the Caspian Sea conflict provides some policy insights useful for finding solutions that can resolve the dispute. In this study, we use methods such as Goal Programming, Compromise Programming, and considering stability of solution the logic of Power Index is used to find a division rule that is stable negotiators. The results of this study shows that the

  1. Acoustic Radiation Optimization Using the Particle Swarm Optimization Algorithm

    NASA Astrophysics Data System (ADS)

    Jeon, Jin-Young; Okuma, Masaaki

    The present paper describes a fundamental study on structural bending design to reduce noise using a new evolutionary population-based heuristic algorithm called the particle swarm optimization algorithm (PSOA). The particle swarm optimization algorithm is a parallel evolutionary computation technique proposed by Kennedy and Eberhart in 1995. This algorithm is based on the social behavior models for bird flocking, fish schooling and other models investigated by zoologists. Optimal structural design problems to reduce noise are highly nonlinear, so that most conventional methods are difficult to apply. The present paper investigates the applicability of PSOA to such problems. Optimal bending design of a vibrating plate using PSOA is performed in order to minimize noise radiation. PSOA can be effectively applied to such nonlinear acoustic radiation optimization.

  2. Optimizing stem cell culture.

    PubMed

    van der Sanden, Boudewijn; Dhobb, Mehdi; Berger, François; Wion, Didier

    2010-11-01

    Stem cells always balance between self-renewal and differentiation. Hence, stem cell culture parameters are critical and need to be continuously refined according to progress in our stem cell biology understanding and the latest technological developments. In the past few years, major efforts have been made to define more precisely the medium composition in which stem cells grow or differentiate. This led to the progressive replacement of ill-defined additives such as serum or feeder cell layers by recombinant cytokines or growth factors. Another example is the control of the oxygen pressure. For many years cell cultures have been done under atmospheric oxygen pressure which is much higher than the one experienced by stem cells in vivo. A consequence of cell metabolism is that cell culture conditions are constantly changing. Therefore, the development of high sensitive monitoring processes and control algorithms is required for ensuring cell culture medium homeostasis. Stem cells also sense the physical constraints of their microenvironment. Rigidity, stiffness, and geometry of the culture substrate influence stem cell fate. Hence, nanotopography is probably as important as medium formulation in the optimization of stem cell culture conditions. Recent advances include the development of synthetic bioinformative substrates designed at the micro- and nanoscale level. On going research in many different fields including stem cell biology, nanotechnology, and bioengineering suggest that our current way to culture cells in Petri dish or flasks will soon be outdated as flying across the Atlantic Ocean in the Lindbergh's plane. PMID:20803548

  3. Cyclone performance and optimization

    SciTech Connect

    Leith, D.

    1989-03-15

    The objectives of this project are: to characterize the gas flow pattern within cyclones, to revise the theory for cyclone performance on the basis of these findings, and to design and test cyclones whose dimensions have been optimized using revised performance theory. This work is important because its successful completion will aid in the technology for combustion of coal in pressurized, fluidized beds. This quarter, we have been hampered somewhat by flow delivery of the bubble generation system and arc lighting system placed on order last fall. This equipment is necessary to map the flow field within cyclones using the techniques described in last quarter's report. Using the bubble generator, we completed this quarter a study of the natural length'' of cyclones of 18 different configurations, each configuration operated at five different gas flows. Results suggest that the equation by Alexander for natural length is incorrect; natural length as measured with the bubble generation system is always below the bottom of the cyclones regardless of the cyclone configuration or gas flow, within the limits of the experimental cyclones tested. This finding is important because natural length is a term in equations used to predict cyclone efficiency. 1 tab.

  4. The optimal coyness game.

    PubMed

    McNamara, John M; Fromhage, Lutz; Barta, Zoltan; Houston, Alasdair I

    2009-03-01

    In many animal species, females will benefit if they can secure their mate's help in raising their young. It has been suggested that they can achieve this by being coy (i.e. reluctant to mate) when courted, because this gives them time to assess a prospective mate's helpfulness and hence allows them to reject non-helpful males. According to this view, coyness should (i) reflect a trade-off between information gain and time lost on the part of the female, and (ii) be subject to an evolutionary feedback between optimal female coyness and male helping behaviour. Previous theory has considered each of these aspects in isolation. By contrast, here we present a comprehensive game theory model of this situation, leading to qualitatively new insights. We predict that a high degree of coyness should be associated with a high encounter rate during mate search, with an intermediate rate of information gain during mate inspection and with an intermediate dependence of reproduction on male help. Strongly biased sex ratios, however, preclude coyness. Due to the mutual feedback between coyness and helpfulness in our model, alternatively stable evolutionary outcomes (with or without coyness) are possible under broad conditions. We also discuss alternative interpretations of coyness. PMID:19129134

  5. Aluminum laser welding optimization

    NASA Astrophysics Data System (ADS)

    Chmelíčková, Hana; Halenka, Viktor; Lapšanská, Hana; Havelková, Martina

    2007-04-01

    Pulsed Nd:YAG laser with maximal power 150 W is used in our laboratory to cut, drill and weld metal and non-metal thin materials to thickness 2 mm. Welding is realized by fixed processing head or movable fiber one with beam diameter 0,6 mm in focus plane. Welding of stainless and low-carbon steel was tested before and results are publicized and used in practice. Now the goal of our experiment was optimization of process parameters for aluminum that has other physical properties than steels, lower density, higher heat conductivity and surface reflexivity. Pure alumina specimen 0,8 mm and Al-Mg-Si alloy 0,5 mm prepared for butt welds. Problem with surface layer of Al IIO 3 was overcome by sanding and chemical cleaning with grinding paste. Critical parameters for good weld shape are specimen position from beam focus plane, pulse length and energy, pulse frequency and the motion velocity that determines percentage of pulse overlap. Argon as protective gas was used with speed 6 liters per second. Thermal distribution in material can be modeled by numerical simulation. Software tool SYSWELD makes possible to fit laser as surface heat source, define weld geometry, and make meshing of specimen to finite elements and compute heat conduction during process. Color isotherms, vectors, mechanical deformations and others results can be study in post-processing.

  6. Optimized System Identification

    NASA Technical Reports Server (NTRS)

    Juang, Jer-Nan; Longman, Richard W.

    1999-01-01

    In system identification, one usually cares most about finding a model whose outputs are as close as possible to the true system outputs when the same input is applied to both. However, most system identification algorithms do not minimize this output error. Often they minimize model equation error instead, as in typical least-squares fits using a finite-difference model, and it is seen here that this distinction is significant. Here, we develop a set of system identification algorithms that minimize output error for multi-input/multi-output and multi-input/single-output systems. This is done with sequential quadratic programming iterations on the nonlinear least-squares problems, with an eigendecomposition to handle indefinite second partials. This optimization minimizes a nonlinear function of many variables, and hence can converge to local minima. To handle this problem, we start the iterations from the OKID (Observer/Kalman Identification) algorithm result. Not only has OKID proved very effective in practice, it minimizes an output error of an observer which has the property that as the data set gets large, it converges to minimizing the criterion of interest here. Hence, it is a particularly good starting point for the nonlinear iterations here. Examples show that the methods developed here eliminate the bias that is often observed using any system identification methods of either over-estimating or under-estimating the damping of vibration modes in lightly damped structures.

  7. Optimal Phase Oscillatory Network

    NASA Astrophysics Data System (ADS)

    Follmann, Rosangela

    2013-03-01

    Important topics as preventive detection of epidemics, collective self-organization, information flow and systemic robustness in clusters are typical examples of processes that can be studied in the context of the theory of complex networks. It is an emerging theory in a field, which has recently attracted much interest, involving the synchronization of dynamical systems associated to nodes, or vertices, of the network. Studies have shown that synchronization in oscillatory networks depends not only on the individual dynamics of each element, but also on the combination of the topology of the connections as well as on the properties of the interactions of these elements. Moreover, the response of the network to small damages, caused at strategic points, can enhance the global performance of the whole network. In this presentation we explore an optimal phase oscillatory network altered by an additional term in the coupling function. The application to associative-memory network shows improvement on the correct information retrieval as well as increase of the storage capacity. The inclusion of some small deviations on the nodes, when solutions are attracted to a false state, results in additional enhancement of the performance of the associative-memory network. Supported by FAPESP - Sao Paulo Research Foundation, grant number 2012/12555-4

  8. Optimizing haemodialysate composition

    PubMed Central

    Locatelli, Francesco; La Milia, Vincenzo; Violo, Leano; Del Vecchio, Lucia; Di Filippo, Salvatore

    2015-01-01

    Survival and quality of life of dialysis patients are strictly dependent on the quality of the haemodialysis (HD) treatment. In this respect, dialysate composition, including water purity, plays a crucial role. A major aim of HD is to normalize predialysis plasma electrolyte and mineral concentrations, while minimizing wide swings in the patient's intradialytic plasma concentrations. Adequate sodium (Na) and water removal is critical for preventing intra- and interdialytic hypotension and pulmonary edema. Avoiding both hyper- and hypokalaemia prevents life-threatening cardiac arrhythmias. Optimal calcium (Ca) and magnesium (Mg) dialysate concentrations may protect the cardiovascular system and the bones, preventing extraskeletal calcifications, severe secondary hyperparathyroidism and adynamic bone disease. Adequate bicarbonate concentration [HCO3−] maintains a stable pH in the body fluids for appropriate protein and membrane functioning and also protects the bones. An adequate dialysate glucose concentration prevents severe hyperglycaemia and life-threating hypoglycaemia, which can lead to severe cardiovascular complications and a worsening of diabetic comorbidities. PMID:26413285

  9. Optimally Squeezed Spin States

    NASA Astrophysics Data System (ADS)

    Rojo, Alberto

    2004-03-01

    We consider optimally spin-squeezed states that maximize the sensitivity of the Ramsey spectroscopy, and for which the signal to noise ratio scales as the number of particles N. Using the variational principle we prove that these states are eigensolutions of the Hamiltonian H(λ)=λ S_z^2-S_x, and that, for large N, the states become equivalent to the quadrature squeezed states of the harmonic oscillator. We present numerical results that illustrate the validity of the equivalence. We also present results of spin squeezing via atom-field interactions within the context of the Tavis-Cummings model. An ensemble of N two-level atoms interacts with a quantized cavity field. For all the atoms initially in their ground states, it is shown that spin squeezing of both the atoms and the field can be achieved provided the initial state of the cavity field has coherence between number states differing by 2. Most of the discussion is restricted to the case of a cavity field initially in a coherent state, but initial squeezed states for the field are also discussed. An analytic solution is found that is valid in the limit that the number of atoms is much greater than unity. References: A. G. Rojo, Phys. Rev A, 68, 013807 (2003); Claudiu Genes, P. R. Berman, and A. G. Rojo Phys. Rev. A 68, 043809 (2003).

  10. Sweeping Jet Optimization Studies

    NASA Technical Reports Server (NTRS)

    Melton, LaTunia Pack; Koklu, Mehti; Andino, Marlyn; Lin, John C.; Edelman, Louis

    2016-01-01

    Progress on experimental efforts to optimize sweeping jet actuators for active flow control (AFC) applications with large adverse pressure gradients is reported. Three sweeping jet actuator configurations, with the same orifice size but di?erent internal geometries, were installed on the flap shoulder of an unswept, NACA 0015 semi-span wing to investigate how the output produced by a sweeping jet interacts with the separated flow and the mechanisms by which the flow separation is controlled. For this experiment, the flow separation was generated by deflecting the wing's 30% chord trailing edge flap to produce an adverse pressure gradient. Steady and unsteady pressure data, Particle Image Velocimetry data, and force and moment data were acquired to assess the performance of the three actuator configurations. The actuator with the largest jet deflection angle, at the pressure ratios investigated, was the most efficient at controlling flow separation on the flap of the model. Oil flow visualization studies revealed that the flow field controlled by the sweeping jets was more three-dimensional than expected. The results presented also show that the actuator spacing was appropriate for the pressure ratios examined.

  11. Optimizing management of glycaemia.

    PubMed

    Chatterjee, Sudesna; Khunti, Kamlesh; Davies, Melanie J

    2016-06-01

    The global epidemic of type 2 diabetes (T2DM) continues largely unabated due to an increasingly sedentary lifestyle and obesogenic environment. A cost-effective patient-centred approach, incorporating glucose-lowering therapy and modification of cardiovascular risk factors, could help prevent the inevitable development and progression of macrovascular and microvascular complications. Glycaemic optimization requires patient structured education, self-management and empowerment, and psychological support along with early and proactive use of glucose lowering therapies, which should be delivered in a system of care as shown by the Chronic Care Model. From diagnosis, intensive glycaemic control and individualised care is aimed at reducing complications. In older people, the goal is maintaining quality of life and minimizing morbidity, especially as overtreatment increases hypoglycaemia risk. Maintaining durable glycaemic control is challenging and complex to achieve without hypoglycaemia, weight gain and other significant adverse effects. Overcoming patient and physician barriers can help ensure adequate treatment initiation and intensification. Cardiovascular safety studies with newer glucose-lowering agents are now mandatory, with a sodium glucose co-transporter-2 inhibitor (empagliflozin), and two glucagon like peptide-1 receptor agonists (liraglutide and semaglutide) being the first to demonstrate superior CV outcomes compared with placebo. PMID:27432074

  12. Optimal hemicube sampling

    SciTech Connect

    Max, N. |

    1992-12-17

    Radiosity algorithms for global illumination, either ``gathering`` or ``shooting`` versions, depend on the calculation of form factors. It is possible to calculate the form factors analytically, but this is difficult when occlusion is involved, so sampling methods are usually preferred. The necessary visibility information can be obtained by ray tracing in the sampled directions. However, area coherence makes it more efficient to project and scan-convert the scene onto a number of planes, for example, the faces of a hemicube. The hemicube faces have traditionally been divided into equal square pixels, but more general subdivisions are practical, and can reduce the variance of the form factor estimates. The hemicube estimates of form factors are based on a finite set of sample directions. We obtain several optimal arrangements of sample directions, which minimize the variance of this estimate. Four approaches are changing the size of the pixels, the shape of the pixels, the shape of the hemicube, or using non-uniform pixel grids. The best approach reduces the variance by 43%. The variance calculation is based on the assumption that the errors in the estimate are caused by the projections of single edges of polygonal patches, and that the positions and orientations of these edges are random.

  13. Optimal hemicube sampling

    SciTech Connect

    Max, N. California Univ., Davis, CA )

    1992-12-17

    Radiosity algorithms for global illumination, either gathering'' or shooting'' versions, depend on the calculation of form factors. It is possible to calculate the form factors analytically, but this is difficult when occlusion is involved, so sampling methods are usually preferred. The necessary visibility information can be obtained by ray tracing in the sampled directions. However, area coherence makes it more efficient to project and scan-convert the scene onto a number of planes, for example, the faces of a hemicube. The hemicube faces have traditionally been divided into equal square pixels, but more general subdivisions are practical, and can reduce the variance of the form factor estimates. The hemicube estimates of form factors are based on a finite set of sample directions. We obtain several optimal arrangements of sample directions, which minimize the variance of this estimate. Four approaches are changing the size of the pixels, the shape of the pixels, the shape of the hemicube, or using non-uniform pixel grids. The best approach reduces the variance by 43%. The variance calculation is based on the assumption that the errors in the estimate are caused by the projections of single edges of polygonal patches, and that the positions and orientations of these edges are random.

  14. Optimization of SRF Linacs

    SciTech Connect

    Powers, Tom

    2013-09-01

    This work describes preliminary results of a new software tool that allows one to vary parameters and understand the effects on the optimized costs of construction plus 10 year operations of an SRF linac, the associated cryogenic facility, and controls, where operations includes the cost of the electrical utilities but not the labor or other costs. It derives from collaborative work done with staff from Accelerator Science and Technology Centre, Daresbury, UK several years ago while they were in the process of developing a conceptual design for the New Light Source project.[1] The initial goal was to convert a spread sheet format to a graphical interface to allow the ability to sweep different parameter sets. The tools also allow one to compare the cost of the different facets of the machine design and operations so as to better understand the tradeoffs. The work was first published in an ICFA Beam Dynamics News Letter.[2] More recent additions to the software include the ability to save and restore input parameters as well as to adjust the Qo versus E parameters in order to explore the potential costs savings associated with doing so. Additionally, program changes now allow one to model the costs associated with a linac that makes use of energy recovery mode of operation.

  15. Industrial cogeneration optimization program

    SciTech Connect

    Not Available

    1980-01-01

    The purpose of this program was to identify up to 10 good near-term opportunities for cogeneration in 5 major energy-consuming industries which produce food, textiles, paper, chemicals, and refined petroleum; select, characterize, and optimize cogeneration systems for these identified opportunities to achieve maximum energy savings for minimum investment using currently available components of cogenerating systems; and to identify technical, institutional, and regulatory obstacles hindering the use of industrial cogeneration systems. The analysis methods used and results obtained are described. Plants with fuel demands from 100,000 Btu/h to 3 x 10/sup 6/ Btu/h were considered. It was concluded that the major impediments to industrial cogeneration are financial, e.g., high capital investment and high charges by electric utilities during short-term cogeneration facility outages. In the plants considered an average energy savings from cogeneration of 15 to 18% compared to separate generation of process steam and electric power was calculated. On a national basis for the 5 industries considered, this extrapolates to saving 1.3 to 1.6 quads per yr or between 630,000 to 750,000 bbl/d of oil. Properly applied, federal activity can do much to realize a substantial fraction of this potential by lowering the barriers to cogeneration and by stimulating wider implementation of this technology. (LCL)

  16. Supply-Chain Optimization Template

    NASA Technical Reports Server (NTRS)

    Quiett, William F.; Sealing, Scott L.

    2009-01-01

    The Supply-Chain Optimization Template (SCOT) is an instructional guide for identifying, evaluating, and optimizing (including re-engineering) aerospace- oriented supply chains. The SCOT was derived from the Supply Chain Council s Supply-Chain Operations Reference (SCC SCOR) Model, which is more generic and more oriented toward achieving a competitive advantage in business.

  17. Aerodynamic design using numerical optimization

    NASA Technical Reports Server (NTRS)

    Murman, E. M.; Chapman, G. T.

    1983-01-01

    The procedure of using numerical optimization methods coupled with computational fluid dynamic (CFD) codes for the development of an aerodynamic design is examined. Several approaches that replace wind tunnel tests, develop pressure distributions and derive designs, or fulfill preset design criteria are presented. The method of Aerodynamic Design by Numerical Optimization (ADNO) is described and illustrated with examples.

  18. A Problem on Optimal Transportation

    ERIC Educational Resources Information Center

    Cechlarova, Katarina

    2005-01-01

    Mathematical optimization problems are not typical in the classical curriculum of mathematics. In this paper we show how several generalizations of an easy problem on optimal transportation were solved by gifted secondary school pupils in a correspondence mathematical seminar, how they can be used in university courses of linear programming and…

  19. Query Evaluation: Strategies and Optimizations.

    ERIC Educational Resources Information Center

    Turtle, Howard; Flood, James

    1995-01-01

    Discusses two query evaluation strategies used in large text retrieval systems: (1) term-at-a-time; and (2) document-at-a-time. Describes optimization techniques that can reduce query evaluation costs. Presents simulation results that compare the performance of these optimization techniques when applied to natural language query evaluation. (JMV)

  20. Optimizing Medical Kits for Spaceflight

    NASA Technical Reports Server (NTRS)

    Keenan, A. B,; Foy, Millennia; Myers, G.

    2014-01-01

    The Integrated Medical Model (IMM) is a probabilistic model that estimates medical event occurrences and mission outcomes for different mission profiles. IMM simulation outcomes describing the impact of medical events on the mission may be used to optimize the allocation of resources in medical kits. Efficient allocation of medical resources, subject to certain mass and volume constraints, is crucial to ensuring the best outcomes of in-flight medical events. We implement a new approach to this medical kit optimization problem. METHODS We frame medical kit optimization as a modified knapsack problem and implement an algorithm utilizing a dynamic programming technique. Using this algorithm, optimized medical kits were generated for 3 different mission scenarios with the goal of minimizing the probability of evacuation and maximizing the Crew Health Index (CHI) for each mission subject to mass and volume constraints. Simulation outcomes using these kits were also compared to outcomes using kits optimized..RESULTS The optimized medical kits generated by the algorithm described here resulted in predicted mission outcomes more closely approached the unlimited-resource scenario for Crew Health Index (CHI) than the implementation in under all optimization priorities. Furthermore, the approach described here improves upon in reducing evacuation when the optimization priority is minimizing the probability of evacuation. CONCLUSIONS This algorithm provides an efficient, effective means to objectively allocate medical resources for spaceflight missions using the Integrated Medical Model.

  1. Optimal dynamic detection of explosives

    SciTech Connect

    Moore, David Steven; Mcgrane, Shawn D; Greenfield, Margo T; Scharff, R J; Rabitz, Herschel A; Roslund, J

    2009-01-01

    The detection of explosives is a notoriously difficult problem, especially at stand-off distances, due to their (generally) low vapor pressure, environmental and matrix interferences, and packaging. We are exploring optimal dynamic detection to exploit the best capabilities of recent advances in laser technology and recent discoveries in optimal shaping of laser pulses for control of molecular processes to significantly enhance the standoff detection of explosives. The core of the ODD-Ex technique is the introduction of optimally shaped laser pulses to simultaneously enhance sensitivity of explosives signatures while reducing the influence of noise and the signals from background interferents in the field (increase selectivity). These goals are being addressed by operating in an optimal nonlinear fashion, typically with a single shaped laser pulse inherently containing within it coherently locked control and probe sub-pulses. With sufficient bandwidth, the technique is capable of intrinsically providing orthogonal broad spectral information for data fusion, all from a single optimal pulse.

  2. Optimized layout generator for microgyroscope

    NASA Astrophysics Data System (ADS)

    Tay, Francis E.; Li, Shifeng; Logeeswaran, V. J.; Ng, David C.

    2000-10-01

    This paper presents an optimized out-of-plane microgyroscope layout generator using AutoCAD R14 and MS ExcelTM as a first attempt to automating the design of resonant micro- inertial sensors. The out-of-plane microgyroscope with two degrees of freedom lumped parameter model was chosen as the synthesis topology. Analytical model for the open loop operating has been derived for the gyroscope performance characteristics. Functional performance parameters such as sensitivity are ensured to be satisfied while simultaneously optimizing a design objective such as minimum area. A single algorithm will optimize the microgyroscope dimensions, while simultaneously maximizing or minimizing the objective functions: maximum sensitivity and minimum area. The multi- criteria objective function and optimization methodology was implemented using the Generalized Reduced Gradient algorithm. For data conversion a DXF to GDS converter was used. The optimized theoretical design performance parameters show good agreement with finite element analysis.

  3. Optimal Distinctiveness Signals Membership Trust.

    PubMed

    Leonardelli, Geoffrey J; Loyd, Denise Lewin

    2016-07-01

    According to optimal distinctiveness theory, sufficiently small minority groups are associated with greater membership trust, even among members otherwise unknown, because the groups are seen as optimally distinctive. This article elaborates on the prediction's motivational and cognitive processes and tests whether sufficiently small minorities (defined by relative size; for example, 20%) are associated with greater membership trust relative to mere minorities (45%), and whether such trust is a function of optimal distinctiveness. Two experiments, examining observers' perceptions of minority and majority groups and using minimal groups and (in Experiment 2) a trust game, revealed greater membership trust in minorities than majorities. In Experiment 2, participants also preferred joining minorities over more powerful majorities. Both effects occurred only when minorities were 20% rather than 45%. In both studies, perceptions of optimal distinctiveness mediated effects. Discussion focuses on the value of relative size and optimal distinctiveness, and when membership trust manifests. PMID:27140657

  4. Optimized design for PIGMI

    SciTech Connect

    Hansborough, L.; Hamm, R.; Stovall, J.; Swenson, D.

    1980-01-01

    PIGMI (Pion Generator for Medical Irradiations) is a compact linear proton accelerator design, optimized for pion production and cancer treatment use in a hospital environment. Technology developed during a four-year PIGMI Prototype experimental program allows the design of smaller, less expensive, and more reliable proton linacs. A new type of low-energy accelerating structure, the radio-frequency quadrupole (RFQ) has been tested; it produces an exceptionally good-quality beam and allows the use of a simple 30-kV injector. Average axial electric-field gradients of over 9 MV/m have been demonstrated in a drift-tube linac (DTL) structure. Experimental work is underway to test the disk-and-washer (DAW) structure, another new type of accelerating structure for use in the high-energy coupled-cavity linac (CCL). Sufficient experimental and developmental progress has been made to closely define an actual PIGMI. It will consist of a 30-kV injector, and RFQ linac to a proton energy of 2.5 MeV, a DTL linac to 125 MeV, and a CCL linac to the final energy of 650 MeV. The total length of the accelerator is 133 meters. The RFQ and DTL will be driven by a single 440-MHz klystron; the CCL will be driven by six 1320-MHz klystrons. The peak beam current is 28 mA. The beam pulse length is 60 ..mu..s at a 60-Hz repetition rate, resulting in a 100-..mu..A average beam current. The total cost of the accelerator is estimated to be approx. $10 million.

  5. Optimal multiobjective design of digital filters using spiral optimization technique.

    PubMed

    Ouadi, Abderrahmane; Bentarzi, Hamid; Recioui, Abdelmadjid

    2013-01-01

    The multiobjective design of digital filters using spiral optimization technique is considered in this paper. This new optimization tool is a metaheuristic technique inspired by the dynamics of spirals. It is characterized by its robustness, immunity to local optima trapping, relative fast convergence and ease of implementation. The objectives of filter design include matching some desired frequency response while having minimum linear phase; hence, reducing the time response. The results demonstrate that the proposed problem solving approach blended with the use of the spiral optimization technique produced filters which fulfill the desired characteristics and are of practical use. PMID:24083108

  6. Optimality criteria solution strategies in multiple constraint design optimization

    NASA Technical Reports Server (NTRS)

    Levy, R.; Parzynski, W.

    1981-01-01

    Procedures and solution strategies are described to solve the conventional structural optimization problem using the Lagrange multiplier technique. The multipliers, obtained through solution of an auxiliary nonlinear optimization problem, lead to optimality criteria to determine the design variables. It is shown that this procedure is essentially equivalent to an alternative formulation using a dual method Lagrangian function objective. Although mathematical formulations are straight-forward, successful applications and computational efficiency depend upon execution procedure strategies. Strategies examined, with application examples, include selection of active constraints, move limits, line search procedures, and side constraint boundaries.

  7. Optimal Flow Control Design

    NASA Technical Reports Server (NTRS)

    Allan, Brian; Owens, Lewis

    2010-01-01

    In support of the Blended-Wing-Body aircraft concept, a new flow control hybrid vane/jet design has been developed for use in a boundary-layer-ingesting (BLI) offset inlet in transonic flows. This inlet flow control is designed to minimize the engine fan-face distortion levels and the first five Fourier harmonic half amplitudes while maximizing the inlet pressure recovery. This concept represents a potentially enabling technology for quieter and more environmentally friendly transport aircraft. An optimum vane design was found by minimizing the engine fan-face distortion, DC60, and the first five Fourier harmonic half amplitudes, while maximizing the total pressure recovery. The optimal vane design was then used in a BLI inlet wind tunnel experiment at NASA Langley's 0.3-meter transonic cryogenic tunnel. The experimental results demonstrated an 80-percent decrease in DPCPavg, the reduction in the circumferential distortion levels, at an inlet mass flow rate corresponding to the middle of the operational range at the cruise condition. Even though the vanes were designed at a single inlet mass flow rate, they performed very well over the entire inlet mass flow range tested in the wind tunnel experiment with the addition of a small amount of jet flow control. While the circumferential distortion was decreased, the radial distortion on the outer rings at the aerodynamic interface plane (AIP) increased. This was a result of the large boundary layer being distributed from the bottom of the AIP in the baseline case to the outer edges of the AIP when using the vortex generator (VG) vane flow control. Experimental results, as already mentioned, showed an 80-percent reduction of DPCPavg, the circumferential distortion level at the engine fan-face. The hybrid approach leverages strengths of vane and jet flow control devices, increasing inlet performance over a broader operational range with significant reduction in mass flow requirements. Minimal distortion level requirements

  8. Optimizing WFIRST Coronagraph Science

    NASA Astrophysics Data System (ADS)

    Macintosh, Bruce

    We propose an in-depth scientific investigation that will define how the WFIRST coronagraphic instrument will discover and characterize nearby planetary systems and how it will use observations of planets and disks to probe the diversity of their compositions, dynamics, and formation. Given the enormous diversity of known planetary systems it is not enough to optimize a coronagraph mission plan for the characterization of solar system analogs. Instead, we must design a mission to characterize a wide variety of planets, from gas and ice giant planets at a range of separations to mid-sized planets with no analogs in our solar system. We must consider updated planet distributions based on the results of the Kepler mission, long-term radial velocity (RV) surveys and updated luminosity distributions of exo-zodiacal dust from interferometric thermal infrared surveys of nearby stars. The properties of all these objects must be informed by our best models of planets and disks, and the process of using WFIRST observations to measure fundamental planetary properties such as composition must derive from rigorous methods. Our team brings a great depth of expertise to inform and accomplish these and all of the other tasks enumerated in the SIT proposal call. We will perform end-to-end modeling that starts with model spectra of planets and images of disks, simulates WFIRST data using these models, accounts for geometries of specific star / planet / disk systems, and incorporates detailed instrument performance models. We will develop and implement data analysis techniques to extract well-calibrated astrophysical signals from complex data, and propose observing plans that maximize the mission's scientific yield. We will work with the community to build observing programs and target lists, inform them of WFIRSTs capabilities, and supply simulated scientific observations for data challenges. Our work will be informed by the experience we have gained from building and observing with

  9. A Novel Particle Swarm Optimization Algorithm for Global Optimization

    PubMed Central

    Wang, Chun-Feng; Liu, Kui

    2016-01-01

    Particle Swarm Optimization (PSO) is a recently developed optimization method, which has attracted interest of researchers in various areas due to its simplicity and effectiveness, and many variants have been proposed. In this paper, a novel Particle Swarm Optimization algorithm is presented, in which the information of the best neighbor of each particle and the best particle of the entire population in the current iteration is considered. Meanwhile, to avoid premature, an abandoned mechanism is used. Furthermore, for improving the global convergence speed of our algorithm, a chaotic search is adopted in the best solution of the current iteration. To verify the performance of our algorithm, standard test functions have been employed. The experimental results show that the algorithm is much more robust and efficient than some existing Particle Swarm Optimization algorithms. PMID:26955387

  10. A Novel Particle Swarm Optimization Algorithm for Global Optimization.

    PubMed

    Wang, Chun-Feng; Liu, Kui

    2016-01-01

    Particle Swarm Optimization (PSO) is a recently developed optimization method, which has attracted interest of researchers in various areas due to its simplicity and effectiveness, and many variants have been proposed. In this paper, a novel Particle Swarm Optimization algorithm is presented, in which the information of the best neighbor of each particle and the best particle of the entire population in the current iteration is considered. Meanwhile, to avoid premature, an abandoned mechanism is used. Furthermore, for improving the global convergence speed of our algorithm, a chaotic search is adopted in the best solution of the current iteration. To verify the performance of our algorithm, standard test functions have been employed. The experimental results show that the algorithm is much more robust and efficient than some existing Particle Swarm Optimization algorithms. PMID:26955387