Satellite mission scheduling algorithm based on GA
NASA Astrophysics Data System (ADS)
Sun, Baolin; Mao, Lifei; Wang, Wenxiang; Xie, Xing; Qin, Qianqing
2007-11-01
The Satellite Mission Scheduling problem (SMS) involves scheduling tasks to be performed by a satellite, where new task requests can arrive at any time, non-deterministically, and must be scheduled in real-time. This paper describes a new Satellite Mission Scheduling problem based on Genetic Algorithm (SMSGA). In this paper, it investigates algorithmic approaches for determining an optimal or near-optimal sequence of tasks, allocated to a satellite payload over time, with dynamic tasking considerations. The simulation results show that the proposed approach is effective and efficient in applications to the real problems.
Genetic Algorithm Approaches for Actuator Placement
NASA Technical Reports Server (NTRS)
Crossley, William A.
2000-01-01
This research investigated genetic algorithm approaches for smart actuator placement to provide aircraft maneuverability without requiring hinged flaps or other control surfaces. The effort supported goals of the Multidisciplinary Design Optimization focus efforts in NASA's Aircraft au program. This work helped to properly identify various aspects of the genetic algorithm operators and parameters that allow for placement of discrete control actuators/effectors. An improved problem definition, including better definition of the objective function and constraints, resulted from this research effort. The work conducted for this research used a geometrically simple wing model; however, an increasing number of potential actuator placement locations were incorporated to illustrate the ability of the GA to determine promising actuator placement arrangements. This effort's major result is a useful genetic algorithm-based approach to assist in the discrete actuator/effector placement problem.
A Test Scheduling Algorithm Based on Two-Stage GA
NASA Astrophysics Data System (ADS)
Yu, Y.; Peng, X. Y.; Peng, Y.
2006-10-01
In this paper, we present a new algorithm to co-optimize the core wrapper design and the SOC test scheduling. The SOC test scheduling problem is first formulated into a twodimension floorplan problem and a sequence pair architecture is used to represent it. Then we propose a two-stage GA (Genetic Algorithm) to solve the SOC test scheduling problem. Experiments on ITC'02 benchmark show that our algorithm can effectively reduce test time so as to decrease SOC test cost.
GA-based discrete dynamic programming approach for scheduling in FMS environments.
Yang, J B
2001-01-01
The paper presents a new genetic algorithm (GA)-based discrete dynamic programming (DDP) approach for generating static schedules in a flexible manufacturing system (FMS) environment. This GA-DDP approach adopts a sequence-dependent schedule generation strategy, where a GA is employed to generate feasible job sequences and a series of discrete dynamic programs are constructed to generate legal schedules for a given sequence of jobs. In formulating the GA, different performance criteria could be easily included. The developed DDF algorithm is capable of identifying locally optimized partial schedules and shares the computation efficiency of dynamic programming. The algorithm is designed In such a way that it does not suffer from the state explosion problem inherent in pure dynamic programming approaches in FMS scheduling. Numerical examples are reported to illustrate the approach. PMID:18244848
A genetic algorithm approach to recognition and data mining
Punch, W.F.; Goodman, E.D.; Min, Pei
1996-12-31
We review here our use of genetic algorithm (GA) and genetic programming (GP) techniques to perform {open_quotes}data mining,{close_quotes} the discovery of particular/important data within large datasets, by finding optimal data classifications using known examples. Our first experiments concentrated on the use of a K-nearest neighbor algorithm in combination with a GA. The GA selected weights for each feature so as to optimize knn classification based on a linear combination of features. This combined GA-knn approach was successfully applied to both generated and real-world data. We later extended this work by substituting a GP for the GA. The GP-knn could not only optimize data classification via linear combinations of features but also determine functional relationships among the features. This allowed for improved performance and new information on important relationships among features. We review the effectiveness of the overall approach on examples from biology and compare the effectiveness of the GA and GP.
Ameliorated GA approach for base station planning
NASA Astrophysics Data System (ADS)
Wang, Andong; Sun, Hongyue; Wu, Xiaomin
2011-10-01
In this paper, we aim at locating base station (BS) rationally to satisfy the most customs by using the least BSs. An ameliorated GA is proposed to search for the optimum solution. In the algorithm, we mesh the area to be planned according to least overlap length derived from coverage radius, bring into isometric grid encoding method to represent BS distribution as well as its number and develop select, crossover and mutation operators to serve our unique necessity. We also construct our comprehensive object function after synthesizing coverage ratio, overlap ratio, population and geographical conditions. Finally, after importing an electronic map of the area to be planned, a recommended strategy draft would be exported correspondingly. We eventually import HongKong, China to simulate and yield a satisfactory solution.
Genetic Algorithm (GA)-Based Inclinometer Layout Optimization.
Liang, Weijie; Zhang, Ping; Chen, Xianping; Cai, Miao; Yang, Daoguo
2015-01-01
This paper presents numerical simulation results of an airflow inclinometer with sensitivity studies and thermal optimization of the printed circuit board (PCB) layout for an airflow inclinometer based on a genetic algorithm (GA). Due to the working principle of the gas sensor, the changes of the ambient temperature may cause dramatic voltage drifts of sensors. Therefore, eliminating the influence of the external environment for the airflow is essential for the performance and reliability of an airflow inclinometer. In this paper, the mechanism of an airflow inclinometer and the influence of different ambient temperatures on the sensitivity of the inclinometer will be examined by the ANSYS-FLOTRAN CFD program. The results show that with changes of the ambient temperature on the sensing element, the sensitivity of the airflow inclinometer is inversely proportional to the ambient temperature and decreases when the ambient temperature increases. GA is used to optimize the PCB thermal layout of the inclinometer. The finite-element simulation method (ANSYS) is introduced to simulate and verify the results of our optimal thermal layout, and the results indicate that the optimal PCB layout greatly improves (by more than 50%) the sensitivity of the inclinometer. The study may be useful in the design of PCB layouts that are related to sensitivity improvement of gas sensors. PMID:25897500
Genetic Algorithm (GA)-Based Inclinometer Layout Optimization
Liang, Weijie; Zhang, Ping; Chen, Xianping; Cai, Miao; Yang, Daoguo
2015-01-01
This paper presents numerical simulation results of an airflow inclinometer with sensitivity studies and thermal optimization of the printed circuit board (PCB) layout for an airflow inclinometer based on a genetic algorithm (GA). Due to the working principle of the gas sensor, the changes of the ambient temperature may cause dramatic voltage drifts of sensors. Therefore, eliminating the influence of the external environment for the airflow is essential for the performance and reliability of an airflow inclinometer. In this paper, the mechanism of an airflow inclinometer and the influence of different ambient temperatures on the sensitivity of the inclinometer will be examined by the ANSYS-FLOTRAN CFD program. The results show that with changes of the ambient temperature on the sensing element, the sensitivity of the airflow inclinometer is inversely proportional to the ambient temperature and decreases when the ambient temperature increases. GA is used to optimize the PCB thermal layout of the inclinometer. The finite-element simulation method (ANSYS) is introduced to simulate and verify the results of our optimal thermal layout, and the results indicate that the optimal PCB layout greatly improves (by more than 50%) the sensitivity of the inclinometer. The study may be useful in the design of PCB layouts that are related to sensitivity improvement of gas sensors. PMID:25897500
Economic Dispatch Using Genetic Algorithm Based Hybrid Approach
Tahir Nadeem Malik; Aftab Ahmad; Shahab Khushnood
2006-07-01
Power Economic Dispatch (ED) is vital and essential daily optimization procedure in the system operation. Present day large power generating units with multi-valves steam turbines exhibit a large variation in the input-output characteristic functions, thus non-convexity appears in the characteristic curves. Various mathematical and optimization techniques have been developed, applied to solve economic dispatch (ED) problem. Most of these are calculus-based optimization algorithms that are based on successive linearization and use the first and second order differentiations of objective function and its constraint equations as the search direction. They usually require heat input, power output characteristics of generators to be of monotonically increasing nature or of piecewise linearity. These simplifying assumptions result in an inaccurate dispatch. Genetic algorithms have used to solve the economic dispatch problem independently and in conjunction with other AI tools and mathematical programming approaches. Genetic algorithms have inherent ability to reach the global minimum region of search space in a short time, but then take longer time to converge the solution. GA based hybrid approaches get around this problem and produce encouraging results. This paper presents brief survey on hybrid approaches for economic dispatch, an architecture of extensible computational framework as common environment for conventional, genetic algorithm and hybrid approaches based solution for power economic dispatch, the implementation of three algorithms in the developed framework. The framework tested on standard test systems for its performance evaluation. (authors)
Calibration of visual model for space manipulator with a hybrid LM-GA algorithm
NASA Astrophysics Data System (ADS)
Jiang, Wensong; Wang, Zhongyu
2016-01-01
A hybrid LM-GA algorithm is proposed to calibrate the camera system of space manipulator to improve its locational accuracy. This algorithm can dynamically fuse the Levenberg-Marqurdt (LM) algorithm and Genetic Algorithm (GA) together to minimize the error of nonlinear camera model. LM algorithm is called to optimize the initial camera parameters that are generated by genetic process previously. Iteration should be stopped if the optimized camera parameters meet the accuracy requirements. Otherwise, new populations are generated again by GA and optimized afresh by LM algorithm until the optimal solutions meet the accuracy requirements. A novel measuring machine of space manipulator is designed to on-orbit dynamic simulation and precision test. The camera system of space manipulator, calibrated by hybrid LM-GA algorithm, is used for locational precision test in this measuring instrument. The experimental results show that the mean composite errors are 0.074 mm for hybrid LM-GA camera calibration model, 1.098 mm for LM camera calibration model, and 1.202 mm for GA camera calibration model. Furthermore, the composite standard deviations are 0.103 mm for the hybrid LM-GA camera calibration model, 1.227 mm for LM camera calibration model, and 1.351 mm for GA camera calibration model. The accuracy of hybrid LM-GA camera calibration model is more than 10 times higher than that of other two methods. All in all, the hybrid LM-GA camera calibration model is superior to both the LM camera calibration model and GA camera calibration model.
Longest jobs first algorithm in solving job shop scheduling using adaptive genetic algorithm (GA)
NASA Astrophysics Data System (ADS)
Alizadeh Sahzabi, Vahid; Karimi, Iman; Alizadeh Sahzabi, Navid; Mamaani Barnaghi, Peiman
2011-12-01
In this paper, genetic algorithm was used to solve job shop scheduling problems. One example discussed in JSSP (Job Shop Scheduling Problem) and I described how we can solve such these problems by genetic algorithm. The goal in JSSP is to gain the shortest process time. Furthermore I proposed a method to obtain best performance on performing all jobs in shortest time. The method mainly, is according to Genetic algorithm (GA) and crossing over between parents always follows the rule which the longest process is at the first in the job queue. In the other word chromosomes is suggested to sorts based on the longest processes to shortest i.e. "longest job first" says firstly look which machine contains most processing time during its performing all its jobs and that is the bottleneck. Secondly, start sort those jobs which are belonging to that specific machine descending. Based on the achieved results," longest jobs first" is the optimized status in job shop scheduling problems. In our results the accuracy would grow up to 94.7% for total processing time and the method improved 4% the accuracy of performing all jobs in the presented example.
Longest jobs first algorithm in solving job shop scheduling using adaptive genetic algorithm (GA)
NASA Astrophysics Data System (ADS)
Alizadeh Sahzabi, Vahid; Karimi, Iman; Alizadeh Sahzabi, Navid; Mamaani Barnaghi, Peiman
2012-01-01
In this paper, genetic algorithm was used to solve job shop scheduling problems. One example discussed in JSSP (Job Shop Scheduling Problem) and I described how we can solve such these problems by genetic algorithm. The goal in JSSP is to gain the shortest process time. Furthermore I proposed a method to obtain best performance on performing all jobs in shortest time. The method mainly, is according to Genetic algorithm (GA) and crossing over between parents always follows the rule which the longest process is at the first in the job queue. In the other word chromosomes is suggested to sorts based on the longest processes to shortest i.e. "longest job first" says firstly look which machine contains most processing time during its performing all its jobs and that is the bottleneck. Secondly, start sort those jobs which are belonging to that specific machine descending. Based on the achieved results," longest jobs first" is the optimized status in job shop scheduling problems. In our results the accuracy would grow up to 94.7% for total processing time and the method improved 4% the accuracy of performing all jobs in the presented example.
System engineering approach to GPM retrieval algorithms
Rose, C. R.; Chandrasekar, V.
2004-01-01
System engineering principles and methods are very useful in large-scale complex systems for developing the engineering requirements from end-user needs. Integrating research into system engineering is a challenging task. The proposed Global Precipitation Mission (GPM) satellite will use a dual-wavelength precipitation radar to measure and map global precipitation with unprecedented accuracy, resolution and areal coverage. The satellite vehicle, precipitation radars, retrieval algorithms, and ground validation (GV) functions are all critical subsystems of the overall GPM system and each contributes to the success of the mission. Errors in the radar measurements and models can adversely affect the retrieved output values. Ground validation (GV) systems are intended to provide timely feedback to the satellite and retrieval algorithms based on measured data. These GV sites will consist of radars and DSD measurement systems and also have intrinsic constraints. One of the retrieval algorithms being studied for use with GPM is the dual-wavelength DSD algorithm that does not use the surface reference technique (SRT). The underlying microphysics of precipitation structures and drop-size distributions (DSDs) dictate the types of models and retrieval algorithms that can be used to estimate precipitation. Many types of dual-wavelength algorithms have been studied. Meneghini (2002) analyzed the performance of single-pass dual-wavelength surface-reference-technique (SRT) based algorithms. Mardiana (2003) demonstrated that a dual-wavelength retrieval algorithm could be successfully used without the use of the SRT. It uses an iterative approach based on measured reflectivities at both wavelengths and complex microphysical models to estimate both No and Do at each range bin. More recently, Liao (2004) proposed a solution to the Do ambiguity problem in rain within the dual-wavelength algorithm and showed a possible melting layer model based on stratified spheres. With the No and Do
NASA Astrophysics Data System (ADS)
Kim, Eunsu; Kim, Manseok; Kim, Jong-Wook
In this paper, a humanoid is simulated and implemented to walk up and down a staircase using the blending polynomial and genetic algorithm (GA). Both ascending and descending a staircase are scheduled by four steps. Each step mimics natural gait of human being and is easy to analyze and implement. Optimal trajectories of ten motors in a lower extremity of a humanoid are rigorously computed to simultaneously satisfy stability condition, walking constraints, and energy efficiency requirements. As an optimization method, GA is applied to search optimal trajectory parameters in blending polynomials. The feasibility of this approach will be validated by simulation with a small humanoid robot.
NASA Astrophysics Data System (ADS)
Igeta, Hideki; Hasegawa, Mikio
Chaotic dynamics have been effectively applied to improve various heuristic algorithms for combinatorial optimization problems in many studies. Currently, the most used chaotic optimization scheme is to drive heuristic solution search algorithms applicable to large-scale problems by chaotic neurodynamics including the tabu effect of the tabu search. Alternatively, meta-heuristic algorithms are used for combinatorial optimization by combining a neighboring solution search algorithm, such as tabu, gradient, or other search method, with a global search algorithm, such as genetic algorithms (GA), ant colony optimization (ACO), or others. In these hybrid approaches, the ACO has effectively optimized the solution of many benchmark problems in the quadratic assignment problem library. In this paper, we propose a novel hybrid method that combines the effective chaotic search algorithm that has better performance than the tabu search and global search algorithms such as ACO and GA. Our results show that the proposed chaotic hybrid algorithm has better performance than the conventional chaotic search and conventional hybrid algorithms. In addition, we show that chaotic search algorithm combined with ACO has better performance than when combined with GA.
Naresh-Kumar, G. Trager-Cowan, C.; Vilalta-Clemente, A.; Morales, M.; Ruterana, P.; Pandey, S.; Cavallini, A.; Cavalcoli, D.; Skuridina, D.; Vogt, P.; Kneissl, M.; Behmenburg, H.; Giesen, C.; Heuken, M.; Gamarra, P.; Di Forte-Poisson, M. A.; Patriarche, G.; Vickridge, I.
2014-12-15
We report on our multi–pronged approach to understand the structural and electrical properties of an InAl(Ga)N(33nm barrier)/Al(Ga)N(1nm interlayer)/GaN(3μm)/ AlN(100nm)/Al{sub 2}O{sub 3} high electron mobility transistor (HEMT) heterostructure grown by metal organic vapor phase epitaxy (MOVPE). In particular we reveal and discuss the role of unintentional Ga incorporation in the barrier and also in the interlayer. The observation of unintentional Ga incorporation by using energy dispersive X–ray spectroscopy analysis in a scanning transmission electron microscope is supported with results obtained for samples with a range of AlN interlayer thicknesses grown under both the showerhead as well as the horizontal type MOVPE reactors. Poisson–Schrödinger simulations show that for high Ga incorporation in the Al(Ga)N interlayer, an additional triangular well with very small depth may be exhibited in parallel to the main 2–DEG channel. The presence of this additional channel may cause parasitic conduction and severe issues in device characteristics and processing. Producing a HEMT structure with InAlGaN as the barrier and AlGaN as the interlayer with appropriate alloy composition may be a possible route to optimization, as it might be difficult to avoid Ga incorporation while continuously depositing the layers using the MOVPE growth method. Our present work shows the necessity of a multicharacterization approach to correlate structural and electrical properties to understand device structures and their performance.
Ancestral genome inference using a genetic algorithm approach.
Gao, Nan; Yang, Ning; Tang, Jijun
2013-01-01
Recent advancement of technologies has now made it routine to obtain and compare gene orders within genomes. Rearrangements of gene orders by operations such as reversal and transposition are rare events that enable researchers to reconstruct deep evolutionary histories. An important application of genome rearrangement analysis is to infer gene orders of ancestral genomes, which is valuable for identifying patterns of evolution and for modeling the evolutionary processes. Among various available methods, parsimony-based methods (including GRAPPA and MGR) are the most widely used. Since the core algorithms of these methods are solvers for the so called median problem, providing efficient and accurate median solver has attracted lots of attention in this field. The "double-cut-and-join" (DCJ) model uses the single DCJ operation to account for all genome rearrangement events. Because mathematically it is much simpler than handling events directly, parsimony methods using DCJ median solvers has better speed and accuracy. However, the DCJ median problem is NP-hard and although several exact algorithms are available, they all have great difficulties when given genomes are distant. In this paper, we present a new algorithm that combines genetic algorithm (GA) with genomic sorting to produce a new method which can solve the DCJ median problem in limited time and space, especially in large and distant datasets. Our experimental results show that this new GA-based method can find optimal or near optimal results for problems ranging from easy to very difficult. Compared to existing parsimony methods which may severely underestimate the true number of evolutionary events, the sorting-based approach can infer ancestral genomes which are much closer to their true ancestors. The code is available at http://phylo.cse.sc.edu. PMID:23658708
Algorithmic approach in the diagnosis of uveitis
Rathinam, S R; Babu, Manohar
2013-01-01
Uveitis is caused by disorders of diverse etiologies including wide spectrum of infectious and non-infectious causes. Often clinical signs are less specific and shared by different diseases. On several occasions, uveitis represents diseases that are developing elsewhere in the body and ocular signs may be the first evidence of such systemic diseases. Uveitis specialists need to have a thorough knowledge of all entities and their work up has to be systematic and complete including systemic and ocular examinations. Creating an algorithmic approach on critical steps to be taken would help the ophthalmologist in arriving at the etiological diagnosis. PMID:23803476
DeMAID/GA USER'S GUIDE Design Manager's Aid for Intelligent Decomposition with a Genetic Algorithm
NASA Technical Reports Server (NTRS)
Rogers, James L.
1996-01-01
Many companies are looking for new tools and techniques to aid a design manager in making decisions that can reduce the time and cost of a design cycle. One tool that is available to aid in this decision making process is the Design Manager's Aid for Intelligent Decomposition (DeMAID). Since the initial release of DEMAID in 1989, numerous enhancements have been added to aid the design manager in saving both cost and time in a design cycle. The key enhancement is a genetic algorithm (GA) and the enhanced version is called DeMAID/GA. The GA orders the sequence of design processes to minimize the cost and time to converge to a solution. These enhancements as well as the existing features of the original version of DEMAID are described. Two sample problems are used to show how these enhancements can be applied to improve the design cycle. This report serves as a user's guide for DeMAID/GA.
The royal road for genetic algorithms: Fitness landscapes and GA performance
Mitchell, M.; Holland, J.H. ); Forrest, S. . Dept. of Computer Science)
1991-01-01
Genetic algorithms (GAs) play a major role in many artificial-life systems, but there is often little detailed understanding of why the GA performs as it does, and little theoretical basis on which to characterize the types of fitness landscapes that lead to successful GA performance. In this paper we propose a strategy for addressing these issues. Our strategy consists of defining a set of features of fitness landscapes that are particularly relevant to the GA, and experimentally studying how various configurations of these features affect the GA's performance along a number of dimensions. In this paper we informally describe an initial set of proposed feature classes, describe in detail one such class ( Royal Road'' functions), and present some initial experimental results concerning the role of crossover and building blocks'' on landscapes constructed from features of this class. 27 refs., 1 fig., 5 tabs.
Evolutionary Algorithms Approach to the Solution of Damage Detection Problems
NASA Astrophysics Data System (ADS)
Salazar Pinto, Pedro Yoajim; Begambre, Oscar
2010-09-01
In this work is proposed a new Self-Configured Hybrid Algorithm by combining the Particle Swarm Optimization (PSO) and a Genetic Algorithm (GA). The aim of the proposed strategy is to increase the stability and accuracy of the search. The central idea is the concept of Guide Particle, this particle (the best PSO global in each generation) transmits its information to a particle of the following PSO generation, which is controlled by the GA. Thus, the proposed hybrid has an elitism feature that improves its performance and guarantees the convergence of the procedure. In different test carried out in benchmark functions, reported in the international literature, a better performance in stability and accuracy was observed; therefore the new algorithm was used to identify damage in a simple supported beam using modal data. Finally, it is worth noting that the algorithm is independent of the initial definition of heuristic parameters.
The mGA1.0: A common LISP implementation of a messy genetic algorithm
NASA Technical Reports Server (NTRS)
Goldberg, David E.; Kerzic, Travis
1990-01-01
Genetic algorithms (GAs) are finding increased application in difficult search, optimization, and machine learning problems in science and engineering. Increasing demands are being placed on algorithm performance, and the remaining challenges of genetic algorithm theory and practice are becoming increasingly unavoidable. Perhaps the most difficult of these challenges is the so-called linkage problem. Messy GAs were created to overcome the linkage problem of simple genetic algorithms by combining variable-length strings, gene expression, messy operators, and a nonhomogeneous phasing of evolutionary processing. Results on a number of difficult deceptive test functions are encouraging with the mGA always finding global optima in a polynomial number of function evaluations. Theoretical and empirical studies are continuing, and a first version of a messy GA is ready for testing by others. A Common LISP implementation called mGA1.0 is documented and related to the basic principles and operators developed by Goldberg et. al. (1989, 1990). Although the code was prepared with care, it is not a general-purpose code, only a research version. Important data structures and global variations are described. Thereafter brief function descriptions are given, and sample input data are presented together with sample program output. A source listing with comments is also included.
DFT algorithms for bit-serial GaAs array processor architectures
NASA Technical Reports Server (NTRS)
Mcmillan, Gary B.
1988-01-01
Systems and Processes Engineering Corporation (SPEC) has developed an innovative array processor architecture for computing Fourier transforms and other commonly used signal processing algorithms. This architecture is designed to extract the highest possible array performance from state-of-the-art GaAs technology. SPEC's architectural design includes a high performance RISC processor implemented in GaAs, along with a Floating Point Coprocessor and a unique Array Communications Coprocessor, also implemented in GaAs technology. Together, these data processors represent the latest in technology, both from an architectural and implementation viewpoint. SPEC has examined numerous algorithms and parallel processing architectures to determine the optimum array processor architecture. SPEC has developed an array processor architecture with integral communications ability to provide maximum node connectivity. The Array Communications Coprocessor embeds communications operations directly in the core of the processor architecture. A Floating Point Coprocessor architecture has been defined that utilizes Bit-Serial arithmetic units, operating at very high frequency, to perform floating point operations. These Bit-Serial devices reduce the device integration level and complexity to a level compatible with state-of-the-art GaAs device technology.
Global Precipitation Measurement: GPM Microwave Imager (GMI) Algorithm Development Approach
NASA Technical Reports Server (NTRS)
Stocker, Erich Franz
2009-01-01
This slide presentation reviews the approach to the development of the Global Precipitation Measurement algorithm. This presentation includes information about the responsibilities for the development of the algorithm, and the calibration. Also included is information about the orbit, and the sun angle. The test of the algorithm code will be done with synthetic data generated from the Precipitation Processing System (PPS).
An algorithmic approach to crustal deformation analysis
NASA Technical Reports Server (NTRS)
Iz, Huseyin Baki
1987-01-01
In recent years the analysis of crustal deformation measurements has become important as a result of current improvements in geodetic methods and an increasing amount of theoretical and observational data provided by several earth sciences. A first-generation data analysis algorithm which combines a priori information with current geodetic measurements was proposed. Relevant methods which can be used in the algorithm were discussed. Prior information is the unifying feature of this algorithm. Some of the problems which may arise through the use of a priori information in the analysis were indicated and preventive measures were demonstrated. The first step in the algorithm is the optimal design of deformation networks. The second step in the algorithm identifies the descriptive model of the deformation field. The final step in the algorithm is the improved estimation of deformation parameters. Although deformation parameters are estimated in the process of model discrimination, they can further be improved by the use of a priori information about them. According to the proposed algorithm this information must first be tested against the estimates calculated using the sample data only. Null-hypothesis testing procedures were developed for this purpose. Six different estimators which employ a priori information were examined. Emphasis was put on the case when the prior information is wrong and analytical expressions for possible improvements under incompatible prior information were derived.
A Genetic Algorithm and Fuzzy Logic Approach for Video Shot Boundary Detection
Thounaojam, Dalton Meitei; Khelchandra, Thongam; Singh, Kh. Manglem; Roy, Sudipta
2016-01-01
This paper proposed a shot boundary detection approach using Genetic Algorithm and Fuzzy Logic. In this, the membership functions of the fuzzy system are calculated using Genetic Algorithm by taking preobserved actual values for shot boundaries. The classification of the types of shot transitions is done by the fuzzy system. Experimental results show that the accuracy of the shot boundary detection increases with the increase in iterations or generations of the GA optimization process. The proposed system is compared to latest techniques and yields better result in terms of F1score parameter. PMID:27127500
A Genetic Algorithm and Fuzzy Logic Approach for Video Shot Boundary Detection.
Thounaojam, Dalton Meitei; Khelchandra, Thongam; Manglem Singh, Kh; Roy, Sudipta
2016-01-01
This paper proposed a shot boundary detection approach using Genetic Algorithm and Fuzzy Logic. In this, the membership functions of the fuzzy system are calculated using Genetic Algorithm by taking preobserved actual values for shot boundaries. The classification of the types of shot transitions is done by the fuzzy system. Experimental results show that the accuracy of the shot boundary detection increases with the increase in iterations or generations of the GA optimization process. The proposed system is compared to latest techniques and yields better result in terms of F1score parameter. PMID:27127500
NASA Astrophysics Data System (ADS)
Sulaiman, Salina; Bade, Abdullah; Lee, Rechard; Tanalol, Siti Hasnah
2014-07-01
Mass Spring Model (MSM) is a highly efficient model in terms of calculations and easy implementation. Mass, spring stiffness coefficient and damping constant are three major components of MSM. This paper focuses on identifying the coefficients of spring stiffness and damping constant using automated tuning method by optimization in generating human liver model capable of responding quickly. To achieve the objective two heuristic approaches are used, namely Simulated Annealing (SA) and Genetic Algorithm (GA) on the human liver model data set. The properties of the mechanical heart, which are taken into consideration, are anisotropy and viscoelasticity. Optimization results from SA and GA are then implemented into the MSM to model two human hearts, each with its SA or GA construction parameters. These techniques are implemented while making FEM construction parameters as benchmark. Step size response of both models are obtained after MSMs were solved using Fourth Order Runge-Kutta (RK4) to compare the elasticity response of both models. Remodelled time using manual calculation methods was compared against heuristic optimization methods of SA and GA in showing that model with automatic construction is more realistic in terms of realtime interaction response time. Liver models generated using SA and GA optimization techniques are compared with liver model from manual calculation. It shows that the reconstruction time required for 1000 repetitions of SA and GA is faster than the manual method. Meanwhile comparison between construction time of SA and GA model indicates that model SA is faster than GA with varying rates of time 0.110635 seconds/1000 repetitions. Real-time interaction of mechanical properties is dependent on rate of time and speed of remodelling process. Thus, the SA and GA have proven to be suitable in enhancing realism of simulated real-time interaction in liver remodelling.
A coupled model tree (MT) genetic algorithm (GA) scheme for biofouling assessment in pipelines.
Opher, Tamar; Ostfeld, Avi
2011-11-15
A computerized learning algorithm was developed for assessing the extent of biofouling formations on the inner surfaces of water supply pipelines. Four identical pipeline experimental systems with four different types of inlet waters were set up as part of a large cooperative project between academia and industry in Israel on biofouling modeling, prediction, and prevention in pipeline systems. Samples were taken periodically for hydraulic, chemical, and biological analyses. Biofilm sampling was done using Robbins devices, carrying stainless steel coupons. An MT-GA, a hybrid model combining model trees (MTs) and genetic algorithms (GAs) in which the sampled input data are selected by the proposed methodology, was developed. The method outcome is a set of empirical linear rules which form a model tree, iteratively optimized by a GA and verified using the dataset resulting from the empirical field studies. Good correlations were achieved between modeled and observed cell coverage area within the biofilm. Sensitivity analysis was conducted by testing the model's response to changes in: (1) the biofilm measure used as output (target) variable; (2) variability of GA parameters; and (3) input attributes. The proposed methodology provides a new tool for biofouling assessment in pipelines. PMID:21978570
A random walk approach to quantum algorithms.
Kendon, Vivien M
2006-12-15
The development of quantum algorithms based on quantum versions of random walks is placed in the context of the emerging field of quantum computing. Constructing a suitable quantum version of a random walk is not trivial; pure quantum dynamics is deterministic, so randomness only enters during the measurement phase, i.e. when converting the quantum information into classical information. The outcome of a quantum random walk is very different from the corresponding classical random walk owing to the interference between the different possible paths. The upshot is that quantum walkers find themselves further from their starting point than a classical walker on average, and this forms the basis of a quantum speed up, which can be exploited to solve problems faster. Surprisingly, the effect of making the walk slightly less than perfectly quantum can optimize the properties of the quantum walk for algorithmic applications. Looking to the future, even with a small quantum computer available, the development of quantum walk algorithms might proceed more rapidly than it has, especially for solving real problems. PMID:17090467
Algorithmic crystal chemistry: A cellular automata approach
Krivovichev, S. V.
2012-01-15
Atomic-molecular mechanisms of crystal growth can be modeled based on crystallochemical information using cellular automata (a particular case of finite deterministic automata). In particular, the formation of heteropolyhedral layered complexes in uranyl selenates can be modeled applying a one-dimensional three-colored cellular automaton. The use of the theory of calculations (in particular, the theory of automata) in crystallography allows one to interpret crystal growth as a computational process (the realization of an algorithm or program with a finite number of steps).
Machine learning algorithms for damage detection: Kernel-based approaches
NASA Astrophysics Data System (ADS)
Santos, Adam; Figueiredo, Eloi; Silva, M. F. M.; Sales, C. S.; Costa, J. C. W. A.
2016-02-01
This paper presents four kernel-based algorithms for damage detection under varying operational and environmental conditions, namely based on one-class support vector machine, support vector data description, kernel principal component analysis and greedy kernel principal component analysis. Acceleration time-series from an array of accelerometers were obtained from a laboratory structure and used for performance comparison. The main contribution of this study is the applicability of the proposed algorithms for damage detection as well as the comparison of the classification performance between these algorithms and other four ones already considered as reliable approaches in the literature. All proposed algorithms revealed to have better classification performance than the previous ones.
Beyond Hydrodynamics via a Fluid Element PIC algorithm, GaPH
NASA Astrophysics Data System (ADS)
Bateson, William; Hewett, Dennis; Lambert, Michael
1996-11-01
For strongly-driven gas and plasma systems, issues of interpenetration and turbulence have led to difficulties with fluid models. For example, a Maxwell distribution within the finite volume could miss the interpenetration and shear regions between two fluids. To address these and other issues, we have extended our Grid and Particle Hydrodynamics (GaPH), a fluid element PIC code, beyond the initial high-precision, 1-D collisionless solutions[2] to 2-D with both binary and viscous drag collisions. The GaPH algorithm still aggressively probes for emerging phase space features by fitting new "particles" to the "hydrodynamic" evolution of individual particles and aggressively merges to preserves economy if interesting features fail to materialize. Recent extensions add collisonal diffusion to the hydrodynamics. Through these and other extensions, GaPH approximates Boltzmann transport thus leaving the fluid model assumption of a local Maxwell distribution behind. [1] This work was performed under the auspices of the U.S. Department of Energy by Lawrence Livermore National Laboratory under Contract W-7405-Eng-48 and by Sandia National Laboratory under Contract DE-AC04-94AL85000. [2] "Beyond Hydrodynamics via Fluid Element Particle-In-Cell", WB Bateson and DW Hewett, (submitted J. Comp. Phys. July 1996).
Genetic Algorithm Approaches to Prebiobiotic Chemistry Modeling
NASA Technical Reports Server (NTRS)
Lohn, Jason; Colombano, Silvano
1997-01-01
We model an artificial chemistry comprised of interacting polymers by specifying two initial conditions: a distribution of polymers and a fixed set of reversible catalytic reactions. A genetic algorithm is used to find a set of reactions that exhibit a desired dynamical behavior. Such a technique is useful because it allows an investigator to determine whether a specific pattern of dynamics can be produced, and if it can, the reaction network found can be then analyzed. We present our results in the context of studying simplified chemical dynamics in theorized protocells - hypothesized precursors of the first living organisms. Our results show that given a small sample of plausible protocell reaction dynamics, catalytic reaction sets can be found. We present cases where this is not possible and also analyze the evolved reaction sets.
Zhang, Wenyu; Yu, Dejian
2015-01-01
As E-government continues to develop with ever-increasing speed, the requirement to enhance traditional government systems and affairs with electronic methods that are more effective and efficient is becoming critical. As a new product of information technology, E-tendering is becoming an inevitable reality owing to its efficiency, fairness, transparency, and accountability. Thus, developing and promoting government E-tendering (GeT) is imperative. This paper presents a hybrid approach combining genetic algorithm (GA) and Technique for Order Preference by Similarity to an Ideal Solution (TOPSIS) to enable GeT to search for the optimal tenderer efficiently and fairly under circumstances where the attributes of the tenderers are expressed as fuzzy number intuitionistic fuzzy sets (FNIFSs). GA is applied to obtain the optimal weights of evaluation criteria of tenderers automatically. TOPSIS is employed to search for the optimal tenderer. A prototype system is built and validated with an illustrative example from GeT to verify the feasibility and availability of the proposed approach. PMID:26147468
Wang, Yan; Xi, Chengyu; Zhang, Shuai; Zhang, Wenyu; Yu, Dejian
2015-01-01
As E-government continues to develop with ever-increasing speed, the requirement to enhance traditional government systems and affairs with electronic methods that are more effective and efficient is becoming critical. As a new product of information technology, E-tendering is becoming an inevitable reality owing to its efficiency, fairness, transparency, and accountability. Thus, developing and promoting government E-tendering (GeT) is imperative. This paper presents a hybrid approach combining genetic algorithm (GA) and Technique for Order Preference by Similarity to an Ideal Solution (TOPSIS) to enable GeT to search for the optimal tenderer efficiently and fairly under circumstances where the attributes of the tenderers are expressed as fuzzy number intuitionistic fuzzy sets (FNIFSs). GA is applied to obtain the optimal weights of evaluation criteria of tenderers automatically. TOPSIS is employed to search for the optimal tenderer. A prototype system is built and validated with an illustrative example from GeT to verify the feasibility and availability of the proposed approach. PMID:26147468
A new algorithmic approach for fingers detection and identification
NASA Astrophysics Data System (ADS)
Mubashar Khan, Arslan; Umar, Waqas; Choudhary, Taimoor; Hussain, Fawad; Haroon Yousaf, Muhammad
2013-03-01
Gesture recognition is concerned with the goal of interpreting human gestures through mathematical algorithms. Gestures can originate from any bodily motion or state but commonly originate from the face or hand. Hand gesture detection in a real time environment, where the time and memory are important issues, is a critical operation. Hand gesture recognition largely depends on the accurate detection of the fingers. This paper presents a new algorithmic approach to detect and identify fingers of human hand. The proposed algorithm does not depend upon the prior knowledge of the scene. It detects the active fingers and Metacarpophalangeal (MCP) of the inactive fingers from an already detected hand. Dynamic thresholding technique and connected component labeling scheme are employed for background elimination and hand detection respectively. Algorithm proposed a new approach for finger identification in real time environment keeping the memory and time constraint as low as possible.
A mild reduction phosphidation approach to nanocrystalline GaP
NASA Astrophysics Data System (ADS)
Chen, Luyang; Luo, Tao; Huang, Mingxing; Gu, Yunle; Shi, Liang; Qian, Yitai
2004-12-01
Nanocrystalline gallium phosphide (GaP) has been prepared through a reduction-phosphidation by using Ga, PCl 3 as gallium and phosphorus sources and metallic sodium as reductant at 350 °C. The XRD pattern can be indexed as cublic GaP with the lattice constant of a=5.446 Å. The TEM image shows particle-like polycrystals and flake-like single crystals. The PL spectrum exhibits one peak at 330 nm for the as-prepared nanocrystalline GaP.
Silva, Leonardo W. T.; Barros, Vitor F.; Silva, Sandro G.
2014-01-01
In launching operations, Rocket Tracking Systems (RTS) process the trajectory data obtained by radar sensors. In order to improve functionality and maintenance, radars can be upgraded by replacing antennas with parabolic reflectors (PRs) with phased arrays (PAs). These arrays enable the electronic control of the radiation pattern by adjusting the signal supplied to each radiating element. However, in projects of phased array radars (PARs), the modeling of the problem is subject to various combinations of excitation signals producing a complex optimization problem. In this case, it is possible to calculate the problem solutions with optimization methods such as genetic algorithms (GAs). For this, the Genetic Algorithm with Maximum-Minimum Crossover (GA-MMC) method was developed to control the radiation pattern of PAs. The GA-MMC uses a reconfigurable algorithm with multiple objectives, differentiated coding and a new crossover genetic operator. This operator has a different approach from the conventional one, because it performs the crossover of the fittest individuals with the least fit individuals in order to enhance the genetic diversity. Thus, GA-MMC was successful in more than 90% of the tests for each application, increased the fitness of the final population by more than 20% and reduced the premature convergence. PMID:25196013
An Airborne Conflict Resolution Approach Using a Genetic Algorithm
NASA Technical Reports Server (NTRS)
Mondoloni, Stephane; Conway, Sheila
2001-01-01
An airborne conflict resolution approach is presented that is capable of providing flight plans forecast to be conflict-free with both area and traffic hazards. This approach is capable of meeting constraints on the flight plan such as required times of arrival (RTA) at a fix. The conflict resolution algorithm is based upon a genetic algorithm, and can thus seek conflict-free flight plans meeting broader flight planning objectives such as minimum time, fuel or total cost. The method has been applied to conflicts occurring 6 to 25 minutes in the future in climb, cruise and descent phases of flight. The conflict resolution approach separates the detection, trajectory generation and flight rules function from the resolution algorithm. The method is capable of supporting pilot-constructed resolutions, cooperative and non-cooperative maneuvers, and also providing conflict resolution on trajectories forecast by an onboard FMC.
Combined Simulated Annealing and Genetic Algorithm Approach to Bus Network Design
NASA Astrophysics Data System (ADS)
Liu, Li; Olszewski, Piotr; Goh, Pong-Chai
A new method - combined simulated annealing (SA) and genetic algorithm (GA) approach is proposed to solve the problem of bus route design and frequency setting for a given road network with fixed bus stop locations and fixed travel demand. The method involves two steps: a set of candidate routes is generated first and then the best subset of these routes is selected by the combined SA and GA procedure. SA is the main process to search for a better solution to minimize the total system cost, comprising user and operator costs. GA is used as a sub-process to generate new solutions. Bus demand assignment on two alternative paths is performed at the solution evaluation stage. The method was implemented on four theoretical grid networks of different size and a benchmark network. Several GA operators (crossover and mutation) were utilized and tested for their effectiveness. The results show that the proposed method can efficiently converge to the optimal solution on a small network but computation time increases significantly with network size. The method can also be used for other transport operation management problems.
DNA Microarray Data Analysis: A Novel Biclustering Algorithm Approach
NASA Astrophysics Data System (ADS)
Tchagang, Alain B.; Tewfik, Ahmed H.
2006-12-01
Biclustering algorithms refer to a distinct class of clustering algorithms that perform simultaneous row-column clustering. Biclustering problems arise in DNA microarray data analysis, collaborative filtering, market research, information retrieval, text mining, electoral trends, exchange analysis, and so forth. When dealing with DNA microarray experimental data for example, the goal of biclustering algorithms is to find submatrices, that is, subgroups of genes and subgroups of conditions, where the genes exhibit highly correlated activities for every condition. In this study, we develop novel biclustering algorithms using basic linear algebra and arithmetic tools. The proposed biclustering algorithms can be used to search for all biclusters with constant values, biclusters with constant values on rows, biclusters with constant values on columns, and biclusters with coherent values from a set of data in a timely manner and without solving any optimization problem. We also show how one of the proposed biclustering algorithms can be adapted to identify biclusters with coherent evolution. The algorithms developed in this study discover all valid biclusters of each type, while almost all previous biclustering approaches will miss some.
Oxidation of GaN: An ab initio thermodynamic approach
NASA Astrophysics Data System (ADS)
Jackson, Adam J.; Walsh, Aron
2013-10-01
GaN is a wide-band-gap semiconductor used in high-efficiency light-emitting diodes and solar cells. The solid is produced industrially at high chemical purities by deposition from a vapor phase, and oxygen may be included at this stage. Oxidation represents a potential path for tuning its properties without introducing more exotic elements or extreme processing conditions. In this work, ab initio computational methods are used to examine the energy potentials and electronic properties of different extents of oxidation in GaN. Solid-state vibrational properties of Ga, GaN, Ga2O3, and a single substitutional oxygen defect have been studied using the harmonic approximation with supercells. A thermodynamic model is outlined which combines the results of ab initio calculations with data from experimental literature. This model allows free energies to be predicted for arbitrary reaction conditions within a wide process envelope. It is shown that complete oxidation is favorable for all industrially relevant conditions, while the formation of defects can be opposed by the use of high temperatures and a high N2:O2 ratio.
An algorithmic approach for clinical management of chronic spinal pain.
Manchikanti, Laxmaiah; Helm, Standiford; Singh, Vijay; Benyamin, Ramsin M; Datta, Sukdeb; Hayek, Salim M; Fellows, Bert; Boswell, Mark V
2009-01-01
Interventional pain management, and the interventional techniques which are an integral part of that specialty, are subject to widely varying definitions and practices. How interventional techniques are applied by various specialties is highly variable, even for the most common procedures and conditions. At the same time, many payors, publications, and guidelines are showing increasing interest in the performance and costs of interventional techniques. There is a lack of consensus among interventional pain management specialists with regards to how to diagnose and manage spinal pain and the type and frequency of spinal interventional techniques which should be utilized to treat spinal pain. Therefore, an algorithmic approach is proposed, providing a step-by-step procedure for managing chronic spinal pain patients based upon evidence-based guidelines. The algorithmic approach is developed based on the best available evidence regarding the epidemiology of various identifiable sources of chronic spinal pain. Such an approach to spinal pain includes an appropriate history, examination, and medical decision making in the management of low back pain, neck pain and thoracic pain. This algorithm also provides diagnostic and therapeutic approaches to clinical management utilizing case examples of cervical, lumbar, and thoracic spinal pain. An algorithm for investigating chronic low back pain without disc herniation commences with a clinical question, examination and imaging findings. If there is evidence of radiculitis, spinal stenosis, or other demonstrable causes resulting in radiculitis, one may proceed with diagnostic or therapeutic epidural injections. In the algorithmic approach, facet joints are entertained first in the algorithm because of their commonality as a source of chronic low back pain followed by sacroiliac joint blocks if indicated and provocation discography as the last step. Based on the literature, in the United States, in patients without disc
Stall Recovery Guidance Algorithms Based on Constrained Control Approaches
NASA Technical Reports Server (NTRS)
Stepanyan, Vahram; Krishnakumar, Kalmanje; Kaneshige, John; Acosta, Diana
2016-01-01
Aircraft loss-of-control, in particular approach to stall or fully developed stall, is a major factor contributing to aircraft safety risks, which emphasizes the need to develop algorithms that are capable of assisting the pilots to identify the problem and providing guidance to recover the aircraft. In this paper we present several stall recovery guidance algorithms, which are implemented in the background without interfering with flight control system and altering the pilot's actions. They are using input and state constrained control methods to generate guidance signals, which are provided to the pilot in the form of visual cues. It is the pilot's decision to follow these signals. The algorithms are validated in the pilot-in-the loop medium fidelity simulation experiment.
Modeling and Algorithmic Approaches to Constitutively-Complex, Microstructured Fluids
Miller, Gregory H.; Forest, Gregory
2011-12-22
We present a new multiscale model for complex uids based on three scales: microscopic, kinetic, and continuum. We choose the microscopic level as Kramers' bead-rod model for polymers, which we describe as a system of stochastic di erential equations with an implicit constraint formulation. The associated Fokker-Planck equation is then derived, and adiabatic elimination removes the fast momentum coordinates. Approached in this way, the kinetic level reduces to a dispersive drift equation. The continuum level is modeled with a nite volume Godunov-projection algorithm. We demonstrate computation of viscoelastic stress divergence using this multiscale approach.
NASA Astrophysics Data System (ADS)
Javad Kazemzadeh-Parsi, Mohammad; Daneshmand, Farhang; Ahmadfard, Mohammad Amin; Adamowski, Jan; Martel, Richard
2015-01-01
In the present study, an optimization approach based on the firefly algorithm (FA) is combined with a finite element simulation method (FEM) to determine the optimum design of pump and treat remediation systems. Three multi-objective functions in which pumping rate and clean-up time are design variables are considered and the proposed FA-FEM model is used to minimize operating costs, total pumping volumes and total pumping rates in three scenarios while meeting water quality requirements. The groundwater lift and contaminant concentration are also minimized through the optimization process. The obtained results show the applicability of the FA in conjunction with the FEM for the optimal design of groundwater remediation systems. The performance of the FA is also compared with the genetic algorithm (GA) and the FA is found to have a better convergence rate than the GA.
NASA Astrophysics Data System (ADS)
Aleardi, Mattia
2015-06-01
Predicting missing log data is a useful capability for geophysicists. Geophysical measurements in boreholes are frequently affected by gaps in the recording of one or more logs. In particular, sonic and shear sonic logs are often recorded over limited intervals along the well path, but the information these logs contain is crucial for many geophysical applications. Estimating missing log intervals from a set of recorded logs is therefore of great interest. In this work, I propose to estimate the data in missing parts of velocity logs using a genetic algorithm (GA) optimisation and I demonstrate that this method is capable of extracting linear or exponential relations that link the velocity to other available logs. The technique was tested on different sets of logs (gamma ray, resistivity, density, neutron, sonic and shear sonic) from three wells drilled in different geological settings and through different lithologies (sedimentary and intrusive rocks). The effectiveness of this methodology is demonstrated by a series of blind tests and by evaluating the correlation coefficients between the true versus predicted velocity values. The combination of GA optimisation with a Gibbs sampler (GS) and subsequent Monte Carlo simulations allows the uncertainties in the final predicted velocities to be reliably quantified. The GA method is also compared with the neural networks (NN) approach and classical multilinear regression. The comparisons show that the GA, NN and multilinear methods provide velocity estimates with the same predictive capability when the relation between the input logs and the seismic velocity is approximately linear. The GA and NN approaches are more robust when the relations are non-linear. However, in all cases, the main advantages of the GA optimisation procedure over the NN approach is that it directly provides an interpretable and simple equation that relates the input and predicted logs. Moreover, the GA method is not affected by the disadvantages
NASA Astrophysics Data System (ADS)
Konak, Abdullah
2014-01-01
This article presents a network design problem with relays considering the two-edge network connectivity. The problem arises in telecommunications and logistic networks where a constraint is imposed on the distance that a commodity can travel on a route without being processed by a relay, and the survivability of the network is critical in case of a component failure. The network design problem involves selecting two-edge disjoint paths between source and destination node pairs and determining the location of relays to minimize the network design cost. The formulated problem is solved by a hybrid approach of a genetic algorithm (GA) and a Lagrangian heuristic such that the GA searches for two-edge disjoint paths for each commodity, and the Lagrangian heuristic is used to determine relays on these paths. The performance of the proposed hybrid approach is compared to the previous approaches from the literature, with promising results.
A new distributed systems scheduling algorithm: a swarm intelligence approach
NASA Astrophysics Data System (ADS)
Haghi Kashani, Mostafa; Sarvizadeh, Raheleh; Jameii, Mahdi
2011-12-01
The scheduling problem in distributed systems is known as an NP-complete problem, and methods based on heuristic or metaheuristic search have been proposed to obtain optimal and suboptimal solutions. The task scheduling is a key factor for distributed systems to gain better performance. In this paper, an efficient method based on memetic algorithm is developed to solve the problem of distributed systems scheduling. With regard to load balancing efficiently, Artificial Bee Colony (ABC) has been applied as local search in the proposed memetic algorithm. The proposed method has been compared to existing memetic-Based approach in which Learning Automata method has been used as local search. The results demonstrated that the proposed method outperform the above mentioned method in terms of communication cost.
Moghri, Mehdi; Omidi, Mostafa; Farahnakian, Masoud
2014-01-01
During the past decade, polymer nanocomposites attracted considerable investment in research and development worldwide. One of the key factors that affect the quality of polymer nanocomposite products in machining is surface roughness. To obtain high quality products and reduce machining costs it is very important to determine the optimal machining conditions so as to achieve enhanced machining performance. The objective of this paper is to develop a predictive model using a combined design of experiments and artificial intelligence approach for optimization of surface roughness in milling of polyamide-6 (PA-6) nanocomposites. A surface roughness predictive model was developed in terms of milling parameters (spindle speed and feed rate) and nanoclay (NC) content using artificial neural network (ANN). As the present study deals with relatively small number of data obtained from full factorial design, application of genetic algorithm (GA) for ANN training is thought to be an appropriate approach for the purpose of developing accurate and robust ANN model. In the optimization phase, a GA is considered in conjunction with the explicit nonlinear function derived from the ANN to determine the optimal milling parameters for minimization of surface roughness for each PA-6 nanocomposite. PMID:24578636
An ab initio-based approach to the stability of GaN(0 0 0 1) surfaces under Ga-rich conditions
NASA Astrophysics Data System (ADS)
Ito, Tomonori; Akiyama, Toru; Nakamura, Kohji
2009-05-01
Structural stability of GaN(0 0 0 1) under Ga-rich conditions is systematically investigated by using our ab initio-based approach. The surface phase diagram for GaN(0 0 0 1) including (2×2) and pseudo-(1×1) is obtained as functions of temperature and Ga beam equivalent pressure by comparing chemical potentials of Ga atom in the gas phase with that on the surface. The calculated results reveal that the pseudo-(1×1) appearing below 684-973 K changes its structure to the (2×2) with Ga adatom at higher temperatures beyond 767-1078 K via the newly found (1×1) with two adlayers of Ga. These results are consistent with the stable temperature range of both the pseudo-(1×1) and (2×2) with Ga adatom obtained experimentally. Furthermore, it should be noted that the structure with another coverage of Ga adatoms between the (1×1) and (2×2)-Ga does not appear as a stable structure of GaN(0 0 0 1). Furthermore, ghost island formation observed by scanning tunneling microscopy is discussed on the basis of the phase diagram.
A genetic algorithm-based approach to flexible flow-line scheduling with variable lot sizes.
Lee, I; Sikora, R; Shaw, M J
1997-01-01
Genetic algorithms (GAs) have been used widely for such combinatorial optimization problems as the traveling salesman problem (TSP), the quadratic assignment problem (QAP), and job shop scheduling. In all of these problems there is usually a well defined representation which GA's use to solve the problem. We present a novel approach for solving two related problems-lot sizing and sequencing-concurrently using GAs. The essence of our approach lies in the concept of using a unified representation for the information about both the lot sizes and the sequence and enabling GAs to evolve the chromosome by replacing primitive genes with good building blocks. In addition, a simulated annealing procedure is incorporated to further improve the performance. We evaluate the performance of applying the above approach to flexible flow line scheduling with variable lot sizes for an actual manufacturing facility, comparing it to such alternative approaches as pair wise exchange improvement, tabu search, and simulated annealing procedures. The results show the efficacy of this approach for flexible flow line scheduling. PMID:18255838
NASA Astrophysics Data System (ADS)
Song, Kaishan; Li, Lin; Li, Shuai; Tedesco, Lenore; Hall, Bob; Li, Zuchuan
2012-08-01
Eagle Creek, Morse and Geist reservoirs, drinking water supply sources for the Indianapolis, Indiana, USA metropolitan region, are experiencing nuisance cyanobacterial blooms. Hyperspectral remote sensing has been proven to be an effective tool for phycocyanin (C-PC) concentration retrieval, a proxy pigment unique to cyanobacteria in freshwater ecosystems. An adaptive model based on genetic algorithm and partial least squares (GA-PLS), together with three-band algorithm (TBA) and other band ratio algorithms were applied to hyperspectral data acquired from in situ (ASD spectrometer) and airborne (AISA sensor) platforms. The results indicated that GA-PLS achieved high correlation between measured and estimated C-PC for GR (RMSE = 16.3 μg/L, RMSE% = 18.2; range (R): 2.6-185.1 μg/L), MR (RMSE = 8.7 μg/L, RMSE% = 15.6; R: 3.3-371.0 μg/L) and ECR (RMSE = 19.3 μg/L, RMSE% = 26.4; R: 0.7-245.0 μg/L) for the in situ datasets. TBA also performed well compared to other band ratio algorithms due to its optimal band tuning process and the reduction of backscattering effects through the third band. GA-PLS (GR: RMSE = 24.1 μg/L, RMSE% = 25.2, R: 25.2-185.1 μg/L; MR: RMSE = 15.7 μg/L, RMSE% = 37.4, R: 2.0-135.1 μg/L) and TBA (GR: RMSE = 28.3 μg/L, RMSE% = 30.1; MR: RMSE = 17.7 μg/L, RMSE% = 41.9) methods results in somewhat lower accuracy using AISA imagery data, which is likely due to atmospheric correction or radiometric resolution. GA-PLS (TBA) obtained an RMSE of 24.82 μg/L (35.8 μg/L), and RMSE% of 31.24 (43.5) between measured and estimated C-PC for aggregated datasets. C-PC maps were generated through GA-PLS using AISA imagery data. The C-PC concentration had an average value of 67.31 ± 44.23 μg/L in MR with a large range of concentration, while the GR had a higher average value 103.17 ± 33.45 μg/L.
A genetic algorithm approach for assessing soil liquefaction potential based on reliability method
NASA Astrophysics Data System (ADS)
Bagheripour, M. H.; Shooshpasha, I.; Afzalirad, M.
2012-02-01
Deterministic approaches are unable to account for the variations in soil's strength properties, earthquake loads, as well as source of errors in evaluations of liquefaction potential in sandy soils which make them questionable against other reliability concepts. Furthermore, deterministic approaches are incapable of precisely relating the probability of liquefaction and the factor of safety (FS). Therefore, the use of probabilistic approaches and especially, reliability analysis is considered since a complementary solution is needed to reach better engineering decisions. In this study, Advanced First-Order Second-Moment (AFOSM) technique associated with genetic algorithm (GA) and its corresponding sophisticated optimization techniques have been used to calculate the reliability index and the probability of liquefaction. The use of GA provides a reliable mechanism suitable for computer programming and fast convergence. A new relation is developed here, by which the liquefaction potential can be directly calculated based on the estimated probability of liquefaction ( P L ), cyclic stress ratio (CSR) and normalized standard penetration test (SPT) blow counts while containing a mean error of less than 10% from the observational data. The validity of the proposed concept is examined through comparison of the results obtained by the new relation and those predicted by other investigators. A further advantage of the proposed relation is that it relates P L and FS and hence it provides possibility of decision making based on the liquefaction risk and the use of deterministic approaches. This could be beneficial to geotechnical engineers who use the common methods of FS for evaluation of liquefaction. As an application, the city of Babolsar which is located on the southern coasts of Caspian Sea is investigated for liquefaction potential. The investigation is based primarily on in situ tests in which the results of SPT are analysed.
NASA Astrophysics Data System (ADS)
White, Ronald P.; Mayne, Howard R.
2000-05-01
An annealing schedule, T(t), is the temperature as function of time whose goal is to bring a system from some initial low-order state to a final high-order state. We use the probability in the lowest energy level as the order parameter, so that an ideally annealed system would have all its population in its ground-state. We consider a model system comprised of discrete energy levels separated by activation barriers. We have carried out annealing calculations on this system for a range of system parameters. In particular, we considered the schedule as a function of the energy level spacing, of the height of the activation barriers, and, in some cases, as a function of degeneracies of the levels. For a given set of physical parameters, and maximum available time, tm, we were able to obtain the optimal schedule by using a genetic algorithm (GA) approach. For the two-level system, analytic solutions are available, and were compared with the GA-optimized results. The agreement was essentially exact. We were able to identify systematic behaviors of the schedules and trends in final probabilities as a function of parameters. We have also carried out Metropolis Monte Carlo (MMC) calculations on simple potential energy functions using the optimal schedules available from the model calculations. Agreement between the model and MMC calculations was excellent.
NASA Astrophysics Data System (ADS)
Jamshidi, Saeid; Boozarjomehry, Ramin Bozorgmehry; Pishvaie, Mahmoud Reza
2009-10-01
In pore network modeling, the void space of a rock sample is represented at the microscopic scale by a network of pores connected by throats. Construction of a reasonable representation of the geometry and topology of the pore space will lead to a reliable prediction of the properties of porous media. Recently, the theory of multi-cellular growth (or L-systems) has been used as a flexible tool for generation of pore network models which do not require any special information such as 2D SEM or 3D pore space images. In general, the networks generated by this method are irregular pore network models which are inherently closer to the complicated nature of the porous media rather than regular lattice networks. In this approach, the construction process is controlled only by the production rules that govern the development process of the network. In this study, genetic algorithm has been used to obtain the optimum values of the uncertain parameters of these production rules to build an appropriate irregular lattice network capable of the prediction of both static and hydraulic information of the target porous medium.
Algorithmic approaches to protein-protein interaction site prediction.
Aumentado-Armstrong, Tristan T; Istrate, Bogdan; Murgita, Robert A
2015-01-01
Interaction sites on protein surfaces mediate virtually all biological activities, and their identification holds promise for disease treatment and drug design. Novel algorithmic approaches for the prediction of these sites have been produced at a rapid rate, and the field has seen significant advancement over the past decade. However, the most current methods have not yet been reviewed in a systematic and comprehensive fashion. Herein, we describe the intricacies of the biological theory, datasets, and features required for modern protein-protein interaction site (PPIS) prediction, and present an integrative analysis of the state-of-the-art algorithms and their performance. First, the major sources of data used by predictors are reviewed, including training sets, evaluation sets, and methods for their procurement. Then, the features employed and their importance in the biological characterization of PPISs are explored. This is followed by a discussion of the methodologies adopted in contemporary prediction programs, as well as their relative performance on the datasets most recently used for evaluation. In addition, the potential utility that PPIS identification holds for rational drug design, hotspot prediction, and computational molecular docking is described. Finally, an analysis of the most promising areas for future development of the field is presented. PMID:25713596
Approach to programming multiprocessing algorithms on the Denelcor HEP
Lusk, E.L.; Overbeek, R.A.
1983-12-01
In the process of learning how to write code for the Denelcor HEP, we have developed an approach that others may well find useful. We believe that the basic synchronization primitives of the HEP (i.e., asynchronous variables), along with the prototypical patterns for their use given in the HEP FORTRAN 77 User's Guide, form too low-level a conceptual basis for the formulation of multiprocessing algorithms. We advocate the use of monitors, which can be easily implemented using the HEP primitives. Attempts to solve substantial problems without introducing higher-level constructs such as monitors can produce code that is unreliable, unintelligible, and restricted to the specific dialect of FORTRAN currently supported on the HEP. Our experience leads us to believe that solutions which are both clear and efficient can be formulated using monitors.
A genetic algorithmic approach to antenna null-steering using a cluster computer.
NASA Astrophysics Data System (ADS)
Recine, Greg; Cui, Hong-Liang
2001-06-01
We apply a genetic algorithm (GA) to the problem of electronically steering the maximums and nulls of an antenna array to desired positions (null toward enemy listener/jammer, max toward friendly listener/transmitter). The antenna pattern itself is computed using NEC2 which is called by the main GA program. Since a GA naturally lends itself to parallelization, this simulation was applied to our new twin 64-node cluster computers (Gemini). Design issues and uses of the Gemini cluster in our group are also discussed.
First-principles approach to investigate toroidal property of magnetoelectric multiferroic GaFeO3
NASA Astrophysics Data System (ADS)
Nie, Yung-mau
2016-01-01
A first-principles approach incorporating the concept of toroidal moments as a measure of the spin vortex is proposed and applied to simulate the toroidization of magnetoelectric multiferroic GaFeO3. The nature of space-inversion and time-reversal violations of ferrotoroidics is reproduced in the simulated magnetic structure of GaFeO3. For undoped GaFeO3, a toroidal moment of -22.38 μB Å per unit cell was obtained, which is the best theoretical estimate till date. Guided by the spin vortex free-energy minimization perturbed by an externally applied field, it was discovered that the minority spin markedly biases the whole toroidization. In summary, this approach not only calculates the toroidal moment but provides a way to understand the toroidal nature of magnetoelectric multiferroics.
Investigation of new approaches for InGaN growth with high indium content for CPV application
Arif, Muhammad; Salvestrini, Jean Paul; Sundaram, Suresh; Streque, Jérémy; Gmili, Youssef El; Puybaret, Renaud; Voss, Paul L.; Belahsene, Sofiane; Ramdane, Abderahim; Martinez, Anthony; Patriarche, Gilles; Fix, Thomas; Slaoui, Abdelillah; Ougazzaden, Abdallah
2015-09-28
We propose to use two new approaches that may overcome the issues of phase separation and high dislocation density in InGaN-based PIN solar cells. The first approach consists in the growth of a thick multi-layered InGaN/GaN absorber. The periodical insertion of the thin GaN interlayers should absorb the In excess and relieve compressive strain. The InGaN layers need to be thin enough to remain fully strained and without phase separation. The second approach consists in the growth of InGaN nano-structures for the achievement of high In content thick InGaN layers. It allows the elimination of the preexisting dislocations in the underlying template. It also allows strain relaxation of InGaN layers without any dislocations, leading to higher In incorporation and reduced piezo-electric effect. The two approaches lead to structural, morphological, and luminescence properties that are significantly improved when compared to those of thick InGaN layers. Corresponding full PIN structures have been realized by growing a p-type GaN layer on the top the half PIN structures. External quantum efficiency, electro-luminescence, and photo-current characterizations have been carried out on the different structures and reveal an enhancement of the performances of the InGaN PIN PV cells when the thick InGaN layer is replaced by either InGaN/GaN multi-layered or InGaN nanorod layer.
Investigation of new approaches for InGaN growth with high indium content for CPV application
NASA Astrophysics Data System (ADS)
Arif, Muhammad; Sundaram, Suresh; Streque, Jérémy; Gmili, Youssef El; Puybaret, Renaud; Belahsene, Sofiane; Ramdane, Abderahim; Martinez, Anthony; Patriarche, Gilles; Fix, Thomas; Slaoui, Abdelillah; Voss, Paul L.; Salvestrini, Jean Paul; Ougazzaden, Abdallah
2015-09-01
We propose to use two new approaches that may overcome the issues of phase separation and high dislocation density in InGaN-based PIN solar cells. The first approach consists in the growth of a thick multi-layered InGaN/GaN absorber. The periodical insertion of the thin GaN interlayers should absorb the In excess and relieve compressive strain. The InGaN layers need to be thin enough to remain fully strained and without phase separation. The second approach consists in the growth of InGaN nano-structures for the achievement of high In content thick InGaN layers. It allows the elimination of the preexisting dislocations in the underlying template. It also allows strain relaxation of InGaN layers without any dislocations, leading to higher In incorporation and reduced piezo-electric effect. The two approaches lead to structural, morphological, and luminescence properties that are significantly improved when compared to those of thick InGaN layers. Corresponding full PIN structures have been realized by growing a p-type GaN layer on the top the half PIN structures. External quantum efficiency, electro-luminescence, and photo-current characterizations have been carried out on the different structures and reveal an enhancement of the performances of the InGaN PIN PV cells when the thick InGaN layer is replaced by either InGaN/GaN multi-layered or InGaN nanorod layer.
A simple algorithm to compute the peak power output of GaAs/Ge solar cells on the Martian surface
Glueck, P.R.; Bahrami, K.A.
1995-12-31
The Jet Propulsion Laboratory`s (JPL`s) Mars Pathfinder Project will deploy a robotic ``microrover`` on the surface of Mars in the summer of 1997. This vehicle will derive primary power from a GaAs/Ge solar array during the day and will ``sleep`` at night. This strategy requires that the rover be able to (1) determine when it is necessary to save the contents of volatile memory late in the afternoon and (2) determine when sufficient power is available to resume operations in the morning. An algorithm was developed that estimates the peak power point of the solar array from the solar array short-circuit current and temperature telemetry, and provides functional redundancy for both measurements using the open-circuit voltage telemetry. The algorithm minimizes vehicle processing and memory utilization by using linear equations instead of look-up tables to estimate peak power with very little loss in accuracy. This paper describes the method used to obtain the algorithm and presents the detailed algorithm design.
Newton Algorithms for Analytic Rotation: An Implicit Function Approach
ERIC Educational Resources Information Center
Boik, Robert J.
2008-01-01
In this paper implicit function-based parameterizations for orthogonal and oblique rotation matrices are proposed. The parameterizations are used to construct Newton algorithms for minimizing differentiable rotation criteria applied to "m" factors and "p" variables. The speed of the new algorithms is compared to that of existing algorithms and to…
Development of a Compound Optimization Approach Based on Imperialist Competitive Algorithm
NASA Astrophysics Data System (ADS)
Wang, Qimei; Yang, Zhihong; Wang, Yong
In this paper, an improved novel approach is developed for the imperialist competitive algorithm to achieve a greater performance. The Nelder-Meand simplex method is applied to execute alternately with the original procedures of the algorithm. The approach is tested on twelve widely-used benchmark functions and is also compared with other relative studies. It is shown that the proposed approach has a faster convergence rate, better search ability, and higher stability than the original algorithm and other relative methods.
Peng, Bin; Liu, Ke-ling; Li, Zhi-min; Wang, Yue-song; Huang, Tu-jiang
2002-06-01
Genetic algorithm (GA) is used in automatic qualitative analysis by a sequential inductively coupled plasma spectrometer (ICP-AES) and a computer program is developed in this paper. No any standard samples are needed, and spectroscopic interferences can be eliminated. All elements and their concentration ranges of an unknown sample can be reported. The replication rate Pr, crossover rate Pc, and mutation rate of the genetic algorithm were adjusted to be 0.6, 0.4 and 0 respectively. The analytical results of GA are in good agreement with the reference values. It indicates that, combined with the intensity information, the GA can be applied to spectroscopic qualitative analysis and expected to become an effective method in qualitative analysis in ICP-AES after further work. PMID:12938334
Approach to complex upper extremity injury: an algorithm.
Ng, Zhi Yang; Askari, Morad; Chim, Harvey
2015-02-01
Patients with complex upper extremity injuries represent a unique subset of the trauma population. In addition to extensive soft tissue defects affecting the skin, bone, muscles and tendons, or the neurovasculature in various combinations, there is usually concomitant involvement of other body areas and organ systems with the potential for systemic compromise due to the underlying mechanism of injury and resultant sequelae. In turn, this has a direct impact on the definitive reconstructive plan. Accurate assessment and expedient treatment is thus necessary to achieve optimal surgical outcomes with the primary goal of limb salvage and functional restoration. Nonetheless, the characteristics of these injuries places such patients at an increased risk of complications ranging from limb ischemia, recalcitrant infections, failure of bony union, intractable pain, and most devastatingly, limb amputation. In this article, the authors present an algorithmic approach toward complex injuries of the upper extremity with due consideration for the various reconstructive modalities and timing of definitive wound closure for the best possible clinical outcomes. PMID:25685098
Lymphocytic panniculitis: an algorithmic approach to lymphocytes in subcutaneous tissue.
Shiau, Carolyn J; Abi Daoud, Marie S; Wong, Se Mang; Crawford, Richard I
2015-12-01
The diagnosis of panniculitis is a relatively rare occurrence for many practising pathologists. The smaller subset of lymphocyte-predominant panniculitis is further complicated by the diagnostic consideration of T cell lymphoma involving the subcutaneous tissue, mimicking inflammatory causes of panniculitis. Accurate classification of the panniculitis is crucial to direct clinical management as treatment options may vary from non-medical therapy to immunosuppressive agents to aggressive chemotherapy. Many diseases show significant overlap in clinical and histological features, making the process of determining a specific diagnosis very challenging. However, with an adequate biopsy including skin and deep subcutaneous tissue, a collaborative effort between clinician and pathologist can often lead to a specific diagnosis. This review provides an algorithmic approach to the diagnosis of lymphocyte-predominant panniculitis, including entities of septal-predominant pattern panniculitis (erythema nodosum, deep necrobiosis lipoidica, morphea profunda and sclerosing panniculitis) and lobular-predominant pattern panniculitis (lupus erythematous panniculitis/lupus profundus, subcutaneous panniculitis-like T cell lymphoma, cutaneous γ-δ T cell lymphoma, Borrelia infection and cold panniculitis). PMID:26602413
A set-membership approach to blind channel equalization algorithm
NASA Astrophysics Data System (ADS)
Li, Yue-ming
2013-03-01
The constant modulus algorithm (CMA) has low computational complexity while presenting slow convergence and possible convergence to local minima, the CMA family of algorithms based on affine projection version (AP-CMA) alleviates the speed limitations of the CMA. However, the computational complexity has been a weak point in the implementation of AP-CMA. To reduce the computational complexity of the adaptive filtering algorithm, a new AP-CMA algorithm based on set membership (SM-AP-CMA) is proposed. The new algorithm combines a bounded error specification on the adaptive filter with the concept of data-reusing. Simulations confirmed that the convergence rate of the proposed algorithm is significantly faster; meanwhile, the excess mean square error can be maintained in a relatively low level and a substantial reduction in the number of updates when compared with its conventional counterpart.
NASA Technical Reports Server (NTRS)
Jenkins, Phillip P.; Hepp, Aloysius F.; Power, Michael B.; Macinnes, Andrew N.; Barron, Andrew R.
1993-01-01
A two order-of-magnitude enhancement of photoluminescence intensity relative to untreated GaAs has been observed for GaAs surfaces coated with chemical vapor-deposited GaS. The increase in photoluminescence intensity can be viewed as an effective reduction in surface recombination velocity and/or band bending. The gallium cluster (/t-Bu/GaS)4 was used as a single-source precursor for the deposition of GaS thin films. The cubane core of the structurally-characterized precursor is retained in the deposited film producing a cubic phase. Furthermore, a near-epitaxial growth is observed for the GaS passivating layer. Films were characterized by transmission electron microscopy, X-ray powder diffraction, and X-ray photoelectron and Rutherford backscattering spectroscopies.
A Functional Programming Approach to AI Search Algorithms
ERIC Educational Resources Information Center
Panovics, Janos
2012-01-01
The theory and practice of search algorithms related to state-space represented problems form the major part of the introductory course of Artificial Intelligence at most of the universities and colleges offering a degree in the area of computer science. Students usually meet these algorithms only in some imperative or object-oriented language…
Planning additional drilling campaign using two-space genetic algorithm: A game theoretical approach
NASA Astrophysics Data System (ADS)
Kumral, Mustafa; Ozer, Umit
2013-03-01
Grade and tonnage are the most important technical uncertainties in mining ventures because of the use of estimations/simulations, which are mostly generated from drill data. Open pit mines are planned and designed on the basis of the blocks representing the entire orebody. Each block has different estimation/simulation variance reflecting uncertainty to some extent. The estimation/simulation realizations are submitted to mine production scheduling process. However, the use of a block model with varying estimation/simulation variances will lead to serious risk in the scheduling. In the medium of multiple simulations, the dispersion variances of blocks can be thought to regard technical uncertainties. However, the dispersion variance cannot handle uncertainty associated with varying estimation/simulation variances of blocks. This paper proposes an approach that generates the configuration of the best additional drilling campaign to generate more homogenous estimation/simulation variances of blocks. In other words, the objective is to find the best drilling configuration in such a way as to minimize grade uncertainty under budget constraint. Uncertainty measure of the optimization process in this paper is interpolation variance, which considers data locations and grades. The problem is expressed as a minmax problem, which focuses on finding the best worst-case performance i.e., minimizing interpolation variance of the block generating maximum interpolation variance. Since the optimization model requires computing the interpolation variances of blocks being simulated/estimated in each iteration, the problem cannot be solved by standard optimization tools. This motivates to use two-space genetic algorithm (GA) approach to solve the problem. The technique has two spaces: feasible drill hole configuration with minimization of interpolation variance and drill hole simulations with maximization of interpolation variance. Two-space interacts to find a minmax solution
Rafiei, Hamid; Khanzadeh, Marziyeh; Mozaffari, Shahla; Bostanifar, Mohammad Hassan; Avval, Zhila Mohajeri; Aalizadeh, Reza; Pourbasheer, Eslam
2016-01-01
Quantitative structure-activity relationship (QSAR) study has been employed for predicting the inhibitory activities of the Hepatitis C virus (HCV) NS5B polymerase inhibitors. A data set consisted of 72 compounds was selected, and then different types of molecular descriptors were calculated. The whole data set was split into a training set (80 % of the dataset) and a test set (20 % of the dataset) using principle component analysis. The stepwise (SW) and the genetic algorithm (GA) techniques were used as variable selection tools. Multiple linear regression method was then used to linearly correlate the selected descriptors with inhibitory activities. Several validation technique including leave-one-out and leave-group-out cross-validation, Y-randomization method were used to evaluate the internal capability of the derived models. The external prediction ability of the derived models was further analyzed using modified r2, concordance correlation coefficient values and Golbraikh and Tropsha acceptable model criteria's. Based on the derived results (GA-MLR), some new insights toward molecular structural requirements for obtaining better inhibitory activity were obtained. PMID:27065774
Flower pollination algorithm: A novel approach for multiobjective optimization
NASA Astrophysics Data System (ADS)
Yang, Xin-She; Karamanoglu, Mehmet; He, Xingshi
2014-09-01
Multiobjective design optimization problems require multiobjective optimization techniques to solve, and it is often very challenging to obtain high-quality Pareto fronts accurately. In this article, the recently developed flower pollination algorithm (FPA) is extended to solve multiobjective optimization problems. The proposed method is used to solve a set of multiobjective test functions and two bi-objective design benchmarks, and a comparison of the proposed algorithm with other algorithms has been made, which shows that the FPA is efficient with a good convergence rate. Finally, the importance for further parametric studies and theoretical analysis is highlighted and discussed.
Genetic algorithms approach for the extraction of the polygonal approximation of planar objects
NASA Astrophysics Data System (ADS)
Erives, Hector; Parra-Loera, Ramon
1996-06-01
A new approach to the extraction of the polygonal approximation is presented. The method obtains a smaller set of the important features by means of an evolutionary algorithm. A genetic approach with some heuristics, improves contour approximation search by starting with a parallel search at various points in the contour. The algorithm uses genetic algorithms to encode a polygonal approximation as a chromosome and evolve it to provide a polygonal approximation. Experimental results are provided.
Motion Cueing Algorithm Development: Human-Centered Linear and Nonlinear Approaches
NASA Technical Reports Server (NTRS)
Houck, Jacob A. (Technical Monitor); Telban, Robert J.; Cardullo, Frank M.
2005-01-01
While the performance of flight simulator motion system hardware has advanced substantially, the development of the motion cueing algorithm, the software that transforms simulated aircraft dynamics into realizable motion commands, has not kept pace. Prior research identified viable features from two algorithms: the nonlinear "adaptive algorithm", and the "optimal algorithm" that incorporates human vestibular models. A novel approach to motion cueing, the "nonlinear algorithm" is introduced that combines features from both approaches. This algorithm is formulated by optimal control, and incorporates a new integrated perception model that includes both visual and vestibular sensation and the interaction between the stimuli. Using a time-varying control law, the matrix Riccati equation is updated in real time by a neurocomputing approach. Preliminary pilot testing resulted in the optimal algorithm incorporating a new otolith model, producing improved motion cues. The nonlinear algorithm vertical mode produced a motion cue with a time-varying washout, sustaining small cues for longer durations and washing out large cues more quickly compared to the optimal algorithm. The inclusion of the integrated perception model improved the responses to longitudinal and lateral cues. False cues observed with the NASA adaptive algorithm were absent. The neurocomputing approach was crucial in that the number of presentations of an input vector could be reduced to meet the real time requirement without degrading the quality of the motion cues.
NASA Astrophysics Data System (ADS)
Lindsay, Anthony; McCloskey, John; Simão, Nuno; Murphy, Shane; Bhloscaidh, Mairead Nic
2014-05-01
Identifying fault sections where slip deficits have accumulated may provide a means for understanding sequences of large megathrust earthquakes. Stress accumulated during the interseismic period on an active megathrust is stored as potential slip, referred to as slip deficit, along locked sections of the fault. Analysis of the spatial distribution of slip during antecedent events along the fault will show where the locked plate has spent its stored slip. Areas of unreleased slip indicate where the potential for large events remain. The location of recent earthquakes and their distribution of slip can be estimated from instrumentally recorded seismic and geodetic data. However, long-term slip-deficit modelling requires detailed information on the size and distribution of slip for pre-instrumental events over hundreds of years covering more than one 'seismic cycle'. This requires the exploitation of proxy sources of data. Coral microatolls, growing in the intertidal zone of the outer island arc of the Sunda trench, present the possibility of reconstructing slip for a number of pre-instrumental earthquakes. Their growth is influenced by tectonic flexing of the continental plate beneath them; they act as long term recorders of the vertical component of deformation. However, the sparse distribution of data available using coral geodesy results in a under determined problem with non-unique solutions. Rather than accepting any one realisation as the definite model satisfying the coral displacement data, a Monte Carlo approach identifies a suite of models consistent with the observations. Using a Genetic Algorithm to accelerate the identification of desirable models, we have developed a Monte Carlo Slip Estimator- Genetic Algorithm (MCSE-GA) which exploits the full range of uncertainty associated with the displacements. Each iteration of the MCSE-GA samples different values from within the spread of uncertainties associated with each coral displacement. The Genetic
A Decision Theoretic Approach to Evaluate Radiation Detection Algorithms
Nobles, Mallory A.; Sego, Landon H.; Cooley, Scott K.; Gosink, Luke J.; Anderson, Richard M.; Hays, Spencer E.; Tardiff, Mark F.
2013-07-01
There are a variety of sensor systems deployed at U.S. border crossings and ports of entry that scan for illicit nuclear material. In this work, we develop a framework for comparing the performance of detection algorithms that interpret the output of these scans and determine when secondary screening is needed. We optimize each algorithm to minimize its risk, or expected loss. We measure an algorithm’s risk by considering its performance over a sample, the probability distribution of threat sources, and the consequence of detection errors. While it is common to optimize algorithms by fixing one error rate and minimizing another, our framework allows one to simultaneously consider multiple types of detection errors. Our framework is flexible and easily adapted to many different assumptions regarding the probability of a vehicle containing illicit material, and the relative consequences of a false positive and false negative errors. Our methods can therefore inform decision makers of the algorithm family and parameter values which best reduce the threat from illicit nuclear material, given their understanding of the environment at any point in time. To illustrate the applicability of our methods, in this paper, we compare the risk from two families of detection algorithms and discuss the policy implications of our results.
A genetic algorithm approach in interface and surface structure optimization
Zhang, Jian
2010-01-01
The thesis is divided into two parts. In the first part a global optimization method is developed for the interface and surface structures optimization. Two prototype systems are chosen to be studied. One is Si[001] symmetric tilted grain boundaries and the other is Ag/Au induced Si(111) surface. It is found that Genetic Algorithm is very efficient in finding lowest energy structures in both cases. Not only existing structures in the experiments can be reproduced, but also many new structures can be predicted using Genetic Algorithm. Thus it is shown that Genetic Algorithm is a extremely powerful tool for the material structures predictions. The second part of the thesis is devoted to the explanation of an experimental observation of thermal radiation from three-dimensional tungsten photonic crystal structures. The experimental results seems astounding and confusing, yet the theoretical models in the paper revealed the physics insight behind the phenomena and can well reproduced the experimental results.
Ocean feature recognition using genetic algorithms with fuzzy fitness functions (GA/F3)
NASA Technical Reports Server (NTRS)
Ankenbrandt, C. A.; Buckles, B. P.; Petry, F. E.; Lybanon, M.
1990-01-01
A model for genetic algorithms with semantic nets is derived for which the relationships between concepts is depicted as a semantic net. An organism represents the manner in which objects in a scene are attached to concepts in the net. Predicates between object pairs are continuous valued truth functions in the form of an inverse exponential function (e sub beta lxl). 1:n relationships are combined via the fuzzy OR (Max (...)). Finally, predicates between pairs of concepts are resolved by taking the average of the combined predicate values of the objects attached to the concept at the tail of the arc representing the predicate in the semantic net. The method is illustrated by applying it to the identification of oceanic features in the North Atlantic.
Test Scheduling for Core-Based SOCs Using Genetic Algorithm Based Heuristic Approach
NASA Astrophysics Data System (ADS)
Giri, Chandan; Sarkar, Soumojit; Chattopadhyay, Santanu
This paper presents a Genetic algorithm (GA) based solution to co-optimize test scheduling and wrapper design for core based SOCs. Core testing solutions are generated as a set of wrapper configurations, represented as rectangles with width equal to the number of TAM (Test Access Mechanism) channels and height equal to the corresponding testing time. A locally optimal best-fit heuristic based bin packing algorithm has been used to determine placement of rectangles minimizing the overall test times, whereas, GA has been utilized to generate the sequence of rectangles to be considered for placement. Experimental result on ITC'02 benchmark SOCs shows that the proposed method provides better solutions compared to the recent works reported in the literature.
NASA Astrophysics Data System (ADS)
Wang, Li-yong; Li, Le; Zhang, Zhi-hua
2016-07-01
Hot compression tests of Ti-6Al-4V alloy in a wide temperature range of 1023-1323 K and strain rate range of 0.01-10 s-1 were conducted by a servo-hydraulic and computer-controlled Gleeble-3500 machine. In order to accurately and effectively characterize the highly nonlinear flow behaviors, support vector regression (SVR) which is a machine learning method was combined with genetic algorithm (GA) for characterizing the flow behaviors, namely, the GA-SVR. The prominent character of GA-SVR is that it with identical training parameters will keep training accuracy and prediction accuracy at a stable level in different attempts for a certain dataset. The learning abilities, generalization abilities, and modeling efficiencies of the mathematical regression model, ANN, and GA-SVR for Ti-6Al-4V alloy were detailedly compared. Comparison results show that the learning ability of the GA-SVR is stronger than the mathematical regression model. The generalization abilities and modeling efficiencies of these models were shown as follows in ascending order: the mathematical regression model < ANN < GA-SVR. The stress-strain data outside experimental conditions were predicted by the well-trained GA-SVR, which improved simulation accuracy of the load-stroke curve and can further improve the related research fields where stress-strain data play important roles, such as speculating work hardening and dynamic recovery, characterizing dynamic recrystallization evolution, and improving processing maps.
An Effective GA-Based Scheduling Algorithm for FlexRay Systems
NASA Astrophysics Data System (ADS)
Ding, Shan; Tomiyama, Hiroyuki; Takada, Hiroaki
An advanced communication system, the FlexRay system, has been developed for future automotive applications. It consists of time-triggered clusters, such as drive-by-wire in cars, in order to meet different requirements and constraints between various sensors, processors, and actuators. In this paper, an approach to static scheduling for FlexRay systems is proposed. Our experimental results show that the proposed scheduling method significantly reduces up to 36.3% of the network traffic compared with a past approach.
A Genetic Algorithm Approach to Nonlinear Least Squares Estimation
ERIC Educational Resources Information Center
Olinsky, Alan D.; Quinn, John T.; Mangiameli, Paul M.; Chen, Shaw K.
2004-01-01
A common type of problem encountered in mathematics is optimizing nonlinear functions. Many popular algorithms that are currently available for finding nonlinear least squares estimators, a special class of nonlinear problems, are sometimes inadequate. They might not converge to an optimal value, or if they do, it could be to a local rather than…
A compensatory algorithm for the slow-down effect on constant-time-separation approaches
NASA Technical Reports Server (NTRS)
Abbott, Terence S.
1991-01-01
In seeking methods to improve airport capacity, the question arose as to whether an electronic display could provide information which would enable the pilot to be responsible for self-separation under instrument conditions to allow for the practical implementation of reduced separation, multiple glide path approaches. A time based, closed loop algorithm was developed and simulator validated for in-trail (one aircraft behind the other) approach and landing. The algorithm was designed to reduce the effects of approach speed reduction prior to landing for the trailing aircraft as well as the dispersion of the interarrival times. The operational task for the validation was an instrument approach to landing while following a single lead aircraft on the same approach path. The desired landing separation was 60 seconds for these approaches. An open loop algorithm, previously developed, was used as a basis for comparison. The results showed that relative to the open loop algorithm, the closed loop one could theoretically provide for a 6 pct. increase in runway throughput. Also, the use of the closed loop algorithm did not affect the path tracking performance and pilot comments indicated that the guidance from the closed loop algorithm would be acceptable from an operational standpoint. From these results, it is concluded that by using a time based, closed loop spacing algorithm, precise interarrival time intervals may be achievable with operationally acceptable pilot workload.
Random Matrix Approach to Quantum Adiabatic Evolution Algorithms
NASA Technical Reports Server (NTRS)
Boulatov, Alexei; Smelyanskiy, Vadier N.
2004-01-01
We analyze the power of quantum adiabatic evolution algorithms (Q-QA) for solving random NP-hard optimization problems within a theoretical framework based on the random matrix theory (RMT). We present two types of the driven RMT models. In the first model, the driving Hamiltonian is represented by Brownian motion in the matrix space. We use the Brownian motion model to obtain a description of multiple avoided crossing phenomena. We show that the failure mechanism of the QAA is due to the interaction of the ground state with the "cloud" formed by all the excited states, confirming that in the driven RMT models. the Landau-Zener mechanism of dissipation is not important. We show that the QAEA has a finite probability of success in a certain range of parameters. implying the polynomial complexity of the algorithm. The second model corresponds to the standard QAEA with the problem Hamiltonian taken from the Gaussian Unitary RMT ensemble (GUE). We show that the level dynamics in this model can be mapped onto the dynamics in the Brownian motion model. However, the driven RMT model always leads to the exponential complexity of the algorithm due to the presence of the long-range intertemporal correlations of the eigenvalues. Our results indicate that the weakness of effective transitions is the leading effect that can make the Markovian type QAEA successful.
An algorithm for fast DNS cavitating flows simulations using homogeneous mixture approach
NASA Astrophysics Data System (ADS)
Žnidarčič, A.; Coutier-Delgosha, O.; Marquillie, M.; Dular, M.
2015-12-01
A new algorithm for fast DNS cavitating flows simulations is developed. The algorithm is based on Kim and Moin projection method form. Homogeneous mixture approach with transport equation for vapour volume fraction is used to model cavitation and various cavitation models can be used. Influence matrix and matrix diagonalisation technique enable fast parallel computations.
Random matrix approach to quantum adiabatic evolution algorithms
Boulatov, A.; Smelyanskiy, V.N.
2005-05-15
We analyze the power of the quantum adiabatic evolution algorithm (QAA) for solving random computationally hard optimization problems within a theoretical framework based on random matrix theory (RMT). We present two types of driven RMT models. In the first model, the driving Hamiltonian is represented by Brownian motion in the matrix space. We use the Brownian motion model to obtain a description of multiple avoided crossing phenomena. We show that nonadiabatic corrections in the QAA are due to the interaction of the ground state with the 'cloud' formed by most of the excited states, confirming that in driven RMT models, the Landau-Zener scenario of pairwise level repulsions is not relevant for the description of nonadiabatic corrections. We show that the QAA has a finite probability of success in a certain range of parameters, implying a polynomial complexity of the algorithm. The second model corresponds to the standard QAA with the problem Hamiltonian taken from the RMT Gaussian unitary ensemble (GUE). We show that the level dynamics in this model can be mapped onto the dynamics in the Brownian motion model. For this reason, the driven GUE model can also lead to polynomial complexity of the QAA. The main contribution to the failure probability of the QAA comes from the nonadiabatic corrections to the eigenstates, which only depend on the absolute values of the transition amplitudes. Due to the mapping between the two models, these absolute values are the same in both cases. Our results indicate that this 'phase irrelevance' is the leading effect that can make both the Markovian- and GUE-type QAAs successful.
A Genetic Algorithm Approach for the TV Self-Promotion Assignment Problem
NASA Astrophysics Data System (ADS)
Pereira, Paulo A.; Fontes, Fernando A. C. C.; Fontes, Dalila B. M. M.
2009-09-01
We report on the development of a Genetic Algorithm (GA), which has been integrated into a Decision Support System to plan the best assignment of the weekly self-promotion space for a TV station. The problem addressed consists on deciding which shows to advertise and when such that the number of viewers, of an intended group or target, is maximized. The GA proposed incorporates a greedy heuristic to find good initial solutions. These solutions, as well as the solutions later obtained through the use of the GA, go then through a repair procedure. This is used with two objectives, which are addressed in turn. Firstly, it checks the solution feasibility and if unfeasible it is fixed by removing some shows. Secondly, it tries to improve the solution by adding some extra shows. Since the problem faced by the commercial TV station is too big and has too many features it cannot be solved exactly. Therefore, in order to test the quality of the solutions provided by the proposed GA we have randomly generated some smaller problem instances. For these problems we have obtained solutions on average within 1% of the optimal solution value.
NASA Astrophysics Data System (ADS)
Anderson, Richard P.
An algorithm for precision approach guidance using GPS and a MicroElectroMechanical Systems/Inertial Navigation System (MEMS/INS) has been developed to meet the Required Navigational Performance (RNP) at a cost that is suitable for General Aviation (GA) applications. This scheme allows for accurate approach guidance (Category I) using Wide Area Augmentation System (WAAS) at locations not served by ILS, MLS or other types of precision landing guidance, thereby greatly expanding the number of useable airports in poor weather. At locations served by a Local Area Augmentation System (LAAS), Category III-like navigation is possible with the novel idea of a Missed Approach Time (MAT) that is similar to a Missed Approach Point (MAP) but not fixed in space. Though certain augmented types of GPS have sufficient precision for approach navigation, its use alone is insufficient to meet RNP due to an inability to monitor loss, degradation or intentional spoofing and meaconing of the GPS signal. A redundant navigation system and a health monitoring system must be added to acquire sufficient reliability, safety and time-to-alert as stated by required navigation performance. An inertial navigation system is the best choice, as it requires no external radio signals and its errors are complementary to GPS. An aiding Kalman filter is used to derive parameters that monitor the correlation between the GPS and MEMS/INS. These approach guidance parameters determines the MAT for a given RNP and provide the pilot or autopilot with proceed/do-not-proceed decision in real time. The enabling technology used to derive the guidance program is a MEMS gyroscope and accelerometer package in conjunction with a single-antenna pseudo-attitude algorithm. To be viable for most GA applications, the hardware must be reasonably priced. The MEMS gyros allows for the first cost-effective INS package to be developed. With lower cost, however, comes higher drift rates and a more dependence on GPS aiding. In
Effective and efficient optics inspection approach using machine learning algorithms
Abdulla, G; Kegelmeyer, L; Liao, Z; Carr, W
2010-11-02
The Final Optics Damage Inspection (FODI) system automatically acquires and utilizes the Optics Inspection (OI) system to analyze images of the final optics at the National Ignition Facility (NIF). During each inspection cycle up to 1000 images acquired by FODI are examined by OI to identify and track damage sites on the optics. The process of tracking growing damage sites on the surface of an optic can be made more effective by identifying and removing signals associated with debris or reflections. The manual process to filter these false sites is daunting and time consuming. In this paper we discuss the use of machine learning tools and data mining techniques to help with this task. We describe the process to prepare a data set that can be used for training and identifying hardware reflections in the image data. In order to collect training data, the images are first automatically acquired and analyzed with existing software and then relevant features such as spatial, physical and luminosity measures are extracted for each site. A subset of these sites is 'truthed' or manually assigned a class to create training data. A supervised classification algorithm is used to test if the features can predict the class membership of new sites. A suite of self-configuring machine learning tools called 'Avatar Tools' is applied to classify all sites. To verify, we used 10-fold cross correlation and found the accuracy was above 99%. This substantially reduces the number of false alarms that would otherwise be sent for more extensive investigation.
Effective and efficient optics inspection approach using machine learning algorithms
NASA Astrophysics Data System (ADS)
Abdulla, Ghaleb M.; Kegelmeyer, Laura Mascio; Liao, Zhi M.; Carr, Wren
2010-11-01
The Final Optics Damage Inspection (FODI) system automatically acquires and utilizes the Optics Inspection (OI) system to analyze images of the final optics at the National Ignition Facility (NIF). During each inspection cycle up to 1000 images acquired by FODI are examined by OI to identify and track damage sites on the optics. The process of tracking growing damage sites on the surface of an optic can be made more effective by identifying and removing signals associated with debris or reflections. The manual process to filter these false sites is daunting and time consuming. In this paper we discuss the use of machine learning tools and data mining techniques to help with this task. We describe the process to prepare a data set that can be used for training and identifying hardware reflections in the image data. In order to collect training data, the images are first automatically acquired and analyzed with existing software and then relevant features such as spatial, physical and luminosity measures are extracted for each site. A subset of these sites is "truthed" or manually assigned a class to create training data. A supervised classification algorithm is used to test if the features can predict the class membership of new sites. A suite of self-configuring machine learning tools called "Avatar Tools" is applied to classify all sites. To verify, we used 10-fold cross correlation and found the accuracy was above 99%. This substantially reduces the number of false alarms that would otherwise be sent for more extensive investigation.
Classification Algorithms for Big Data Analysis, a Map Reduce Approach
NASA Astrophysics Data System (ADS)
Ayma, V. A.; Ferreira, R. S.; Happ, P.; Oliveira, D.; Feitosa, R.; Costa, G.; Plaza, A.; Gamba, P.
2015-03-01
Since many years ago, the scientific community is concerned about how to increase the accuracy of different classification methods, and major achievements have been made so far. Besides this issue, the increasing amount of data that is being generated every day by remote sensors raises more challenges to be overcome. In this work, a tool within the scope of InterIMAGE Cloud Platform (ICP), which is an open-source, distributed framework for automatic image interpretation, is presented. The tool, named ICP: Data Mining Package, is able to perform supervised classification procedures on huge amounts of data, usually referred as big data, on a distributed infrastructure using Hadoop MapReduce. The tool has four classification algorithms implemented, taken from WEKA's machine learning library, namely: Decision Trees, Naïve Bayes, Random Forest and Support Vector Machines (SVM). The results of an experimental analysis using a SVM classifier on data sets of different sizes for different cluster configurations demonstrates the potential of the tool, as well as aspects that affect its performance.
Scheduling language and algorithm development study. Appendix: Study approach and activity summary
NASA Technical Reports Server (NTRS)
1974-01-01
The approach and organization of the study to develop a high level computer programming language and a program library are presented. The algorithm and problem modeling analyses are summarized. The approach used to identify and specify the capabilities required in the basic language is described. Results of the analyses used to define specifications for the scheduling module library are presented.
A Genetic Algorithm Variational Approach to Data Assimilation and Application to Volcanic Emissions
NASA Astrophysics Data System (ADS)
Schmehl, Kerrie J.; Haupt, Sue Ellen; Pavolonis, Michael J.
2012-03-01
Variational data assimilation methods optimize the match between an observed and a predicted field. These methods normally require information on error variances of both the analysis and the observations, which are sometimes difficult to obtain for transport and dispersion problems. Here, the variational problem is set up as a minimization problem that directly minimizes the root mean squared error of the difference between the observations and the prediction. In the context of atmospheric transport and dispersion, the solution of this optimization problem requires a robust technique. A genetic algorithm (GA) is used here for that solution, forming the GA-Variational (GA-Var) technique. The philosophy and formulation of the technique is described here. An advantage of the technique includes that it does not require observation or analysis error covariances nor information about any variables that are not directly assimilated. It can be employed in the context of either a forward assimilation problem or used to retrieve unknown source or meteorological information by solving the inverse problem. The details of the method are reviewed. As an example application, GA-Var is demonstrated for predicting the plume from a volcanic eruption. First the technique is employed to retrieve the unknown emission rate and the steering winds of the volcanic plume. Then that information is assimilated into a forward prediction of its transport and dispersion. Concentration data are derived from satellite data to determine the observed ash concentrations. A case study is made of the March 2009 eruption of Mount Redoubt in Alaska. The GA-Var technique is able to determine a wind speed and direction that matches the observations well and a reasonable emission rate.
Obese and Overweight Children and Adolescents: An Algorithmic Clinical Approach
Khosravi, Shahrzad; Borna, Sima
2013-01-01
Obesity in children and adolescents is a hot issue throughout the world. Numerous complications are related to childhood obesity, such as cardiovascular disease, diabetes, insulin resistance and psychological problems. Therefore, identification and treatment of this problem have an important role in the health system. In this clinical approach, we have provided a general overview of the assessment and management of obesity in children and adolescents, including definitions, history-taking, physical examinations, and laboratory testing for general practitioners and pediatricians. Furthermore, conventional therapies (physical activity, eating habits and behavioral modification) and non-conventional treatments (drugs and surgery options) have been discussed. PMID:24910738
Impact ionization in GaAs: A screened exchange density-functional approach
Picozzi, S.; Asahi, R.; Geller, C. B.; Continenza, A.; Freeman, A. J.
2001-08-13
Results are presented of a fully ab initio calculation of impact ionization rates in GaAs within the density functional theory framework, using a screened-exchange formalism and the highly precise all-electron full-potential linearized augmented plane wave method. The calculated impact ionization rates show a marked orientation dependence in k space, indicating the strong restrictions imposed by the conservation of energy and momentum. This anisotropy diminishes as the impacting electron energy increases. A Keldysh type fit performed on the energy-dependent rate shows a rather soft edge and a threshold energy greater than the direct band gap. The consistency with available Monte Carlo and empirical pseudopotential calculations shows the reliability of our approach and paves the way to ab initio calculations of pair production rates in new and more complex materials.
Forest Height Retrieval Algorithm Using a Complex Visibility Function Approach
NASA Astrophysics Data System (ADS)
Chu, T.; Zebker, H. A.
2011-12-01
Vegetation structure and biomass on earth's terrestrial surface are critical parameters that influences global carbon cycle, habitat, climate, and resources of economic value. Space-borne and air-borne remote sensing instruments are the most practical means of obtaining information such as tree height and biomass on a large scale. SAR (Synthetic aperture radars) especially InSAR (Interferometric SAR) has been utilized in the recent years to quantify vegetation parameters such as height and biomass. However methods used to quantify global vegetation has yet to produce accurate results. It is the goal of this study to develop a signal-processing algorithm through simulation to determine vegetation heights that would lead to accurate height and biomass retrievals. A standard SAR image represents a projection of the 3D distributed backscatter onto a 2D plane. InSAR is capable of determining topography or the height of vegetation. Vegetation height is determined from the mean scattering phase center of all scatterers within a resolution cell. InSAR is capable of generating a 3D height surface, but the distribution of scatters in height is under-determined and cannot be resolved by a single-baseline measurement. One interferogram therefore is insufficient to uniquely determine vertical characteristics of even a simple 3D forest. An aperture synthesis technique in the height or vertical dimension would enable improved resolution capability to distinguish scatterers of different location in the vertical dimension. Repeat pass observations allow us differential interferometry to populate the frequency domain from which we can use the Fourier transform relation to get to the brightness or backscatter domain. Ryle and Hewish first introduced this technique of aperture synthesis in the 1960's for large radio telescope arrays. This technique would allow us to focus the antenna beam pattern in the vertical direction and increase vertical resolving power. It enable us to
Load identification approach based on basis pursuit denoising algorithm
NASA Astrophysics Data System (ADS)
Ginsberg, D.; Ruby, M.; Fritzen, C. P.
2015-07-01
The information of the external loads is of great interest in many fields of structural analysis, such as structural health monitoring (SHM) systems or assessment of damage after extreme events. However, in most cases it is not possible to measure the external forces directly, so they need to be reconstructed. Load reconstruction refers to the problem of estimating an input to a dynamic system when the system output and the impulse response functions are usually the knowns. Generally, this leads to a so called ill-posed inverse problem, which involves solving an underdetermined linear system of equations. For most practical applications it can be assumed that the applied loads are not arbitrarily distributed in time and space, at least some specific characteristics about the external excitation are known a priori. In this contribution this knowledge was used to develop a more suitable force reconstruction method, which allows identifying the time history and the force location simultaneously by employing significantly fewer sensors compared to other reconstruction approaches. The properties of the external force are used to transform the ill-posed problem into a sparse recovery task. The sparse solution is acquired by solving a minimization problem known as basis pursuit denoising (BPDN). The possibility of reconstructing loads based on noisy structural measurement signals will be demonstrated by considering two frequently occurring loading conditions: harmonic excitation and impact events, separately and combined. First a simulation study of a simple plate structure is carried out and thereafter an experimental investigation of a real beam is performed.
Genetic algorithm based approach to optimize phenotypical traits of virtual rice.
Ding, Weilong; Xu, Lifeng; Wei, Yang; Wu, Fuli; Zhu, Defeng; Zhang, Yuping; Max, Nelson
2016-08-21
How to select and combine good traits of rice to get high-production individuals is one of the key points in developing crop ideotype cultivation technologies. Existing cultivation methods for producing ideal plants, such as field trials and crop modeling, have some limits. In this paper, we propose a method based on a genetic algorithm (GA) and a functional-structural plant model (FSPM) to optimize plant types of virtual rice by dynamically adjusting phenotypical traits. In this algorithm, phenotypical traits such as leaf angles, plant heights, the maximum number of tiller, and the angle of tiller are considered as input parameters of our virtual rice model. We evaluate the photosynthetic output as a function of these parameters, and optimized them using a GA. This method has been implemented on GroIMP using the modeling language XL (eXtended L-System) and RGG (Relational Growth Grammar). A double haploid population of rice is adopted as test material in a case study. Our experimental results show that our method can not only optimize the parameters of rice plant type and increase the amount of light absorption, but can also significantly increase crop yield. PMID:27179460
Chen, Hong-Yan; Zhao, Geng-Xing; Li, Xi-Can; Wang, Xiang-Feng; Li, Yu-Ling
2013-11-01
Taking the Qihe County in Shandong Province of East China as the study area, soil samples were collected from the field, and based on the hyperspectral reflectance measurement of the soil samples and the transformation with the first deviation, the spectra were denoised and compressed by discrete wavelet transform (DWT), the variables for the soil alkali hydrolysable nitrogen quantitative estimation models were selected by genetic algorithms (GA), and the estimation models for the soil alkali hydrolysable nitrogen content were built by using partial least squares (PLS) regression. The discrete wavelet transform and genetic algorithm in combining with partial least squares (DWT-GA-PLS) could not only compress the spectrum variables and reduce the model variables, but also improve the quantitative estimation accuracy of soil alkali hydrolysable nitrogen content. Based on the 1-2 levels low frequency coefficients of discrete wavelet transform, and under the condition of large scale decrement of spectrum variables, the calibration models could achieve the higher or the same prediction accuracy as the soil full spectra. The model based on the second level low frequency coefficients had the highest precision, with the model predicting R2 being 0.85, the RMSE being 8.11 mg x kg(-1), and RPD being 2.53, indicating the effectiveness of DWT-GA-PLS method in estimating soil alkali hydrolysable nitrogen content. PMID:24564148
Zhuang, Weibing; Gao, Zhihong; Zhang, Zhen
2013-01-01
Hormones are closely associated with dormancy in deciduous fruit trees, and gibberellins (GAs) are known to be particularly important. In this study, we observed that GA4 treatment led to earlier bud break in Japanese apricot. To understand better the promoting effect of GA4 on the dormancy release of Japanese apricot flower buds, proteomic and transcriptomic approaches were used to analyse the mechanisms of dormancy release following GA4 treatment, based on two-dimensional gel electrophoresis (2-DE) and digital gene expression (DGE) profiling, respectively. More than 600 highly reproducible protein spots (P<0.05) were detected and, following GA4 treatment, 38 protein spots showed more than a 2-fold difference in expression, and 32 protein spots were confidently identified according to the databases. Compared with water treatment, many proteins that were associated with energy metabolism and oxidation–reduction showed significant changes after GA4 treatment, which might promote dormancy release. We observed that genes at the mRNA level associated with energy metabolism and oxidation–reduction also played an important role in this process. Analysis of the functions of the identified proteins and genes and the related metabolic pathways would provide a comprehensive proteomic and transcriptomic view of the coordination of dormancy release after GA4 treatment in Japanese apricot flower buds. PMID:24014872
Protein side-chain packing problem: a maximum edge-weight clique algorithmic approach.
Dukka Bahadur, K C; Tomita, Etsuji; Suzuki, Jun'ichi; Akutsu, Tatsuya
2005-02-01
"Protein Side-chain Packing" has an ever-increasing application in the field of bio-informatics, dating from the early methods of homology modeling to protein design and to the protein docking. However, this problem is computationally known to be NP-hard. In this regard, we have developed a novel approach to solve this problem using the notion of a maximum edge-weight clique. Our approach is based on efficient reduction of protein side-chain packing problem to a graph and then solving the reduced graph to find the maximum clique by applying an efficient clique finding algorithm developed by our co-authors. Since our approach is based on deterministic algorithms in contrast to the various existing algorithms based on heuristic approaches, our algorithm guarantees of finding an optimal solution. We have tested this approach to predict the side-chain conformations of a set of proteins and have compared the results with other existing methods. We have found that our results are favorably comparable or better than the results produced by the existing methods. As our test set contains a protein of 494 residues, we have obtained considerable improvement in terms of size of the proteins and in terms of the efficiency and the accuracy of prediction. PMID:15751115
Self-adaptive parameters in genetic algorithms
NASA Astrophysics Data System (ADS)
Pellerin, Eric; Pigeon, Luc; Delisle, Sylvain
2004-04-01
Genetic algorithms are powerful search algorithms that can be applied to a wide range of problems. Generally, parameter setting is accomplished prior to running a Genetic Algorithm (GA) and this setting remains unchanged during execution. The problem of interest to us here is the self-adaptive parameters adjustment of a GA. In this research, we propose an approach in which the control of a genetic algorithm"s parameters can be encoded within the chromosome of each individual. The parameters" values are entirely dependent on the evolution mechanism and on the problem context. Our preliminary results show that a GA is able to learn and evaluate the quality of self-set parameters according to their degree of contribution to the resolution of the problem. These results are indicative of a promising approach to the development of GAs with self-adaptive parameter settings that do not require the user to pre-adjust parameters at the outset.
An approach to select the appropriate image fusion algorithm for night vision systems
NASA Astrophysics Data System (ADS)
Schwan, Gabriele; Scherer-Negenborn, Norbert
2015-10-01
For many years image fusion has been an important subject in the image processing community. The purpose of image fusion is taking over the relevant information from two or several images to construct one result image. In the past many fusion algorithms were developed and published. Some attempts were made to assess the results from several fusion algorithms automatically with the objective of gaining the best suited output for human observers. But it was shown, that such objective machine-assessment does not always correlate with the observer's subjective perception. In this paper a novel approach is presented, which selects the appropriate fusion algorithm to receive the best image enhancement results for human observers. Assessment of the fusion algorithms' results was done based on the local contrasts. Fusion algorithms are used on a representative data set covering different use cases and image contents. These fusion results of selected data are judged subjectively by some human observers. Then the assessment algorithm with the best fit to the visual perception is used to select the best fusion algorithm for comparable scenarios.
An Approach for Assessing RNA-seq Quantification Algorithms in Replication Studies
Wu, Po-Yen; Phan, John H.; Wang, May D.
2016-01-01
One way to gain a more comprehensive picture of the complex function of a cell is to study the transcriptome. A promising technology for studying the transcriptome is RNA sequencing, an application of which is to quantify elements in the transcriptome and to link quantitative observations to biology. Although numerous quantification algorithms are publicly available, no method of systematically assessing these algorithms has been developed. To meet the need for such an assessment, we present an approach that includes (1) simulated and real datasets, (2) three alignment strategies, and (3) six quantification algorithms. Examining the normalized root-mean-square error, the percentage error of the coefficient of variation, and the distribution of the coefficient of variation, we found that quantification algorithms with the input of sequence alignment reported in the transcriptomic coordinate usually performed better in terms of the multiple metrics proposed in this study.
Pal, Partha S; Kar, R; Mandal, D; Ghoshal, S P
2015-11-01
This paper presents an efficient approach to identify different stable and practically useful Hammerstein models as well as unstable nonlinear process along with its stable closed loop counterpart with the help of an evolutionary algorithm as Colliding Bodies Optimization (CBO) optimization algorithm. The performance measures of the CBO based optimization approach such as precision, accuracy are justified with the minimum output mean square value (MSE) which signifies that the amount of bias and variance in the output domain are also the least. It is also observed that the optimization of output MSE in the presence of outliers has resulted in a very close estimation of the output parameters consistently, which also justifies the effective general applicability of the CBO algorithm towards the system identification problem and also establishes the practical usefulness of the applied approach. Optimum values of the MSEs, computational times and statistical information of the MSEs are all found to be the superior as compared with those of the other existing similar types of stochastic algorithms based approaches reported in different recent literature, which establish the robustness and efficiency of the applied CBO based identification scheme. PMID:26362314
A Fuzzy Genetic Algorithm Approach to an Adaptive Information Retrieval Agent.
ERIC Educational Resources Information Center
Martin-Bautista, Maria J.; Vila, Maria-Amparo; Larsen, Henrik Legind
1999-01-01
Presents an approach to a Genetic Information Retrieval Agent Filter (GIRAF) that filters and ranks documents retrieved from the Internet according to users' preferences by using a Genetic Algorithm and fuzzy set theory to handle the imprecision of users' preferences and users' evaluation of the retrieved documents. (Author/LRW)
ERIC Educational Resources Information Center
Moreno, Julian; Ovalle, Demetrio A.; Vicari, Rosa M.
2012-01-01
Considering that group formation is one of the key processes in collaborative learning, the aim of this paper is to propose a method based on a genetic algorithm approach for achieving inter-homogeneous and intra-heterogeneous groups. The main feature of such a method is that it allows for the consideration of as many student characteristics as…
Zaki, Mohammad Reza; Varshosaz, Jaleh; Fathi, Milad
2015-05-20
Multivariate nature of drug loaded nanospheres manufacturing in term of multiplicity of involved factors makes it a time consuming and expensive process. In this study genetic algorithm (GA) and artificial neural network (ANN), two tools inspired by natural process, were employed to optimize and simulate the manufacturing process of agar nanospheres. The efficiency of GA was evaluated against the response surface methodology (RSM). The studied responses included particle size, poly dispersity index, zeta potential, drug loading and release efficiency. GA predicted greater extremum values for response factors compared to RSM. However, real values showed some deviations from predicted data. Appropriate agreement was found between ANN model predicted and real values for all five response factors with high correlation coefficients. GA was more successful than RSM in optimization and along with ANN were efficient tools in optimizing and modeling the fabrication process of drug loaded in agar nanospheres. PMID:25817674
Semibulk InGaN: A novel approach for thick, single phase, epitaxial InGaN layers grown by MOVPE
NASA Astrophysics Data System (ADS)
Pantzas, K.; El Gmili, Y.; Dickerson, J.; Gautier, S.; Largeau, L.; Mauguin, O.; Patriarche, G.; Suresh, S.; Moudakir, T.; Bishop, C.; Ahaitouf, A.; Rivera, T.; Tanguy, C.; Voss, P. L.; Ougazzaden, A.
2013-05-01
In this paper we demonstrate a solution to systematically obtain thick, single phase InGaN epilayers by MOVPE. The solution consists in periodically inserting ultra-thin GaN interlayers during InGaN growth. Measurements by HAADF-STEM, X-ray diffraction, cathodoluminescence and photoluminescence demonstrate the effective suppression of the three-dimensional sublayer that is shown to spontaneously form in control InGaN epilayers grown without this method. Simulation predicts that tunneling through the GaN barriers is efficient and that carrier transport through this semi-bulk InGaN/GaN structure is similar to that of bulk InGaN. Such structures may be useful for improving the efficiency of InGaN solar cells by allowing thicker, higher quality InGaN absorption layers.
NASA Technical Reports Server (NTRS)
Li, C.-J.; Sun, Q.; Lagowski, J.; Gatos, H. C.
1985-01-01
The microscale characterization of electronic defects in (SI) GaAs has been a challenging issue in connection with materials problems encountered in GaAs IC technology. The main obstacle which limits the applicability of high resolution electron beam methods such as Electron Beam-Induced Current (EBIC) and cathodoluminescence (CL) is the low concentration of free carriers in semiinsulating (SI) GaAs. The present paper provides a new photo-EBIC characterization approach which combines the spectroscopic advantages of optical methods with the high spatial resolution and scanning capability of EBIC. A scanning electron microscope modified for electronic characterization studies is shown schematically. The instrument can operate in the standard SEM mode, in the EBIC modes (including photo-EBIC and thermally stimulated EBIC /TS-EBIC/), and in the cathodo-luminescence (CL) and scanning modes. Attention is given to the use of CL, Photo-EBIC, and TS-EBIC techniques.
NASA Astrophysics Data System (ADS)
Handels, Heinz; Ross, Th; Kreusch, J.; Wolff, H. H.; Poeppl, S. J.
1998-06-01
A new approach to computer supported recognition of melanoma and naevocytic naevi based on high resolution skin surface profiles is presented. Profiles are generated by sampling an area of 4 X 4 mm2 at a resolution of 125 sample points per mm with a laser profilometer at a vertical resolution of 0.1 micrometers . With image analysis algorithms Haralick's texture parameters, Fourier features and features based on fractal analysis are extracted. In order to improve classification performance, a subsequent feature selection process is applied to determine the best possible subset of features. Genetic algorithms are optimized for the feature selection process, and results of different approaches are compared. As quality measure for feature subsets, the error rate of the nearest neighbor classifier estimated with the leaving-one-out method is used. In comparison to heuristic strategies and greedy algorithms, genetic algorithms show the best results for the feature selection problem. After feature selection, several architectures of feed forward neural networks with error back-propagation are evaluated. Classification performance of the neural classifier is optimized using different topologies, learning parameters and pruning algorithms. The best neural classifier achieved an error rate of 4.5% and was found after network pruning. The best result in all with an error rate of 2.3% was obtained with the nearest neighbor classifier.
Evaluation of multi-algorithm optimization approach in multi-objective rainfall-runoff calibration
NASA Astrophysics Data System (ADS)
Shafii, M.; de Smedt, F.
2009-04-01
Calibration of rainfall-runoff models is one of the issues in which hydrologists have been interested over past decades. Because of the multi-objective nature of rainfall-runoff calibration, and due to advances in computational power, population-based optimization techniques are becoming increasingly popular to be applied for multi-objective calibration schemes. Over past recent years, such methods have shown to be powerful search methods for this purpose, especially when there are a large number of calibration parameters. However, application of these methods is always criticised based on the fact that it is not possible to develop a single algorithm which is always efficient for different problems. Therefore, more recent efforts have been focused towards development of simultaneous multiple optimization algorithms to overcome this drawback. This paper involves one of the most recent population-based multi-algorithm approaches, named AMALGAM, for application to multi-objective rainfall-runoff calibration in a distributed hydrological model, WetSpa. This algorithm merges the strengths of different optimization algorithms and it, thus, has proven to be more efficient than other methods. In order to evaluate this issue, comparison between results of this paper and those previously reported using a normal multi-objective evolutionary algorithm would be the next step of this study.
Yanmaz, Ersin; Sarıpınar, Emin; Şahin, Kader; Geçen, Nazmiye; Çopur, Fatih
2011-04-01
4D-QSAR studies were performed on a series of 87 penicillin analogues using the electron conformational-genetic algorithm (EC-GA) method. In this EC-based method, each conformation of the molecular system is described by a matrix (ECMC) with both electron structural parameters and interatomic distances as matrix elements. Multiple comparisons of these matrices within given tolerances for high active and low active penicillin compounds allow one to separate a smaller number of matrix elements (ECSA) which represent the pharmacophore groups. The effect of conformations was investigated building model 1 and 2 based on ensemble of conformers and single conformer, respectively. GA was used to select the most important descriptors and to predict the theoretical activity of the training (74 compounds) and test (13 compounds, commercial penicillins) sets. The model 1 for training and test sets obtained by optimum 12 parameters gave more satisfactory results (R(training)(2)=0.861, SE(training)=0.044, R(test)(2)=0.892, SE(test)=0.099, q(2)=0.702, q(ext1)(2)=0.777 and q(ext2)(2)=0.733) than model 2 (R(training)(2)=0.774, SE(training)=0.056, R(test)(2)=0.840, SE(test)=0.121, q(2)=0.514, q(ext1)(2)=0.641 and q(ext2)(2)=0.570). To estimate the individual influence of each of the molecular descriptors on biological activity, the E statistics technique was applied to the derived EC-GA model. PMID:21419636
One-qubit quantum gates in a circular graphene quantum dot: genetic algorithm approach
NASA Astrophysics Data System (ADS)
Amparán, Gibrán; Rojas, Fernando; Pérez-Garrido, Antonio
2013-05-01
The aim of this work was to design and control, using genetic algorithm (GA) for parameter optimization, one-charge-qubit quantum logic gates σ x, σ y, and σ z, using two bound states as a qubit space, of circular graphene quantum dots in a homogeneous magnetic field. The method employed for the proposed gate implementation is through the quantum dynamic control of the qubit subspace with an oscillating electric field and an onsite (inside the quantum dot) gate voltage pulse with amplitude and time width modulation which introduce relative phases and transitions between states. Our results show that we can obtain values of fitness or gate fidelity close to 1, avoiding the leakage probability to higher states. The system evolution, for the gate operation, is presented with the dynamics of the probability density, as well as a visualization of the current of the pseudospin, characteristic of a graphene structure. Therefore, we conclude that is possible to use the states of the graphene quantum dot (selecting the dot size and magnetic field) to design and control the qubit subspace, with these two time-dependent interactions, to obtain the optimal parameters for a good gate fidelity using GA.
A DNA-based algorithm for minimizing decision rules: a rough sets approach.
Kim, Ikno; Chu, Yu-Yi; Watada, Junzo; Wu, Jui-Yu; Pedrycz, Witold
2011-09-01
Rough sets are often exploited for data reduction and classification. While they are conceptually appealing, the techniques used with rough sets can be computationally demanding. To address this obstacle, the objective of this study is to investigate the use of DNA molecules and associated techniques as an optimization vehicle to support algorithms of rough sets. In particular, we develop a DNA-based algorithm to derive decision rules of minimal length. This new approach can be of value when dealing with a large number of objects and their attributes, in which case the complexity of rough-sets-based methods is NP-hard. The proposed algorithm shows how the essential components involved in the minimization of decision rules in data processing can be realized. PMID:22020105
NASA Astrophysics Data System (ADS)
Reyes-Gómez, E.; Perdomo-Leiva, C. A.; Oliveira, L. E.; de Dios-Leyva, M.
1998-04-01
A theoretical resonant-tunnelling approach is used in a detailed study of the electronic and transmission properties of quasiperiodic Fibonacci GaAs-(Ga,Al)As semiconductor superlattices, under applied electric fields. The theoretical scheme is based upon an exact solution of the corresponding Schroedinger equations in different wells and barriers, through the use of Airy functions, and a transfer-matrix technique. The calculated quasibound resonant energies agree quite well with previous theoretical parameter-based results within a tight-binding scheme, in the particular case of isolated Fibonacci building blocks. Theoretical resonant-tunnelling results for 0953-8984/10/16/009/img6 and 0953-8984/10/16/009/img7 generations of the quasiperiodic Fibonacci superlattice reveal the occurrence of anticrossings of the resonant levels with applied electric fields, together with the conduction- and valence-level wave function localization properties and electric-field-induced migration to specific regions of the semiconductor quasiperiodic heterostructure. Finally, theoretical resonant-tunnelling calculations for the interband transition energies are shown to be in quite good quantitative agreement with previously reported experimental photocurrent measurements.
2011-01-01
Background Position-specific priors (PSP) have been used with success to boost EM and Gibbs sampler-based motif discovery algorithms. PSP information has been computed from different sources, including orthologous conservation, DNA duplex stability, and nucleosome positioning. The use of prior information has not yet been used in the context of combinatorial algorithms. Moreover, priors have been used only independently, and the gain of combining priors from different sources has not yet been studied. Results We extend RISOTTO, a combinatorial algorithm for motif discovery, by post-processing its output with a greedy procedure that uses prior information. PSP's from different sources are combined into a scoring criterion that guides the greedy search procedure. The resulting method, called GRISOTTO, was evaluated over 156 yeast TF ChIP-chip sequence-sets commonly used to benchmark prior-based motif discovery algorithms. Results show that GRISOTTO is at least as accurate as other twelve state-of-the-art approaches for the same task, even without combining priors. Furthermore, by considering combined priors, GRISOTTO is considerably more accurate than the state-of-the-art approaches for the same task. We also show that PSP's improve GRISOTTO ability to retrieve motifs from mouse ChiP-seq data, indicating that the proposed algorithm can be applied to data from a different technology and for a higher eukaryote. Conclusions The conclusions of this work are twofold. First, post-processing the output of combinatorial algorithms by incorporating prior information leads to a very efficient and effective motif discovery method. Second, combining priors from different sources is even more beneficial than considering them separately. PMID:21513505
NASA Astrophysics Data System (ADS)
Zheng, Feifei; Simpson, Angus R.; Zecchin, Aaron C.
2011-08-01
This paper proposes a novel optimization approach for the least cost design of looped water distribution systems (WDSs). Three distinct steps are involved in the proposed optimization approach. In the first step, the shortest-distance tree within the looped network is identified using the Dijkstra graph theory algorithm, for which an extension is proposed to find the shortest-distance tree for multisource WDSs. In the second step, a nonlinear programming (NLP) solver is employed to optimize the pipe diameters for the shortest-distance tree (chords of the shortest-distance tree are allocated the minimum allowable pipe sizes). Finally, in the third step, the original looped water network is optimized using a differential evolution (DE) algorithm seeded with diameters in the proximity of the continuous pipe sizes obtained in step two. As such, the proposed optimization approach combines the traditional deterministic optimization technique of NLP with the emerging evolutionary algorithm DE via the proposed network decomposition. The proposed methodology has been tested on four looped WDSs with the number of decision variables ranging from 21 to 454. Results obtained show the proposed approach is able to find optimal solutions with significantly less computational effort than other optimization techniques.
The Simplified Aircraft-Based Paired Approach With the ALAS Alerting Algorithm
NASA Technical Reports Server (NTRS)
Perry, Raleigh B.; Madden, Michael M.; Torres-Pomales, Wilfredo; Butler, Ricky W.
2013-01-01
This paper presents the results of an investigation of a proposed concept for closely spaced parallel runways called the Simplified Aircraft-based Paired Approach (SAPA). This procedure depends upon a new alerting algorithm called the Adjacent Landing Alerting System (ALAS). This study used both low fidelity and high fidelity simulations to validate the SAPA procedure and test the performance of the new alerting algorithm. The low fidelity simulation enabled a determination of minimum approach distance for the worst case over millions of scenarios. The high fidelity simulation enabled an accurate determination of timings and minimum approach distance in the presence of realistic trajectories, communication latencies, and total system error for 108 test cases. The SAPA procedure and the ALAS alerting algorithm were applied to the 750-ft parallel spacing (e.g., SFO 28L/28R) approach problem. With the SAPA procedure as defined in this paper, this study concludes that a 750-ft application does not appear to be feasible, but preliminary results for 1000-ft parallel runways look promising.
A Graph Algorithmic Approach to Separate Direct from Indirect Neural Interactions.
Wollstadt, Patricia; Meyer, Ulrich; Wibral, Michael
2015-01-01
Network graphs have become a popular tool to represent complex systems composed of many interacting subunits; especially in neuroscience, network graphs are increasingly used to represent and analyze functional interactions between multiple neural sources. Interactions are often reconstructed using pairwise bivariate analyses, overlooking the multivariate nature of interactions: it is neglected that investigating the effect of one source on a target necessitates to take all other sources as potential nuisance variables into account; also combinations of sources may act jointly on a given target. Bivariate analyses produce networks that may contain spurious interactions, which reduce the interpretability of the network and its graph metrics. A truly multivariate reconstruction, however, is computationally intractable because of the combinatorial explosion in the number of potential interactions. Thus, we have to resort to approximative methods to handle the intractability of multivariate interaction reconstruction, and thereby enable the use of networks in neuroscience. Here, we suggest such an approximative approach in the form of an algorithm that extends fast bivariate interaction reconstruction by identifying potentially spurious interactions post-hoc: the algorithm uses interaction delays reconstructed for directed bivariate interactions to tag potentially spurious edges on the basis of their timing signatures in the context of the surrounding network. Such tagged interactions may then be pruned, which produces a statistically conservative network approximation that is guaranteed to contain non-spurious interactions only. We describe the algorithm and present a reference implementation in MATLAB to test the algorithm's performance on simulated networks as well as networks derived from magnetoencephalographic data. We discuss the algorithm in relation to other approximative multivariate methods and highlight suitable application scenarios. Our approach is a
Earthquake—explosion discrimination using genetic algorithm-based boosting approach
NASA Astrophysics Data System (ADS)
Orlic, Niksa; Loncaric, Sven
2010-02-01
An important and challenging problem in seismic data processing is to discriminate between natural seismic events such as earthquakes and artificial seismic events such as explosions. Many automatic techniques for seismogram classification have been proposed in the literature. Most of these methods have a similar approach to seismogram classification: a predefined set of features based on ad-hoc feature selection criteria is extracted from the seismogram waveform or spectral data and these features are used for signal classification. In this paper we propose a novel approach for seismogram classification. A specially formulated genetic algorithm has been employed to automatically search for a near-optimal seismogram feature set, instead of using ad-hoc feature selection criteria. A boosting method is added to the genetic algorithm when searching for multiple features in order to improve classification performance. A learning set of seismogram data is used by the genetic algorithm to discover a near-optimal feature set. The feature set identified by the genetic algorithm is then used for seismogram classification. The described method is developed to classify seismograms in two groups, whereas a brief overview of method extension for multiple group classification is given. For method verification, a learning set consisting of 40 local earthquake seismograms and 40 explosion seismograms was used. The method was validated on seismogram set consisting of 60 local earthquake seismograms and 60 explosion seismograms, with correct classification of 85%.
Wang, Wenliang; Wang, Haiyan; Yang, Weijia; Zhu, Yunnong; Li, Guoqiang
2016-01-01
High-quality GaN epitaxial films have been grown on Si substrates with Al buffer layer by the combination of molecular beam epitaxy (MBE) and pulsed laser deposition (PLD) technologies. MBE is used to grow Al buffer layer at first, and then PLD is deployed to grow GaN epitaxial films on the Al buffer layer. The surface morphology, crystalline quality, and interfacial property of as-grown GaN epitaxial films on Si substrates are studied systematically. The as-grown ~300 nm-thick GaN epitaxial films grown at 850 °C with ~30 nm-thick Al buffer layer on Si substrates show high crystalline quality with the full-width at half-maximum (FWHM) for GaN(0002) and GaN(102) X-ray rocking curves of 0.45° and 0.61°, respectively; very flat GaN surface with the root-mean-square surface roughness of 2.5 nm; as well as the sharp and abrupt GaN/AlGaN/Al/Si hetero-interfaces. Furthermore, the corresponding growth mechanism of GaN epitaxial films grown on Si substrates with Al buffer layer by the combination of MBE and PLD is hence studied in depth. This work provides a novel and simple approach for the epitaxial growth of high-quality GaN epitaxial films on Si substrates. PMID:27101930
NASA Astrophysics Data System (ADS)
Wang, Wenliang; Wang, Haiyan; Yang, Weijia; Zhu, Yunnong; Li, Guoqiang
2016-04-01
High-quality GaN epitaxial films have been grown on Si substrates with Al buffer layer by the combination of molecular beam epitaxy (MBE) and pulsed laser deposition (PLD) technologies. MBE is used to grow Al buffer layer at first, and then PLD is deployed to grow GaN epitaxial films on the Al buffer layer. The surface morphology, crystalline quality, and interfacial property of as-grown GaN epitaxial films on Si substrates are studied systematically. The as-grown ~300 nm-thick GaN epitaxial films grown at 850 °C with ~30 nm-thick Al buffer layer on Si substrates show high crystalline quality with the full-width at half-maximum (FWHM) for GaN(0002) and GaN(102) X-ray rocking curves of 0.45° and 0.61°, respectively; very flat GaN surface with the root-mean-square surface roughness of 2.5 nm as well as the sharp and abrupt GaN/AlGaN/Al/Si hetero-interfaces. Furthermore, the corresponding growth mechanism of GaN epitaxial films grown on Si substrates with Al buffer layer by the combination of MBE and PLD is hence studied in depth. This work provides a novel and simple approach for the epitaxial growth of high-quality GaN epitaxial films on Si substrates.
Wang, Wenliang; Wang, Haiyan; Yang, Weijia; Zhu, Yunnong; Li, Guoqiang
2016-01-01
High-quality GaN epitaxial films have been grown on Si substrates with Al buffer layer by the combination of molecular beam epitaxy (MBE) and pulsed laser deposition (PLD) technologies. MBE is used to grow Al buffer layer at first, and then PLD is deployed to grow GaN epitaxial films on the Al buffer layer. The surface morphology, crystalline quality, and interfacial property of as-grown GaN epitaxial films on Si substrates are studied systematically. The as-grown ~300 nm-thick GaN epitaxial films grown at 850 °C with ~30 nm-thick Al buffer layer on Si substrates show high crystalline quality with the full-width at half-maximum (FWHM) for GaN(0002) and GaN(102) X-ray rocking curves of 0.45° and 0.61°, respectively; very flat GaN surface with the root-mean-square surface roughness of 2.5 nm; as well as the sharp and abrupt GaN/AlGaN/Al/Si hetero-interfaces. Furthermore, the corresponding growth mechanism of GaN epitaxial films grown on Si substrates with Al buffer layer by the combination of MBE and PLD is hence studied in depth. This work provides a novel and simple approach for the epitaxial growth of high-quality GaN epitaxial films on Si substrates. PMID:27101930
The infection algorithm: an artificial epidemic approach for dense stereo correspondence.
Olague, Gustavo; Fernández, Francisco; Pérez, Cynthia B; Lutton, Evelyne
2006-01-01
We present a new bio-inspired approach applied to a problem of stereo image matching. This approach is based on an artificial epidemic process, which we call the infection algorithm. The problem at hand is a basic one in computer vision for 3D scene reconstruction. It has many complex aspects and is known as an extremely difficult one. The aim is to match the contents of two images in order to obtain 3D information that allows the generation of simulated projections from a viewpoint that is different from the ones of the initial photographs. This process is known as view synthesis. The algorithm we propose exploits the image contents in order to produce only the necessary 3D depth information, while saving computational time. It is based on a set of distributed rules, which propagate like an artificial epidemic over the images. Experiments on a pair of real images are presented, and realistic reprojected images have been generated. PMID:16953787
Dlouhy, Brian J; Dahdaleh, Nader S; Menezes, Arnold H
2015-04-01
The craniovertebral junction (CVJ), or the craniocervical junction (CCJ) as it is otherwise known, houses the crossroads of the CNS and is composed of the occipital bone that surrounds the foramen magnum, the atlas vertebrae, the axis vertebrae, and their associated ligaments and musculature. The musculoskeletal organization of the CVJ is unique and complex, resulting in a wide range of congenital, developmental, and acquired pathology. The refinements of the transoral approach to the CVJ by the senior author (A.H.M.) in the late 1970s revolutionized the treatment of CVJ pathology. At the same time, a physiological approach to CVJ management was adopted at the University of Iowa Hospitals and Clinics in 1977 based on the stability and motion dynamics of the CVJ and the site of encroachment, incorporating the transoral approach for irreducible ventral CVJ pathology. Since then, approaches and techniques to treat ventral CVJ lesions have evolved. In the last 40 years at University of Iowa Hospitals and Clinics, multiple approaches to the CVJ have evolved and a better understanding of CVJ pathology has been established. In addition, new reduction strategies that have diminished the need to perform ventral decompressive approaches have been developed and implemented. In this era of surgical subspecialization, to properly treat complex CVJ pathology, the CVJ specialist must be trained in skull base transoral and endoscopic endonasal approaches, pediatric and adult CVJ spine surgery, and must understand and be able to treat the complex CSF dynamics present in CVJ pathology to provide the appropriate, optimal, and tailored treatment strategy for each individual patient, both child and adult. This is a comprehensive review of the history and evolution of the transoral approaches, extended transoral approaches, endoscopie assisted transoral approaches, endoscopie endonasal approaches, and CVJ reduction strategies. Incorporating these advancements, the authors update the
NASA Astrophysics Data System (ADS)
Abdulsattar, Mudar Ahmed
2016-05-01
Wurtzite nanocrystals of gallium nitride are approached using wurtzoid molecular building blocks. Structural and vibrational properties are investigated for both bare and hydrogen passivated GaN molecules and small nanocrystals. Wurtzoids are bundles of capped (3, 0) nanotubes that form the wurtzite phase when they reach nanocrystal or bulk sizes. Results show that experimental bulk gap is generally confined between bare and H passivated wurtzoids. Structural parameters such as bond lengths and bond angles are in good agreement with experimental bulk values. Results of longitudinal optical (LO) vibrational frequencies of present molecules are red shifted with respect to experimental bulk in agreement with previous studies for other materials. Presently modeled GaN wurtzite nanocrystals and molecules are found suitable for the description of hydrogen sensing in ambient conditions in agreement with experimental findings. N sites in GaN wurtzoid are found responsible for the detection of hydrogen molecules. The Ga sites are found to be either oxidized or permanently connected via van der Waals' forces to nitrogen or hydrogen molecules.
NASA Technical Reports Server (NTRS)
Sanyal, Soumya; Jain, Amit; Das, Sajal K.; Biswas, Rupak
2003-01-01
In this paper, we propose a distributed approach for mapping a single large application to a heterogeneous grid environment. To minimize the execution time of the parallel application, we distribute the mapping overhead to the available nodes of the grid. This approach not only provides a fast mapping of tasks to resources but is also scalable. We adopt a hierarchical grid model and accomplish the job of mapping tasks to this topology using a scheduler tree. Results show that our three-phase algorithm provides high quality mappings, and is fast and scalable.
NASA Astrophysics Data System (ADS)
Hu, Chuanmin; Lee, Zhongping; Franz, Bryan
2012-01-01
A new empirical algorithm is proposed to estimate surface chlorophyll a (Chl) concentrations in the global ocean for Chl ≤ 0.25 mg m-3(˜78% of the global ocean area). The algorithm is based on a color index (CI), defined as the difference between remote-sensing reflectance (Rrs, sr-1) in the green and a reference formed linearly between Rrsin the blue and red. For low-Chl waters, in situ data showed a tighter (and therefore better) relationship between CI and Chl than between traditional band ratios and Chl, which was further validated using global data collected concurrently by ship-borne and Sea-viewing Wide Field-of-view Sensor (SeaWiFS) and Moderate Resolution Imaging Spectroradiometer (MODIS)/Aqua instruments. Model simulations showed that for low-Chl waters, compared with the band-ratio algorithm, the CI-based algorithm (CIA) was more tolerant to changes in chlorophyll-specific backscattering coefficient and performed similarly for different relative contributions of nonphytoplankton absorption. Simulations using existing atmospheric correction approaches further demonstrated that the CIA was much less sensitive than band-ratio algorithms to various errors induced by instrument noise and imperfect atmospheric correction (including sun glint and whitecap corrections). Image and time series analyses of SeaWiFS and MODIS/Aqua data also showed improved performance in terms of reduced image noise, more coherent spatial and temporal patterns, and better consistency between the two sensors. The reduction in noise and other errors is particularly useful to improve the detection of various ocean features such as eddies. Preliminary tests over Medium-Resolution Imaging Spectrometer and Coastal Zone Color Scanner data indicate that the new approach should be generally applicable to all past, current, and future ocean color instruments.
Jin, Cong; Jin, Shu-Wei
2016-06-01
A number of different gene selection approaches based on gene expression profiles (GEP) have been developed for tumour classification. A gene selection approach selects the most informative genes from the whole gene space, which is an important process for tumour classification using GEP. This study presents an improved swarm intelligent optimisation algorithm to select genes for maintaining the diversity of the population. The most essential characteristic of the proposed approach is that it can automatically determine the number of the selected genes. On the basis of the gene selection, the authors construct a variety of the tumour classifiers, including the ensemble classifiers. Four gene datasets are used to evaluate the performance of the proposed approach. The experimental results confirm that the proposed classifiers for tumour classification are indeed effective. PMID:27187989
NASA Astrophysics Data System (ADS)
Rocha, M. C.; Saraiva, J. T.
2012-10-01
The basic objective of Transmission Expansion Planning (TEP) is to schedule a number of transmission projects along an extended planning horizon minimizing the network construction and operational costs while satisfying the requirement of delivering power safely and reliably to load centres along the horizon. This principle is quite simple, but the complexity of the problem and the impact on society transforms TEP on a challenging issue. This paper describes a new approach to solve the dynamic TEP problem, based on an improved discrete integer version of the Evolutionary Particle Swarm Optimization (EPSO) meta-heuristic algorithm. The paper includes sections describing in detail the EPSO enhanced approach, the mathematical formulation of the TEP problem, including the objective function and the constraints, and a section devoted to the application of the developed approach to this problem. Finally, the use of the developed approach is illustrated using a case study based on the IEEE 24 bus 38 branch test system.
A novel ROC approach for performance evaluation of target detection algorithms
NASA Astrophysics Data System (ADS)
Ganapathy, Priya; Skipper, Julie A.
2007-04-01
Receiver operator characteristic (ROC) analysis is an emerging automated target recognition system performance assessment tool. The ROC metric, area under the curve (AUC), is a universally accepted measure of classifying accuracy. In the presented approach, the detection algorithm output, i.e., a response plane (RP), must consist of grayscale values wherein a maximum value (e.g. 255) corresponds to highest probability of target locations. AUC computation involves the comparison of the RP and the ground truth to classify RP pixels as true positives (TP), true negatives (TN), false positives (FP), or false negatives (FN). Ideally, the background and all objects other than targets are TN. Historically, evaluation methods have excluded the background, and only a few spoof objects likely to be considered as a hit by detection algorithms were a priori demarcated as TN. This can potentially exaggerate the algorithm's performance. Here, a new ROC approach has been developed that divides the entire image into mutually exclusive target (TP) and background (TN) grid squares with adjustable size. Based on the overlap of the thresholded RP with the TP and TN grids, the FN and FP fractions are computed. Variation of the grid square size can bias the ROC results by artificially altering specificity, so an assessment of relative performance under a constant grid square size is adopted in our approach. A pilot study was performed to assess the method's ability to capture RP changes under three different detection algorithm parameter settings on ten images with different backgrounds and target orientations. An ANOVA-based comparison of the AUCs for the three settings showed a significant difference (p<0.001) at 95% confidence interval.
Review and Analysis of Algorithmic Approaches Developed for Prognostics on CMAPSS Dataset
NASA Technical Reports Server (NTRS)
Ramasso, Emannuel; Saxena, Abhinav
2014-01-01
Benchmarking of prognostic algorithms has been challenging due to limited availability of common datasets suitable for prognostics. In an attempt to alleviate this problem several benchmarking datasets have been collected by NASA's prognostic center of excellence and made available to the Prognostics and Health Management (PHM) community to allow evaluation and comparison of prognostics algorithms. Among those datasets are five C-MAPSS datasets that have been extremely popular due to their unique characteristics making them suitable for prognostics. The C-MAPSS datasets pose several challenges that have been tackled by different methods in the PHM literature. In particular, management of high variability due to sensor noise, effects of operating conditions, and presence of multiple simultaneous fault modes are some factors that have great impact on the generalization capabilities of prognostics algorithms. More than 70 publications have used the C-MAPSS datasets for developing data-driven prognostic algorithms. The C-MAPSS datasets are also shown to be well-suited for development of new machine learning and pattern recognition tools for several key preprocessing steps such as feature extraction and selection, failure mode assessment, operating conditions assessment, health status estimation, uncertainty management, and prognostics performance evaluation. This paper summarizes a comprehensive literature review of publications using C-MAPSS datasets and provides guidelines and references to further usage of these datasets in a manner that allows clear and consistent comparison between different approaches.
Modeling and Algorithmic Approaches to Constitutively-Complex, Micro-structured Fluids
Forest, Mark Gregory
2014-05-06
The team for this Project made significant progress on modeling and algorithmic approaches to hydrodynamics of fluids with complex microstructure. Our advances are broken down into modeling and algorithmic approaches. In experiments a driven magnetic bead in a complex fluid accelerates out of the Stokes regime and settles into another apparent linear response regime. The modeling explains the take-off as a deformation of entanglements, and the longtime behavior is a nonlinear, far-from-equilibrium property. Furthermore, the model has predictive value, as we can tune microstructural properties relative to the magnetic force applied to the bead to exhibit all possible behaviors. Wave-theoretic probes of complex fluids have been extended in two significant directions, to small volumes and the nonlinear regime. Heterogeneous stress and strain features that lie beyond experimental capability were studied. It was shown that nonlinear penetration of boundary stress in confined viscoelastic fluids is not monotone, indicating the possibility of interlacing layers of linear and nonlinear behavior, and thus layers of variable viscosity. Models, algorithms, and codes were developed and simulations performed leading to phase diagrams of nanorod dispersion hydrodynamics in parallel shear cells and confined cavities representative of film and membrane processing conditions. Hydrodynamic codes for polymeric fluids are extended to include coupling between microscopic and macroscopic models, and to the strongly nonlinear regime.
MRPack: Multi-Algorithm Execution Using Compute-Intensive Approach in MapReduce.
Idris, Muhammad; Hussain, Shujaat; Siddiqi, Muhammad Hameed; Hassan, Waseem; Syed Muhammad Bilal, Hafiz; Lee, Sungyoung
2015-01-01
Large quantities of data have been generated from multiple sources at exponential rates in the last few years. These data are generated at high velocity as real time and streaming data in variety of formats. These characteristics give rise to challenges in its modeling, computation, and processing. Hadoop MapReduce (MR) is a well known data-intensive distributed processing framework using the distributed file system (DFS) for Big Data. Current implementations of MR only support execution of a single algorithm in the entire Hadoop cluster. In this paper, we propose MapReducePack (MRPack), a variation of MR that supports execution of a set of related algorithms in a single MR job. We exploit the computational capability of a cluster by increasing the compute-intensiveness of MapReduce while maintaining its data-intensive approach. It uses the available computing resources by dynamically managing the task assignment and intermediate data. Intermediate data from multiple algorithms are managed using multi-key and skew mitigation strategies. The performance study of the proposed system shows that it is time, I/O, and memory efficient compared to the default MapReduce. The proposed approach reduces the execution time by 200% with an approximate 50% decrease in I/O cost. Complexity and qualitative results analysis shows significant performance improvement. PMID:26305223
MRPack: Multi-Algorithm Execution Using Compute-Intensive Approach in MapReduce
2015-01-01
Large quantities of data have been generated from multiple sources at exponential rates in the last few years. These data are generated at high velocity as real time and streaming data in variety of formats. These characteristics give rise to challenges in its modeling, computation, and processing. Hadoop MapReduce (MR) is a well known data-intensive distributed processing framework using the distributed file system (DFS) for Big Data. Current implementations of MR only support execution of a single algorithm in the entire Hadoop cluster. In this paper, we propose MapReducePack (MRPack), a variation of MR that supports execution of a set of related algorithms in a single MR job. We exploit the computational capability of a cluster by increasing the compute-intensiveness of MapReduce while maintaining its data-intensive approach. It uses the available computing resources by dynamically managing the task assignment and intermediate data. Intermediate data from multiple algorithms are managed using multi-key and skew mitigation strategies. The performance study of the proposed system shows that it is time, I/O, and memory efficient compared to the default MapReduce. The proposed approach reduces the execution time by 200% with an approximate 50% decrease in I/O cost. Complexity and qualitative results analysis shows significant performance improvement. PMID:26305223
Metamorphic approach to single quantum dot emission at 1.55 {mu}m on GaAs substrate
Semenova, E. S.; Hostein, R.; Patriarche, G.; Mauguin, O.; Largeau, L.; Robert-Philip, I.; Beveratos, A.; Lemaitre, A.
2008-05-15
We report on the fabrication and the characterization of InAs quantum dots (QDs) embedded in an indium rich In{sub 0.42}Ga{sub 0.58}As metamorphic matrix grown on a GaAs substrate. Growth conditions were chosen so as to minimize the number of threading dislocations and other defects produced during the plastic relaxation. Sharp and bright lines, originating from the emission of a few isolated single quantum dots, were observed in microphotoluminescence around 1.55 {mu}m at 5 K. They exhibit, in particular, a characteristic exciton/biexciton behavior. These QDs could offer an interesting alternative to other approaches as InAs/InP QDs for the realization of single photon emitters at telecom wavelengths.
Tumuluru, J.S.; Sokhansanj, Shahabaddine
2008-12-01
Abstract In the present study, response surface method (RSM) and genetic algorithm (GA) were used to study the effects of process variables like screw speed, rpm (x1), L/D ratio (x2), barrel temperature ( C; x3), and feed mix moisture content (%; x4), on flow rate of biomass during single-screw extrusion cooking. A second-order regression equation was developed for flow rate in terms of the process variables. The significance of the process variables based on Pareto chart indicated that screw speed and feed mix moisture content had the most influence followed by L/D ratio and barrel temperature on the flow rate. RSM analysis indicated that a screw speed>80 rpm, L/D ratio> 12, barrel temperature>80 C, and feed mix moisture content>20% resulted in maximum flow rate. Increase in screw speed and L/D ratio increased the drag flow and also the path of traverse of the feed mix inside the extruder resulting in more shear. The presence of lipids of about 35% in the biomass feed mix might have induced a lubrication effect and has significantly influenced the flow rate. The second-order regression equations were further used as the objective function for optimization using genetic algorithm. A population of 100 and iterations of 100 have successfully led to convergence the optimum. The maximum and minimum flow rates obtained using GA were 13.19 10 7 m3/s (x1=139.08 rpm, x2=15.90, x3=99.56 C, and x4=59.72%) and 0.53 10 7 m3/s (x1=59.65 rpm, x2= 11.93, x3=68.98 C, and x4=20.04%).
Development of an Aircraft Approach and Departure Atmospheric Profile Generation Algorithm
NASA Technical Reports Server (NTRS)
Buck, Bill K.; Velotas, Steven G.; Rutishauser, David K. (Technical Monitor)
2004-01-01
In support of NASA Virtual Airspace Modeling and Simulation (VAMS) project, an effort was initiated to develop and test techniques for extracting meteorological data from landing and departing aircraft, and for building altitude based profiles for key meteorological parameters from these data. The generated atmospheric profiles will be used as inputs to NASA s Aircraft Vortex Spacing System (AVOLSS) Prediction Algorithm (APA) for benefits and trade analysis. A Wake Vortex Advisory System (WakeVAS) is being developed to apply weather and wake prediction and sensing technologies with procedures to reduce current wake separation criteria when safe and appropriate to increase airport operational efficiency. The purpose of this report is to document the initial theory and design of the Aircraft Approach Departure Atmospheric Profile Generation Algorithm.
Branch-pipe-routing approach for ships using improved genetic algorithm
NASA Astrophysics Data System (ADS)
Sui, Haiteng; Niu, Wentie
2016-05-01
Branch-pipe routing plays fundamental and critical roles in ship-pipe design. The branch-pipe-routing problem is a complex combinatorial optimization problem and is thus difficult to solve when depending only on human experts. A modified genetic-algorithm-based approach is proposed in this paper to solve this problem. The simplified layout space is first divided into threedimensional (3D) grids to build its mathematical model. Branch pipes in layout space are regarded as a combination of several two-point pipes, and the pipe route between two connection points is generated using an improved maze algorithm. The coding of branch pipes is then defined, and the genetic operators are devised, especially the complete crossover strategy that greatly accelerates the convergence speed. Finally, simulation tests demonstrate the performance of proposed method.
The new approach for infrared target tracking based on the particle filter algorithm
NASA Astrophysics Data System (ADS)
Sun, Hang; Han, Hong-xia
2011-08-01
to further improve tracking performance. Experimental results show that this algorithm can compensate shortcoming of the particle filter has too much computation, and can effectively overcome the fault that mean shift is easy to fall into local extreme value instead of global maximum value .Last because of the gray and fusion target motion information, this approach also inhibit interference from the background, ultimately improve the stability and the real-time of the target track.
Single-shot x-ray phase contrast imaging with an algorithmic approach using spectral detection
NASA Astrophysics Data System (ADS)
Das, Mini; Park, Chan-Soo; Fredette, Nathaniel R.
2016-04-01
X-ray phase contrast imaging has been investigated during the last two decades for potential benefits in soft tissue imaging. Long imaging time, high radiation dose and general measurement complexity involving motion of x-ray optical components have prevented the clinical translation of these methods. In all existing popular phase contrast imaging methods, multiple measurements per projection angle involving motion of optical components are required to achieve quantitatively accurate estimation of absorption, phase and differential phase. Recently we proposed an algorithmic approach to use spectral detection data in a phase contrast imaging setup to obtain absorption, phase and differential phase in a single-step. Our generic approach has been shown via simulations in all three types of phase contrast imaging: propagation, coded aperture and grating interferometry. While other groups have used spectral detector in phase contrast imaging setups, our proposed method is unique in outlining an approach to use this spectral data to simplify phase contrast imaging. In this abstract we show the first experimental proof of our single-shot phase retrieval using a Medipix3 photon counting detector in an edge illumination aperture (also referred to as coded aperture) phase contrast set up as well as for a free space propagation setup. Our preliminary results validate our new transport equation for edge illumination PCI and our spectral phase retrieval algorithm for both PCI methods being investigated. Comparison with simulations also point to excellent performance of Medipix3 built-in charge sharing correction mechanism.
A genetic algorithms approach for altering the membership functions in fuzzy logic controllers
NASA Technical Reports Server (NTRS)
Shehadeh, Hana; Lea, Robert N.
1992-01-01
Through previous work, a fuzzy control system was developed to perform translational and rotational control of a space vehicle. This problem was then re-examined to determine the effectiveness of genetic algorithms on fine tuning the controller. This paper explains the problems associated with the design of this fuzzy controller and offers a technique for tuning fuzzy logic controllers. A fuzzy logic controller is a rule-based system that uses fuzzy linguistic variables to model human rule-of-thumb approaches to control actions within a given system. This 'fuzzy expert system' features rules that direct the decision process and membership functions that convert the linguistic variables into the precise numeric values used for system control. Defining the fuzzy membership functions is the most time consuming aspect of the controller design. One single change in the membership functions could significantly alter the performance of the controller. This membership function definition can be accomplished by using a trial and error technique to alter the membership functions creating a highly tuned controller. This approach can be time consuming and requires a great deal of knowledge from human experts. In order to shorten development time, an iterative procedure for altering the membership functions to create a tuned set that used a minimal amount of fuel for velocity vector approach and station-keep maneuvers was developed. Genetic algorithms, search techniques used for optimization, were utilized to solve this problem.
A Genetic Algorithm Approach to Motion Sensor Placement in Smart Environments
Thomas, Brian L.; Crandall, Aaron S.; Cook, Diane J.
2016-01-01
Smart environments and ubiquitous computing technologies hold great promise for a wide range of real world applications. The medical community is particularly interested in high quality measurement of activities of daily living. With accurate computer modeling of older adults, decision support tools may be built to assist care providers. One aspect of effectively deploying these technologies is determining where the sensors should be placed in the home to effectively support these end goals. This work introduces and evaluates a set of approaches for generating sensor layouts in the home. These approaches range from the gold standard of human intuition-based placement to more advanced search algorithms, including Hill Climbing and Genetic Algorithms. The generated layouts are evaluated based on their ability to detect activities while minimizing the number of needed sensors. Sensor-rich environments can provide valuable insights about adults as they go about their lives. These sensors, once in place, provide information on daily behavior that can facilitate an aging-in-place approach to health care. PMID:27453810
Liu, Yuangang; Guo, Qingsheng; Sun, Yageng; Ma, Xiaoya
2014-01-01
Scale reduction from source to target maps inevitably leads to conflicts of map symbols in cartography and geographic information systems (GIS). Displacement is one of the most important map generalization operators and it can be used to resolve the problems that arise from conflict among two or more map objects. In this paper, we propose a combined approach based on constraint Delaunay triangulation (CDT) skeleton and improved elastic beam algorithm for automated building displacement. In this approach, map data sets are first partitioned. Then the displacement operation is conducted in each partition as a cyclic and iterative process of conflict detection and resolution. In the iteration, the skeleton of the gap spaces is extracted using CDT. It then serves as an enhanced data model to detect conflicts and construct the proximity graph. Then, the proximity graph is adjusted using local grouping information. Under the action of forces derived from the detected conflicts, the proximity graph is deformed using the improved elastic beam algorithm. In this way, buildings are displaced to find an optimal compromise between related cartographic constraints. To validate this approach, two topographic map data sets (i.e., urban and suburban areas) were tested. The results were reasonable with respect to each constraint when the density of the map was not extremely high. In summary, the improvements include (1) an automated parameter-setting method for elastic beams, (2) explicit enforcement regarding the positional accuracy constraint, added by introducing drag forces, (3) preservation of local building groups through displacement over an adjusted proximity graph, and (4) an iterative strategy that is more likely to resolve the proximity conflicts than the one used in the existing elastic beam algorithm. PMID:25470727
Liu, Yuangang; Guo, Qingsheng; Sun, Yageng; Ma, Xiaoya
2014-01-01
Scale reduction from source to target maps inevitably leads to conflicts of map symbols in cartography and geographic information systems (GIS). Displacement is one of the most important map generalization operators and it can be used to resolve the problems that arise from conflict among two or more map objects. In this paper, we propose a combined approach based on constraint Delaunay triangulation (CDT) skeleton and improved elastic beam algorithm for automated building displacement. In this approach, map data sets are first partitioned. Then the displacement operation is conducted in each partition as a cyclic and iterative process of conflict detection and resolution. In the iteration, the skeleton of the gap spaces is extracted using CDT. It then serves as an enhanced data model to detect conflicts and construct the proximity graph. Then, the proximity graph is adjusted using local grouping information. Under the action of forces derived from the detected conflicts, the proximity graph is deformed using the improved elastic beam algorithm. In this way, buildings are displaced to find an optimal compromise between related cartographic constraints. To validate this approach, two topographic map data sets (i.e., urban and suburban areas) were tested. The results were reasonable with respect to each constraint when the density of the map was not extremely high. In summary, the improvements include (1) an automated parameter-setting method for elastic beams, (2) explicit enforcement regarding the positional accuracy constraint, added by introducing drag forces, (3) preservation of local building groups through displacement over an adjusted proximity graph, and (4) an iterative strategy that is more likely to resolve the proximity conflicts than the one used in the existing elastic beam algorithm. PMID:25470727
A genetic algorithm approach to probing the evolution of self-organized nanostructured systems.
Siepmann, Peter; Martin, Christopher P; Vancea, Ioan; Moriarty, Philip J; Krasnogor, Natalio
2007-07-01
We present a new methodology, based on a combination of genetic algorithms and image morphometry, for matching the outcome of a Monte Carlo simulation to experimental observations of a far-from-equilibrium nanosystem. The Monte Carlo model used simulates a colloidal solution of nanoparticles drying on a solid substrate and has previously been shown to produce patterns very similar to those observed experimentally. Our approach enables the broad parameter space associated with simulated nanoparticle self-organization to be searched effectively for a given experimental target morphology. PMID:17552572
New approach for motion coordination of a mobile manipulator using fuzzy behavioral algorithms
NASA Astrophysics Data System (ADS)
Haeusler, Kurt; Klement, Erich P.; Zeichen, Gerfried
1998-10-01
In this paper a new approach for the coordination of the motion axes of a mobile manipulator based on fuzzy behavioral algorithms and its implementation on a physical demonstrator is presented. The kinematic redundancy of the overall system (consisting of a 7 DOF manipulator and a 3 DOF mobile robot) will be used for autonomous and reactive motion of the mobile manipulator within poorly structured and even dynamically changing surroundings. Sensors around the mobile and along the manipulator will provide the necessary information for navigation purposes and perception of the environment.
NASA Technical Reports Server (NTRS)
Blackwell, William C., Jr.
2004-01-01
In this paper space is modeled as a lattice of Compton wave oscillators (CWOs) of near- Planck size. It is shown that gravitation and special relativity emerge from the interaction between particles Compton waves. To develop this CWO model an algorithmic approach was taken, incorporating simple rules of interaction at the Planck-scale developed using well known physical laws. This technique naturally leads to Newton s law of gravitation and a new form of doubly special relativity. The model is in apparent agreement with the holographic principle, and it predicts a cutoff energy for ultrahigh-energy cosmic rays that is consistent with observational data.
Omidi-Kashani, Farzad; Ebrahimzadeh, Mohamad Hossein; Salari, Saman
2014-12-01
Lumbar spondylolysis and spondylolisthesis are common spinal disorders that most of the times are incidental findings or respond favorably to conservative treatment. In a small percentage of the patients, surgical intervention becomes necessary. Because too much attention has been paid to novel surgical techniques and new modern spinal implants, some of fundamental concepts have been forgotten. Identifying that small but important number of patients with lumbar spondylolysis or spondylolisthesis who would really benefit from lumbar surgery is one of those forgotten concepts. In this paper, we have developed an algorithmic approach to determine who is a good candidate for surgery due to lumbar spondylolysis or spondylolisthesis. PMID:25558333
Compromise Approach-Based Genetic Algorithm for Constrained Multiobjective Portfolio Selection Model
NASA Astrophysics Data System (ADS)
Li, Jun
In this paper, fuzzy set theory is incorporated into a multiobjective portfolio selection model for investors’ taking into three criteria: return, risk and liquidity. The cardinality constraint, the buy-in threshold constraint and the round-lots constraints are considered in the proposed model. To overcome the difficulty of evaluation a large set of efficient solutions and selection of the best one on non-dominated surface, a compromise approach-based genetic algorithm is presented to obtain a compromised solution for the proposed constrained multiobjective portfolio selection model.
A Heuristic Approach Based on Clarke-Wright Algorithm for Open Vehicle Routing Problem
2013-01-01
We propose a heuristic approach based on the Clarke-Wright algorithm (CW) to solve the open version of the well-known capacitated vehicle routing problem in which vehicles are not required to return to the depot after completing service. The proposed CW has been presented in four procedures composed of Clarke-Wright formula modification, open-route construction, two-phase selection, and route postimprovement. Computational results show that the proposed CW is competitive and outperforms classical CW in all directions. Moreover, the best known solution is also obtained in 97% of tested instances (60 out of 62). PMID:24382948
Ebrahimzadeh, Mohamad Hossein; Salari, Saman
2014-01-01
Lumbar spondylolysis and spondylolisthesis are common spinal disorders that most of the times are incidental findings or respond favorably to conservative treatment. In a small percentage of the patients, surgical intervention becomes necessary. Because too much attention has been paid to novel surgical techniques and new modern spinal implants, some of fundamental concepts have been forgotten. Identifying that small but important number of patients with lumbar spondylolysis or spondylolisthesis who would really benefit from lumbar surgery is one of those forgotten concepts. In this paper, we have developed an algorithmic approach to determine who is a good candidate for surgery due to lumbar spondylolysis or spondylolisthesis. PMID:25558333
Brasier, Martin D; Antcliffe, Jonathan; Saunders, Martin; Wacey, David
2015-04-21
New analytical approaches and discoveries are demanding fresh thinking about the early fossil record. The 1.88-Ga Gunflint chert provides an important benchmark for the analysis of early fossil preservation. High-resolution analysis of Gunflintia shows that microtaphonomy can help to resolve long-standing paleobiological questions. Novel 3D nanoscale reconstructions of the most ancient complex fossil Eosphaera reveal features hitherto unmatched in any crown-group microbe. While Eosphaera may preserve a symbiotic consortium, a stronger conclusion is that multicellular morphospace was differently occupied in the Paleoproterozoic. The 3.46-Ga Apex chert provides a test bed for claims of biogenicity of cell-like structures. Mapping plus focused ion beam milling combined with transmission electron microscopy data demonstrate that microfossil-like taxa, including species of Archaeoscillatoriopsis and Primaevifilum, are pseudofossils formed from vermiform phyllosilicate grains during hydrothermal alteration events. The 3.43-Ga Strelley Pool Formation shows that plausible early fossil candidates are turning up in unexpected environmental settings. Our data reveal how cellular clusters of unexpectedly large coccoids and tubular sheath-like envelopes were trapped between sand grains and entombed within coatings of dripstone beach-rock silica cement. These fossils come from Earth's earliest known intertidal to supratidal shoreline deposit, accumulated under aerated but oxygen poor conditions. PMID:25901305
A discrete twin-boundary approach for simulating the magneto-mechanical response of Ni–Mn–Ga
NASA Astrophysics Data System (ADS)
Faran, Eilon; Shilo, Doron
2016-09-01
The design and optimization of ferromagnetic shape memory alloys (FSMA)-based devices require quantitative understanding of the dynamics of twin boundaries within these materials. Here, we present a discrete twin boundary modeling approach for simulating the behavior of an FSMA Ni–Mn–Ga crystal under combined magneto-mechanical loading conditions. The model is based on experimentally measured kinetic relations that describe the motion of individual twin boundaries over a wide range of velocities. The resulting calculations capture the dynamic response of Ni–Mn–Ga and reveal the relations between fundamental material parameters and actuation performance at different frequencies of the magnetic field. In particular, we show that at high field rates, the magnitude of the lattice barrier that resists twin boundary motion is the important property that determines the level of actuation strain, while the contribution of twinning stress property is minor. Consequently, type II twin boundaries, whose lattice barrier is smaller compared to type I, are expected to show better actuation performance at high rates, irrespective of the differences in the twinning stress property between the two boundary types. In addition, the simulation enables optimization of the actuation strain of a Ni–Mn–Ga crystal by adjusting the magnitude of the bias mechanical stress, thus providing direct guidelines for the design of actuating devices. Finally, we show that the use of a linear kinetic law for simulating the twinning-based response is inadequate and results in incorrect predictions.
A conflict-free, path-level parallelization approach for sequential simulation algorithms
NASA Astrophysics Data System (ADS)
Rasera, Luiz Gustavo; Machado, Péricles Lopes; Costa, João Felipe C. L.
2015-07-01
Pixel-based simulation algorithms are the most widely used geostatistical technique for characterizing the spatial distribution of natural resources. However, sequential simulation does not scale well for stochastic simulation on very large grids, which are now commonly found in many petroleum, mining, and environmental studies. With the availability of multiple-processor computers, there is an opportunity to develop parallelization schemes for these algorithms to increase their performance and efficiency. Here we present a conflict-free, path-level parallelization strategy for sequential simulation. The method consists of partitioning the simulation grid into a set of groups of nodes and delegating all available processors for simulation of multiple groups of nodes concurrently. An automated classification procedure determines which groups are simulated in parallel according to their spatial arrangement in the simulation grid. The major advantage of this approach is that it does not require conflict resolution operations, and thus allows exact reproduction of results. Besides offering a large performance gain when compared to the traditional serial implementation, the method provides efficient use of computational resources and is generic enough to be adapted to several sequential algorithms.
An algorithmic and information-theoretic approach to multimetric index construction
Schoolmaster, Donald R., Jr.; Grace, James B.; Schweiger, E. William; Guntenspergen, Glenn R.; Mitchell, Brian R.; Miller, Kathryn M.; Little, Amanda M.
2013-01-01
The use of multimetric indices (MMIs), such as the widely used index of biological integrity (IBI), to measure, track, summarize and infer the overall impact of human disturbance on biological communities has been steadily growing in recent years. Initially, MMIs were developed for aquatic communities using pre-selected biological metrics as indicators of system integrity. As interest in these bioassessment tools has grown, so have the types of biological systems to which they are applied. For many ecosystem types the appropriate biological metrics to use as measures of biological integrity are not known a priori. As a result, a variety of ad hoc protocols for selecting metrics empirically has developed. However, the assumptions made by proposed protocols have not be explicitly described or justified, causing many investigators to call for a clear, repeatable methodology for developing empirically derived metrics and indices that can be applied to any biological system. An issue of particular importance that has not been sufficiently addressed is the way that individual metrics combine to produce an MMI that is a sensitive composite indicator of human disturbance. In this paper, we present and demonstrate an algorithm for constructing MMIs given a set of candidate metrics and a measure of human disturbance. The algorithm uses each metric to inform a candidate MMI, and then uses information-theoretic principles to select MMIs that capture the information in the multidimensional system response from among possible MMIs. Such an approach can be used to create purely empirical (data-based) MMIs or can, optionally, be influenced by expert opinion or biological theory through the use of a weighting vector to create value-weighted MMIs. We demonstrate the algorithm with simulated data to demonstrate the predictive capacity of the final MMIs and with real data from wetlands from Acadia and Rocky Mountain National Parks. For the Acadia wetland data, the algorithm identified
Combined mixed approach algorithm for in-line phase-contrast x-ray imaging
De Caro, Liberato; Scattarella, Francesco; Giannini, Cinzia; Tangaro, Sabina; Rigon, Luigi; Longo, Renata; Bellotti, Roberto
2010-07-15
Purpose: In the past decade, phase-contrast imaging (PCI) has been applied to study different kinds of tissues and human body parts, with an increased improvement of the image quality with respect to simple absorption radiography. A technique closely related to PCI is phase-retrieval imaging (PRI). Indeed, PCI is an imaging modality thought to enhance the total contrast of the images through the phase shift introduced by the object (human body part); PRI is a mathematical technique to extract the quantitative phase-shift map from PCI. A new phase-retrieval algorithm for the in-line phase-contrast x-ray imaging is here proposed. Methods: The proposed algorithm is based on a mixed transfer-function and transport-of-intensity approach (MA) and it requires, at most, an initial approximate estimate of the average phase shift introduced by the object as prior knowledge. The accuracy in the initial estimate determines the convergence speed of the algorithm. The proposed algorithm retrieves both the object phase and its complex conjugate in a combined MA (CMA). Results: Although slightly less computationally effective with respect to other mixed-approach algorithms, as two phases have to be retrieved, the results obtained by the CMA on simulated data have shown that the obtained reconstructed phase maps are characterized by particularly low normalized mean square errors. The authors have also tested the CMA on noisy experimental phase-contrast data obtained by a suitable weakly absorbing sample consisting of a grid of submillimetric nylon fibers as well as on a strongly absorbing object made of a 0.03 mm thick lead x-ray resolution star pattern. The CMA has shown a good efficiency in recovering phase information, also in presence of noisy data, characterized by peak-to-peak signal-to-noise ratios down to a few dBs, showing the possibility to enhance with phase radiography the signal-to-noise ratio for features in the submillimetric scale with respect to the attenuation
An algorithmic strategy for selecting a surgical approach in cervical deformity correction.
Hann, Shannon; Chalouhi, Nohra; Madineni, Ravichandra; Vaccaro, Alexander R; Albert, Todd J; Harrop, James; Heller, Joshua E
2014-05-01
Adult degenerative cervical kyphosis is a debilitating disease that often requires complex surgical management. Young spine surgeons, residents, and fellows are often confused as to which surgical approach to choose due to lack of experience, absence of a systematic method of surgical management, and today's plethora of information regarding surgical techniques. Although surgeons may be able to perform anterior, posterior, or combined (360°) approaches to the cervical spine, many struggle to rationally choose an appropriate approach for deformity correction. The authors introduce an algorithm based on morphology and pathology of adult cervical kyphosis to help the surgeon select the appropriate approach when performing cervical deformity surgery. Cervical deformities are categorized into 5 different prevalent morphological types encountered in clinical settings. A surgical approach tailored to each category/type of deformity is then discussed, with a concrete case illustration provided for each. Preoperative assessment of kyphosis, determination of the goal for surgery, and the complications associated with cervical deformity correction are also summarized. This article's goal is to assist with understanding the big picture for surgical management in cervical spinal deformity. PMID:24785487
NASA Astrophysics Data System (ADS)
Okawa, S.; Yamamoto, H.; Miwa, Y.; Yamada, Y.
2011-07-01
Fluorescence diffuse optical tomography (FDOT) based on the total light approach is developed. The continuous wave light is used for excitation in this system. The reconstruction algorithm is based on the total light approach that reconstructs the absorption coefficients increased by the fluorophore. Additionally we propose noise reduction using the algebraic reconstruction technique (ART) incorporating the truncated singular value decomposition (TSVD). Numerical and phantom experiments show that the developed system successfully reconstructs the fluorophore concentration in the biological media, and the ART with TSVD alleviates the influence of noises. In vivo experiment demonstrated that the developed FDOT system localized the fluorescent agent which was concentrated in the cancer transplanted into a kidney in a mouse.
NASA Astrophysics Data System (ADS)
Shyue, Keh-Ming; Xiao, Feng
2014-07-01
We describe a novel interface-sharpening approach for efficient numerical resolution of a compressible homogeneous two-phase flow governed by a quasi-conservative five-equation model of Allaire et al. (2001) [1]. The algorithm uses a semi-discrete wave propagation method to find approximate solution of this model numerically. In the algorithm, in regions near the interfaces where two different fluid components are present within a cell, the THINC (Tangent of Hyperbola for INterface Capturing) scheme is used as a basis for the reconstruction of a sub-grid discontinuity of volume fractions at each cell edge, and it is complemented by a homogeneous-equilibrium-consistent technique that is derived to ensure a consistent modeling of the other interpolated physical variables in the model. In regions away from the interfaces where the flow is single phase, standard reconstruction scheme such as MUSCL or WENO can be used for obtaining high-order interpolated states. These reconstructions are then used as the initial data for Riemann problems, and the resulting fluctuations form the basis for the spatial discretization. Time integration of the algorithm is done by employing a strong stability-preserving Runge-Kutta method. Numerical results are shown for sample problems with the Mie-Grüneisen equation of state for characterizing the materials of interests in both one and two space dimensions that demonstrate the feasibility of the proposed method for interface-sharpening of compressible two-phase flow. To demonstrate the competitiveness of our approach, we have also included results obtained using the anti-diffusion interface sharpening method.
Soft tissue balancing in varus total knee arthroplasty: an algorithmic approach.
Verdonk, Peter C M; Pernin, Jerome; Pinaroli, Alban; Ait Si Selmi, Tarik; Neyret, Philippe
2009-06-01
We present an algorithmic release approach to the varus knee, including a novel pie crust release technique of the superficial MCL, in 359 total knee arthroplasty patients and report the clinical and radiological outcome. Medio-lateral stability was evaluated as normal in 97% of group 0 (deep MCL), 95% of group 1 (pie crust superficial MCL) and 83% of group 2 (distal superficial MCL). The mean preoperative hip-knee angle was 174.0, 172.1, and 169.5 and was corrected postoperatively to 179.1, 179.2, and 177.6 for groups 0, 1, and 2, respectively. A satisfactory correction in the coronal plane was achieved in 82.9% of all-comers falling within the 180 degrees +/- 3 degrees interval. An algorithmic release approach can be beneficial for soft tissue balancing. In all patients, the deep medial collateral ligament should be released and otseophytes removed. The novel pie crust technique of the superficial MCL is safe, efficient and reliable, provided a medial release of 6-8 mm or less is required. The release of the superficial MCL on the distal tibia is advocated in severe varus knees. Preoperative coronal alignment is an important predictor for the release technique, but should be combined with other parameters such as reducibility of the deformity and the obtained gap asymmetry. PMID:19290507
Vertical and lateral flight optimization algorithm and missed approach cost calculation
NASA Astrophysics Data System (ADS)
Murrieta Mendoza, Alejandro
Flight trajectory optimization is being looked as a way of reducing flight costs, fuel burned and emissions generated by the fuel consumption. The objective of this work is to find the optimal trajectory between two points. To find the optimal trajectory, the parameters of weight, cost index, initial coordinates, and meteorological conditions along the route are provided to the algorithm. This algorithm finds the trajectory where the global cost is the most economical. The global cost is a compromise between fuel burned and flight time, this is determined using a cost index that assigns a cost in terms of fuel to the flight time. The optimization is achieved by calculating a candidate optimal cruise trajectory profile from all the combinations available in the aircraft performance database. With this cruise candidate profile, more cruises profiles are calculated taken into account the climb and descend costs. During cruise, step climbs are evaluated to optimize the trajectory. The different trajectories are compared and the most economical one is defined as the optimal vertical navigation profile. From the optimal vertical navigation profile, different lateral routes are tested. Taking advantage of the meteorological influence, the algorithm looks for the lateral navigation trajectory where the global cost is the most economical. That route is then selected as the optimal lateral navigation profile. The meteorological data was obtained from environment Canada. The new way of obtaining data from the grid from environment Canada proposed in this work resulted in an important computation time reduction compared against other methods such as bilinear interpolation. The algorithm developed here was evaluated in two different aircraft: the Lockheed L-1011 and the Sukhoi Russian regional jet. The algorithm was developed in MATLAB, and the validation was performed using Flight-Sim by Presagis and the FMS CMA-9000 by CMC Electronics -- Esterline. At the end of this work a
NASA Astrophysics Data System (ADS)
D'Avezac, Mayeul; Zunger, Alex
2007-03-01
In many problems in molecular and solid state structures one needs to determine the energy-minimizing decoration of sites by different atom-types (i. e.configuration). The sheer size of this configurational space can be horrendous even if the underlying lattice-type is known. The ab-initio total-energy surface for different (relaxed) configurations can often be parameterized by a spin-like Hamiltonian (Cluster-Expansion) with discrete spin -variables denoting the type of atom occupying each site. We compare two search strategies for the energy-minimizing configuration: (i) A discrete-variable genetic-algorithm approach( S. V. Dudiy and A. Zunger, PRL 97, 046401 (2006) ) and (ii) a continuous-variable approach (M. Wang et al, J. Am. Chem. Soc. 128, 3228 (2006) ) where the discrete-spin functional is mapped onto a continuous-spin functional (virtual atoms) and the search is guided by local gradients with respect to each spin. We compare their efficiency at locating the ground-state configurations of fcc Au-Pd Alloy in terms of number of calls to the functional. We show that a GA approach with diversity-enhancing constraints and reciprocal-space mating easily outperforms the VA approach.
Balima, O.; Favennec, Y.; Rousse, D.
2013-10-15
Highlights: •New strategies to improve the accuracy of the reconstruction through mesh and finite element parameterization. •Use of gradient filtering through an alternative inner product within the adjoint method. •An integral form of the cost function is used to make the reconstruction compatible with all finite element formulations, continuous and discontinuous. •Gradient-based algorithm with the adjoint method is used for the reconstruction. -- Abstract: Optical tomography is mathematically treated as a non-linear inverse problem where the optical properties of the probed medium are recovered through the minimization of the errors between the experimental measurements and their predictions with a numerical model at the locations of the detectors. According to the ill-posed behavior of the inverse problem, some regularization tools must be performed and the Tikhonov penalization type is the most commonly used in optical tomography applications. This paper introduces an optimized approach for optical tomography reconstruction with the finite element method. An integral form of the cost function is used to take into account the surfaces of the detectors and make the reconstruction compatible with all finite element formulations, continuous and discontinuous. Through a gradient-based algorithm where the adjoint method is used to compute the gradient of the cost function, an alternative inner product is employed for preconditioning the reconstruction algorithm. Moreover, appropriate re-parameterization of the optical properties is performed. These regularization strategies are compared with the classical Tikhonov penalization one. It is shown that both the re-parameterization and the use of the Sobolev cost function gradient are efficient for solving such an ill-posed inverse problem.
Armañanzas, Rubén; Saeys, Yvan; Inza, Iñaki; García-Torres, Miguel; Bielza, Concha; van de Peer, Yves; Larrañaga, Pedro
2011-01-01
Progress is continuously being made in the quest for stable biomarkers linked to complex diseases. Mass spectrometers are one of the devices for tackling this problem. The data profiles they produce are noisy and unstable. In these profiles, biomarkers are detected as signal regions (peaks), where control and disease samples behave differently. Mass spectrometry (MS) data generally contain a limited number of samples described by a high number of features. In this work, we present a novel class of evolutionary algorithms, estimation of distribution algorithms (EDA), as an efficient peak selector in this MS domain. There is a trade-of f between the reliability of the detected biomarkers and the low number of samples for analysis. For this reason, we introduce a consensus approach, built upon the classical EDA scheme, that improves stability and robustness of the final set of relevant peaks. An entire data workflow is designed to yield unbiased results. Four publicly available MS data sets (two MALDI-TOF and another two SELDI-TOF) are analyzed. The results are compared to the original works, and a new plot (peak frequential plot) for graphically inspecting the relevant peaks is introduced. A complete online supplementary page, which can be found at http://www.sc.ehu.es/ccwbayes/members/ruben/ms, includes extended info and results, in addition to Matlab scripts and references. PMID:21393653
A Decompositional Approach to Executing Quality Data Model Algorithms on the i2b2 Platform.
Mo, Huan; Jiang, Guoqian; Pacheco, Jennifer A; Kiefer, Richard; Rasmussen, Luke V; Pathak, Jyotishman; Denny, Joshua C; Thompson, William K
2016-01-01
The Quality Data Model (QDM) is an established standard for representing electronic clinical quality measures on electronic health record (EHR) repositories. The Informatics for Integrated Biology and the Bedside (i2b2) is a widely used platform for implementing clinical data repositories. However, translation from QDM to i2b2 is challenging, since QDM allows for complex queries beyond the capability of single i2b2 messages. We have developed an approach to decompose complex QDM algorithms into workflows of single i2b2 messages, and execute them on the KNIME data analytics platform. Each workflow operation module is composed of parameter lists, a template for the i2b2 message, an mechanism to create parameter updates, and a web service call to i2b2. The communication between workflow modules relies on passing keys ofi2b2 result sets. As a demonstration of validity, we describe the implementation and execution of a type 2 diabetes mellitus phenotype algorithm against an i2b2 data repository. PMID:27570665
Du, Hubing; Gao, Honghong
2016-08-20
Affected by the height dependent effects, the phase-shifting shadow moiré can only be implemented in an approximate way. In the technique, a fixed phase step around π/2 rad between two adjacent frames is usually introduced by a grating translation in its own plane. So the method is not flexible in some situations. Additionally, because the shadow moiré fringes have a complex intensity distribution, computing the introduced phase shift from the existing arccosine function or arcsine function-based phase shift extraction algorithm always exhibits instability. To solve it, we developed a Gram-Schmidt orthonormalization approach based on a three-frame self-calibration phase-shifting algorithm with equal but unknown phase steps. The proposed method using the arctangent function is fast and can be implemented robustly in many applications. We also do optical experiments to demonstrate the correction of the proposed method by referring to the result of the conventional five-step phase-shifting shadow moiré. The results show the correctness of the proposed method. PMID:27556993
A Decompositional Approach to Executing Quality Data Model Algorithms on the i2b2 Platform
Mo, Huan; Jiang, Guoqian; Pacheco, Jennifer A.; Kiefer, Richard; Rasmussen, Luke V.; Pathak, Jyotishman; Denny, Joshua C.; Thompson, William K.
2016-01-01
The Quality Data Model (QDM) is an established standard for representing electronic clinical quality measures on electronic health record (EHR) repositories. The Informatics for Integrated Biology and the Bedside (i2b2) is a widely used platform for implementing clinical data repositories. However, translation from QDM to i2b2 is challenging, since QDM allows for complex queries beyond the capability of single i2b2 messages. We have developed an approach to decompose complex QDM algorithms into workflows of single i2b2 messages, and execute them on the KNIME data analytics platform. Each workflow operation module is composed of parameter lists, a template for the i2b2 message, an mechanism to create parameter updates, and a web service call to i2b2. The communication between workflow modules relies on passing keys ofi2b2 result sets. As a demonstration of validity, we describe the implementation and execution of a type 2 diabetes mellitus phenotype algorithm against an i2b2 data repository. PMID:27570665
A Dynamic Health Assessment Approach for Shearer Based on Artificial Immune Algorithm.
Wang, Zhongbin; Xu, Xihua; Si, Lei; Ji, Rui; Liu, Xinhua; Tan, Chao
2016-01-01
In order to accurately identify the dynamic health of shearer, reducing operating trouble and production accident of shearer and improving coal production efficiency further, a dynamic health assessment approach for shearer based on artificial immune algorithm was proposed. The key technologies such as system framework, selecting the indicators for shearer dynamic health assessment, and health assessment model were provided, and the flowchart of the proposed approach was designed. A simulation example, with an accuracy of 96%, based on the collected data from industrial production scene was provided. Furthermore, the comparison demonstrated that the proposed method exhibited higher classification accuracy than the classifiers based on back propagation-neural network (BP-NN) and support vector machine (SVM) methods. Finally, the proposed approach was applied in an engineering problem of shearer dynamic health assessment. The industrial application results showed that the paper research achievements could be used combining with shearer automation control system in fully mechanized coal face. The simulation and the application results indicated that the proposed method was feasible and outperforming others. PMID:27123002
A Dynamic Health Assessment Approach for Shearer Based on Artificial Immune Algorithm
Wang, Zhongbin; Xu, Xihua; Si, Lei; Ji, Rui; Liu, Xinhua; Tan, Chao
2016-01-01
In order to accurately identify the dynamic health of shearer, reducing operating trouble and production accident of shearer and improving coal production efficiency further, a dynamic health assessment approach for shearer based on artificial immune algorithm was proposed. The key technologies such as system framework, selecting the indicators for shearer dynamic health assessment, and health assessment model were provided, and the flowchart of the proposed approach was designed. A simulation example, with an accuracy of 96%, based on the collected data from industrial production scene was provided. Furthermore, the comparison demonstrated that the proposed method exhibited higher classification accuracy than the classifiers based on back propagation-neural network (BP-NN) and support vector machine (SVM) methods. Finally, the proposed approach was applied in an engineering problem of shearer dynamic health assessment. The industrial application results showed that the paper research achievements could be used combining with shearer automation control system in fully mechanized coal face. The simulation and the application results indicated that the proposed method was feasible and outperforming others. PMID:27123002
GA-ANFIS Expert System Prototype for Prediction of Dermatological Diseases.
Begic Fazlic, Lejla; Avdagic, Korana; Omanovic, Samir
2015-01-01
This paper presents novel GA-ANFIS expert system prototype for dermatological disease detection by using dermatological features and diagnoses collected in real conditions. Nine dermatological features are used as inputs to classifiers that are based on Adaptive Neuro-Fuzzy Inference Systems (ANFIS) for the first level of fuzzy model optimization. After that, they are used as inputs in Genetic Algorithm (GA) for the second level of fuzzy model optimization within GA-ANFIS system. GA-ANFIS system performs optimization in two steps. Modelling and validation of the novel GA-ANFIS system approach is performed in MATLAB environment by using validation set of data. Some conclusions concerning the impacts of features on the detection of dermatological diseases were obtained through analysis of the GA-ANFIS. We compared GA-ANFIS and ANFIS results. The results confirmed that the proposed GA-ANFIS model achieved accuracy rates which are higher than the ones we got by ANFIS model. PMID:25991223
Shang, J.S.; Andrienko, D.A.; Huang, P.G.; Surzhikov, S.T.
2014-06-01
An efficient computational capability for nonequilibrium radiation simulation via the ray tracing technique has been accomplished. The radiative rate equation is iteratively coupled with the aerodynamic conservation laws including nonequilibrium chemical and chemical–physical kinetic models. The spectral properties along tracing rays are determined by a space partition algorithm of the nearest neighbor search process, and the numerical accuracy is further enhanced by a local resolution refinement using the Gauss–Lobatto polynomial. The interdisciplinary governing equations are solved by an implicit delta formulation through the diminishing residual approach. The axisymmetric radiating flow fields over the reentry RAM-CII probe have been simulated and verified with flight data and previous solutions by traditional methods. A computational efficiency gain nearly forty times is realized over that of the existing simulation procedures.
Combinatorial Multiobjective Optimization Using Genetic Algorithms
NASA Technical Reports Server (NTRS)
Crossley, William A.; Martin. Eric T.
2002-01-01
The research proposed in this document investigated multiobjective optimization approaches based upon the Genetic Algorithm (GA). Several versions of the GA have been adopted for multiobjective design, but, prior to this research, there had not been significant comparisons of the most popular strategies. The research effort first generalized the two-branch tournament genetic algorithm in to an N-branch genetic algorithm, then the N-branch GA was compared with a version of the popular Multi-Objective Genetic Algorithm (MOGA). Because the genetic algorithm is well suited to combinatorial (mixed discrete / continuous) optimization problems, the GA can be used in the conceptual phase of design to combine selection (discrete variable) and sizing (continuous variable) tasks. Using a multiobjective formulation for the design of a 50-passenger aircraft to meet the competing objectives of minimizing takeoff gross weight and minimizing trip time, the GA generated a range of tradeoff designs that illustrate which aircraft features change from a low-weight, slow trip-time aircraft design to a heavy-weight, short trip-time aircraft design. Given the objective formulation and analysis methods used, the results of this study identify where turboprop-powered aircraft and turbofan-powered aircraft become more desirable for the 50 seat passenger application. This aircraft design application also begins to suggest how a combinatorial multiobjective optimization technique could be used to assist in the design of morphing aircraft.
Fabrication of AlGaN/GaN Ω-shaped nanowire fin-shaped FETs by a top-down approach
NASA Astrophysics Data System (ADS)
Im, Ki-Sik; Sindhuri, Vodapally; Jo, Young-Woo; Son, Dong-Hyeok; Lee, Jae-Hoon; Cristoloveanu, Sorin; Lee, Jung-Hee
2015-06-01
An AlGaN/GaN-based Ω-shaped nanowire fin-shaped FET (FinFET) with a fin width of 50 nm was fabricated using tetramethylammonium hydroxide (TMAH)-based lateral wet etching. An atomic layer deposited (ALD) HfO2 side-wall layer served as the etching mask. ALD Al2O3 and TiN layers were used as the gate dielectric and gate metal, respectively. The Ω-shaped gate structure fully depletes the active fin body and almost completely separates the depleted fin from the underlying thick GaN buffer layer, resulting in superior device performance. The top-down processing proposed in this work provides a viable pathway towards gate-all-around devices for III-nitride semiconductors.
Fist Principles Approach to the Magneto Caloric Effect: Application to Ni2MnGa
NASA Astrophysics Data System (ADS)
Odbadrakh, Khorgolkhuu; Nicholson, Don; Rusanu, Aurelian; Eisenbach, Markus; Brown, Gregory; Evans, Boyd, III
2011-03-01
The magneto-caloric effect (MCE) has potential application in heating and cooling technologies. In this work, we present calculated magnetic structure of a candidate MCE material, Ni 2 MnGa. The magnetic configurations of a 144 atom supercell is first explored using first-principle, the results are then used to fit exchange parameters of a Heisenberg Hamiltonian. The Wang-Landau method is used to calculate the magnetic density of states of the Heisenberg Hamiltonian. Based on this classical estimate, the magnetic density of states is calculated using the Wang Landau method with energies obtained from the first principles method. The Currie temperature and other thermodynamic properties are calculated using the density of states. The relationships between the density of magnetic states and the field induced adiabatic temperature change and isothermal entropy change are discussed. This work was sponsored by the Laboratory Directed Research and Development Program (ORNL), by the Mathematical, Information, and Computational Sciences Division; Office of Advanced Scientific Computing Research (US DOE), and by the Materials Sciences and Engineering Division; Office of Basic Energy Sciences (US DOE).
Processing approach towards the formation of thin-film Cu(In,Ga)Se2
Beck, Markus E.; Noufi, Rommel
2003-01-01
A two-stage method of producing thin-films of group IB-IIIA-VIA on a substrate for semiconductor device applications includes a first stage of depositing an amorphous group IB-IIIA-VIA precursor onto an unheated substrate, wherein the precursor contains all of the group IB and group IIIA constituents of the semiconductor thin-film to be produced in the stoichiometric amounts desired for the final product, and a second stage which involves subjecting the precursor to a short thermal treatment at 420.degree. C.-550.degree. C. in a vacuum or under an inert atmosphere to produce a single-phase, group IB-III-VIA film. Preferably the precursor also comprises the group VIA element in the stoichiometric amount desired for the final semiconductor thin-film. The group IB-IIIA-VIA semiconductor films may be, for example, Cu(In,Ga)(Se,S).sub.2 mixed-metal chalcogenides. The resultant supported group IB-IIIA-VIA semiconductor film is suitable for use in photovoltaic applications.
ERIC Educational Resources Information Center
Reese, Debbie Denise; Tabachnick, Barbara G.
2010-01-01
In this paper, the authors summarize a quantitative analysis demonstrating that the CyGaMEs toolset for embedded assessment of learning within instructional games measures growth in conceptual knowledge by quantifying player behavior. CyGaMEs stands for Cyberlearning through GaME-based, Metaphor Enhanced Learning Objects. Some scientists of…
A data mining approach to optimize pellets manufacturing process based on a decision tree algorithm.
Ronowicz, Joanna; Thommes, Markus; Kleinebudde, Peter; Krysiński, Jerzy
2015-06-20
The present study is focused on the thorough analysis of cause-effect relationships between pellet formulation characteristics (pellet composition as well as process parameters) and the selected quality attribute of the final product. The shape using the aspect ratio value expressed the quality of pellets. A data matrix for chemometric analysis consisted of 224 pellet formulations performed by means of eight different active pharmaceutical ingredients and several various excipients, using different extrusion/spheronization process conditions. The data set contained 14 input variables (both formulation and process variables) and one output variable (pellet aspect ratio). A tree regression algorithm consistent with the Quality by Design concept was applied to obtain deeper understanding and knowledge of formulation and process parameters affecting the final pellet sphericity. The clear interpretable set of decision rules were generated. The spehronization speed, spheronization time, number of holes and water content of extrudate have been recognized as the key factors influencing pellet aspect ratio. The most spherical pellets were achieved by using a large number of holes during extrusion, a high spheronizer speed and longer time of spheronization. The described data mining approach enhances knowledge about pelletization process and simultaneously facilitates searching for the optimal process conditions which are necessary to achieve ideal spherical pellets, resulting in good flow characteristics. This data mining approach can be taken into consideration by industrial formulation scientists to support rational decision making in the field of pellets technology. PMID:25835791
Diffuse lung disease of infancy: a pattern-based, algorithmic approach to histological diagnosis.
Armes, Jane E; Mifsud, William; Ashworth, Michael
2015-02-01
Diffuse lung disease (DLD) of infancy has multiple aetiologies and the spectrum of disease is substantially different from that seen in older children and adults. In many cases, a specific diagnosis renders a dire prognosis for the infant, with profound management implications. Two recently published series of DLD of infancy, collated from the archives of specialist centres, indicate that the majority of their cases were referred, implying that the majority of biopsies taken for DLD of infancy are first received by less experienced pathologists. The current literature describing DLD of infancy takes a predominantly aetiological approach to classification. We present an algorithmic, histological, pattern-based approach to diagnosis of DLD of infancy, which, with the aid of appropriate multidisciplinary input, including clinical and radiological expertise and ancillary diagnostic studies, may lead to an accurate and useful interim report, with timely exclusion of inappropriate diagnoses. Subsequent referral to a specialist centre for confirmatory diagnosis will be dependent on the individual case and the decision of the multidisciplinary team. PMID:25477529
Diffuse lung disease of infancy: a pattern-based, algorithmic approach to histological diagnosis
Armes, Jane E; Mifsud, William; Ashworth, Michael
2015-01-01
Diffuse lung disease (DLD) of infancy has multiple aetiologies and the spectrum of disease is substantially different from that seen in older children and adults. In many cases, a specific diagnosis renders a dire prognosis for the infant, with profound management implications. Two recently published series of DLD of infancy, collated from the archives of specialist centres, indicate that the majority of their cases were referred, implying that the majority of biopsies taken for DLD of infancy are first received by less experienced pathologists. The current literature describing DLD of infancy takes a predominantly aetiological approach to classification. We present an algorithmic, histological, pattern-based approach to diagnosis of DLD of infancy, which, with the aid of appropriate multidisciplinary input, including clinical and radiological expertise and ancillary diagnostic studies, may lead to an accurate and useful interim report, with timely exclusion of inappropriate diagnoses. Subsequent referral to a specialist centre for confirmatory diagnosis will be dependent on the individual case and the decision of the multidisciplinary team. PMID:25477529
A Parallel Approach To Optimum Actuator Selection With a Genetic Algorithm
NASA Technical Reports Server (NTRS)
Rogers, James L.
2000-01-01
Recent discoveries in smart technologies have created a variety of aerodynamic actuators which have great potential to enable entirely new approaches to aerospace vehicle flight control. For a revolutionary concept such as a seamless aircraft with no moving control surfaces, there is a large set of candidate locations for placing actuators, resulting in a substantially larger number of combinations to examine in order to find an optimum placement satisfying the mission requirements. The placement of actuators on a wing determines the control effectiveness of the airplane. One approach to placement Maximizes the moments about the pitch, roll, and yaw axes, while minimizing the coupling. Genetic algorithms have been instrumental in achieving good solutions to discrete optimization problems, such as the actuator placement problem. As a proof of concept, a genetic has been developed to find the minimum number of actuators required to provide uncoupled pitch, roll, and yaw control for a simplified, untapered, unswept wing model. To find the optimum placement by searching all possible combinations would require 1,100 hours. Formulating the problem and as a multi-objective problem and modifying it to take advantage of the parallel processing capabilities of a multi-processor computer, reduces the optimization time to 22 hours.
Embedding SAS approach into conjugate gradient algorithms for asymmetric 3D elasticity problems
Chen, Hsin-Chu; Warsi, N.A.; Sameh, A.
1996-12-31
In this paper, we present two strategies to embed the SAS (symmetric-and-antisymmetric) scheme into conjugate gradient (CG) algorithms to make solving 3D elasticity problems, with or without global reflexive symmetry, more efficient. The SAS approach is physically a domain decomposition scheme that takes advantage of reflexive symmetry of discretized physical problems, and algebraically a matrix transformation method that exploits special reflexivity properties of the matrix resulting from discretization. In addition to offering large-grain parallelism, which is valuable in a multiprocessing environment, the SAS scheme also has the potential for reducing arithmetic operations in the numerical solution of a reasonably wide class of scientific and engineering problems. This approach can be applied directly to problems that have global reflexive symmetry, yielding smaller and independent subproblems to solve, or indirectly to problems with partial symmetry, resulting in loosely coupled subproblems. The decomposition is achieved by separating the reflexive subspace from the antireflexive one, possessed by a special class of matrices A, A {element_of} C{sup n x n} that satisfy the relation A = PAP where P is a reflection matrix (symmetric signed permutation matrix).
Optimal management of substrates in anaerobic co-digestion: An ant colony algorithm approach.
Verdaguer, Marta; Molinos-Senante, María; Poch, Manel
2016-04-01
Sewage sludge (SWS) is inevitably produced in urban wastewater treatment plants (WWTPs). The treatment of SWS on site at small WWTPs is not economical; therefore, the SWS is typically transported to an alternative SWS treatment center. There is increased interest in the use of anaerobic digestion (AnD) with co-digestion as an SWS treatment alternative. Although the availability of different co-substrates has been ignored in most of the previous studies, it is an essential issue for the optimization of AnD co-digestion. In a pioneering approach, this paper applies an Ant-Colony-Optimization (ACO) algorithm that maximizes the generation of biogas through AnD co-digestion in order to optimize the discharge of organic waste from different waste sources in real-time. An empirical application is developed based on a virtual case study that involves organic waste from urban WWTPs and agrifood activities. The results illustrate the dominate role of toxicity levels in selecting contributions to the AnD input. The methodology and case study proposed in this paper demonstrate the usefulness of the ACO approach in supporting a decision process that contributes to improving the sustainability of organic waste and SWS management. PMID:26868846
One-year results of an algorithmic approach to managing failed back surgery syndrome
Avellanal, Martín; Diaz-Reganon, Gonzalo; Orts, Alejandro; Soto, Silvia
2014-01-01
BACKGROUND: Failed back surgery syndrome (FBSS) is a major clinical problem. Different etiologies with different incidence rates have been proposed. There are currently no standards regarding the management of these patients. Epiduroscopy is an endoscopic technique that may play a role in the management of FBSS. OBJECTIVE: To evaluate an algorithm for management of severe FBSS including epiduroscopy as a diagnostic and therapeutic tool. METHODS: A total of 133 patients with severe symptoms of FBSS (visual analogue scale score ≥7) and no response to pharmacological treatment and physical therapy were included. A six-step management algorithm was applied. Data, including patient demographics, pain and surgical procedure, were analyzed. In all cases, one or more objective causes of pain were established. Treatment success was defined as ≥50% long-term pain relief maintained during the first year of follow-up. Final allocation of patients was registered: good outcome with conservative treatment, surgical reintervention and palliative treatment with implantable devices. RESULTS: Of 122 patients enrolled, 59.84% underwent instrumented surgery and 40.16% a noninstrumented procedure. Most (64.75%) experienced significant pain relief with conventional pain clinic treatments; 15.57% required surgical treatment. Palliative spinal cord stimulation and spinal analgesia were applied in 9.84% and 2.46% of the cases, respectively. The most common diagnosis was epidural fibrosis, followed by disc herniation, global or lateral stenosis, and foraminal stenosis. CONCLUSIONS: A new six-step ladder approach to severe FBSS management that includes epiduroscopy was analyzed. Etiologies are accurately described and a useful role of epiduroscopy was confirmed. PMID:25222573
NASA Astrophysics Data System (ADS)
Mojarab, Masoud; Kossobokov, Vladimir; Memarian, Hossein; Zare, Mehdi
2015-07-01
On 23rd October 2011, an M7.3 earthquake near the Turkish city of Van, killed more than 600 people, injured over 4000, and left about 60,000 homeless. It demolished hundreds of buildings and caused great damages to thousand others in Van, Ercis, Muradiye, and Çaldıran. The earthquake's epicenter is located about 70 km from a preceding M7.3 earthquake that occurred in November 1976 and destroyed several villages near the Turkey-Iran border and killed thousands of people. This study, by means of retrospective application of the M8 algorithm, checks to see if the 2011 Van earthquake could have been predicted. The algorithm is based on pattern recognition of Times of Increased Probability (TIP) of a target earthquake from the transient seismic sequence at lower magnitude ranges in a Circle of Investigation (CI). Specifically, we applied a modified M8 algorithm adjusted to a rather low level of earthquake detection in the region following three different approaches to determine seismic transients. In the first approach, CI centers are distributed on intersections of morphostructural lineaments recognized as prone to magnitude 7 + earthquakes. In the second approach, centers of CIs are distributed on local extremes of the seismic density distribution, and in the third approach, CI centers were distributed uniformly on the nodes of a 1∘×1∘ grid. According to the results of the M8 algorithm application, the 2011 Van earthquake could have been predicted in any of the three approaches. We noted that it is possible to consider the intersection of TIPs instead of their union to improve the certainty of the prediction results. Our study confirms the applicability of a modified version of the M8 algorithm for predicting earthquakes at the Iranian-Turkish plateau, as well as for mitigation of damages in seismic events in which pattern recognition algorithms may play an important role.
Pitschner, H F; Berkowitsch, A
2001-01-01
Symbolic dynamics as a non linear method and computation of the normalized algorithmic complexity (C alpha) was applied to basket-catheter mapping of atrial fibrillation (AF) in the right human atrium. The resulting different degrees of organisation of AF have been compared to conventional classification of Wells. Short time temporal and spatial distribution of the C alpha during AF and effects of propafenone on this distribution have been investigated in 30 patients. C alpha was calculated for a moving window. Generated C alpha was analyzed within 10 minutes before and after administration of propafenone. The inter-regional C alpha distribution was statistically analyzed. Inter-regional C alpha differences were found in all patients (p < 0.001). The right atrium could be divided in high- and low complexity areas according to individual patterns. A significant C alpha increase in cranio-caudal direction was confirmed inter-individually (p < 0.01). The administration of propafenone enlarged the areas of low complexity. PMID:11889958
NASA Technical Reports Server (NTRS)
Hoang, TY
1994-01-01
A real-time, high-rate precision navigation Kalman filter algorithm is developed and analyzed. This Navigation algorithm blends various navigation data collected during terminal area approach of an instrumented helicopter. Navigation data collected include helicopter position and velocity from a global position system in differential mode (DGPS) as well as helicopter velocity and attitude from an inertial navigation system (INS). The goal of the Navigation algorithm is to increase the DGPS accuracy while producing navigational data at the 64 Hertz INS update rate. It is important to note that while the data was post flight processed, the Navigation algorithm was designed for real-time analysis. The design of the Navigation algorithm resulted in a nine-state Kalman filter. The Kalman filter's state matrix contains position, velocity, and velocity bias components. The filter updates positional readings with DGPS position, INS velocity, and velocity bias information. In addition, the filter incorporates a sporadic data rejection scheme. This relatively simple model met and exceeded the ten meter absolute positional requirement. The Navigation algorithm results were compared with truth data derived from a laser tracker. The helicopter flight profile included terminal glideslope angles of 3, 6, and 9 degrees. Two flight segments extracted during each terminal approach were used to evaluate the Navigation algorithm. The first segment recorded small dynamic maneuver in the lateral plane while motion in the vertical plane was recorded by the second segment. The longitudinal, lateral, and vertical averaged positional accuracies for all three glideslope approaches are as follows (mean plus or minus two standard deviations in meters): longitudinal (-0.03 plus or minus 1.41), lateral (-1.29 plus or minus 2.36), and vertical (-0.76 plus or minus 2.05).
A novel stochastic optimization algorithm.
Li, B; Jiang, W
2000-01-01
This paper presents a new stochastic approach SAGACIA based on proper integration of simulated annealing algorithm (SAA), genetic algorithm (GA), and chemotaxis algorithm (CA) for solving complex optimization problems. SAGACIA combines the advantages of SAA, GA, and CA together. It has the following features: (1) it is not the simple mix of SAA, GA, and CA; (2) it works from a population; (3) it can be easily used to solve optimization problems either with continuous variables or with discrete variables, and it does not need coding and decoding,; and (4) it can easily escape from local minima and converge quickly. Good solutions can be obtained in a very short time. The search process of SAGACIA can be explained with Markov chains. In this paper, it is proved that SAGACIA has the property of global asymptotical convergence. SAGACIA has been applied to solve such problems as scheduling, the training of artificial neural networks, and the optimizing of complex functions. In all the test cases, the performance of SAGACIA is better than that of SAA, GA, and CA. PMID:18244742
NASA Astrophysics Data System (ADS)
Della Mora, S.; Boschi, L.; Becker, T. W.; Giardini, D.
2010-12-01
The wavelength spectrum of three-dimensional (3D) heterogeneity naturally reflects the nature of Earth dynamics, and is in its own right an important constraint for geodynamical modeling. The Earth's spectrum has been usually evaluated indirectly, on the basis of previously derived tomographic models. If the geographic distribution of seismic heterogeneities is neglected, however, one can invert global seismic data directly to find the spectrum of the Earth. Inverting for the spectrum is in principle (fewer unknowns) cheaper and robust than inverting for the 3D structure of a planet: this should allow us to constrain planetary structure at smaller scales than by current 3D models. Based on the work of Gudmundsson and coworkers in the early 1990s, we have developed a linear algorithm for surface waves. The spectra we obtain are in qualitative agreement with results from 3D tomography, but the resolving power is generally lower, due to the simplifications required to linearise the ``spectral'' inversion. To overcome this problem, we performed full nonlinear inversions of synthetically generated and real datasets, and compare the obtained spectra with the input and tomographic models respectively. The inversions are calculated on a distributed memory parallel nodes cluster, employing the MPI package. An evolutionary strategy approach is used to explore the parameter space, using the PIKAIA software. The first preliminary results show a resolving power higher than that of linearised inversion. This confirms that the approximations required in the linear formulation affect the solution quality, and suggests that the nonlinear approach might effectively help to constrain the heterogeneity spectrum more robustly than currently possible.
A novel approach for accurate identification of splice junctions based on hybrid algorithms.
Mandal, Indrajit
2015-01-01
The precise prediction of splice junctions as 'exon-intron' or 'intron-exon' boundaries in a given DNA sequence is an important task in Bioinformatics. The main challenge is to determine the splice sites in the coding region. Due to the intrinsic complexity and the uncertainty in gene sequence, the adoption of data mining methods is increasingly becoming popular. There are various methods developed on different strategies in this direction. This article focuses on the construction of new hybrid machine learning ensembles that solve the splice junction task more effectively. A novel supervised feature reduction technique is developed using entropy-based fuzzy rough set theory optimized by greedy hill-climbing algorithm. The average prediction accuracy achieved is above 98% with 95% confidence interval. The performance of the proposed methods is evaluated using various metrics to establish the statistical significance of the results. The experiments are conducted using various schemes with human DNA sequence data. The obtained results are highly promising as compared with the state-of-the-art approaches in literature. PMID:25203504
Sezgin, Billur; Ayhan, Suhan; Tuncer, Serhan; Sencan, Ayse; Aral, Mubin
2012-10-01
Despite appropriate surgical technique and follow-up, flap failures can be encountered for which no valid reason is evident. Current literature states that these unpredictable flap failures can be caused by unknown patient factors, such as undiagnosed hypercoagulability. Our approach and experience utilizing an algorithm to minimize unpredictable failures in microvascular breast reconstruction by predetermining hypercoagulation risk factors in preoperative patients is presented. A prospective assessment of microsurgical breast reconstruction candidates between October 2007 and December 2010 was conducted. Patients were questioned about their tendency toward hypercoagulation. A thrombophilia panel was requested for patients confirming any risk factors. Appropriate surgical planning was conducted according to results of the panel. Of the 60 patients thoroughly questioned about hypercoagulation tendency, 21 (35%) confirmed having prothrombotic tendency and were referred to the thrombophilia testing. The results indicated hypercoagulation in 9 (15%) patients. The primary reconstruction plan of utilizing free flaps was abandoned for these patients and pedicled flaps or implants were preferred for reconstruction. These percentages emphasize the value of questioning risk factors and testing for hypercoagulation in patients seeking microsurgical breast reconstruction. We believe that detailed preoperative questioning of risk factors and appropriate testing according to prothrombotic tendency is beneficial in minimizing unpredictable flap failures and increasing rates of success. PMID:22744893
A multi-layer cellular automata approach for algorithmic generation of virtual case studies: VIBe.
Sitzenfrei, R; Fach, S; Kinzel, H; Rauch, W
2010-01-01
Analyses of case studies are used to evaluate new or existing technologies, measures or strategies with regard to their impact on the overall process. However, data availability is limited and hence, new technologies, measures or strategies can only be tested on a limited number of case studies. Owing to the specific boundary conditions and system properties of each single case study, results can hardly be generalized or transferred to other boundary conditions. virtual infrastructure benchmarking (VIBe) is a software tool which algorithmically generates virtual case studies (VCSs) for urban water systems. System descriptions needed for evaluation are extracted from VIBe whose parameters are based on real world case studies and literature. As a result VIBe writes Input files for water simulation software as EPANET and EPA SWMM. With such input files numerous simulations can be performed and the results can be benchmarked and analysed stochastically at a city scale. In this work the approach of VIBe is applied with parameters according to a section of the Inn valley and therewith 1,000 VCSs are generated and evaluated. A comparison of the VCSs with data of real world case studies shows that the real world case studies fit within the parameter ranges of the VCSs. Consequently, VIBe tackles the problem of limited availability of case study data. PMID:20057089
Self-reconfigurable approach for computation-intensive motion estimation algorithm in H.264/AVC
NASA Astrophysics Data System (ADS)
Lee, Jooheung; Ryu, Chul; Kim, Soontae
2012-04-01
The authors propose a self-reconfigurable approach to perform H.264/AVC variable block size motion estimation computation on field-programmable gate arrays. We use dynamic partial reconfiguration to change the hardware architecture of motion estimation during run-time. Hardware adaptation to meet the real-time computing requirements for the given video resolutions and frame rates is performed through self-reconfiguration. An embedded processor is used to control the reconfiguration of partial bitstreams of motion estimation adaptively. The partial bitstreams for different motion estimation computation arrays are compressed using LZSS algorithm. On-chip BlockRAM is used as a cache to pre-store the partial bitstreams so that run-time reconfiguration can be fully utilized. We designed a hardware module to fetch the pre-stored partial bitstream from BlockRAM to an internal configuration access port. Comparison results show that our motion estimation architecture improves toward data reuse, and the memory bandwidth overhead is reduced. Using our self-reconfigurable platform, the reconfiguration overhead can be removed and 367 MB/sec reconfiguration rate can be achieved. The experimental results show that the external memory accesses are reduced by 62.4% and it can operate at a frequency of 91.7 MHz.
NASA Astrophysics Data System (ADS)
Alexandre, E.; Cuadra, L.; Nieto-Borge, J. C.; Candil-García, G.; del Pino, M.; Salcedo-Sanz, S.
2015-08-01
Wave parameters computed from time series measured by buoys (significant wave height Hs, mean wave period, etc.) play a key role in coastal engineering and in the design and operation of wave energy converters. Storms or navigation accidents can make measuring buoys break down, leading to missing data gaps. In this paper we tackle the problem of locally reconstructing Hs at out-of-operation buoys by using wave parameters from nearby buoys, based on the spatial correlation among values at neighboring buoy locations. The novelty of our approach for its potential application to problems in coastal engineering is twofold. On one hand, we propose a genetic algorithm hybridized with an extreme learning machine that selects, among the available wave parameters from the nearby buoys, a subset FnSP with nSP parameters that minimizes the Hs reconstruction error. On the other hand, we evaluate to what extent the selected parameters in subset FnSP are good enough in assisting other machine learning (ML) regressors (extreme learning machines, support vector machines and gaussian process regression) to reconstruct Hs. The results show that all the ML method explored achieve a good Hs reconstruction in the two different locations studied (Caribbean Sea and West Atlantic).
Garbuzov, D.Z.; Martinelli, R.U.; Khalfin, V.; Lee, H.; Morris, N.A.; Taylor, G.C.; Connolly, J.C.; Charache, G.W.; DePoy, D.M.
1997-10-01
Heterojunction n-Al{sub 0.25}Ga{sub 0.75}As{sub 0.02}Sb{sub 098}/p-In{sub 0.16}Ga{sub 0.84}As{sub 0.04}Sb{sub 0.96} thermophotovoltaic (TPV) cells were grown by molecular-beam epitaxy on n-GaSb-substrates. In the spectral range from 1 {micro}m to 2.1 {micro}m these cells, as well as homojunction n-p-In{sub 0.16}Ga{sub 0.84}As{sub 0.04}Sb{sub 0.96} cells, have demonstrated internal quantum efficiencies exceeding 80%, despite about a 200 meV barrier in the conduction band at the heterointerface. Estimation shows that the thermal emission of the electrons photogenerated in p-region over this barrier can provide high efficiency for hetero-cells if the electron recombination time in p-In{sub 0.16}Ga{sub 0.84}As{sub 0.04}Sb{sub 0.96}is longer than 10 ns. Keeping the same internal efficiency as homojunction cells, hetero-cells provide a unique opportunity to decrease the dark forward current and thereby increase open circuit voltage (V{sub {proportional_to}}) and fill factor at a given illumination level. It is shown that the decrease of the forward current in hetero-cells is due to the lower recombination rate in n-type wider-bandgap space-charge region and to the suppression of the hole component of the forward current. The improvement in V{sub {proportional_to}} reaches 100% at illumination level equivalent to 1 mA/cm{sup 2} and it decreases to 5% at the highest illumination levels (2--3 A/cm{sup 2}), where the electron current component dominates in both the homo- and heterojunction cells. Values of V{sub {proportional_to}} as high as 310 meV have been obtained for a hetero-cell at illumination levels of 3 A/cm{sup 2}. Under this condition, the expected fill factor value is about 72% for a hetero-cell with improved series resistance. The heterojunction concept provides excellent prospects for further reduction of the dark forward current in TPV cells.
Prediction of Heart Attack Risk Using GA-ANFIS Expert System Prototype.
Begic Fazlic, Lejla; Avdagic, Aja; Besic, Ingmar
2015-01-01
The aim of this research is to develop a novel GA-ANFIS expert system prototype for classifying heart disease degree of a patient by using heart diseases attributes (features) and diagnoses taken in the real conditions. Thirteen attributes have been used as inputs to classifiers being based on Adaptive Neuro-Fuzzy Inference Systems (ANFIS) for the first level of fuzzy model optimization. They are used as inputs in Genetic Algorithm (GA) for the second level of fuzzy model optimization within GA-ANFIS system. GA-ANFIS system performs optimization in two steps. Modelling and validating of the novel GA-ANFIS system approach is performed in MATLAB environment. We compared GA-ANFIS and ANFIS results. The proposed GA-ANFIS model with the predicted value technique is more efficient when diagnosis of heart disease is concerned, as well the earlier method we got by ANFIS model. PMID:25980885
New Approach for IIR Adaptive Lattice Filter Structure Using Simultaneous Perturbation Algorithm
NASA Astrophysics Data System (ADS)
Martinez, Jorge Ivan Medina; Nakano, Kazushi; Higuchi, Kohji
Adaptive infinite impulse response (IIR), or recursive, filters are less attractive mainly because of the stability and the difficulties associated with their adaptive algorithms. Therefore, in this paper the adaptive IIR lattice filters are studied in order to devise algorithms that preserve the stability of the corresponding direct-form schemes. We analyze the local properties of stationary points, a transformation achieving this goal is suggested, which gives algorithms that can be efficiently implemented. Application to the Steiglitz-McBride (SM) and Simple Hyperstable Adaptive Recursive Filter (SHARF) algorithms is presented. Also a modified version of Simultaneous Perturbation Stochastic Approximation (SPSA) is presented in order to get the coefficients in a lattice form more efficiently and with a lower computational cost and complexity. The results are compared with previous lattice versions of these algorithms. These previous lattice versions may fail to preserve the stability of stationary points.
NASA Astrophysics Data System (ADS)
Zhang, Jingzhao; Zhang, Yiou; Tse, Kinfai; Deng, Bei; Xu, Hu; Zhu, Junyi
2016-05-01
The accurate absolute surface energies of (0001)/(000 1 ¯ ) surfaces of wurtzite structures are crucial in determining the thin film growth mode of important energy materials. However, the surface energies still remain to be solved due to the intrinsic difficulty of calculating the dangling bond energy of asymmetrically bonded surface atoms. In this study, we used a pseudo-hydrogen passivation method to estimate the dangling bond energy and calculate the polar surfaces of ZnO and GaN. The calculations were based on the pseudo chemical potentials obtained from a set of tetrahedral clusters or simple pseudo-molecules, using density functional theory approaches. The surface energies of (0001)/(000 1 ¯ ) surfaces of wurtzite ZnO and GaN that we obtained showed relatively high self-consistencies. A wedge structure calculation with a new bottom surface passivation scheme of group-I and group-VII elements was also proposed and performed to show converged absolute surface energy of wurtzite ZnO polar surfaces, and these results were also compared with the above method. The calculated results generally show that the surface energies of GaN are higher than those of ZnO, suggesting that ZnO tends to wet the GaN substrate, while GaN is unlikely to wet ZnO. Therefore, it will be challenging to grow high quality GaN thin films on ZnO substrates; however, high quality ZnO thin film on GaN substrate would be possible. These calculations and comparisons may provide important insights into crystal growth of the above materials, thereby leading to significant performance enhancements in semiconductor devices.
An Algorithmic Framework for Multiobjective Optimization
Ganesan, T.; Elamvazuthi, I.; Shaari, Ku Zilati Ku; Vasant, P.
2013-01-01
Multiobjective (MO) optimization is an emerging field which is increasingly being encountered in many fields globally. Various metaheuristic techniques such as differential evolution (DE), genetic algorithm (GA), gravitational search algorithm (GSA), and particle swarm optimization (PSO) have been used in conjunction with scalarization techniques such as weighted sum approach and the normal-boundary intersection (NBI) method to solve MO problems. Nevertheless, many challenges still arise especially when dealing with problems with multiple objectives (especially in cases more than two). In addition, problems with extensive computational overhead emerge when dealing with hybrid algorithms. This paper discusses these issues by proposing an alternative framework that utilizes algorithmic concepts related to the problem structure for generating efficient and effective algorithms. This paper proposes a framework to generate new high-performance algorithms with minimal computational overhead for MO optimization. PMID:24470795
An algorithmic framework for multiobjective optimization.
Ganesan, T; Elamvazuthi, I; Shaari, Ku Zilati Ku; Vasant, P
2013-01-01
Multiobjective (MO) optimization is an emerging field which is increasingly being encountered in many fields globally. Various metaheuristic techniques such as differential evolution (DE), genetic algorithm (GA), gravitational search algorithm (GSA), and particle swarm optimization (PSO) have been used in conjunction with scalarization techniques such as weighted sum approach and the normal-boundary intersection (NBI) method to solve MO problems. Nevertheless, many challenges still arise especially when dealing with problems with multiple objectives (especially in cases more than two). In addition, problems with extensive computational overhead emerge when dealing with hybrid algorithms. This paper discusses these issues by proposing an alternative framework that utilizes algorithmic concepts related to the problem structure for generating efficient and effective algorithms. This paper proposes a framework to generate new high-performance algorithms with minimal computational overhead for MO optimization. PMID:24470795
Dalzini, Annalisa; Bergamini, Christian; Biondi, Barbara; De Zotti, Marta; Panighel, Giacomo; Fato, Romana; Peggion, Cristina; Bortolus, Marco; Maniero, Anna Lisa
2016-01-01
Peptaibols are peculiar peptides produced by fungi as weapons against other microorganisms. Previous studies showed that peptaibols are promising peptide-based drugs because they act against cell membranes rather than a specific target, thus lowering the possibility of the onset of multi-drug resistance, and they possess non-coded α-amino acid residues that confer proteolytic resistance. Trichogin GA IV (TG) is a short peptaibol displaying antimicrobial and cytotoxic activity. In the present work, we studied thirteen TG analogues, adopting a multidisciplinary approach. We showed that the cytotoxicity is tuneable by single amino-acids substitutions. Many analogues maintain the same level of non-selective cytotoxicity of TG and three analogues are completely non-toxic. Two promising lead compounds, characterized by the introduction of a positively charged unnatural amino-acid in the hydrophobic face of the helix, selectively kill T67 cancer cells without affecting healthy cells. To explain the determinants of the cytotoxicity, we investigated the structural parameters of the peptides, their cell-binding properties, cell localization, and dynamics in the membrane, as well as the cell membrane composition. We show that, while cytotoxicity is governed by the fine balance between the amphipathicity and hydrophobicity, the selectivity depends also on the expression of negatively charged phospholipids on the cell surface. PMID:27039838
Dalzini, Annalisa; Bergamini, Christian; Biondi, Barbara; De Zotti, Marta; Panighel, Giacomo; Fato, Romana; Peggion, Cristina; Bortolus, Marco; Maniero, Anna Lisa
2016-01-01
Peptaibols are peculiar peptides produced by fungi as weapons against other microorganisms. Previous studies showed that peptaibols are promising peptide-based drugs because they act against cell membranes rather than a specific target, thus lowering the possibility of the onset of multi-drug resistance, and they possess non-coded α-amino acid residues that confer proteolytic resistance. Trichogin GA IV (TG) is a short peptaibol displaying antimicrobial and cytotoxic activity. In the present work, we studied thirteen TG analogues, adopting a multidisciplinary approach. We showed that the cytotoxicity is tuneable by single amino-acids substitutions. Many analogues maintain the same level of non-selective cytotoxicity of TG and three analogues are completely non-toxic. Two promising lead compounds, characterized by the introduction of a positively charged unnatural amino-acid in the hydrophobic face of the helix, selectively kill T67 cancer cells without affecting healthy cells. To explain the determinants of the cytotoxicity, we investigated the structural parameters of the peptides, their cell-binding properties, cell localization, and dynamics in the membrane, as well as the cell membrane composition. We show that, while cytotoxicity is governed by the fine balance between the amphipathicity and hydrophobicity, the selectivity depends also on the expression of negatively charged phospholipids on the cell surface. PMID:27039838
NASA Astrophysics Data System (ADS)
Yan, H.; Zheng, M. J.; Zhu, D. Y.; Wang, H. T.; Chang, W. S.
2015-07-01
When using clutter suppression interferometry (CSI) algorithm to perform signal processing in a three-channel wide-area surveillance radar system, the primary concern is to effectively suppress the ground clutter. However, a portion of moving target's energy is also lost in the process of channel cancellation, which is often neglected in conventional applications. In this paper, we firstly investigate the two-dimensional (radial velocity dimension and squint angle dimension) residual amplitude of moving targets after channel cancellation with CSI algorithm. Then, a new approach is proposed to increase the two-dimensional detection probability of moving targets by reserving the maximum value of the three channel cancellation results in non-uniformly spaced channel system. Besides, theoretical expression of the false alarm probability with the proposed approach is derived in the paper. Compared with the conventional approaches in uniformly spaced channel system, simulation results validate the effectiveness of the proposed approach. To our knowledge, it is the first time that the two-dimensional detection probability of CSI algorithm is studied.
2012-01-01
RA is a syndrome consisting of different pathogenetic subsets in which distinct molecular mechanisms may drive common final pathways. Recent work has provided proof of principle that biomarkers may be identified predictive of the response to targeted therapy. Based on new insights, an initial treatment algorithm is presented that may be used to guide treatment decisions in patients who have failed one TNF inhibitor. Key questions in this algorithm relate to the question whether the patient is a primary vs a secondary non-responder to TNF blockade and whether the patient is RF and/or anti-citrullinated peptide antibody positive. This preliminary algorithm may contribute to more cost-effective treatment of RA, and provides the basis for more extensive algorithms when additional data become available. PMID:21890615
Asgari, Mohammad; Soltani, Nasim Yahya; Riahi, Ali
2010-01-01
There are varieties of wideband direction-of-arrival (DOA) estimation algorithms. Their structure comprises a number of narrowband ones, each performs in one frequency in a given bandwidth, and then different responses should be combined in a proper way to yield true DOAs. Hence, wideband algorithms are always complex and so non-real-time. This paper investigates a method to derive a flat response of narrowband multiple signal classification (MUSIC) [R. O. Schmidt, IEEE Trans. Antennas Propag., 34, 276-280 (1986)] algorithm in the whole frequencies of given band. Therefore, required conditions of applying narrowband algorithm on wideband impinging signals will be given through a concrete analysis. It could be found out that array sensor locations are able to compensate the frequency variations to reach a flat response of DOAs in a specified wideband frequency. PMID:20058975
Ahirwal, M K; Kumar, Anil; Singh, G K
2013-01-01
This paper explores the migration of adaptive filtering with swarm intelligence/evolutionary techniques employed in the field of electroencephalogram/event-related potential noise cancellation or extraction. A new approach is proposed in the form of controlled search space to stabilize the randomness of swarm intelligence techniques especially for the EEG signal. Swarm-based algorithms such as Particles Swarm Optimization, Artificial Bee Colony, and Cuckoo Optimization Algorithm with their variants are implemented to design optimized adaptive noise canceler. The proposed controlled search space technique is tested on each of the swarm intelligence techniques and is found to be more accurate and powerful. Adaptive noise canceler with traditional algorithms such as least-mean-square, normalized least-mean-square, and recursive least-mean-square algorithms are also implemented to compare the results. ERP signals such as simulated visual evoked potential, real visual evoked potential, and real sensorimotor evoked potential are used, due to their physiological importance in various EEG studies. Average computational time and shape measures of evolutionary techniques are observed 8.21E-01 sec and 1.73E-01, respectively. Though, traditional algorithms take negligible time consumption, but are unable to offer good shape preservation of ERP, noticed as average computational time and shape measure difference, 1.41E-02 sec and 2.60E+00, respectively. PMID:24407307
A new approach to optic disc detection in human retinal images using the firefly algorithm.
Rahebi, Javad; Hardalaç, Fırat
2016-03-01
There are various methods and algorithms to detect the optic discs in retinal images. In recent years, much attention has been given to the utilization of the intelligent algorithms. In this paper, we present a new automated method of optic disc detection in human retinal images using the firefly algorithm. The firefly intelligent algorithm is an emerging intelligent algorithm that was inspired by the social behavior of fireflies. The population in this algorithm includes the fireflies, each of which has a specific rate of lighting or fitness. In this method, the insects are compared two by two, and the less attractive insects can be observed to move toward the more attractive insects. Finally, one of the insects is selected as the most attractive, and this insect presents the optimum response to the problem in question. Here, we used the light intensity of the pixels of the retinal image pixels instead of firefly lightings. The movement of these insects due to local fluctuations produces different light intensity values in the images. Because the optic disc is the brightest area in the retinal images, all of the insects move toward brightest area and thus specify the location of the optic disc in the image. The results of implementation show that proposed algorithm could acquire an accuracy rate of 100 % in DRIVE dataset, 95 % in STARE dataset, and 94.38 % in DiaRetDB1 dataset. The results of implementation reveal high capability and accuracy of proposed algorithm in the detection of the optic disc from retinal images. Also, recorded required time for the detection of the optic disc in these images is 2.13 s for DRIVE dataset, 2.81 s for STARE dataset, and 3.52 s for DiaRetDB1 dataset accordingly. These time values are average value. PMID:26093773
Symbolic integration of a class of algebraic functions. [by an algorithmic approach
NASA Technical Reports Server (NTRS)
Ng, E. W.
1974-01-01
An algorithm is presented for the symbolic integration of a class of algebraic functions. This class consists of functions made up of rational expressions of an integration variable x and square roots of polynomials, trigonometric and hyperbolic functions of x. The algorithm is shown to consist of the following components:(1) the reduction of input integrands to conical form; (2) intermediate internal representations of integrals; (3) classification of outputs; and (4) reduction and simplification of outputs to well-known functions.
NASA Astrophysics Data System (ADS)
Lau, Erin-Ee-Lin; Chung, Wan-Young
A novel RSSI (Received Signal Strength Indication) refinement algorithm is proposed to enhance the resolution for indoor and outdoor real-time location tracking system. The proposed refinement algorithm is implemented in two separate phases. During the first phase, called the pre-processing step, RSSI values at different static locations are collected and processed to build a calibrated model for each reference node. Different measurement campaigns pertinent to each parameter in the model are implemented to analyze the sensitivity of RSSI. The propagation models constructed for each reference nodes are needed by the second phase. During the next phase, called the runtime process, real-time tracking is performed. Smoothing algorithm is proposed to minimize the dynamic fluctuation of radio signal received from each reference node when the mobile target is moving. Filtered RSSI values are converted to distances using formula calibrated in the first phase. Finally, an iterative trilateration algorithm is used for position estimation. Experiments relevant to the optimization algorithm are carried out in both indoor and outdoor environments and the results validated the feasibility of proposed algorithm in reducing the dynamic fluctuation for more accurate position estimation.
NASA Astrophysics Data System (ADS)
Dalzell, B. J.; Gassman, P. W.; Kling, C.
2015-12-01
In the Minnesota River Basin, sediments originating from failing stream banks and bluffs account for the majority of the riverine load and contribute to water quality impairments in the Minnesota River as well as portions of the Mississippi River upstream of Lake Pepin. One approach for mitigating this problem may be targeted wetland restoration in Minnesota River Basin tributaries in order to reduce the magnitude and duration of peak flow events which contribute to bluff and stream bank failures. In order to determine effective arrangements and properties of wetlands to achieve peak flow reduction, we are employing a genetic algorithm approach coupled with a SWAT model of the Cottonwood River, a tributary of the Minnesota River. The genetic algorithm approach will evaluate combinations of basic wetland features as represented by SWAT: surface area, volume, contributing area, and hydraulic conductivity of the wetland bottom. These wetland parameters will be weighed against economic considerations associated with land use trade-offs in this agriculturally productive landscape. Preliminary results show that the SWAT model is capable of simulating daily hydrology very well and genetic algorithm evaluation of wetland scenarios is ongoing. Anticipated results will include (1) combinations of wetland parameters that are most effective for reducing peak flows, and (2) evaluation of economic trade-offs between wetland restoration, water quality, and agricultural productivity in the Cottonwood River watershed.
NASA Technical Reports Server (NTRS)
Idris, Husni; Vivona, Robert A.; Al-Wakil, Tarek
2009-01-01
This document describes exploratory research on a distributed, trajectory oriented approach for traffic complexity management. The approach is to manage traffic complexity based on preserving trajectory flexibility and minimizing constraints. In particular, the document presents metrics for trajectory flexibility; a method for estimating these metrics based on discrete time and degree of freedom assumptions; a planning algorithm using these metrics to preserve flexibility; and preliminary experiments testing the impact of preserving trajectory flexibility on traffic complexity. The document also describes an early demonstration capability of the trajectory flexibility preservation function in the NASA Autonomous Operations Planner (AOP) platform.
Results with an Algorithmic Approach to Hybrid Repair of the Aortic Arch
Andersen, Nicholas D.; Williams, Judson B.; Hanna, Jennifer M.; Shah, Asad A.; McCann, Richard L.; Hughes, G. Chad
2013-01-01
Objective Hybrid repair of the transverse aortic arch may allow for aortic arch repair with reduced morbidity in patients who are suboptimal candidates for conventional open surgery. Here, we present our results with an algorithmic approach to hybrid arch repair, based upon the extent of aortic disease and patient comorbidities. Methods Between August 2005 and January 2012, 87 patients underwent hybrid arch repair by three principal procedures: zone 1 endograft coverage with extra-anatomic left carotid revascularization (zone 1, n=19), zone 0 endograft coverage with aortic arch debranching (zone 0, n=48), or total arch replacement with staged stented elephant trunk completion (stented elephant trunk, n=20). Results The mean patient age was 64 years and the mean expected in-hospital mortality rate was 16.3% as calculated by the EuroSCORE II. 22% (n=19) of operations were non-elective. Sternotomy, cardiopulmonary bypass, and deep hypothermic circulatory arrest were required in 78% (n=68), 45% (n=39), and 31% (n=27) of patients, respectively, to allow for total arch replacement, arch debranching, or other concomitant cardiac procedures, including ascending ± hemi-arch replacement in 17% (n=8) of patients undergoing zone 0 repair. All stented elephant trunk procedures (n=20) and 19% (n=9) of zone 0 procedures were staged, with 41% (n=12) of patients undergoing staged repair during a single hospitalization. The 30-day/in-hospital rates of stroke and permanent paraplegia/paraparesis were 4.6% (n=4) and 1.2% (n=1), respectively. Three of 27 (11.1%) patients with native ascending aorta zone 0 proximal landing zone experienced retrograde type A dissection following endograft placement. The overall in-hospital mortality rate was 5.7% (n=5), however, 30-day/in-hospital mortality increased to 14.9% (n=13) due to eight 30-day out-of-hospital deaths. Native ascending aorta zone 0 endograft placement was found to be the only univariate predictor of 30-day/in-hospital mortality
NASA Technical Reports Server (NTRS)
Hu, Chuanmin; Lee, Zhongping; Franz, Bryan
2011-01-01
A new empirical algorithm is proposed to estimate surface chlorophyll-a concentrations (Chl) in the global ocean for Chl less than or equal to 0.25 milligrams per cubic meters (approximately 77% of the global ocean area). The algorithm is based on a color index (CI), defined as the difference between remote sensing reflectance (R(sub rs), sr(sup -1) in the green and a reference formed linearly between R(sub rs) in the blue and red. For low Chl waters, in situ data showed a tighter (and therefore better) relationship between CI and Chl than between traditional band-ratios and Chl, which was further validated using global data collected concurrently by ship-borne and SeaWiFS satellite instruments. Model simulations showed that for low Chl waters, compared with the band-ratio algorithm, the CI-based algorithm (CIA) was more tolerant to changes in chlorophyll-specific backscattering coefficient, and performed similarly for different relative contributions of non-phytoplankton absorption. Simulations using existing atmospheric correction approaches further demonstrated that the CIA was much less sensitive than band-ratio algorithms to various errors induced by instrument noise and imperfect atmospheric correction (including sun glint and whitecap corrections). Image and time-series analyses of SeaWiFS and MODIS/Aqua data also showed improved performance in terms of reduced image noise, more coherent spatial and temporal patterns, and consistency between the two sensors. The reduction in noise and other errors is particularly useful to improve the detection of various ocean features such as eddies. Preliminary tests over MERIS and CZCS data indicate that the new approach should be generally applicable to all existing and future ocean color instruments.
Scheduling Earth Observing Fleets Using Evolutionary Algorithms: Problem Description and Approach
NASA Technical Reports Server (NTRS)
Globus, Al; Crawford, James; Lohn, Jason; Morris, Robert; Clancy, Daniel (Technical Monitor)
2002-01-01
We describe work in progress concerning multi-instrument, multi-satellite scheduling. Most, although not all, Earth observing instruments currently in orbit are unique. In the relatively near future, however, we expect to see fleets of Earth observing spacecraft, many carrying nearly identical instruments. This presents a substantially new scheduling challenge. Inspired by successful commercial applications of evolutionary algorithms in scheduling domains, this paper presents work in progress regarding the use of evolutionary algorithms to solve a set of Earth observing related model problems. Both the model problems and the software are described. Since the larger problems will require substantial computation and evolutionary algorithms are embarrassingly parallel, we discuss our parallelization techniques using dedicated and cycle-scavenged workstations.
The M-OLAP Cube Selection Problem: A Hyper-polymorphic Algorithm Approach
NASA Astrophysics Data System (ADS)
Loureiro, Jorge; Belo, Orlando
OLAP systems depend heavily on the materialization of multidimensional structures to speed-up queries, whose appropriate selection constitutes the cube selection problem. However, the recently proposed distribution of OLAP structures emerges to answer new globalization's requirements, capturing the known advantages of distributed databases. But this hardens the search for solutions, especially due to the inherent heterogeneity, imposing an extra characteristic of the algorithm that must be used: adaptability. Here the emerging concept known as hyper-heuristic can be a solution. In fact, having an algorithm where several (meta-)heuristics may be selected under the control of a heuristic has an intrinsic adaptive behavior. This paper presents a hyper-heuristic polymorphic algorithm used to solve the extended cube selection and allocation problem generated in M-OLAP architectures.
Gokhale, Nikhil S
2016-01-01
Vernal keratoconjunctivitis is an ocular allergy that is common in the pediatric age group. It is often chronic, severe, and nonresponsive to the available treatment options. Management of these children is difficult and often a dilemma for the practitioner. There is a need to simplify and standardize its management. To achieve this goal, we require a grading system to judge the severity of inflammation and an algorithm to select the appropriate medications. This article provides a simple and practically useful grading system and a stepladder algorithm for systematic treatment of these patients. Use of appropriate treatment modalities can reduce treatment and disease-related complications. PMID:27050351
Genetic algorithm optimization of atomic clusters
Morris, J.R.; Deaven, D.M.; Ho, K.M.; Wang, C.Z.; Pan, B.C.; Wacker, J.G.; Turner, D.E. |
1996-12-31
The authors have been using genetic algorithms to study the structures of atomic clusters and related problems. This is a problem where local minima are easy to locate, but barriers between the many minima are large, and the number of minima prohibit a systematic search. They use a novel mating algorithm that preserves some of the geometrical relationship between atoms, in order to ensure that the resultant structures are likely to inherit the best features of the parent clusters. Using this approach, they have been able to find lower energy structures than had been previously obtained. Most recently, they have been able to turn around the building block idea, using optimized structures from the GA to learn about systematic structural trends. They believe that an effective GA can help provide such heuristic information, and (conversely) that such information can be introduced back into the algorithm to assist in the search process.
NASA Astrophysics Data System (ADS)
Meyer, Ulrich; Negoescu, Andrei; Weichert, Volker
Despite disillusioning worst-case behavior, classic algorithms for single-source shortest-paths (SSSP) like Bellman-Ford are still being used in practice, especially due to their simple data structures. However, surprisingly little is known about the average-case complexity of these approaches. We provide new theoretical and experimental results for the performance of classic label-correcting SSSP algorithms on graph classes with non-negative random edge weights. In particular, we prove a tight lower bound of Ω(n 2) for the running times of Bellman-Ford on a class of sparse graphs with O(n) nodes and edges; the best previous bound was Ω(n 4/3 - ɛ ). The same improvements are shown for Pallottino's algorithm. We also lift a lower bound for the approximate bucket implementation of Dijkstra's algorithm from Ω(n logn / loglogn) to Ω(n 1.2 - ɛ ). Furthermore, we provide an experimental evaluation of our new graph classes in comparison with previously used test inputs.
Premaladha, J; Ravichandran, K S
2016-04-01
Dermoscopy is a technique used to capture the images of skin, and these images are useful to analyze the different types of skin diseases. Malignant melanoma is a kind of skin cancer whose severity even leads to death. Earlier detection of melanoma prevents death and the clinicians can treat the patients to increase the chances of survival. Only few machine learning algorithms are developed to detect the melanoma using its features. This paper proposes a Computer Aided Diagnosis (CAD) system which equips efficient algorithms to classify and predict the melanoma. Enhancement of the images are done using Contrast Limited Adaptive Histogram Equalization technique (CLAHE) and median filter. A new segmentation algorithm called Normalized Otsu's Segmentation (NOS) is implemented to segment the affected skin lesion from the normal skin, which overcomes the problem of variable illumination. Fifteen features are derived and extracted from the segmented images are fed into the proposed classification techniques like Deep Learning based Neural Networks and Hybrid Adaboost-Support Vector Machine (SVM) algorithms. The proposed system is tested and validated with nearly 992 images (malignant & benign lesions) and it provides a high classification accuracy of 93 %. The proposed CAD system can assist the dermatologists to confirm the decision of the diagnosis and to avoid excisional biopsies. PMID:26872778
Wang, Shuaiqun; Aorigele; Kong, Wei; Zeng, Weiming; Hong, Xiaomin
2016-01-01
Gene expression data composed of thousands of genes play an important role in classification platforms and disease diagnosis. Hence, it is vital to select a small subset of salient features over a large number of gene expression data. Lately, many researchers devote themselves to feature selection using diverse computational intelligence methods. However, in the progress of selecting informative genes, many computational methods face difficulties in selecting small subsets for cancer classification due to the huge number of genes (high dimension) compared to the small number of samples, noisy genes, and irrelevant genes. In this paper, we propose a new hybrid algorithm HICATS incorporating imperialist competition algorithm (ICA) which performs global search and tabu search (TS) that conducts fine-tuned search. In order to verify the performance of the proposed algorithm HICATS, we have tested it on 10 well-known benchmark gene expression classification datasets with dimensions varying from 2308 to 12600. The performance of our proposed method proved to be superior to other related works including the conventional version of binary optimization algorithm in terms of classification accuracy and the number of selected genes. PMID:27579323
A Low-Tech, Hands-On Approach To Teaching Sorting Algorithms to Working Students.
ERIC Educational Resources Information Center
Dios, R.; Geller, J.
1998-01-01
Focuses on identifying the educational effects of "activity oriented" instructional techniques. Examines which instructional methods produce enhanced learning and comprehension. Discusses the problem of learning "sorting algorithms," a major topic in every Computer Science curriculum. Presents a low-tech, hands-on teaching method for sorting…
Point process algorithm: a new Bayesian approach for TPF-I planet signal extraction
NASA Technical Reports Server (NTRS)
Velusamy, T.; Marsh, K. A.; Ware, B.
2005-01-01
TPF-I capability for planetary signal extraction, including both detection and spectral characterization, can be optimized by taking proper account of instrumental characteristics and astrophysical prior information. We have developed the Point Process Algorithm, a Bayesian technique for estracting planetary signals using the sine/cosine chopped outputs of a dual nulling interferometer.
Aorigele; Zeng, Weiming; Hong, Xiaomin
2016-01-01
Gene expression data composed of thousands of genes play an important role in classification platforms and disease diagnosis. Hence, it is vital to select a small subset of salient features over a large number of gene expression data. Lately, many researchers devote themselves to feature selection using diverse computational intelligence methods. However, in the progress of selecting informative genes, many computational methods face difficulties in selecting small subsets for cancer classification due to the huge number of genes (high dimension) compared to the small number of samples, noisy genes, and irrelevant genes. In this paper, we propose a new hybrid algorithm HICATS incorporating imperialist competition algorithm (ICA) which performs global search and tabu search (TS) that conducts fine-tuned search. In order to verify the performance of the proposed algorithm HICATS, we have tested it on 10 well-known benchmark gene expression classification datasets with dimensions varying from 2308 to 12600. The performance of our proposed method proved to be superior to other related works including the conventional version of binary optimization algorithm in terms of classification accuracy and the number of selected genes. PMID:27579323
Application of genetic algorithms to tuning fuzzy control systems
NASA Technical Reports Server (NTRS)
Espy, Todd; Vombrack, Endre; Aldridge, Jack
1993-01-01
Real number genetic algorithms (GA) were applied for tuning fuzzy membership functions of three controller applications. The first application is our 'Fuzzy Pong' demonstration, a controller that controls a very responsive system. The performance of the automatically tuned membership functions exceeded that of manually tuned membership functions both when the algorithm started with randomly generated functions and with the best manually-tuned functions. The second GA tunes input membership functions to achieve a specified control surface. The third application is a practical one, a motor controller for a printed circuit manufacturing system. The GA alters the positions and overlaps of the membership functions to accomplish the tuning. The applications, the real number GA approach, the fitness function and population parameters, and the performance improvements achieved are discussed. Directions for further research in tuning input and output membership functions and in tuning fuzzy rules are described.
NASA Astrophysics Data System (ADS)
Abed, Azher M.; Abed, Issa Ahmed; Majdi, Hasan Sh.; Al-Shamani, Ali Najah; Sopian, K.
2016-02-01
This study proposes a new procedure for optimal design of shell and tube heat exchangers. The electromagnetism-like algorithm is applied to save on heat exchanger capital cost and designing a compact, high performance heat exchanger with effective use of the allowable pressure drop (cost of the pump). An optimization algorithm is then utilized to determine the optimal values of both geometric design parameters and maximum allowable pressure drop by pursuing the minimization of a total cost function. A computer code is developed for the optimal shell and tube heat exchangers. Different test cases are solved to demonstrate the effectiveness and ability of the proposed algorithm. Results are also compared with those obtained by other approaches available in the literature. The comparisons indicate that a proposed design procedure can be successfully applied in the optimal design of shell and tube heat exchangers. In particular, in the examined cases a reduction of total costs up to 30, 29, and 56.15 % compared with the original design and up to 18, 5.5 and 7.4 % compared with other approaches for case study 1, 2 and 3 respectively, are observed. In this work, economic optimization resulting from the proposed design procedure are relevant especially when the size/volume is critical for high performance and compact unit, moderate volume and cost are needed.
NASA Technical Reports Server (NTRS)
Fitzjerrell, D. G.; Grounds, D. J.; Leonard, J. I.
1975-01-01
Using a whole body algorithm simulation model, a wide variety and large number of stresses as well as different stress levels were simulated including environmental disturbances, metabolic changes, and special experimental situations. Simulation of short term stresses resulted in simultaneous and integrated responses from the cardiovascular, respiratory, and thermoregulatory subsystems and the accuracy of a large number of responding variables was verified. The capability of simulating significantly longer responses was demonstrated by validating a four week bed rest study. In this case, the long term subsystem model was found to reproduce many experimentally observed changes in circulatory dynamics, body fluid-electrolyte regulation, and renal function. The value of systems analysis and the selected design approach for developing a whole body algorithm was demonstrated.
A nondamaging electron microscopy approach to map In distribution in InGaN light-emitting diodes
NASA Astrophysics Data System (ADS)
Özdöl, V. B.; Koch, C. T.; van Aken, P. A.
2010-09-01
Dark-field inline electron holography and, for comparison, high-resolution transmission electron microscopy are used to investigate the distribution of indium in GaN-based commercial high-efficiency green light-emitting diodes consisting of InGaN multiquantum wells (QWs). Owing to the low electron doses used in inline holography measurements; this technique allows to map the indium distribution without introducing any noticeable electron beam-induced damage which is hardly avoidable in other quantitative transmission electron microscopy methods. Combining the large field of view with a spatial resolution better than 1 nm, we show that the InGaN QWs exhibit random alloy nature without any evidence of nanometer scale gross indium clustering in the whole active region.
Multi-objective optimization of lithium-ion battery model using genetic algorithm approach
NASA Astrophysics Data System (ADS)
Zhang, Liqiang; Wang, Lixin; Hinds, Gareth; Lyu, Chao; Zheng, Jun; Li, Junfu
2014-12-01
A multi-objective parameter identification method for modeling of Li-ion battery performance is presented. Terminal voltage and surface temperature curves at 15 °C and 30 °C are used as four identification objectives. The Pareto fronts of two types of Li-ion battery are obtained using the modified multi-objective genetic algorithm NSGA-II and the final identification results are selected using the multiple criteria decision making method TOPSIS. The simulated data using the final identification results are in good agreement with experimental data under a range of operating conditions. The validation results demonstrate that the modified NSGA-II and TOPSIS algorithms can be used as robust and reliable tools for identifying parameters of multi-physics models for many types of Li-ion batteries.
Ganzetti, Marco; Wenderoth, Nicole; Mantini, Dante
2016-01-01
Intensity non-uniformity (INU) in magnetic resonance (MR) imaging is a major issue when conducting analyses of brain structural properties. An inaccurate INU correction may result in qualitative and quantitative misinterpretations. Several INU correction methods exist, whose performance largely depend on the specific parameter settings that need to be chosen by the user. Here we addressed the question of how to select the best input parameters for a specific INU correction algorithm. Our investigation was based on the INU correction algorithm implemented in SPM, but this can be in principle extended to any other algorithm requiring the selection of input parameters. We conducted a comprehensive comparison of indirect metrics for the assessment of INU correction performance, namely the coefficient of variation of white matter (CVWM), the coefficient of variation of gray matter (CVGM), and the coefficient of joint variation between white matter and gray matter (CJV). Using simulated MR data, we observed the CJV to be more accurate than CVWM and CVGM, provided that the noise level in the INU-corrected image was controlled by means of spatial smoothing. Based on the CJV, we developed a data-driven approach for selecting INU correction parameters, which could effectively work on actual MR images. To this end, we implemented an enhanced procedure for the definition of white and gray matter masks, based on which the CJV was calculated. Our approach was validated using actual T1-weighted images collected with 1.5 T, 3 T, and 7 T MR scanners. We found that our procedure can reliably assist the selection of valid INU correction algorithm parameters, thereby contributing to an enhanced inhomogeneity correction in MR images. PMID:27014050
Ganzetti, Marco; Wenderoth, Nicole; Mantini, Dante
2016-01-01
Intensity non-uniformity (INU) in magnetic resonance (MR) imaging is a major issue when conducting analyses of brain structural properties. An inaccurate INU correction may result in qualitative and quantitative misinterpretations. Several INU correction methods exist, whose performance largely depend on the specific parameter settings that need to be chosen by the user. Here we addressed the question of how to select the best input parameters for a specific INU correction algorithm. Our investigation was based on the INU correction algorithm implemented in SPM, but this can be in principle extended to any other algorithm requiring the selection of input parameters. We conducted a comprehensive comparison of indirect metrics for the assessment of INU correction performance, namely the coefficient of variation of white matter (CVWM), the coefficient of variation of gray matter (CVGM), and the coefficient of joint variation between white matter and gray matter (CJV). Using simulated MR data, we observed the CJV to be more accurate than CVWM and CVGM, provided that the noise level in the INU-corrected image was controlled by means of spatial smoothing. Based on the CJV, we developed a data-driven approach for selecting INU correction parameters, which could effectively work on actual MR images. To this end, we implemented an enhanced procedure for the definition of white and gray matter masks, based on which the CJV was calculated. Our approach was validated using actual T1-weighted images collected with 1.5 T, 3 T, and 7 T MR scanners. We found that our procedure can reliably assist the selection of valid INU correction algorithm parameters, thereby contributing to an enhanced inhomogeneity correction in MR images. PMID:27014050
NASA Astrophysics Data System (ADS)
Riha, Stefan; Krawczyk, Harald
2011-11-01
Water quality monitoring in the Baltic Sea is of high ecological importance for all its neighbouring countries. They are highly interested in a regular monitoring of water quality parameters of their regional zones. A special attention is paid to the occurrence and dissemination of algae blooms. Among the appearing blooms the possibly toxicological or harmful cyanobacteria cultures are a special case of investigation, due to their specific optical properties and due to the negative influence on the ecological state of the aquatic system. Satellite remote sensing, with its high temporal and spatial resolution opportunities, allows the frequent observations of large areas of the Baltic Sea with special focus on its two seasonal algae blooms. For a better monitoring of the cyanobacteria dominated summer blooms, adapted algorithms are needed which take into account the special optical properties of blue-green algae. Chlorophyll-a standard algorithms typically fail in a correct recognition of these occurrences. To significantly improve the opportunities of observation and propagation of the cyanobacteria blooms, the Marine Remote Sensing group of DLR has started the development of a model based inversion algorithm that includes a four component bio-optical water model for Case2 waters, which extends the commonly calculated parameter set chlorophyll, Suspended Matter and CDOM with an additional parameter for the estimation of phycocyanin absorption. It was necessary to carry out detailed optical laboratory measurements with different cyanobacteria cultures, occurring in the Baltic Sea, for the generation of a specific bio-optical model. The inversion of satellite remote sensing data is based on an artificial Neural Network technique. This is a model based multivariate non-linear inversion approach. The specifically designed Neural Network is trained with a comprehensive dataset of simulated reflectance values taking into account the laboratory obtained specific optical
NASA Astrophysics Data System (ADS)
Li, Hong; Zhang, Li; Jiao, Yong-Chang
2016-07-01
This paper presents an interactive approach based on a discrete differential evolution algorithm to solve a class of integer bilevel programming problems, in which integer decision variables are controlled by an upper-level decision maker and real-value or continuous decision variables are controlled by a lower-level decision maker. Using the Karush--Kuhn-Tucker optimality conditions in the lower-level programming, the original discrete bilevel formulation can be converted into a discrete single-level nonlinear programming problem with the complementarity constraints, and then the smoothing technique is applied to deal with the complementarity constraints. Finally, a discrete single-level nonlinear programming problem is obtained, and solved by an interactive approach. In each iteration, for each given upper-level discrete variable, a system of nonlinear equations including the lower-level variables and Lagrange multipliers is solved first, and then a discrete nonlinear programming problem only with inequality constraints is handled by using a discrete differential evolution algorithm. Simulation results show the effectiveness of the proposed approach.
Taheri, Shahrooz; Mat Saman, Muhamad Zameri; Wong, Kuan Yew
2013-01-01
One of the cost-intensive issues in managing warehouses is the order picking problem which deals with the retrieval of items from their storage locations in order to meet customer requests. Many solution approaches have been proposed in order to minimize traveling distance in the process of order picking. However, in practice, customer orders have to be completed by certain due dates in order to avoid tardiness which is neglected in most of the related scientific papers. Consequently, we proposed a novel solution approach in order to minimize tardiness which consists of four phases. First of all, weighted association rule mining has been used to calculate associations between orders with respect to their due date. Next, a batching model based on binary integer programming has been formulated to maximize the associations between orders within each batch. Subsequently, the order picking phase will come up which used a Genetic Algorithm integrated with the Traveling Salesman Problem in order to identify the most suitable travel path. Finally, the Genetic Algorithm has been applied for sequencing the constructed batches in order to minimize tardiness. Illustrative examples and comparisons are presented to demonstrate the proficiency and solution quality of the proposed approach. PMID:23864823
NASA Astrophysics Data System (ADS)
Sans, J. A.; Sanchez-Royo, J. F.; Tobias-Rossell, G.; Canadell-Casanova, E.; Segura, A.
2010-01-01
Previous results show that the discrepancy between the experimental measurements and standard model theoretical calculations can be related with our assumption of unperturbed dispersion in the conduction band of the material. For overcoming this limitation, it has been proposed the band anti-crossing (BAC) model, based on the interaction of an antibonding state of Ga-O with the conduction band. Extending this model to other doping elements from III-group (Al, Ga and In), coherent results for the optical band-gap energy were obtained. These results have been supported by theoretical calculus of the electronic band structure, carried out by numerical atomic orbitals density functional theory (DFT) approach. This method is designed for efficient calculations in large systems and implemented in the SIESTA code.
ERIC Educational Resources Information Center
Uno, Mariko
2016-01-01
This study investigates the emergence and development of the discourse-pragmatic functions of the Japanese subject markers "wa" and "ga" from a usage-based perspective (Tomasello, 2000). The use of each marker in longitudinal speech data for four Japanese children from 1;0 to 3;1 and their parents available in the CHILDES…
NASA Astrophysics Data System (ADS)
Han, Zheng; Chen, Guangqi; Li, Yange; Wang, Wei; Zhang, Hong
2015-07-01
The estimation of debris-flow velocity in a cross-section is of primary importance due to its correlation to impact force, run up and superelevation. However, previous methods sometimes neglect the observed asymmetric velocity distribution, and consequently underestimate the debris-flow velocity. This paper presents a new approach for exploring the debris-flow velocity distribution in a cross-section. The presented approach uses an iteration algorithm based on the Riemann integral method to search an approximate solution to the unknown flow surface. The established laws for vertical velocity profile are compared and subsequently integrated to analyze the velocity distribution in the cross-section. The major benefit of the presented approach is that natural channels typically with irregular beds and superelevations can be taken into account, and the resulting approximation by the approach well replicates the direct integral solution. The approach is programmed in MATLAB environment, and the code is open to the public. A well-documented debris-flow event in Sichuan Province, China, is used to demonstrate the presented approach. Results show that the solutions of the flow surface and the mean velocity well reproduce the investigated results. Discussion regarding the model sensitivity and the source of errors concludes the paper.
A lake detection algorithm (LDA) using Landsat 8 data: A comparative approach in glacial environment
NASA Astrophysics Data System (ADS)
Bhardwaj, Anshuman; Singh, Mritunjay Kumar; Joshi, P. K.; Snehmani; Singh, Shaktiman; Sam, Lydia; Gupta, R. D.; Kumar, Rajesh
2015-06-01
Glacial lakes show a wide range of turbidity. Owing to this, the normalized difference water indices (NDWIs) as proposed by many researchers, do not give appropriate results in case of glacial lakes. In addition, the sub-pixel proportion of water and use of different optical band combinations are also reported to produce varying results. In the wake of the changing climate and increasing GLOFs (glacial lake outburst floods), there is a need to utilize wide optical and thermal capabilities of Landsat 8 data for the automated detection of glacial lakes. In the present study, the optical and thermal bandwidths of Landsat 8 data were explored along with the terrain slope parameter derived from Advanced Spaceborne Thermal Emission and Reflection Radiometer Global Digital Elevation Model Version2 (ASTER GDEM V2), for detecting and mapping glacial lakes. The validation of the algorithm was performed using manually digitized and subsequently field corrected lake boundaries. The pre-existing NDWIs were also evaluated to determine the supremacy and the stability of the proposed algorithm for glacial lake detection. Two new parameters, LDI (lake detection index) and LF (lake fraction) were proposed to comment on the performances of the indices. The lake detection algorithm (LDA) performed best in case of both, mixed lake pixels and pure lake pixels with no false detections (LDI = 0.98) and very less areal underestimation (LF = 0.73). The coefficient of determination (R2) between areal extents of lake pixels, extracted using the LDA and the actual lake area, was very high (0.99). With understanding of the terrain conditions and slight threshold adjustments, this work can be replicated for any mountainous region of the world.
Implementation of genetic algorithm for distribution systems loss minimum re-configuration
Nara, K.; Shiose, A. ); Kitagawa, M.; Ishihara, T. )
1992-08-01
In this paper, a distribution systems loss minimum reconfiguration method by genetic algorithm is proposed. The problem is a complex mixed integer programming problem and is very difficult to solve by a mathematical programming approach. A genetic algorithm (GA) is a search or optimization algorithm based on the mechanics of natural selection and natural genetics. Since GA is suitable to solve combinatorial optimization problems, it can be successfully applied to problems of loss minimum in distribution systems. Numerical examples demonstrate the validity and effectiveness of the proposed methodology.
Chaos-based image encryption using a hybrid genetic algorithm and a DNA sequence
NASA Astrophysics Data System (ADS)
Enayatifar, Rasul; Abdullah, Abdul Hanan; Isnin, Ismail Fauzi
2014-05-01
The paper studies a recently developed evolutionary-based image encryption algorithm. A novel image encryption algorithm based on a hybrid model of deoxyribonucleic acid (DNA) masking, a genetic algorithm (GA) and a logistic map is proposed. This study uses DNA and logistic map functions to create the number of initial DNA masks and applies GA to determine the best mask for encryption. The significant advantage of this approach is improving the quality of DNA masks to obtain the best mask that is compatible with plain images. The experimental results and computer simulations both confirm that the proposed scheme not only demonstrates excellent encryption but also resists various typical attacks.
Berkolaiko, G.; Kuipers, J.
2013-12-15
Electronic transport through chaotic quantum dots exhibits universal behaviour which can be understood through the semiclassical approximation. Within the approximation, calculation of transport moments reduces to codifying classical correlations between scattering trajectories. These can be represented as ribbon graphs and we develop an algorithmic combinatorial method to generate all such graphs with a given genus. This provides an expansion of the linear transport moments for systems both with and without time reversal symmetry. The computational implementation is then able to progress several orders further than previous semiclassical formulae as well as those derived from an asymptotic expansion of random matrix results. The patterns observed also suggest a general form for the higher orders.
NASA Astrophysics Data System (ADS)
Shen, Chong; Cao, Huiliang; Li, Jie; Tang, Jun; Zhang, Xiaoming; Shi, Yunbo; Yang, Wei; Liu, Jun
2016-03-01
A noise reduction algorithm based on an improved empirical mode decomposition (EMD) and forward linear prediction (FLP) is proposed for the fiber optic gyroscope (FOG). Referred to as the EMD-FLP algorithm, it was developed to decompose the FOG outputs into a number of intrinsic mode functions (IMFs) after which mode manipulations are performed to select noise-only IMFs, mixed IMFs, and residual IMFs. The FLP algorithm is then employed to process the mixed IMFs, from which the refined IMFs components are reconstructed to produce the final de-noising results. This hybrid approach is applied to, and verified using, both simulated signals and experimental FOG outputs. The results from the applications show that the method eliminates noise more effectively than the conventional EMD or FLP methods and decreases the standard deviations of the FOG outputs after de-noising from 0.17 to 0.026 under sweep frequency vibration and from 0.22 to 0.024 under fixed frequency vibration.
NASA Astrophysics Data System (ADS)
Inclan, Eric; Lassester, Jack; Geohegan, David; Yoon, Mina
Research in TiO2 materials is highly relevant to energy and device applications, however, precise control of their morphologies and characterization are still a grand challenge in the field. We developed and applied a hybrid optimization algorithm to explore configuration spaces of energetically metastable TiO2. Our approach was to minimize the total energy of TiO2 lusters in order to identify the energy landscape of plausible (TiO2)n (n = 1-100). The hybrid algorithm retained good agreement with a regression on structures published in literature up to n = 25. Using first-principles density functional theory, we analyze basic properties of the hybrid-algorithm generated TiO2 nanoparticles. Our results show the expected convergence to bulk material characteristics as the cluster size increases in that the band gap varies with respect to the size of the nanocluster. The nanoclusters trended toward compact, low surface area structures that share characteristics of the bulk, namely octahedral microstructures as the nanoclusters increased in size. Our study helps in better identifying and characterizing experimentally observed structures. This work is supported by the U.S. Department of Energy, Office of Science, Basic Energy Sciences, Materials Sciences and Engineering Division.
Shen, Chong; Cao, Huiliang; Li, Jie; Tang, Jun; Zhang, Xiaoming; Shi, Yunbo; Yang, Wei; Liu, Jun
2016-03-01
A noise reduction algorithm based on an improved empirical mode decomposition (EMD) and forward linear prediction (FLP) is proposed for the fiber optic gyroscope (FOG). Referred to as the EMD-FLP algorithm, it was developed to decompose the FOG outputs into a number of intrinsic mode functions (IMFs) after which mode manipulations are performed to select noise-only IMFs, mixed IMFs, and residual IMFs. The FLP algorithm is then employed to process the mixed IMFs, from which the refined IMFs components are reconstructed to produce the final de-noising results. This hybrid approach is applied to, and verified using, both simulated signals and experimental FOG outputs. The results from the applications show that the method eliminates noise more effectively than the conventional EMD or FLP methods and decreases the standard deviations of the FOG outputs after de-noising from 0.17 to 0.026 under sweep frequency vibration and from 0.22 to 0.024 under fixed frequency vibration. PMID:27036770
Anor, Tomer; Madsen, Joseph R; Dupont, Pierre
2011-05-01
We propose a novel systematic approach to optimizing the design of concentric tube robots for neurosurgical procedures. These procedures require that the robot approach specified target sites while navigating and operating within an anatomically constrained work space. The availability of preoperative imaging makes our approach particularly suited for neurosurgery, and we illustrate the method with the example of endoscopic choroid plexus ablation. A novel parameterization of the robot characteristics is used in conjunction with a global pattern search optimization method. The formulation returns the design of the least-complex robot capable of reaching single or multiple target points in a confined space with constrained optimization metrics. A particular advantage of this approach is that it identifies the need for either fixed-curvature versus variable-curvature sections. We demonstrate the performance of the method in four clinically relevant examples. PMID:22270831
Anor, Tomer; Madsen, Joseph R.; Dupont, Pierre
2011-01-01
We propose a novel systematic approach to optimizing the design of concentric tube robots for neurosurgical procedures. These procedures require that the robot approach specified target sites while navigating and operating within an anatomically constrained work space. The availability of preoperative imaging makes our approach particularly suited for neurosurgery, and we illustrate the method with the example of endoscopic choroid plexus ablation. A novel parameterization of the robot characteristics is used in conjunction with a global pattern search optimization method. The formulation returns the design of the least-complex robot capable of reaching single or multiple target points in a confined space with constrained optimization metrics. A particular advantage of this approach is that it identifies the need for either fixed-curvature versus variable-curvature sections. We demonstrate the performance of the method in four clinically relevant examples. PMID:22270831
Instrument design and optimization using genetic algorithms
Hoelzel, Robert; Bentley, Phillip M.; Fouquet, Peter
2006-10-15
This article describes the design of highly complex physical instruments by using a canonical genetic algorithm (GA). The procedure can be applied to all instrument designs where performance goals can be quantified. It is particularly suited to the optimization of instrument design where local optima in the performance figure of merit are prevalent. Here, a GA is used to evolve the design of the neutron spin-echo spectrometer WASP which is presently being constructed at the Institut Laue-Langevin, Grenoble, France. A comparison is made between this artificial intelligence approach and the traditional manual design methods. We demonstrate that the search of parameter space is more efficient when applying the genetic algorithm, and the GA produces a significantly better instrument design. Furthermore, it is found that the GA increases flexibility, by facilitating the reoptimization of the design after changes in boundary conditions during the design phase. The GA also allows the exploration of 'nonstandard' magnet coil geometries. We conclude that this technique constitutes a powerful complementary tool for the design and optimization of complex scientific apparatus, without replacing the careful thought processes employed in traditional design methods.