NASA Technical Reports Server (NTRS)
Moore, J. E.
1975-01-01
An enumeration algorithm is presented for solving a scheduling problem similar to the single machine job shop problem with sequence dependent setup times. The scheduling problem differs from the job shop problem in two ways. First, its objective is to select an optimum subset of the available tasks to be performed during a fixed period of time. Secondly, each task scheduled is constrained to occur within its particular scheduling window. The algorithm is currently being used to develop typical observational timelines for a telescope that will be operated in earth orbit. Computational times associated with timeline development are presented.
JIT single machine scheduling problem with periodic preventive maintenance
NASA Astrophysics Data System (ADS)
Shahriari, Mohammadreza; Shoja, Naghi; Zade, Amir Ebrahimi; Barak, Sasan; Sharifi, Mani
2016-03-01
This article investigates a JIT single machine scheduling problem with a periodic preventive maintenance. Also to maintain the quality of the products, there is a limitation on the maximum number of allowable jobs in each period. The proposed bi-objective mixed integer model minimizes total earliness-tardiness and makespan simultaneously. Due to the computational complexity of the problem, multi-objective particle swarm optimization (MOPSO) algorithm is implemented. Also, as well as MOPSO, two other optimization algorithms are used for comparing the results. Eventually, Taguchi method with metrics analysis is presented to tune the algorithms' parameters and a multiple criterion decision making technique based on the technique for order of preference by similarity to ideal solution is applied to choose the best algorithm. Comparison results confirmed the supremacy of MOPSO to the other algorithms.
Manipulating Tabu List to Handle Machine Breakdowns in Job Shop Scheduling Problems
NASA Astrophysics Data System (ADS)
Nababan, Erna Budhiarti; SalimSitompul, Opim
2011-06-01
Machine breakdowns in a production schedule may occur on a random basis that make the well-known hard combinatorial problem of Job Shop Scheduling Problems (JSSP) becomes more complex. One of popular techniques used to solve the combinatorial problems is Tabu Search. In this technique, moves that will be not allowed to be revisited are retained in a tabu list in order to avoid in gaining solutions that have been obtained previously. In this paper, we propose an algorithm to employ a second tabu list to keep broken machines, in addition to the tabu list that keeps the moves. The period of how long the broken machines will be kept on the list is categorized using fuzzy membership function. Our technique are tested to the benchmark data of JSSP available on the OR library. From the experiment, we found that our algorithm is promising to help a decision maker to face the event of machine breakdowns.
Scheduling of flow shop problems on 3 machines in fuzzy environment with double transport facility
NASA Astrophysics Data System (ADS)
Sathish, Shakeela; Ganesan, K.
2016-06-01
Flow shop scheduling is a decision making problem in production and manufacturing field which has a significant impact on the performance of an organization. When the machines on which jobs are to be processed are placed at different places, the transportation time plays a significant role in production. Further two different transport agents where 1st takes the job from 1st machine to 2nd machine and then returns back to the first machine and the 2nd takes the job from 2nd machine to 3rd machine and then returns back to the 2nd machine are also considered. We propose a method to minimize the total make span; without converting the fuzzy processing time to classical numbers by using a new type of fuzzy arithmetic and a fuzzy ranking method. A numerical example is provided to explain the proposed method.
Some single-machine scheduling problems with learning effects and two competing agents.
Li, Hongjie; Li, Zeyuan; Yin, Yunqiang
2014-01-01
This study considers a scheduling environment in which there are two agents and a set of jobs, each of which belongs to one of the two agents and its actual processing time is defined as a decreasing linear function of its starting time. Each of the two agents competes to process its respective jobs on a single machine and has its own scheduling objective to optimize. The objective is to assign the jobs so that the resulting schedule performs well with respect to the objectives of both agents. The objective functions addressed in this study include the maximum cost, the total weighted completion time, and the discounted total weighted completion time. We investigate three problems arising from different combinations of the objectives of the two agents. The computational complexity of the problems is discussed and solution algorithms where possible are presented. PMID:25009829
Some Single-Machine Scheduling Problems with Learning Effects and Two Competing Agents
Li, Hongjie; Li, Zeyuan; Yin, Yunqiang
2014-01-01
This study considers a scheduling environment in which there are two agents and a set of jobs, each of which belongs to one of the two agents and its actual processing time is defined as a decreasing linear function of its starting time. Each of the two agents competes to process its respective jobs on a single machine and has its own scheduling objective to optimize. The objective is to assign the jobs so that the resulting schedule performs well with respect to the objectives of both agents. The objective functions addressed in this study include the maximum cost, the total weighted completion time, and the discounted total weighted completion time. We investigate three problems arising from different combinations of the objectives of the two agents. The computational complexity of the problems is discussed and solution algorithms where possible are presented. PMID:25009829
Hybrid Metaheuristics for Solving a Fuzzy Single Batch-Processing Machine Scheduling Problem
Molla-Alizadeh-Zavardehi, S.; Tavakkoli-Moghaddam, R.; Lotfi, F. Hosseinzadeh
2014-01-01
This paper deals with a problem of minimizing total weighted tardiness of jobs in a real-world single batch-processing machine (SBPM) scheduling in the presence of fuzzy due date. In this paper, first a fuzzy mixed integer linear programming model is developed. Then, due to the complexity of the problem, which is NP-hard, we design two hybrid metaheuristics called GA-VNS and VNS-SA applying the advantages of genetic algorithm (GA), variable neighborhood search (VNS), and simulated annealing (SA) frameworks. Besides, we propose three fuzzy earliest due date heuristics to solve the given problem. Through computational experiments with several random test problems, a robust calibration is applied on the parameters. Finally, computational results on different-scale test problems are presented to compare the proposed algorithms. PMID:24883359
Hybrid metaheuristics for solving a fuzzy single batch-processing machine scheduling problem.
Molla-Alizadeh-Zavardehi, S; Tavakkoli-Moghaddam, R; Lotfi, F Hosseinzadeh
2014-01-01
This paper deals with a problem of minimizing total weighted tardiness of jobs in a real-world single batch-processing machine (SBPM) scheduling in the presence of fuzzy due date. In this paper, first a fuzzy mixed integer linear programming model is developed. Then, due to the complexity of the problem, which is NP-hard, we design two hybrid metaheuristics called GA-VNS and VNS-SA applying the advantages of genetic algorithm (GA), variable neighborhood search (VNS), and simulated annealing (SA) frameworks. Besides, we propose three fuzzy earliest due date heuristics to solve the given problem. Through computational experiments with several random test problems, a robust calibration is applied on the parameters. Finally, computational results on different-scale test problems are presented to compare the proposed algorithms. PMID:24883359
NASA Astrophysics Data System (ADS)
Li, Kai; Yang, Shan-Lin; Ren, Ming-Lun
2011-10-01
This article considers the single-machine scheduling problem to minimise the total resource consumption under the constraint that the makespan does not exceed a given limit, in which the release date of a job is a linear decreasing continuous function of the resource consumption. This problem is NP-hard in the strong sense. We design a simulated annealing (SA) algorithm to obtain the near-optimal solution with high quality. Two operators, right-move and left-move, are defined and their influences on the resource consumption are analysed. We use two operations, insert and swap, to generate the neighbourhood, and discuss how to calculate the change of total resource consumption. To evaluate the performance of the proposed algorithm, we relax the problem to an assignment problem, and obtain a lower bound by the Hungary method. And then, we improve its quality by Chu's method. Based on the settings that Janiak provided, we generate the random test data in our experiments to simulate the ingot preheating and hot-rolling process in steel mills. The accuracy and efficiency of the proposed SA algorithm are tested based on those data with problem sizes varying from 50 to 200 jobs. The computational results indicate that the SA approach is promising and capable of solving large-scale problems in a reasonable time.
Xu, Zhenzhen; Zou, Yongxing; Kong, Xiangjie
2015-01-01
To our knowledge, this paper investigates the first application of meta-heuristic algorithms to tackle the parallel machines scheduling problem with weighted late work criterion and common due date ([Formula: see text]). Late work criterion is one of the performance measures of scheduling problems which considers the length of late parts of particular jobs when evaluating the quality of scheduling. Since this problem is known to be NP-hard, three meta-heuristic algorithms, namely ant colony system, genetic algorithm, and simulated annealing are designed and implemented, respectively. We also propose a novel algorithm named LDF (largest density first) which is improved from LPT (longest processing time first). The computational experiments compared these meta-heuristic algorithms with LDF, LPT and LS (list scheduling), and the experimental results show that SA performs the best in most cases. However, LDF is better than SA in some conditions, moreover, the running time of LDF is much shorter than SA. PMID:26702371
Gang scheduling a parallel machine
Gorda, B.C.; Brooks, E.D. III.
1991-12-01
Program development on parallel machines can be a nightmare of scheduling headaches. We have developed a portable time sharing mechanism to handle the problem of scheduling gangs of processes. User programs and their gangs of processes are put to sleep and awakened by the gang scheduler to provide a time sharing environment. Time quantum are adjusted according to priority queues and a system of fair share accounting. The initial platform for this software is the 128 processor BBN TC2000 in use in the Massively Parallel Computing Initiative at the Lawrence Livermore National Laboratory.
Gang scheduling a parallel machine
Gorda, B.C.; Brooks, E.D. III.
1991-03-01
Program development on parallel machines can be a nightmare of scheduling headaches. We have developed a portable time sharing mechanism to handle the problem of scheduling gangs of processors. User program and their gangs of processors are put to sleep and awakened by the gang scheduler to provide a time sharing environment. Time quantums are adjusted according to priority queues and a system of fair share accounting. The initial platform for this software is the 128 processor BBN TC2000 in use in the Massively Parallel Computing Initiative at the Lawrence Livermore National Laboratory. 2 refs., 1 fig.
NASA Astrophysics Data System (ADS)
Joo, Cheol Min; Kim, Byung Soo
2012-09-01
This article considers a parallel machine scheduling problem with ready times, due times and sequence-dependent setup times. The objective of this problem is to determine the allocation policy of jobs and the scheduling policy of machines to minimize the weighted sum of setup times, delay times and tardy times. A mathematical model for optimal solution is derived. An in-depth analysis of the model shows that it is very complicated and difficult to obtain optimal solutions as the problem size becomes large. Therefore, two meta-heuristics, genetic algorithm (GA) and a new population-based evolutionary meta-heuristic called self-evolution algorithm (SEA), are proposed. The performances of the meta-heuristic algorithms are evaluated through comparison with optimal solutions using several randomly generated examples.
NASA Astrophysics Data System (ADS)
Liou, Cheng-Dar; Hsieh, Yi-Chih; Chen, Yin-Yann
2013-01-01
This article investigates the two-machine flow-shop group scheduling problem (GSP) with sequence-dependent setup and removal times, and job transportation times between machines. The objective is to minimise the total completion time. As known, this problem is an NP-hard problem and generalises the typical two-machine GSPs. In this article, a new encoding scheme based on permutation representation is proposed to transform a random job permutation to a feasible permutation for GSPs. The proposed encoding scheme simultaneously determines both the sequence of jobs in each group and the sequence of groups. By reasonably combining particle swarm optimisation (PSO) and genetic algorithm (GA), we develop a fast and easily implemented hybrid algorithm (HA) for solving the considered problems. The effectiveness and efficiency of the proposed HA are demonstrated and compared with those of standard PSO and GA by numerical results of various tested instances with group numbers up to 20. In addition, three different lower bounds are developed to evaluate the solution quality of the HA. Limited numerical results indicate that the proposed HA is a viable and effective approach for the studied two-machine flow-shop group scheduling problem.
Single-machine group scheduling problems with deteriorating and learning effect
NASA Astrophysics Data System (ADS)
Xingong, Zhang; Yong, Wang; Shikun, Bai
2016-07-01
The concepts of deteriorating jobs and learning effects have been individually studied in many scheduling problems. However, most studies considering the deteriorating and learning effects ignore the fact that production efficiency can be increased by grouping various parts and products with similar designs and/or production processes. This phenomenon is known as 'group technology' in the literature. In this paper, a new group scheduling model with deteriorating and learning effects is proposed, where learning effect depends not only on job position, but also on the position of the corresponding job group; deteriorating effect depends on its starting time of the job. This paper shows that the makespan and the total completion time problems remain polynomial optimal solvable under the proposed model. In addition, a polynomial optimal solution is also presented to minimise the maximum lateness problem under certain agreeable restriction.
NASA Astrophysics Data System (ADS)
Xu, Ye; Wang, Ling; Wang, Shengyao; Liu, Min
2013-12-01
In this article, an effective shuffled frog-leaping algorithm (SFLA) is proposed to solve the hybrid flow-shop scheduling problem with identical parallel machines (HFSP-IPM). First, some novel heuristic decoding rules for both job order decision and machine assignment are proposed. Then, three hybrid decoding schemes are designed to decode job order sequences to schedules. A special bi-level crossover and multiple local search operators are incorporated in the searching framework of the SFLA to enrich the memetic searching behaviour and to balance the exploration and exploitation capabilities. Meanwhile, some theoretical analysis for the local search operators is provided for guiding the local search. The parameter setting of the algorithm is also investigated based on the Taguchi method of design of experiments. Finally, numerical testing based on well-known benchmarks and comparisons with some existing algorithms are carried out to demonstrate the effectiveness of the proposed algorithm.
NASA Astrophysics Data System (ADS)
Hasani, Keramat; Kravchenko, Svetlana A.; Werner, Frank
2016-01-01
This article considers the problem of scheduling a given set of n jobs on two identical parallel machines with a single server. Each job must be processed on one of the machines. Before processing, the server has to set up the relevant machine. The objective is to minimize the makespan. For this unary NP-hard problem, two fast constructive algorithms with a complexity of O(n2) are presented. The performance of these algorithms is evaluated for instances with up to 10,000 jobs. Computational results indicate that the algorithms have an excellent performance for very large instances so that the obtained objective function values are very close to a lower bound, and in many cases even an optimal solution is achieved. Superiority over all existing algorithms is obtained by sequencing the jobs on the two machines so that the machine idle time and the server waiting time are minimized. In doing so, the characteristics of an optimal solution resulting from its relevant lower bound are taken into account.
Gang scheduling a parallel machine. Revision 1
Gorda, B.C.; Brooks, E.D. III
1991-12-01
Program development on parallel machines can be a nightmare of scheduling headaches. We have developed a portable time sharing mechanism to handle the problem of scheduling gangs of processes. User programs and their gangs of processes are put to sleep and awakened by the gang scheduler to provide a time sharing environment. Time quantum are adjusted according to priority queues and a system of fair share accounting. The initial platform for this software is the 128 processor BBN TC2000 in use in the Massively Parallel Computing Initiative at the Lawrence Livermore National Laboratory.
NASA Astrophysics Data System (ADS)
Kaplan, Sezgin; Rabadi, Ghaith
2013-01-01
This article addresses the aerial refuelling scheduling problem (ARSP), where a set of fighter jets (jobs) with certain ready times must be refuelled from tankers (machines) by their due dates; otherwise, they reach a low fuel level (deadline) incurring a high cost. ARSP is an identical parallel machine scheduling problem with release times and due date-to-deadline windows to minimize the total weighted tardiness. A simulated annealing (SA) and metaheuristic for randomized priority search (Meta-RaPS) with the newly introduced composite dispatching rule, apparent piecewise tardiness cost with ready times (APTCR), are applied to the problem. Computational experiments compared the algorithms' solutions to optimal solutions for small problems and to each other for larger problems. To obtain optimal solutions, a mixed integer program with a piecewise weighted tardiness objective function was solved for up to 12 jobs. The results show that Meta-RaPS performs better in terms of average relative error but SA is more efficient.
Parallel machine scheduling with a common server
Hall, N.; Sriskandarajah, C.; Potts, C.
1994-12-31
This paper considers the nonpreemptive scheduling of a given set of jobs on several identical, parallel machines. Each job must be processed on one of the machines. Prior to processing, a job must be loaded (setup) by a single server onto the relevant machine. The server may be a human operator, a robot, or a piece of specialized equipment. We study a number of classical scheduling objectives in this environment, including makespan, maximum lateness, the sum of completion times, the number of late jobs, and total tardiness, as well as weighted versions of some of these. The number of machines may be constant or arbitrary. Setup times may be unit, equal, or arbitrary. Processing times may be unit or arbitrary. For each problem considered, we attempt to provide either an efficient algorithm, or a proof that such an algorithm is unlikely to exist. Our results provide a mapping of the computational complexity of these problems. Included in these results are generalizations of the classical algorithms of Moore, Lawler and Moore and Lawler. In addition, we describe two heuristics for makespan scheduling in this environment, and provide an exact analysis of their worst-case performance.
NASA Astrophysics Data System (ADS)
He, Yong; Sun, Li
2015-05-01
In this paper, we introduce a group scheduling model with general deteriorating jobs and learning effects in which deteriorating jobs and learning effects are both considered simultaneously. This means that the actual processing time of a job depends not only on the processing time of the jobs already processed, but also on its scheduled position. In our model, the group setup times are general linear functions of their starting times and the jobs in the same group have general position-dependent learning effects and time-dependent deterioration. The objective of scheduling problems is to minimise the makespan and the sum of completion times, respectively. We show that the problems remain solvable in polynomial time under the proposed model.
A Study on Machine Maintenance Scheduling Using Distributed Cooperative Approach
NASA Astrophysics Data System (ADS)
Tsujibe, Akihisa; Kaihara, Toshiya; Fujii, Nobutada; Nonaka, Youichi
In this study, we propose a distributed cooperative scheduling method, and apply the method into a machine maintenance scheduling problem in re-entrant production systems. As one of the distributed cooperative scheduling methods, we focus on Lagrangian decomposition and coordination (LDC) method, and formulate the machine maintenance scheduling problem with LDC so as to improve computational efficiency by decomposing an original scheduling problem into several sub-problems. The derived solutions by solving the decomposed dual problem are converted into feasible solutions with a heuristic procedure applied in this study. The proposed approach regards maintenance as job with starting and finishing time constraints, so that product and maintenance schedule can realize proper maintenance operations without losing productivity. We show the effectiveness of the proposed method in several simulation experiments.
Static scheduling for dynamic dataflow machines
Beck, M.; Pingali, K.K. ); Nicolau, A. )
1990-12-01
Dynamic dataflow machines exploit parallelism among loop iterations by loop unraveling: all iterations of the loop are started together and operations in various iterations execute when their input data are present. Unbounded loop unraveling can strain the resources available on the machine and, in extreme cases, deadlock can occur due to overcommitment of resources. Previous efforts to address this problem have focused mainly on run-time mechanisms of debatable utility. Loop bounding, a compile-time technique, controls parallelism by permitting a fixed number of iterations to execute at one time. In this paper, the authors argue that loop bounding can lead to inefficient use of resources, and we propose an alternative way of compiling loops of overlapped execution of loop iterations. The authors introduce the notion of a stage decomposition of a loop, which defines a partition of the operations in a loop iteration into stages, and they show that the problem of choosing a stage decomposition for a particular loop can be tackled by applying static scheduling techniques like the ones used in generating code for VLIW machines. These techniques permit the compiler to allocate resources more skillfully than with loop bounding. The practical utility of stage decomposition remains to be tested on a real dataflow machine. In the absence of one, the authors describe how the schema could be implemented on the Monsoon dataflow machine being built at MIT.
Integrated network design and scheduling problems :
Nurre, Sarah G.; Carlson, Jeffrey J.
2014-01-01
We consider the class of integrated network design and scheduling problems. These problems focus on selecting and scheduling operations that will change the characteristics of a network, while being speci cally concerned with the performance of the network over time. Motivating applications of INDS problems include infrastructure restoration after extreme events and building humanitarian distribution supply chains. While similar models have been proposed, no one has performed an extensive review of INDS problems from their complexity, network and scheduling characteristics, information, and solution methods. We examine INDS problems under a parallel identical machine scheduling environment where the performance of the network is evaluated by solving classic network optimization problems. We classify that all considered INDS problems as NP-Hard and propose a novel heuristic dispatching rule algorithm that selects and schedules sets of arcs based on their interactions in the network. We present computational analysis based on realistic data sets representing the infrastructures of coastal New Hanover County, North Carolina, lower Manhattan, New York, and a realistic arti cial community CLARC County. These tests demonstrate the importance of a dispatching rule to arrive at near-optimal solutions during real-time decision making activities. We extend INDS problems to incorporate release dates which represent the earliest an operation can be performed and exible release dates through the introduction of specialized machine(s) that can perform work to move the release date earlier in time. An online optimization setting is explored where the release date of a component is not known.
A simple neural network scheduler for real-time machine task scheduling
Gritzo, R.E.
1991-01-01
The recent development of a new generation of automated radionuclide assay equipment in our facility requires embedded software at each machine for the scheduling of tasks. The execution time requirements of real-time embedded software limit the complexity of the scheduler design. By representing the scheduling problem properly, a simple backpropagation neural network performs the scheduling function within the imposed requirements. Operational tests have demonstrated that the neural network scheduler has met all development goals and is superior to the previous approaches. 3 refs., 1 tab.
A threshold scheduling strategy for Sisal on distributed memory machines
Pande, S.S.; Agrawal, D.P.; Mauney, J. )
1994-05-01
The problem of scheduling tasks on distributed memory machines is known to be NP-complete in the strong sense, ruling out the possibility of a pseudo-polynomial algorithm. This paper introduces a new heuristic algorithm for scheduling Sisal (Streams and Iterations In a Single Assignment Language) programs on a distributed memory machine, Intel Touchstone i860. The compile time scheduling method works on IF-2, an intermediate form based on the dataflow parallelism in the program. The authors initially carry out a dependence analysis, to bind the implicit dependencies across IF-2 graph boundaries, followed by a cost assignment based on Intel Touchstone i860 timings. The scheduler works in two phases. The first phase of the scheduler finds the earliest and latest completion times of each task given by the shortest and longest paths from root task to the given task respectively. A threshold defined as the difference between the latest and the earliest start times of the task, is found. The scheduler varies the value of the allowable threshold, and determines the best value for minimal schedule length. In the second phase of the scheduler, the authors merge the processors to generate schedule to match the available number of processors. Schedule results for several benchmark programs have been included to demonstrate the effectiveness of their approach.
The Microchp Scheduling Problem
NASA Astrophysics Data System (ADS)
Bosman, M. G. C.; Bakker, V.; Molderink, A.; Hurink, J. L.; Smit, G. J. M.
2009-08-01
The increasing penetration of renewable energy sources, the demand for more energy efficient electricity production and the increase in distributed electricity generation causes a shift in the way electricity is produced and consumed. The downside of these changes in the electricity grid is that network stability and controllability becomes more difficult compared to the old situation. The new network has to accommodate various means of production, consumption and buffering and needs to offer control over the energy flows between these three elements. In order to offer such a control mechanism we need to know more about the individual aspects. In this paper we focus on the modelling of distributed production. Especially we look at the use of microCHP (Combined Heat and Power) appliances in a group of houses. The problem of planning the production runs of the microCHP is modelled via an ILP formulation both for a single house and for a group of houses.
Bounded Parallel-Batch Scheduling on Unrelated Parallel Machines
NASA Astrophysics Data System (ADS)
Miao, Cuixia; Zhang, Yuzhong; Wang, Chengfei
In this paper, we consider the bounded parallel-batch scheduling problem on unrelated parallel machines. Problems R m |B|F are NP-hard for any objective function F. For this reason, we discuss the special case with p ij = p i for i = 1, 2, ⋯ , m , j = 1, 2, ⋯ , n. We give optimal algorithms for the general scheduling to minimize total weighted completion time, makespan and the number of tardy jobs. And we design pseudo-polynomial time algorithms for the case with rejection penalty to minimize the makespan and the total weighted completion time plus the total penalty of the rejected jobs, respectively.
A simple neural network scheduler for real-time machine task scheduling
Gritzo, R.E.
1992-06-01
The recent development of a new generation of automated radionuclide assay equipment in our facility requires embedded software at each machine for the scheduling of sample assay tasks. The execution time requirements of real-time embedded software limit the complexity of the schedular software. By representing the scheduling problem properly, a simple backpropagation neural network performs the scheduling function within the imposed requirements. Operational tests have demonstrated that the neural network schedular has met all development goals and is superior to the previous approaches. This paper describes the design and development of the neural network task scheduler. In addition, several aspects of the practical application of neural networks to real-world problems are discussed.
A simple neural network scheduler for real-time machine task scheduling
Gritzo, R.E.
1992-01-01
The recent development of a new generation of automated radionuclide assay equipment in our facility requires embedded software at each machine for the scheduling of sample assay tasks. The execution time requirements of real-time embedded software limit the complexity of the schedular software. By representing the scheduling problem properly, a simple backpropagation neural network performs the scheduling function within the imposed requirements. Operational tests have demonstrated that the neural network schedular has met all development goals and is superior to the previous approaches. This paper describes the design and development of the neural network task scheduler. In addition, several aspects of the practical application of neural networks to real-world problems are discussed.
Job shop scheduling problem with late work criterion
NASA Astrophysics Data System (ADS)
Piroozfard, Hamed; Wong, Kuan Yew
2015-05-01
Scheduling is considered as a key task in many industries, such as project based scheduling, crew scheduling, flight scheduling, machine scheduling, etc. In the machine scheduling area, the job shop scheduling problems are considered to be important and highly complex, in which they are characterized as NP-hard. The job shop scheduling problems with late work criterion and non-preemptive jobs are addressed in this paper. Late work criterion is a fairly new objective function. It is a qualitative measure and concerns with late parts of the jobs, unlike classical objective functions that are quantitative measures. In this work, simulated annealing was presented to solve the scheduling problem. In addition, operation based representation was used to encode the solution, and a neighbourhood search structure was employed to search for the new solutions. The case studies are Lawrence instances that were taken from the Operations Research Library. Computational results of this probabilistic meta-heuristic algorithm were compared with a conventional genetic algorithm, and a conclusion was made based on the algorithm and problem.
Two-machine flow shop scheduling integrated with preventive maintenance planning
NASA Astrophysics Data System (ADS)
Wang, Shijin; Liu, Ming
2016-02-01
This paper investigates an integrated optimisation problem of production scheduling and preventive maintenance (PM) in a two-machine flow shop with time to failure of each machine subject to a Weibull probability distribution. The objective is to find the optimal job sequence and the optimal PM decisions before each job such that the expected makespan is minimised. To investigate the value of integrated scheduling solution, computational experiments on small-scale problems with different configurations are conducted with total enumeration method, and the results are compared with those of scheduling without maintenance but with machine degradation, and individual job scheduling combined with independent PM planning. Then, for large-scale problems, four genetic algorithm (GA) based heuristics are proposed. The numerical results with several large problem sizes and different configurations indicate the potential benefits of integrated scheduling solution and the results also show that proposed GA-based heuristics are efficient for the integrated problem.
Bicriteria single machine scheduling with setup times and learning effects
NASA Astrophysics Data System (ADS)
Soroush, H. M.
2012-11-01
We study a bicriteria single machine scheduling problem with job-dependent and past-sequence-dependent (psd) setup time and job-dependent learning effects. The goal is to find the optimal sequence that minimizes a linear combination of a pair of performance criteria consisting of the makespan, the total completion time, and the total absolute differences in completion times. We show that special cases of the resulting three problems are solvable polynomially. However, the general cases cannot be solved in polynomial time; thus, branch-and-bound (B&B) methods are proposed to derive optimal sequences. Computational results demonstrate that the B&B methods solve relatively large problem instances in reasonable amounts of time.
NASA Astrophysics Data System (ADS)
Afzalirad, Mojtaba; Rezaeian, Javad
2016-04-01
This study involves an unrelated parallel machine scheduling problem in which sequence-dependent set-up times, different release dates, machine eligibility and precedence constraints are considered to minimize total late works. A new mixed-integer programming model is presented and two efficient hybrid meta-heuristics, genetic algorithm and ant colony optimization, combined with the acceptance strategy of the simulated annealing algorithm (Metropolis acceptance rule), are proposed to solve this problem. Manifestly, the precedence constraints greatly increase the complexity of the scheduling problem to generate feasible solutions, especially in a parallel machine environment. In this research, a new corrective algorithm is proposed to obtain the feasibility in all stages of the algorithms. The performance of the proposed algorithms is evaluated in numerical examples. The results indicate that the suggested hybrid ant colony optimization statistically outperformed the proposed hybrid genetic algorithm in solving large-size test problems.
Group scheduling problems in directional sensor networks
NASA Astrophysics Data System (ADS)
Singh, Alok; Rossi, André
2015-12-01
This article addresses the problem of scheduling a set of groups of directional sensors arising as a result of applying an exact or a heuristic approach for solving a problem involving directional sensors. The problem seeks a schedule for these groups that minimizes the total energy consumed in switching from one group to the next group in the schedule. In practice, when switching from a group to the next one, active sensors in the new group have to rotate in order to face their working direction. These rotations consume energy, and the problem is to schedule the groups so as to minimize the total amount of energy consumed by all the sensor rotations, knowing the initial angular positions of all the sensors. In this article, it is assumed that energy consumption is proportional to the angular movement for all the sensors. Another problem version is also investigated that seeks to minimize the total time during which the sensor network cannot cover all the targets because active sensors are rotating. Both problems are proved to be ?-hard, and a lower bound for the first problem is presented. A greedy heuristic and a genetic algorithm are also proposed for addressing the problem of minimizing total rotation in the general case. Finally, a local search is also proposed to improve the solutions obtained through a genetic algorithm.
Prediction based proactive thermal virtual machine scheduling in green clouds.
Kinger, Supriya; Kumar, Rajesh; Sharma, Anju
2014-01-01
Cloud computing has rapidly emerged as a widely accepted computing paradigm, but the research on Cloud computing is still at an early stage. Cloud computing provides many advanced features but it still has some shortcomings such as relatively high operating cost and environmental hazards like increasing carbon footprints. These hazards can be reduced up to some extent by efficient scheduling of Cloud resources. Working temperature on which a machine is currently running can be taken as a criterion for Virtual Machine (VM) scheduling. This paper proposes a new proactive technique that considers current and maximum threshold temperature of Server Machines (SMs) before making scheduling decisions with the help of a temperature predictor, so that maximum temperature is never reached. Different workload scenarios have been taken into consideration. The results obtained show that the proposed system is better than existing systems of VM scheduling, which does not consider current temperature of nodes before making scheduling decisions. Thus, a reduction in need of cooling systems for a Cloud environment has been obtained and validated. PMID:24737962
Prediction Based Proactive Thermal Virtual Machine Scheduling in Green Clouds
Kinger, Supriya; Kumar, Rajesh; Sharma, Anju
2014-01-01
Cloud computing has rapidly emerged as a widely accepted computing paradigm, but the research on Cloud computing is still at an early stage. Cloud computing provides many advanced features but it still has some shortcomings such as relatively high operating cost and environmental hazards like increasing carbon footprints. These hazards can be reduced up to some extent by efficient scheduling of Cloud resources. Working temperature on which a machine is currently running can be taken as a criterion for Virtual Machine (VM) scheduling. This paper proposes a new proactive technique that considers current and maximum threshold temperature of Server Machines (SMs) before making scheduling decisions with the help of a temperature predictor, so that maximum temperature is never reached. Different workload scenarios have been taken into consideration. The results obtained show that the proposed system is better than existing systems of VM scheduling, which does not consider current temperature of nodes before making scheduling decisions. Thus, a reduction in need of cooling systems for a Cloud environment has been obtained and validated. PMID:24737962
Solving Stochastic Flexible Flow Shop Scheduling Problems with a Decomposition-Based Approach
NASA Astrophysics Data System (ADS)
Wang, K.; Choi, S. H.
2010-06-01
Real manufacturing is dynamic and tends to suffer a lot of uncertainties. Research on production scheduling under uncertainty has recently received much attention. Although various approaches have been developed for scheduling under uncertainty, this problem is still difficult to tackle by any single approach, because of its inherent difficulties. This chapter describes a decomposition-based approach (DBA) for makespan minimisation of a flexible flow shop (FFS) scheduling problem with stochastic processing times. The DBA decomposes an FFS into several machine clusters which can be solved more easily by different approaches. A neighbouring K-means clustering algorithm is developed to firstly group the machines of an FFS into an appropriate number of machine clusters, based on a weighted cluster validity index. A back propagation network (BPN) is then adopted to assign either the Shortest Processing Time (SPT) Algorithm or the Genetic Algorithm (GA) to generate a sub-schedule for each machine cluster. After machine grouping and approach assignment, an overall schedule is generated by integrating the sub-schedules of the machine clusters. Computation results reveal that the DBA is superior to SPT and GA alone for FFS scheduling under stochastic processing times, and that it can be easily adapted to schedule FFS under other uncertainties.
Job shop scheduling model for non-identic machine with fixed delivery time to minimize tardiness
NASA Astrophysics Data System (ADS)
Kusuma, K. K.; Maruf, A.
2016-02-01
Scheduling non-identic machines problem with low utilization characteristic and fixed delivery time are frequent in manufacture industry. This paper propose a mathematical model to minimize total tardiness for non-identic machines in job shop environment. This model will be categorized as an integer linier programming model and using branch and bound algorithm as the solver method. We will use fixed delivery time as main constraint and different processing time to process a job. The result of this proposed model shows that the utilization of production machines can be increase with minimal tardiness using fixed delivery time as constraint.
Optimum Production Control and Workforce Scheduling of Machining Project
NASA Astrophysics Data System (ADS)
Lan, Tian-Syung; Lo, Chih-Yao; Hou, Cheng-I.
Through the proposed model in this study, the production control with the consideration of workforce scheduling for advanced manufacturing systems becomes realistically and concretely solvable. This study not only meditates the concept of balancing machine productivity and human ability into the objective, but also implements Calculus of Variations to optimize the profit for a deterministic production quantity. In addition, the optimum solutions of dynamic productivity control and workforce scheduling are comprehensively provided. Moreover, the decision criteria for selecting the optimum solution and the sensitivity analysis of the critical variables are fully discussed. This study definitely contributes the applicable strategy to control the productivity and workforce in manufacturing and provides the valuable tool to conclusively optimize the profit of a machining project for operations research in today`s manufacturing industry with profound insight.
Optimal pre-scheduling of problem remappings
NASA Technical Reports Server (NTRS)
Nicol, David M.; Saltz, Joel H.
1987-01-01
A large class of scientific computational problems can be characterized as a sequence of steps where a significant amount of computation occurs each step, but the work performed at each step is not necessarily identical. Two good examples of this type of computation are: (1) regridding methods which change the problem discretization during the course of the computation, and (2) methods for solving sparse triangular systems of linear equations. Recent work has investigated a means of mapping such computations onto parallel processors; the method defines a family of static mappings with differing degrees of importance placed on the conflicting goals of good load balance and low communication/synchronization overhead. The performance tradeoffs are controllable by adjusting the parameters of the mapping method. To achieve good performance it may be necessary to dynamically change these parameters at run-time, but such changes can impose additional costs. If the computation's behavior can be determined prior to its execution, it can be possible to construct an optimal parameter schedule using a low-order-polynomial-time dynamic programming algorithm. Since the latter can be expensive, the performance is studied of the effect of a linear-time scheduling heuristic on one of the model problems, and it is shown to be effective and nearly optimal.
NASA Astrophysics Data System (ADS)
Wang, Li-Chih; Chen, Yin-Yann; Chen, Tzu-Li; Cheng, Chen-Yang; Chang, Chin-Wei
2014-10-01
This paper studies a solar cell industry scheduling problem, which is similar to traditional hybrid flowshop scheduling (HFS). In a typical HFS problem, the allocation of machine resources for each order should be scheduled in advance. However, the challenge in solar cell manufacturing is the number of machines that can be adjusted dynamically to complete the job. An optimal production scheduling model is developed to explore these issues, considering the practical characteristics, such as hybrid flowshop, parallel machine system, dedicated machines, sequence independent job setup times and sequence dependent job setup times. The objective of this model is to minimise the makespan and to decide the processing sequence of the orders/lots in each stage, lot-splitting decisions for the orders and the number of machines used to satisfy the demands in each stage. From the experimental results, lot-splitting has significant effect on shortening the makespan, and the improvement effect is influenced by the processing time and the setup time of orders. Therefore, the threshold point to improve the makespan can be identified. In addition, the model also indicates that more lot-splitting approaches, that is, the flexibility of allocating orders/lots to machines is larger, will result in a better scheduling performance.
A hybrid dynamic harmony search algorithm for identical parallel machines scheduling
NASA Astrophysics Data System (ADS)
Chen, Jing; Pan, Quan-Ke; Wang, Ling; Li, Jun-Qing
2012-02-01
In this article, a dynamic harmony search (DHS) algorithm is proposed for the identical parallel machines scheduling problem with the objective to minimize makespan. First, an encoding scheme based on a list scheduling rule is developed to convert the continuous harmony vectors to discrete job assignments. Second, the whole harmony memory (HM) is divided into multiple small-sized sub-HMs, and each sub-HM performs evolution independently and exchanges information with others periodically by using a regrouping schedule. Third, a novel improvisation process is applied to generate a new harmony by making use of the information of harmony vectors in each sub-HM. Moreover, a local search strategy is presented and incorporated into the DHS algorithm to find promising solutions. Simulation results show that the hybrid DHS (DHS_LS) is very competitive in comparison to its competitors in terms of mean performance and average computational time.
Frutos, M; Méndez, M; Tohmé, F; Broz, D
2013-01-01
Many of the problems that arise in production systems can be handled with multiobjective techniques. One of those problems is that of scheduling operations subject to constraints on the availability of machines and buffer capacity. In this paper we analyze different Evolutionary multiobjective Algorithms (MOEAs) for this kind of problems. We consider an experimental framework in which we schedule production operations for four real world Job-Shop contexts using three algorithms, NSGAII, SPEA2, and IBEA. Using two performance indexes, Hypervolume and R2, we found that SPEA2 and IBEA are the most efficient for the tasks at hand. On the other hand IBEA seems to be a better choice of tool since it yields more solutions in the approximate Pareto frontier. PMID:24489502
Frutos, M.; Méndez, M.; Tohmé, F.; Broz, D.
2013-01-01
Many of the problems that arise in production systems can be handled with multiobjective techniques. One of those problems is that of scheduling operations subject to constraints on the availability of machines and buffer capacity. In this paper we analyze different Evolutionary multiobjective Algorithms (MOEAs) for this kind of problems. We consider an experimental framework in which we schedule production operations for four real world Job-Shop contexts using three algorithms, NSGAII, SPEA2, and IBEA. Using two performance indexes, Hypervolume and R2, we found that SPEA2 and IBEA are the most efficient for the tasks at hand. On the other hand IBEA seems to be a better choice of tool since it yields more solutions in the approximate Pareto frontier. PMID:24489502
Run-time scheduling and execution of loops on message passing machines
NASA Technical Reports Server (NTRS)
Crowley, Kay; Saltz, Joel; Mirchandaney, Ravi; Berryman, Harry
1989-01-01
Sparse system solvers and general purpose codes for solving partial differential equations are examples of the many types of problems whose irregularity can result in poor performance on distributed memory machines. Often, the data structures used in these problems are very flexible. Crucial details concerning loop dependences are encoded in these structures rather than being explicitly represented in the program. Good methods for parallelizing and partitioning these types of problems require assignment of computations in rather arbitrary ways. Naive implementations of programs on distributed memory machines requiring general loop partitions can be extremely inefficient. Instead, the scheduling mechanism needs to capture the data reference patterns of the loops in order to partition the problem. First, the indices assigned to each processor must be locally numbered. Next, it is necessary to precompute what information is needed by each processor at various points in the computation. The precomputed information is then used to generate an execution template designed to carry out the computation, communication, and partitioning of data, in an optimized manner. The design is presented for a general preprocessor and schedule executer, the structures of which do not vary, even though the details of the computation and of the type of information are problem dependent.
Run-time scheduling and execution of loops on message passing machines
NASA Technical Reports Server (NTRS)
Saltz, Joel; Crowley, Kathleen; Mirchandaney, Ravi; Berryman, Harry
1990-01-01
Sparse system solvers and general purpose codes for solving partial differential equations are examples of the many types of problems whose irregularity can result in poor performance on distributed memory machines. Often, the data structures used in these problems are very flexible. Crucial details concerning loop dependences are encoded in these structures rather than being explicitly represented in the program. Good methods for parallelizing and partitioning these types of problems require assignment of computations in rather arbitrary ways. Naive implementations of programs on distributed memory machines requiring general loop partitions can be extremely inefficient. Instead, the scheduling mechanism needs to capture the data reference patterns of the loops in order to partition the problem. First, the indices assigned to each processor must be locally numbered. Next, it is necessary to precompute what information is needed by each processor at various points in the computation. The precomputed information is then used to generate an execution template designed to carry out the computation, communication, and partitioning of data, in an optimized manner. The design is presented for a general preprocessor and schedule executer, the structures of which do not vary, even though the details of the computation and of the type of information are problem dependent.
Eroglu, Duygu Yilmaz; Ozmutlu, H Cenk
2014-01-01
We developed mixed integer programming (MIP) models and hybrid genetic-local search algorithms for the scheduling problem of unrelated parallel machines with job sequence and machine-dependent setup times and with job splitting property. The first contribution of this paper is to introduce novel algorithms which make splitting and scheduling simultaneously with variable number of subjobs. We proposed simple chromosome structure which is constituted by random key numbers in hybrid genetic-local search algorithm (GAspLA). Random key numbers are used frequently in genetic algorithms, but it creates additional difficulty when hybrid factors in local search are implemented. We developed algorithms that satisfy the adaptation of results of local search into the genetic algorithms with minimum relocation operation of genes' random key numbers. This is the second contribution of the paper. The third contribution of this paper is three developed new MIP models which are making splitting and scheduling simultaneously. The fourth contribution of this paper is implementation of the GAspLAMIP. This implementation let us verify the optimality of GAspLA for the studied combinations. The proposed methods are tested on a set of problems taken from the literature and the results validate the effectiveness of the proposed algorithms. PMID:24977204
Hidri, Lotfi; Gharbi, Anis; Louly, Mohamed Aly
2014-01-01
We focus on the two-center hybrid flow shop scheduling problem with identical parallel machines and removal times. The job removal time is the required duration to remove it from a machine after its processing. The objective is to minimize the maximum completion time (makespan). A heuristic and a lower bound are proposed for this NP-Hard problem. These procedures are based on the optimal solution of the parallel machine scheduling problem with release dates and delivery times. The heuristic is composed of two phases. The first one is a constructive phase in which an initial feasible solution is provided, while the second phase is an improvement one. Intensive computational experiments have been conducted to confirm the good performance of the proposed procedures. PMID:25610911
A Solution Method of Scheduling Problem with Worker Allocation by a Genetic Algorithm
NASA Astrophysics Data System (ADS)
Osawa, Akira; Ida, Kenichi
In a scheduling problem with worker allocation (SPWA) proposed by Iima et al, the worker's skill level to each machine is all the same. However, each worker has a different skill level for each machine in the real world. For that reason, we propose a new model of SPWA in which a worker has the different skill level to each machine. To solve the problem, we propose a new GA for SPWA consisting of the following new three procedures, shortening of idle time, modifying infeasible solution to feasible solution, and a new selection method for GA. The effectiveness of the proposed algorithm is clarified by numerical experiments using benchmark problems for job-shop scheduling.
Machine Learning Based Online Performance Prediction for Runtime Parallelization and Task Scheduling
Li, J; Ma, X; Singh, K; Schulz, M; de Supinski, B R; McKee, S A
2008-10-09
With the emerging many-core paradigm, parallel programming must extend beyond its traditional realm of scientific applications. Converting existing sequential applications as well as developing next-generation software requires assistance from hardware, compilers and runtime systems to exploit parallelism transparently within applications. These systems must decompose applications into tasks that can be executed in parallel and then schedule those tasks to minimize load imbalance. However, many systems lack a priori knowledge about the execution time of all tasks to perform effective load balancing with low scheduling overhead. In this paper, we approach this fundamental problem using machine learning techniques first to generate performance models for all tasks and then applying those models to perform automatic performance prediction across program executions. We also extend an existing scheduling algorithm to use generated task cost estimates for online task partitioning and scheduling. We implement the above techniques in the pR framework, which transparently parallelizes scripts in the popular R language, and evaluate their performance and overhead with both a real-world application and a large number of synthetic representative test scripts. Our experimental results show that our proposed approach significantly improves task partitioning and scheduling, with maximum improvements of 21.8%, 40.3% and 22.1% and average improvements of 15.9%, 16.9% and 4.2% for LMM (a real R application) and synthetic test cases with independent and dependent tasks, respectively.
Application of a hybrid generation/utility assessment heuristic to a class of scheduling problems
NASA Technical Reports Server (NTRS)
Heyward, Ann O.
1989-01-01
A two-stage heuristic solution approach for a class of multiobjective, n-job, 1-machine scheduling problems is described. Minimization of job-to-job interference for n jobs is sought. The first stage generates alternative schedule sequences by interchanging pairs of schedule elements. The set of alternative sequences can represent nodes of a decision tree; each node is reached via decision to interchange job elements. The second stage selects the parent node for the next generation of alternative sequences through automated paired comparison of objective performance for all current nodes. An application of the heuristic approach to communications satellite systems planning is presented.
Solution of the NP-hard total tardiness minimization problem in scheduling theory
NASA Astrophysics Data System (ADS)
Lazarev, A. A.
2007-06-01
The classical NP-hard (in the ordinary sense) problem of scheduling jobs in order to minimize the total tardiness for a single machine 1‖Σ T j is considered. An NP-hard instance of the problem is completely analyzed. A procedure for partitioning the initial set of jobs into subsets is proposed. Algorithms are constructed for finding an optimal schedule depending on the number of subsets. The complexity of the algorithms is O( n 2Σ p j ), where n is the number of jobs and p j is the processing time of the jth job ( j = 1, 2, …, n).
Shi, Xuefei; Xu, Dehua
2014-01-01
We consider a single machine scheduling problem with multiple maintenance activities, where the maintenance duration function is of the linear form f(t) = a+bt with a ≥ 0 and b > 1. We propose an approximation algorithm named FFD-LS2I with a worst-case bound of 2 for problem. We also show that there is no polynomial time approximation algorithm with a worst-case bound less than 2 for the problem with b ≥ 0 unless P = NP, which implies that the FFD-LS2I algorithm is the best possible algorithm for the case b > 1 and that the FFD-LS algorithm, which is proposed in the literature, is the best possible algorithm for the case b ≤ 1 both from the worst-case bound point of view. PMID:24701177
AI techniques for a space application scheduling problem
NASA Technical Reports Server (NTRS)
Thalman, N.; Sparn, T.; Jaffres, L.; Gablehouse, D.; Judd, D.; Russell, C.
1991-01-01
Scheduling is a very complex optimization problem which can be categorized as an NP-complete problem. NP-complete problems are quite diverse, as are the algorithms used in searching for an optimal solution. In most cases, the best solutions that can be derived for these combinatorial explosive problems are near-optimal solutions. Due to the complexity of the scheduling problem, artificial intelligence (AI) can aid in solving these types of problems. Some of the factors are examined which make space application scheduling problems difficult and presents a fairly new AI-based technique called tabu search as applied to a real scheduling application. the specific problem is concerned with scheduling application. The specific problem is concerned with scheduling solar and stellar observations for the SOLar-STellar Irradiance Comparison Experiment (SOLSTICE) instrument in a constrained environment which produces minimum impact on the other instruments and maximizes target observation times. The SOLSTICE instrument will gly on-board the Upper Atmosphere Research Satellite (UARS) in 1991, and a similar instrument will fly on the earth observing system (Eos).
Handling Deafness Problem of Scheduled Multi-Channel Polling MACs
NASA Astrophysics Data System (ADS)
Jiang, Fulong; Liu, Hao; Shi, Longxing
Combining scheduled channel polling with channel diversity is a promising way for a MAC protocol to achieve high energy efficiency and performance under both light and heavy traffic conditions. However, the deafness problem may cancel out the benefit of channel diversity. In this paper, we first investigate the deafness problem of scheduled multi-channel polling MACs with experiments. Then we propose and evaluate two schemes to handle the deafness problem. Our experiment shows that deafness is a significant reason for performance degradation in scheduled multi-channel polling MACs. A proper scheme should be chosen depending on the traffic pattern and the design objective.
The application of artificial intelligence to astronomical scheduling problems
NASA Technical Reports Server (NTRS)
Johnston, Mark D.
1992-01-01
Efficient utilization of expensive space- and ground-based observatories is an important goal for the astronomical community; the cost of modern observing facilities is enormous, and the available observing time is much less than the demand from astronomers around the world. The complexity and variety of scheduling constraints and goals has led several groups to investigate how artificial intelligence (AI) techniques might help solve these kinds of problems. The earliest and most successful of these projects was started at Space Telescope Science Institute in 1987 and has led to the development of the Spike scheduling system to support the scheduling of Hubble Space Telescope (HST). The aim of Spike at STScI is to allocate observations to timescales of days to a week observing all scheduling constraints and maximizing preferences that help ensure that observations are made at optimal times. Spike has been in use operationally for HST since shortly after the observatory was launched in Apr. 1990. Although developed specifically for HST scheduling, Spike was carefully designed to provide a general framework for similar (activity-based) scheduling problems. In particular, the tasks to be scheduled are defined in the system in general terms, and no assumptions about the scheduling timescale are built in. The mechanisms for describing, combining, and propagating temporal and other constraints and preferences are quite general. The success of this approach has been demonstrated by the application of Spike to the scheduling of other satellite observatories: changes to the system are required only in the specific constraints that apply, and not in the framework itself. In particular, the Spike framework is sufficiently flexible to handle both long-term and short-term scheduling, on timescales of years down to minutes or less. This talk will discuss recent progress made in scheduling search techniques, the lessons learned from early HST operations, the application of Spike
Analysis of Feeder Bus Network Design and Scheduling Problems
Almasi, Mohammad Hadi; Karim, Mohamed Rehan
2014-01-01
A growing concern for public transit is its inability to shift passenger's mode from private to public transport. In order to overcome this problem, a more developed feeder bus network and matched schedules will play important roles. The present paper aims to review some of the studies performed on Feeder Bus Network Design and Scheduling Problem (FNDSP) based on three distinctive parts of the FNDSP setup, namely, problem description, problem characteristics, and solution approaches. The problems consist of different subproblems including data preparation, feeder bus network design, route generation, and feeder bus scheduling. Subsequently, descriptive analysis and classification of previous works are presented to highlight the main characteristics and solution methods. Finally, some of the issues and trends for future research are identified. This paper is targeted at dealing with the FNDSP to exhibit strategic and tactical goals and also contributes to the unification of the field which might be a useful complement to the few existing reviews. PMID:24526890
NASA Astrophysics Data System (ADS)
Konno, Yohko; Suzuki, Keiji
This paper describes an approach to development of a solution algorithm of a general-purpose for large scale problems using “Local Clustering Organization (LCO)” as a new solution for Job-shop scheduling problem (JSP). Using a performance effective large scale scheduling in the study of usual LCO, a solving JSP keep stability induced better solution is examined. In this study for an improvement of a performance of a solution for JSP, processes to a optimization by LCO is examined, and a scheduling solution-structure is extended to a new solution-structure based on machine-division. A solving method introduced into effective local clustering for the solution-structure is proposed as an extended LCO. An extended LCO has an algorithm which improves scheduling evaluation efficiently by clustering of parallel search which extends over plural machines. A result verified by an application of extended LCO on various scale of problems proved to conduce to minimizing make-span and improving on the stable performance.
Code of Federal Regulations, 2014 CFR
2014-07-01
... machine cards available from Federal Supply Schedule contracts. 101-26.509-1 Section 101-26.509-1 Public....509-1 Requisitioning tabulating machine cards available from Federal Supply Schedule contracts... electrical and mechanical contact tabulating machines, including aperture cards and copy cards....
Code of Federal Regulations, 2012 CFR
2012-07-01
... tabulating machine cards available from Federal Supply Schedule contracts. 101-26.509-1 Section 101-26.509-1... Programs § 101-26.509-1 Requisitioning tabulating machine cards available from Federal Supply Schedule... applicable to electrical and mechanical contact tabulating machines, including aperture cards and copy...
Code of Federal Regulations, 2013 CFR
2013-07-01
... machine cards available from Federal Supply Schedule contracts. 101-26.509-1 Section 101-26.509-1 Public....509-1 Requisitioning tabulating machine cards available from Federal Supply Schedule contracts... electrical and mechanical contact tabulating machines, including aperture cards and copy cards....
Code of Federal Regulations, 2011 CFR
2011-07-01
... machine cards available from Federal Supply Schedule contracts. 101-26.509-1 Section 101-26.509-1 Public....509-1 Requisitioning tabulating machine cards available from Federal Supply Schedule contracts... electrical and mechanical contact tabulating machines, including aperture cards and copy cards....
Meta-RaPS Algorithm for the Aerial Refueling Scheduling Problem
NASA Technical Reports Server (NTRS)
Kaplan, Sezgin; Arin, Arif; Rabadi, Ghaith
2011-01-01
The Aerial Refueling Scheduling Problem (ARSP) can be defined as determining the refueling completion times for each fighter aircraft (job) on multiple tankers (machines). ARSP assumes that jobs have different release times and due dates, The total weighted tardiness is used to evaluate schedule's quality. Therefore, ARSP can be modeled as a parallel machine scheduling with release limes and due dates to minimize the total weighted tardiness. Since ARSP is NP-hard, it will be more appropriate to develop a pproimate or heuristic algorithm to obtain solutions in reasonable computation limes. In this paper, Meta-Raps-ATC algorithm is implemented to create high quality solutions. Meta-RaPS (Meta-heuristic for Randomized Priority Search) is a recent and promising meta heuristic that is applied by introducing randomness to a construction heuristic. The Apparent Tardiness Rule (ATC), which is a good rule for scheduling problems with tardiness objective, is used to construct initial solutions which are improved by an exchanging operation. Results are presented for generated instances.
NASA Astrophysics Data System (ADS)
Guo, Peng; Cheng, Wenming; Wang, Yi
2015-11-01
This article considers the parallel machine scheduling problem with step-deteriorating jobs and sequence-dependent setup times. The objective is to minimize the total tardiness by determining the allocation and sequence of jobs on identical parallel machines. In this problem, the processing time of each job is a step function dependent upon its starting time. An individual extended time is penalized when the starting time of a job is later than a specific deterioration date. The possibility of deterioration of a job makes the parallel machine scheduling problem more challenging than ordinary ones. A mixed integer programming model for the optimal solution is derived. Due to its NP-hard nature, a hybrid discrete cuckoo search algorithm is proposed to solve this problem. In order to generate a good initial swarm, a modified Biskup-Hermann-Gupta (BHG) heuristic called MBHG is incorporated into the population initialization. Several discrete operators are proposed in the random walk of Lévy flights and the crossover search. Moreover, a local search procedure based on variable neighbourhood descent is integrated into the algorithm as a hybrid strategy in order to improve the quality of elite solutions. Computational experiments are executed on two sets of randomly generated test instances. The results show that the proposed hybrid algorithm can yield better solutions in comparison with the commercial solver CPLEX® with a one hour time limit, the discrete cuckoo search algorithm and the existing variable neighbourhood search algorithm.
NASA Astrophysics Data System (ADS)
Chang, Yung-Chia; Li, Vincent C.; Chiang, Chia-Ju
2014-04-01
Make-to-order or direct-order business models that require close interaction between production and distribution activities have been adopted by many enterprises in order to be competitive in demanding markets. This article considers an integrated production and distribution scheduling problem in which jobs are first processed by one of the unrelated parallel machines and then distributed to corresponding customers by capacitated vehicles without intermediate inventory. The objective is to find a joint production and distribution schedule so that the weighted sum of total weighted job delivery time and the total distribution cost is minimized. This article presents a mathematical model for describing the problem and designs an algorithm using ant colony optimization. Computational experiments illustrate that the algorithm developed is capable of generating near-optimal solutions. The computational results also demonstrate the value of integrating production and distribution in the model for the studied problem.
NASA Astrophysics Data System (ADS)
Sodan, Angela C.
Virtual machines have become an important approach to provide performance isolation and performance guarantees (QoS) on cluster servers and on many-core SMP servers. Many-core CPUs are a current trend in CPU design and require jobs to be parallel for exploitation of the performance potential. Very promising for batch job scheduling with virtual machines on both cluster servers and many-core SMP servers is adaptive scheduling which can adjust sizes of parallel jobs to consider different load situations and different resource availability. Then, the resource allocation and resource partitioning can be determined at virtual-machine level and be propagated down to the job sizes. The paper investigates job re-sizing and virtual-machine resizing, and the effects which the efficiency curve of the jobs has on the resulting performance. Additionally, the paper presents a simple, yet effective queuing-model approach for predicting performance under different resource allocation.
Innately Split Model for Job-shop Scheduling Problem
NASA Astrophysics Data System (ADS)
Ikeda, Kokolo; Kobayashi, Sigenobu
Job-shop Scheduling Problem (JSP) is one of the most difficult benchmark problems. GA approaches often fail searching the global optimum because of the deception UV-structure of JSPs. In this paper, we introduce a novel framework model of GA, Innately Split Model (ISM) which prevents UV-phenomenon, and discuss on its power particularly. Next we analyze the structure of JSPs with the help of the UV-structure hypothesys, and finally we show ISM's excellent performance on JSP.
Extended precedence preservative crossover for job shop scheduling problems
NASA Astrophysics Data System (ADS)
Ong, Chung Sin; Moin, Noor Hasnah; Omar, Mohd
2013-04-01
Job shop scheduling problems (JSSP) is one of difficult combinatorial scheduling problems. A wide range of genetic algorithms based on the two parents crossover have been applied to solve the problem but multi parents (more than two parents) crossover in solving the JSSP is still lacking. This paper proposes the extended precedence preservative crossover (EPPX) which uses multi parents for recombination in the genetic algorithms. EPPX is a variation of the precedence preservative crossover (PPX) which is one of the crossovers that perform well to find the solutions for the JSSP. EPPX is based on a vector to determine the gene selected in recombination for the next generation. Legalization of children (offspring) can be eliminated due to the JSSP representation encoded by using permutation with repetition that guarantees the feasibility of chromosomes. The simulations are performed on a set of benchmarks from the literatures and the results are compared to ensure the sustainability of multi parents recombination in solving the JSSP.
Solving Open Job-Shop Scheduling Problems by SAT Encoding
NASA Astrophysics Data System (ADS)
Koshimura, Miyuki; Nabeshima, Hidetomo; Fujita, Hiroshi; Hasegawa, Ryuzo
This paper tries to solve open Job-Shop Scheduling Problems (JSSP) by translating them into Boolean Satisfiability Testing Problems (SAT). The encoding method is essentially the same as the one proposed by Crawford and Baker. The open problems are ABZ8, ABZ9, YN1, YN2, YN3, and YN4. We proved that the best known upper bounds 678 of ABZ9 and 884 of YN1 are indeed optimal. We also improved the upper bound of YN2 and lower bounds of ABZ8, YN2, YN3 and YN4.
A Heuristic Approach for International Crude Oil Transportation Scheduling Problems
NASA Astrophysics Data System (ADS)
Yin, Sisi; Nishi, Tatsushi; Izuno, Tsukasa
In this paper, we propose a heuristic algorithm to solve a practical ship scheduling problem for international crude oil transportation. The problem is considered as a vehicle routing problem with split deliveries. The objective of this paper is to find an optimal assignment of tankers, a sequence of visiting and loading volume simultaneously in order to minimize the total distance satisfying the capacity of tankers. A savings-based meta-heuristic algorithm with lot sizing parameters and volume assignment heuristic is developed. The proposed method is applied to solve a case study with real data. Computational results demonstrate the effectiveness of the heuristic algorithm compared with that of human operators.
Artificial immune algorithm for multi-depot vehicle scheduling problems
NASA Astrophysics Data System (ADS)
Wu, Zhongyi; Wang, Donggen; Xia, Linyuan; Chen, Xiaoling
2008-10-01
In the fast-developing logistics and supply chain management fields, one of the key problems in the decision support system is that how to arrange, for a lot of customers and suppliers, the supplier-to-customer assignment and produce a detailed supply schedule under a set of constraints. Solutions to the multi-depot vehicle scheduling problems (MDVRP) help in solving this problem in case of transportation applications. The objective of the MDVSP is to minimize the total distance covered by all vehicles, which can be considered as delivery costs or time consumption. The MDVSP is one of nondeterministic polynomial-time hard (NP-hard) problem which cannot be solved to optimality within polynomial bounded computational time. Many different approaches have been developed to tackle MDVSP, such as exact algorithm (EA), one-stage approach (OSA), two-phase heuristic method (TPHM), tabu search algorithm (TSA), genetic algorithm (GA) and hierarchical multiplex structure (HIMS). Most of the methods mentioned above are time consuming and have high risk to result in local optimum. In this paper, a new search algorithm is proposed to solve MDVSP based on Artificial Immune Systems (AIS), which are inspirited by vertebrate immune systems. The proposed AIS algorithm is tested with 30 customers and 6 vehicles located in 3 depots. Experimental results show that the artificial immune system algorithm is an effective and efficient method for solving MDVSP problems.
Code of Federal Regulations, 2012 CFR
2012-10-01
... 49 Transportation 4 2012-10-01 2012-10-01 false On-track roadway maintenance machines; inspection... RAILROAD WORKPLACE SAFETY On-Track Roadway Maintenance Machines and Hi-Rail Vehicles § 214.527 On-track roadway maintenance machines; inspection for compliance and schedule for repairs. (a) The operator of...
Code of Federal Regulations, 2014 CFR
2014-10-01
... 49 Transportation 4 2014-10-01 2014-10-01 false On-track roadway maintenance machines; inspection... RAILROAD WORKPLACE SAFETY On-Track Roadway Maintenance Machines and Hi-Rail Vehicles § 214.527 On-track roadway maintenance machines; inspection for compliance and schedule for repairs. (a) The operator of...
Code of Federal Regulations, 2011 CFR
2011-10-01
... 49 Transportation 4 2011-10-01 2011-10-01 false On-track roadway maintenance machines; inspection... RAILROAD WORKPLACE SAFETY On-Track Roadway Maintenance Machines and Hi-Rail Vehicles § 214.527 On-track roadway maintenance machines; inspection for compliance and schedule for repairs. (a) The operator of...
Code of Federal Regulations, 2013 CFR
2013-10-01
... 49 Transportation 4 2013-10-01 2013-10-01 false On-track roadway maintenance machines; inspection... RAILROAD WORKPLACE SAFETY On-Track Roadway Maintenance Machines and Hi-Rail Vehicles § 214.527 On-track roadway maintenance machines; inspection for compliance and schedule for repairs. (a) The operator of...
Code of Federal Regulations, 2010 CFR
2010-10-01
... 49 Transportation 4 2010-10-01 2010-10-01 false On-track roadway maintenance machines; inspection... RAILROAD WORKPLACE SAFETY On-Track Roadway Maintenance Machines and Hi-Rail Vehicles § 214.527 On-track roadway maintenance machines; inspection for compliance and schedule for repairs. (a) The operator of...
Automated problem scheduling and reduction of synchronization delay effects
NASA Technical Reports Server (NTRS)
Saltz, Joel H.
1987-01-01
It is anticipated that in order to make effective use of many future high performance architectures, programs will have to exhibit at least a medium grained parallelism. A framework is presented for partitioning very sparse triangular systems of linear equations that is designed to produce favorable preformance results in a wide variety of parallel architectures. Efficient methods for solving these systems are of interest because: (1) they provide a useful model problem for use in exploring heuristics for the aggregation, mapping and scheduling of relatively fine grained computations whose data dependencies are specified by directed acrylic graphs, and (2) because such efficient methods can find direct application in the development of parallel algorithms for scientific computation. Simple expressions are derived that describe how to schedule computational work with varying degrees of granularity. The Encore Multimax was used as a hardware simulator to investigate the performance effects of using the partitioning techniques presented in shared memory architectures with varying relative synchronization costs.
Problems of vibroacoustical diagnostics of textile machines
NASA Technical Reports Server (NTRS)
Ivanov, L. N.; Pobol, O. N.; Gevorkyan, G. T.
1973-01-01
The application of a vibroacoustical method for analyzing the vibration characteristics of textile machines is discussed. A drawing of a typical mechanism is provided to show the sources of vibration. The conditions under which the machine operates are defined. An accelerometer is used as a sensor. The point of installation is selected so that a dip in the frequency characteristic of the vibrating part, on which the sensor is installed, in the region of the maximum acoustical signal. This allows for a drop in the vibrations for those values of the gap in the mechanism which amount to 10 to 20 db.
Efficiently Scheduling Multi-core Guest Virtual Machines on Multi-core Hosts in Network Simulation
Yoginath, Srikanth B; Perumalla, Kalyan S
2011-01-01
Virtual machine (VM)-based simulation is a method used by network simulators to incorporate realistic application behaviors by executing actual VMs as high-fidelity surrogates for simulated end-hosts. A critical requirement in such a method is the simulation time-ordered scheduling and execution of the VMs. Prior approaches such as time dilation are less efficient due to the high degree of multiplexing possible when multiple multi-core VMs are simulated on multi-core host systems. We present a new simulation time-ordered scheduler to efficiently schedule multi-core VMs on multi-core real hosts, with a virtual clock realized on each virtual core. The distinguishing features of our approach are: (1) customizable granularity of the VM scheduling time unit on the simulation time axis, (2) ability to take arbitrary leaps in virtual time by VMs to maximize the utilization of host (real) cores when guest virtual cores idle, and (3) empirically determinable optimality in the tradeoff between total execution (real) time and time-ordering accuracy levels. Experiments show that it is possible to get nearly perfect time-ordered execution, with a slight cost in total run time, relative to optimized non-simulation VM schedulers. Interestingly, with our time-ordered scheduler, it is also possible to reduce the time-ordering error from over 50% of non-simulation scheduler to less than 1% realized by our scheduler, with almost the same run time efficiency as that of the highly efficient non-simulation VM schedulers.
An Improved Differential Evolution Solution for Software Project Scheduling Problem
Biju, A. C.; Victoire, T. Aruldoss Albert; Mohanasundaram, Kumaresan
2015-01-01
This paper proposes a differential evolution (DE) method for the software project scheduling problem (SPSP). The interest on finding a more efficient solution technique for SPSP is always a topic of interest due to the fact of ever growing challenges faced by the software industry. The curse of dimensionality is introduced in the scheduling problem by ever increasing software assignments and the number of staff who handles it. Thus the SPSP is a class of NP-hard problem, which requires a rigorous solution procedure which guarantees a reasonably better solution. Differential evolution is a direct search stochastic optimization technique that is fairly fast and reasonably robust. It is also capable of handling nondifferentiable, nonlinear, and multimodal objective functions like SPSP. This paper proposes a refined DE where a new mutation mechanism is introduced. The superiority of the proposed method is experimented and demonstrated by solving the SPSP on 50 random instances and the results are compared with some of the techniques in the literature. PMID:26495419
Petri Net Modeling and Decomposition Method for Solving Production Scheduling Problems
NASA Astrophysics Data System (ADS)
Nishi, Tatsushi; Maeno, Ryota
Considering the need to develop general scheduling problem solver, the recent integration of Petri Nets as modeling tools into effective optimization methods for scheduling problems is very promising. The paper addresses a Petri Net modeling and decomposition method for solving a wide variety of scheduling problems. The scheduling problems are represented as the optimal transition firing sequence problems for timed Petri Nets. The Petri Net is decomposed into several subnets in which each subproblem can be easily solved by Dijkstra' algorithm. The approach is applied to a flowshop scheduling problem. The performance of the proposed algorithm is compared with that of a simulated annealing method.
Problem of technological inheritance in machine engineering
NASA Astrophysics Data System (ADS)
Blumenstein, Valery; Rakhimyanov, Kharis; Heifetz, Mikhail; Kleptzov, Alexander
2016-01-01
This article demonstrates the importance of the research study with regard to the technological inheritance of the properties, which characterize the surface layer, at different stages of a part's life cycle. It looks back at the major achievements and gives the findings relating to the technological inheritance of the parameters of the surface layer strength and quality as well as to how they affect the performance properties of machine parts. It demonstrates that high rates of machine engineering development, occurrence of new materials and more complicated machine operation environment require a shorter period for design-to-manufacture facility by reducing experiments and increasing design work. That, in its turn, generates the necessity in more complex but also more accurate models of metal behavior under stressing. It is especially critical for strengthening treatment. Among them are the models developed within the mechanics of technological inheritance. It is assumed that at the stages of a part's life cycle deformation accumulates on a continuous basis and the plasticity reserve of the metal, which the surface layer is made of, depletes. The research study of technological inheritance and the discovery of physical patterns of the evolution and degradation of the structures in a thin surface layer, which occur during machining and operational stressing of parts made from existing and unique including nanopatterned metals, is a crucial scientific challenge. This leads to the acquisition of new knowledge in the plasticity of state-of-the-art metals in the conditions of complex non monotonous stressing and to the development of efficient integrated and combined methods of technological impact.
Applications of dynamic scheduling technique to space related problems: Some case studies
NASA Technical Reports Server (NTRS)
Nakasuka, Shinichi; Ninomiya, Tetsujiro
1994-01-01
The paper discusses the applications of 'Dynamic Scheduling' technique, which has been invented for the scheduling of Flexible Manufacturing System, to two space related scheduling problems: operation scheduling of a future space transportation system, and resource allocation in a space system with limited resources such as space station or space shuttle.
Applications of dynamic scheduling technique to space related problems: Some case studies
NASA Astrophysics Data System (ADS)
Nakasuka, Shinichi; Ninomiya, Tetsujiro
1994-10-01
The paper discusses the applications of 'Dynamic Scheduling' technique, which has been invented for the scheduling of Flexible Manufacturing System, to two space related scheduling problems: operation scheduling of a future space transportation system, and resource allocation in a space system with limited resources such as space station or space shuttle.
Optimized Hypervisor Scheduler for Parallel Discrete Event Simulations on Virtual Machine Platforms
Yoginath, Srikanth B; Perumalla, Kalyan S
2013-01-01
With the advent of virtual machine (VM)-based platforms for parallel computing, it is now possible to execute parallel discrete event simulations (PDES) over multiple virtual machines, in contrast to executing in native mode directly over hardware as is traditionally done over the past decades. While mature VM-based parallel systems now offer new, compelling benefits such as serviceability, dynamic reconfigurability and overall cost effectiveness, the runtime performance of parallel applications can be significantly affected. In particular, most VM-based platforms are optimized for general workloads, but PDES execution exhibits unique dynamics significantly different from other workloads. Here we first present results from experiments that highlight the gross deterioration of the runtime performance of VM-based PDES simulations when executed using traditional VM schedulers, quantitatively showing the bad scaling properties of the scheduler as the number of VMs is increased. The mismatch is fundamental in nature in the sense that any fairness-based VM scheduler implementation would exhibit this mismatch with PDES runs. We also present a new scheduler optimized specifically for PDES applications, and describe its design and implementation. Experimental results obtained from running PDES benchmarks (PHOLD and vehicular traffic simulations) over VMs show over an order of magnitude improvement in the run time of the PDES-optimized scheduler relative to the regular VM scheduler, with over 20 reduction in run time of simulations using up to 64 VMs. The observations and results are timely in the context of emerging systems such as cloud platforms and VM-based high performance computing installations, highlighting to the community the need for PDES-specific support, and the feasibility of significantly reducing the runtime overhead for scalable PDES on VM platforms.
Assessment of New Load Schedules for the Machine Calibration of a Force Balance
NASA Technical Reports Server (NTRS)
Ulbrich, N.; Gisler, R.; Kew, R.
2015-01-01
New load schedules for the machine calibration of a six-component force balance are currently being developed and evaluated at the NASA Ames Balance Calibration Laboratory. One of the proposed load schedules is discussed in the paper. It has a total of 2082 points that are distributed across 16 load series. Several criteria were applied to define the load schedule. It was decided, for example, to specify the calibration load set in force balance format as this approach greatly simplifies the definition of the lower and upper bounds of the load schedule. In addition, all loads are assumed to be applied in a calibration machine by using the one-factor-at-a-time approach. At first, all single-component loads are applied in six load series. Then, three two-component load series are applied. They consist of the load pairs (N1, N2), (S1, S2), and (RM, AF). Afterwards, four three-component load series are applied. They consist of the combinations (N1, N2, AF), (S1, S2, AF), (N1, N2, RM), and (S1, S2, RM). In the next step, one four-component load series is applied. It is the load combination (N1, N2, S1, S2). Finally, two five-component load series are applied. They are the load combination (N1, N2, S1, S2, AF) and (N1, N2, S1, S2, RM). The maximum difference between loads of two subsequent data points of the load schedule is limited to 33 % of capacity. This constraint helps avoid unwanted load "jumps" in the load schedule that can have a negative impact on the performance of a calibration machine. Only loadings of the single- and two-component load series are loaded to 100 % of capacity. This approach was selected because it keeps the total number of calibration points to a reasonable limit while still allowing for the application of some of the more complex load combinations. Data from two of NASA's force balances is used to illustrate important characteristics of the proposed 2082-point calibration load schedule.
A Solution Method of Job-shop Scheduling Problems by the Idle Time Shortening Type Genetic Algorithm
NASA Astrophysics Data System (ADS)
Ida, Kenichi; Osawa, Akira
In this paper, we propose a new idle time shortening method for Job-shop scheduling problems (JSPs). We insert its method into a genetic algorithm (GA). The purpose of JSP is to find a schedule with the minimum makespan. We suppose that it is effective to reduce idle time of a machine in order to improve the makespan. The left shift is a famous algorithm in existing algorithms for shortening idle time. The left shift can not arrange the work to idle time. For that reason, some idle times are not shortened by the left shift. We propose two kinds of algorithms which shorten such idle time. Next, we combine these algorithms and the reversal of a schedule. We apply GA with its algorithm to benchmark problems and we show its effectiveness.
NASA Astrophysics Data System (ADS)
Ausaf, Muhammad Farhan; Gao, Liang; Li, Xinyu
2015-12-01
For increasing the overall performance of modern manufacturing systems, effective integration of process planning and scheduling functions has been an important area of consideration among researchers. Owing to the complexity of handling process planning and scheduling simultaneously, most of the research work has been limited to solving the integrated process planning and scheduling (IPPS) problem for a single objective function. As there are many conflicting objectives when dealing with process planning and scheduling, real world problems cannot be fully captured considering only a single objective for optimization. Therefore considering multi-objective IPPS (MOIPPS) problem is inevitable. Unfortunately, only a handful of research papers are available on solving MOIPPS problem. In this paper, an optimization algorithm for solving MOIPPS problem is presented. The proposed algorithm uses a set of dispatching rules coupled with priority assignment to optimize the IPPS problem for various objectives like makespan, total machine load, total tardiness, etc. A fixed sized external archive coupled with a crowding distance mechanism is used to store and maintain the non-dominated solutions. To compare the results with other algorithms, a C-matric based method has been used. Instances from four recent papers have been solved to demonstrate the effectiveness of the proposed algorithm. The experimental results show that the proposed method is an efficient approach for solving the MOIPPS problem.
Nurse Scheduling System based on Dynamic Weighted Maximal Constraint Satisfaction Problem
NASA Astrophysics Data System (ADS)
Hattori, Hiromitsu; Isomura, Atsushi; Ito, Takayuki; Ozono, Tadachika; Shintani, Toramatsu
Scheduling has been an important research field in Artificial Intelligence. Because typical scheduling problems could be modeled as a Constraint Satisfaction Problem(CSP), several constraint satisfaction techniques have been proposed. In order to handle the different levels of importance of the constraints, solving a problem as a Weighted Maximal Constraint Satisfaction Problem(W-MaxCSP) is an promising approach. However, there exists the case where unexpected events are added and some sudden changes are required, i.e., the case with dynamic changes in scheduling problems. In this paper, we describe such dynamic scheduling problem as a Dynamic Weighted Maximal Constraint Satisfaction Problem(DW-MaxCSP) in which constraints would changes dynamically. Generally, it is undesirable to determine vastly modified schedule even if re-scheduling is needed. A new schedule should be close to the current one as much as possible. In order to obtain stable solutions, we propose the methodology to maintain portions of the current schedule using the provisional soft constraints, which explicitly penalize the changes from the current schedule. We have experimentally confirmed the efficacy of re-scheduling based on our method with provisional constraints. In this paper, we construct the nurse scheduling system for applying the proposed scheduling method.
Genetic-Annealing Algorithm in Grid Environment for Scheduling Problems
NASA Astrophysics Data System (ADS)
Cruz-Chávez, Marco Antonio; Rodríguez-León, Abelardo; Ávila-Melgar, Erika Yesenia; Juárez-Pérez, Fredy; Cruz-Rosales, Martín H.; Rivera-López, Rafael
This paper presents a parallel hybrid evolutionary algorithm executed in a grid environment. The algorithm executes local searches using simulated annealing within a Genetic Algorithm to solve the job shop scheduling problem. Experimental results of the algorithm obtained in the "Tarantula MiniGrid" are shown. Tarantula was implemented by linking two clusters from different geographic locations in Mexico (Morelos-Veracruz). The technique used to link the two clusters and configure the Tarantula MiniGrid is described. The effects of latency in communication between the two clusters are discussed. It is shown that the evolutionary algorithm presented is more efficient working in Grid environments because it can carry out major exploration and exploitation of the solution space.
Application of TRIZ approach to machine vibration condition monitoring problems
NASA Astrophysics Data System (ADS)
Cempel, Czesław
2013-12-01
Up to now machine condition monitoring has not been seriously approached by TRIZ1TRIZ= Russian acronym for Inventive Problem Solving System, created by G. Altshuller ca 50 years ago. users, and the knowledge of TRIZ methodology has not been applied there intensively. However, there are some introductory papers of present author posted on Diagnostic Congress in Cracow (Cempel, in press [11]), and Diagnostyka Journal as well. But it seems to be further need to make such approach from different sides in order to see, if some new knowledge and technology will emerge. In doing this we need at first to define the ideal final result (IFR) of our innovation problem. As a next we need a set of parameters to describe the problems of system condition monitoring (CM) in terms of TRIZ language and set of inventive principles possible to apply, on the way to IFR. This means we should present the machine CM problem by means of contradiction and contradiction matrix. When specifying the problem parameters and inventive principles, one should use analogy and metaphorical thinking, which by definition is not exact but fuzzy, and leads sometimes to unexpected results and outcomes. The paper undertakes this important problem again and brings some new insight into system and machine CM problems. This may mean for example the minimal dimensionality of TRIZ engineering parameter set for the description of machine CM problems, and the set of most useful inventive principles applied to given engineering parameter and contradictions of TRIZ.
NASA Technical Reports Server (NTRS)
Smith, Stephen F.; Pathak, Dhiraj K.
1991-01-01
In this paper, we report work aimed at applying concepts of constraint-based problem structuring and multi-perspective scheduling to an over-subscribed scheduling problem. Previous research has demonstrated the utility of these concepts as a means for effectively balancing conflicting objectives in constraint-relaxable scheduling problems, and our goal here is to provide evidence of their similar potential in the context of HST observation scheduling. To this end, we define and experimentally assess the performance of two time-bounded heuristic scheduling strategies in balancing the tradeoff between resource setup time minimization and satisfaction of absolute time constraints. The first strategy considered is motivated by dispatch-based manufacturing scheduling research, and employs a problem decomposition that concentrates local search on minimizing resource idle time due to setup activities. The second is motivated by research in opportunistic scheduling and advocates a problem decomposition that focuses attention on the goal activities that have the tightest temporal constraints. Analysis of experimental results gives evidence of differential superiority on the part of each strategy in different problem solving circumstances. A composite strategy based on recognition of characteristics of the current problem solving state is then defined and tested to illustrate the potential benefits of constraint-based problem structuring and multi-perspective scheduling in over-subscribe scheduling problems.
NASA Astrophysics Data System (ADS)
Tang, Dunbing; Dai, Min
2015-09-01
The traditional production planning and scheduling problems consider performance indicators like time, cost and quality as optimization objectives in manufacturing processes. However, environmentally-friendly factors like energy consumption of production have not been completely taken into consideration. Against this background, this paper addresses an approach to modify a given schedule generated by a production planning and scheduling system in a job shop floor, where machine tools can work at different cutting speeds. It can adjust the cutting speeds of the operations while keeping the original assignment and processing sequence of operations of each job fixed in order to obtain energy savings. First, the proposed approach, based on a mixed integer programming mathematical model, changes the total idle time of the given schedule to minimize energy consumption in the job shop floor while accepting the optimal solution of the scheduling objective, makespan. Then, a genetic-simulated annealing algorithm is used to explore the optimal solution due to the fact that the problem is strongly NP-hard. Finally, the effectiveness of the approach is performed smalland large-size instances, respectively. The experimental results show that the approach can save 5%-10% of the average energy consumption while accepting the optimal solution of the makespan in small-size instances. In addition, the average maximum energy saving ratio can reach to 13%. And it can save approximately 1%-4% of the average energy consumption and approximately 2.4% of the average maximum energy while accepting the near-optimal solution of the makespan in large-size instances. The proposed research provides an interesting point to explore an energy-aware schedule optimization for a traditional production planning and scheduling problem.
Problems in modeling man machine control behavior in biodynamic environments
NASA Technical Reports Server (NTRS)
Jex, H. R.
1972-01-01
Reviewed are some current problems in modeling man-machine control behavior in a biodynamic environment. It is given in two parts: (1) a review of the models which are appropriate for manual control behavior and the added elements necessary to deal with biodynamic interfaces; and (2) a review of some biodynamic interface pilot/vehicle problems which have occurred, been solved, or need to be solved.
Exploiting the ALICE HLT for PROOF by scheduling of Virtual Machines
NASA Astrophysics Data System (ADS)
Meoni, Marco; Boettger, Stefan; Zelnicek, Pierre; Lindenstruth, Volker; Kebschull, Udo
2011-12-01
The HLT (High-Level Trigger) group of the ALICE experiment at the LHC has prepared a virtual Parallel ROOT Facility (PROOF) enabled cluster (HAF - HLT Analysis Facility) for fast physics analysis, detector calibration and reconstruction of data samples. The HLT-Cluster currently consists of 2860 CPU cores and 175TB of storage. Its purpose is the online filtering of the relevant part of data produced by the particle detector. However, data taking is not running continuously and exploiting unused cluster resources for other applications is highly desirable and improves the usage-cost ratio of the HLT cluster. As such, unused computing resources are dedicated to a PROOF-enabled virtual cluster available to the entire collaboration. This setup is especially aimed at the prototyping phase of analyses that need a high number of development iterations and a short response time, e.g. tuning of analysis cuts, calibration and alignment. HAF machines are enabled and disabled upon user request to start or complete analysis tasks. This is achieved by a virtual machine scheduling framework which dynamically assigns and migrates virtual machines running PROOF workers to unused physical resources. Using this approach we extend the HLT usage scheme to running both online and offline computing, thereby optimizing the resource usage.
Full Glowworm Swarm Optimization Algorithm for Whole-Set Orders Scheduling in Single Machine
Yu, Zhang; Yang, Xiaomei
2013-01-01
By analyzing the characteristics of whole-set orders problem and combining the theory of glowworm swarm optimization, a new glowworm swarm optimization algorithm for scheduling is proposed. A new hybrid-encoding schema combining with two-dimensional encoding and random-key encoding is given. In order to enhance the capability of optimal searching and speed up the convergence rate, the dynamical changed step strategy is integrated into this algorithm. Furthermore, experimental results prove its feasibility and efficiency. PMID:24294135
Discrete Bat Algorithm for Optimal Problem of Permutation Flow Shop Scheduling
Luo, Qifang; Zhou, Yongquan; Xie, Jian; Ma, Mingzhi; Li, Liangliang
2014-01-01
A discrete bat algorithm (DBA) is proposed for optimal permutation flow shop scheduling problem (PFSP). Firstly, the discrete bat algorithm is constructed based on the idea of basic bat algorithm, which divide whole scheduling problem into many subscheduling problems and then NEH heuristic be introduced to solve subscheduling problem. Secondly, some subsequences are operated with certain probability in the pulse emission and loudness phases. An intensive virtual population neighborhood search is integrated into the discrete bat algorithm to further improve the performance. Finally, the experimental results show the suitability and efficiency of the present discrete bat algorithm for optimal permutation flow shop scheduling problem. PMID:25243220
NASA Astrophysics Data System (ADS)
Li, Wenjie; Li, Shisheng
2015-03-01
We study the online batch scheduling of equal-length jobs on two identical batch machines. Each batch machine can process up to b jobs simultaneously as a batch (where b is called the capacity of the machines). The goal is to determine a schedule that maximises the (weighted) number of early jobs. For the non-preemptive model, we first present an upper bound that depends on the machine capacity b, and then we provide a greedy online algorithm with a competitive ratio of 1/(b + 1). For the preemption-restart model with b = ∞, we first show that no online algorithm has a competitive ratio greater than 0.595, and then we design an online algorithm with a competitive ratio of ?.
An Optimization Model for Scheduling Problems with Two-Dimensional Spatial Resource Constraint
NASA Technical Reports Server (NTRS)
Garcia, Christopher; Rabadi, Ghaith
2010-01-01
Traditional scheduling problems involve determining temporal assignments for a set of jobs in order to optimize some objective. Some scheduling problems also require the use of limited resources, which adds another dimension of complexity. In this paper we introduce a spatial resource-constrained scheduling problem that can arise in assembly, warehousing, cross-docking, inventory management, and other areas of logistics and supply chain management. This scheduling problem involves a twodimensional rectangular area as a limited resource. Each job, in addition to having temporal requirements, has a width and a height and utilizes a certain amount of space inside the area. We propose an optimization model for scheduling the jobs while respecting all temporal and spatial constraints.
Code of Federal Regulations, 2010 CFR
2010-07-01
... 41 Public Contracts and Property Management 2 2010-07-01 2010-07-01 true Requisitioning tabulating machine cards available from Federal Supply Schedule contracts. 101-26.509-1 Section 101-26.509-1 Public Contracts and Property Management Federal Property Management Regulations System FEDERAL PROPERTY MANAGEMENT REGULATIONS SUPPLY AND...
Code of Federal Regulations, 2010 CFR
2010-07-01
... 41 Public Contracts and Property Management 2 2010-07-01 2010-07-01 true Requisitioning tabulating machine cards not available from Federal Supply Schedule contracts. 101-26.509-2 Section 101-26.509-2 Public Contracts and Property Management Federal Property Management Regulations System FEDERAL PROPERTY MANAGEMENT REGULATIONS SUPPLY...
NASA Astrophysics Data System (ADS)
Bai, Danyu
2015-08-01
This paper discusses the flow shop scheduling problem to minimise the total quadratic completion time (TQCT) with release dates in offline and online environments. For this NP-hard problem, the investigation is focused on the performance of two online algorithms based on the Shortest Processing Time among Available jobs rule. Theoretical results indicate the asymptotic optimality of the algorithms as the problem scale is sufficiently large. To further enhance the quality of the original solutions, the improvement scheme is provided for these algorithms. A new lower bound with performance guarantee is provided, and computational experiments show the effectiveness of these heuristics. Moreover, several results of the single-machine TQCT problem with release dates are also obtained for the deduction of the main theorem.
Optimizing scheduling problem using an estimation of distribution algorithm and genetic algorithm
NASA Astrophysics Data System (ADS)
Qun, Jiang; Yang, Ou; Dong, Shi-Du
2007-12-01
This paper presents a methodology for using heuristic search methods to optimize scheduling problem. Specifically, an Estimation of Distribution Algorithm (EDA)- Population Based Incremental Learning (PBIL), and Genetic Algorithm (GA) have been applied to finding effective arrangement of curriculum schedule of Universities. To our knowledge, EDAs have been applied to fewer real world problems compared to GAs, and the goal of the present paper is to expand the application domain of this technique. The experimental results indicate a good applicability of PBIL to optimize scheduling problem.
The problem of scheduling for the linear section of a single-track railway
NASA Astrophysics Data System (ADS)
Akimova, Elena N.; Gainanov, Damir N.; Golubev, Oleg A.; Kolmogortsev, Ilya D.; Konygin, Anton V.
2016-06-01
The paper is devoted to the problem of scheduling for the linear section of a single-track railway: how to organize the flow in both directions in the most efficient way. In this paper, the authors propose an algorithm for scheduling, examine the properties of this algorithm and perform the computational experiments.
Electric power scheduling - A distributed problem-solving approach
NASA Technical Reports Server (NTRS)
Mellor, Pamela A.; Dolce, James L.; Krupp, Joseph C.
1990-01-01
Space Station Freedom's power system, along with the spacecraft's other subsystems, needs to carefully conserve its resources and yet strive to maximize overall Station productivity. Due to Freedom's distributed design, each subsystem must work cooperatively within the Station community. There is a need for a scheduling tool which will preserve this distributed structure, allow each subsystem the latitude to satisfy its own constraints, and preserve individual value systems while maintaining Station-wide integrity.
Electric power scheduling: A distributed problem-solving approach
NASA Technical Reports Server (NTRS)
Mellor, Pamela A.; Dolce, James L.; Krupp, Joseph C.
1990-01-01
Space Station Freedom's power system, along with the spacecraft's other subsystems, needs to carefully conserve its resources and yet strive to maximize overall Station productivity. Due to Freedom's distributed design, each subsystem must work cooperatively within the Station community. There is a need for a scheduling tool which will preserve this distributed structure, allow each subsystem the latitude to satisfy its own constraints, and preserve individual value systems while maintaining Station-wide integrity. The value-driven free-market economic model is such a tool.
Solving a real-world problem using an evolving heuristically driven schedule builder.
Hart, E; Ross, P; Nelson, J
1998-01-01
This work addresses the real-life scheduling problem of a Scottish company that must produce daily schedules for the catching and transportation of large numbers of live chickens. The problem is complex and highly constrained. We show that it can be successfully solved by division into two subproblems and solving each using a separate genetic algorithm (GA). We address the problem of whether this produces locally optimal solutions and how to overcome this. We extend the traditional approach of evolving a "permutation + schedule builder" by concentrating on evolving the schedule builder itself. This results in a unique schedule builder being built for each daily scheduling problem, each individually tailored to deal with the particular features of that problem. This results in a robust, fast, and flexible system that can cope with most of the circumstances imaginable at the factory. We also compare the performance of a GA approach to several other evolutionary methods and show that population-based methods are superior to both hill-climbing and simulated annealing in the quality of solutions produced. Population-based methods also have the distinct advantage of producing multiple, equally fit solutions, which is of particular importance when considering the practical aspects of the problem. PMID:10021741
Solving scheduling tournament problems using a new version of CLONALG
NASA Astrophysics Data System (ADS)
Pérez-Cáceres, Leslie; Riff, María Cristina
2015-01-01
The travelling tournament problem (TTP) is an important and well-known problem within the collective sports research community. The problem is NP-hard which makes difficult finding quality solution in short amount of time. Recently a new kind of TTP has been proposed 'The Relaxed Travelling Tournament Problem'. This version of the problem allows teams to have some days off during the tournament. In this paper, we propose an immune algorithm that is able to solve both problem versions. The algorithm uses moves which are based on the team home/away patterns. One of these moves has been specially designed for the relaxed travel tournament instances. We have tested the algorithm using well-known problem benchmarks and the results obtained are very encouraging.
Mapping unstructured grid problems to the connection machine
NASA Technical Reports Server (NTRS)
Hammond, Steven W.; Schreiber, Robert
1990-01-01
We present a highly parallel graph mapping technique that enables one to solve unstructured grid problems on massively parallel computers. Many implicit and explicit methods for solving discretizated partial differential equations require each point in the discretization to exchange data with its neighboring points every time step or iteration. The time spent communicating can limit the high performance promised by massively parallel computing. To eliminate this bottleneck, we map the graph of the irregular problem to the graph representing the interconnection topology of the computer such that the sum of the distances that the messages travel is minimized. We show that, in comparison to a naive assignment of processors, our heuristic mapping algorithm significantly reduces the communication time on the Connection Machine, CM-2.
Three-index Model for Westenberger-Kallrath Benchmark Scheduling Problem
NASA Astrophysics Data System (ADS)
Vooradi, Ramsagar; Shaik, Munawar A.; Gupta, Nikhil M.
2010-10-01
Short-term scheduling of batch operations has become an important research area in the last two decades. Recently Shaik and Floudas (2009) proposed a novel unified model for short-term scheduling using unit-specific event based continuous time representation employing three-index binary and continuous variables. In this work, we extend this three index model to solve a challenging benchmark problem from the scheduling literature that covers most of the features contributing to the complexity of batch process scheduling in industry. In order to implement the problem, new sets of constraints and modifications are incorporated into the three-index model. The different demand instances of the benchmark problem have been solved using the developed model and the results are compared with the literature to demonstrate the effectiveness of the proposed three-index model.
On scheduling models for the frequency interval assignment problem with cumulative interferences
NASA Astrophysics Data System (ADS)
Kiatmanaroj, Kata; Artigues, Christian; Houssin, Laurent
2016-05-01
In this article, models and methods for solving a real-life frequency assignment problem based on scheduling theory are investigated. A realistic frequency assignment problem involving cumulative interference constraints in which the aim is to maximize the number of assigned users is considered. If interferences are assumed to be binary, a multiple carrier frequency assignment problem can be treated as a disjunctive scheduling problem since a user requesting a number of contiguous frequencies can be considered as a non-preemptive task with a processing time, and two interfering users can be modelled through a disjunctive constraint on the corresponding tasks. A binary interference version of the problem is constructed and a disjunctive scheduling model is derived. Based on the binary representation, two models are proposed. The first one relies on an interference matrix and the second one considers maximal cliques. A third, cumulative, model that yields a new class of scheduling problems is also proposed. Computational experiments show that the case-study frequency assignment problem can be solved efficiently with disjunctive scheduling techniques.
A modify ant colony optimization for the grid jobs scheduling problem with QoS requirements
NASA Astrophysics Data System (ADS)
Pu, Xun; Lu, XianLiang
2011-10-01
Job scheduling with customers' quality of service (QoS) requirement is challenging in grid environment. In this paper, we present a modify Ant colony optimization (MACO) for the Job scheduling problem in grid. Instead of using the conventional construction approach to construct feasible schedules, the proposed algorithm employs a decomposition method to satisfy the customer's deadline and cost requirements. Besides, a new mechanism of service instances state updating is embedded to improve the convergence of MACO. Experiments demonstrate the effectiveness of the proposed algorithm.
Neighbourhood generation mechanism applied in simulated annealing to job shop scheduling problems
NASA Astrophysics Data System (ADS)
Cruz-Chávez, Marco Antonio
2015-11-01
This paper presents a neighbourhood generation mechanism for the job shop scheduling problems (JSSPs). In order to obtain a feasible neighbour with the generation mechanism, it is only necessary to generate a permutation of an adjacent pair of operations in a scheduling of the JSSP. If there is no slack time between the adjacent pair of operations that is permuted, then it is proven, through theory and experimentation, that the new neighbour (schedule) generated is feasible. It is demonstrated that the neighbourhood generation mechanism is very efficient and effective in a simulated annealing.
ERIC Educational Resources Information Center
Tsakanikos, Elias; Underwood, Lisa; Sturmey, Peter; Bouras, Nick; McCarthy, Jane
2011-01-01
The present study employed the Disability Assessment Schedule (DAS) to assess problem behaviors in a large sample of adults with ID (N = 568) and evaluate the psychometric properties of this instrument. Although the DAS problem behaviors were found to be internally consistent (Cronbach's [alpha] = 0.87), item analysis revealed one weak item…
NASA Technical Reports Server (NTRS)
Wang, Lui; Valenzuela-Rendon, Manuel
1993-01-01
The Space Station Freedom will require the supply of items in a regular fashion. A schedule for the delivery of these items is not easy to design due to the large span of time involved and the possibility of cancellations and changes in shuttle flights. This paper presents the basic concepts of a genetic algorithm model, and also presents the results of an effort to apply genetic algorithms to the design of propellant resupply schedules. As part of this effort, a simple simulator and an encoding by which a genetic algorithm can find near optimal schedules have been developed. Additionally, this paper proposes ways in which robust schedules, i.e., schedules that can tolerate small changes, can be found using genetic algorithms.
Scheduling Earth Observing Fleets Using Evolutionary Algorithms: Problem Description and Approach
NASA Technical Reports Server (NTRS)
Globus, Al; Crawford, James; Lohn, Jason; Morris, Robert; Clancy, Daniel (Technical Monitor)
2002-01-01
We describe work in progress concerning multi-instrument, multi-satellite scheduling. Most, although not all, Earth observing instruments currently in orbit are unique. In the relatively near future, however, we expect to see fleets of Earth observing spacecraft, many carrying nearly identical instruments. This presents a substantially new scheduling challenge. Inspired by successful commercial applications of evolutionary algorithms in scheduling domains, this paper presents work in progress regarding the use of evolutionary algorithms to solve a set of Earth observing related model problems. Both the model problems and the software are described. Since the larger problems will require substantial computation and evolutionary algorithms are embarrassingly parallel, we discuss our parallelization techniques using dedicated and cycle-scavenged workstations.
Problem-Solving at a Circuit-Board Assembly Machine: A Microanalysis.
ERIC Educational Resources Information Center
Kleifgen, Jo Anne; Frenz-Belken, Patricia
A study described machine operators' problem-solving actions at a computerized circuit-board assembly machine in a small manufacturing plant located on the West Coast. Participants were a machine operator and his supervisor, both from Vietnam, who were building large prototype boards for a major computer corporation. Over a 6.5 minute interval,…
A modified genetic algorithm with fuzzy roulette wheel selection for job-shop scheduling problems
NASA Astrophysics Data System (ADS)
Thammano, Arit; Teekeng, Wannaporn
2015-05-01
The job-shop scheduling problem is one of the most difficult production planning problems. Since it is in the NP-hard class, a recent trend in solving the job-shop scheduling problem is shifting towards the use of heuristic and metaheuristic algorithms. This paper proposes a novel metaheuristic algorithm, which is a modification of the genetic algorithm. This proposed algorithm introduces two new concepts to the standard genetic algorithm: (1) fuzzy roulette wheel selection and (2) the mutation operation with tabu list. The proposed algorithm has been evaluated and compared with several state-of-the-art algorithms in the literature. The experimental results on 53 JSSPs show that the proposed algorithm is very effective in solving the combinatorial optimization problems. It outperforms all state-of-the-art algorithms on all benchmark problems in terms of the ability to achieve the optimal solution and the computational time.
Solving multi-objective job shop scheduling problems using a non-dominated sorting genetic algorithm
NASA Astrophysics Data System (ADS)
Piroozfard, Hamed; Wong, Kuan Yew
2015-05-01
The efforts of finding optimal schedules for the job shop scheduling problems are highly important for many real-world industrial applications. In this paper, a multi-objective based job shop scheduling problem by simultaneously minimizing makespan and tardiness is taken into account. The problem is considered to be more complex due to the multiple business criteria that must be satisfied. To solve the problem more efficiently and to obtain a set of non-dominated solutions, a meta-heuristic based non-dominated sorting genetic algorithm is presented. In addition, task based representation is used for solution encoding, and tournament selection that is based on rank and crowding distance is applied for offspring selection. Swapping and insertion mutations are employed to increase diversity of population and to perform intensive search. To evaluate the modified non-dominated sorting genetic algorithm, a set of modified benchmarking job shop problems obtained from the OR-Library is used, and the results are considered based on the number of non-dominated solutions and quality of schedules obtained by the algorithm.
Active Solution Space and Search on Job-shop Scheduling Problem
NASA Astrophysics Data System (ADS)
Watanabe, Masato; Ida, Kenichi; Gen, Mitsuo
In this paper we propose a new searching method of Genetic Algorithm for Job-shop scheduling problem (JSP). The coding method that represent job number in order to decide a priority to arrange a job to Gannt Chart (called the ordinal representation with a priority) in JSP, an active schedule is created by using left shift. We define an active solution at first. It is solution which can create an active schedule without using left shift, and set of its defined an active solution space. Next, we propose an algorithm named Genetic Algorithm with active solution space search (GA-asol) which can create an active solution while solution is evaluated, in order to search the active solution space effectively. We applied it for some benchmark problems to compare with other method. The experimental results show good performance.
NASA Astrophysics Data System (ADS)
Marwati, Rini; Yulianti, Kartika; Pangestu, Herny Wulandari
2016-02-01
A fuzzy evolutionary algorithm is an integration of an evolutionary algorithm and a fuzzy system. In this paper, we present an application of a genetic algorithm to a fuzzy evolutionary algorithm to detect and to solve chromosomes conflict. A chromosome conflict is identified by existence of any two genes in a chromosome that has the same values as two genes in another chromosome. Based on this approach, we construct an algorithm to solve a lecture scheduling problem. Time codes, lecture codes, lecturer codes, and room codes are defined as genes. They are collected to become chromosomes. As a result, the conflicted schedule turns into chromosomes conflict. Built in the Delphi program, results show that the conflicted lecture schedule problem is solvable by this algorithm.
An information theoretic view of the scheduling problem in whole-body CAD
NASA Astrophysics Data System (ADS)
Zhan, Yiqiang; Zhou, Xiang Sean; Krishnan, Arun
2008-03-01
Emerging whole-body imaging technologies push computer aided detection/diagnosis (CAD) to scale up to a whole-body level, which involves multiple organs or anatomical structure. To be exploited in this paper is the fact that the various tasks in whole-body CAD are often highly dependent (e.g., the localization of the femur heads strongly predicts the position of the iliac bifurcation of the aorta). One way to effectively employ task dependency is to schedule the tasks such that outputs of some tasks are used to guide the others. In this sense, optimal task scheduling is key to improve overall performance of a whole-body CAD system. In this paper, we propose a method for task scheduling that is optimal in an information-theoretic sense. The central idea is to schedule tasks in such an order that each operation achieves maximum expected information gain over all the tasks. The formulation embeds two intuitive principles: (1) a task with higher confidence tends to be scheduled earlier; (2) a task with higher predictive power for other tasks tends to be scheduled earlier. More specifically, task dependency is modeled by conditional probability; the outcome of each task is assumed to be probabilistic as well; and the objective function is based on the reduction of the summed conditional entropy over all tasks. The validation is carried out on a challenging CAD problem, multi-organ localization in whole-body CT. Compared to unscheduled and ad hoc scheduled organ detection/localization, our scheduled execution achieves higher accuracy with much less computation time.
ERIC Educational Resources Information Center
Borrero, Carrie S. W.; Vollmer, Timothy R.; Borrero, John C.; Bourret, Jason C.; Sloman, Kimberly N.; Samaha, Andrew L.; Dallery, Jesse
2010-01-01
This study evaluated how children who exhibited functionally equivalent problem and appropriate behavior allocate responding to experimentally arranged reinforcer rates. Relative reinforcer rates were arranged on concurrent variable-interval schedules and effects on relative response rates were interpreted using the generalized matching equation.…
Manipulating Slot Machine Preference in Problem Gamblers through Contextual Control
ERIC Educational Resources Information Center
Nastally, Becky L.; Dixon, Mark R.; Jackson, James W.
2010-01-01
Pathological and nonpathological gamblers completed a task that assessed preference among 2 concurrently available slot machines. Subsequent assessments of choice were conducted after various attempts to transfer contextual functions associated with irrelevant characteristics of the slot machines. Results indicated that the nonproblem gambling…
Exact and Metaheuristic Approaches for a Bi-Objective School Bus Scheduling Problem
Chen, Xiaopan; Kong, Yunfeng; Dang, Lanxue; Hou, Yane; Ye, Xinyue
2015-01-01
As a class of hard combinatorial optimization problems, the school bus routing problem has received considerable attention in the last decades. For a multi-school system, given the bus trips for each school, the school bus scheduling problem aims at optimizing bus schedules to serve all the trips within the school time windows. In this paper, we propose two approaches for solving the bi-objective school bus scheduling problem: an exact method of mixed integer programming (MIP) and a metaheuristic method which combines simulated annealing with local search. We develop MIP formulations for homogenous and heterogeneous fleet problems respectively and solve the models by MIP solver CPLEX. The bus type-based formulation for heterogeneous fleet problem reduces the model complexity in terms of the number of decision variables and constraints. The metaheuristic method is a two-stage framework for minimizing the number of buses to be used as well as the total travel distance of buses. We evaluate the proposed MIP and the metaheuristic method on two benchmark datasets, showing that on both instances, our metaheuristic method significantly outperforms the respective state-of-the-art methods. PMID:26176764
Exact and Metaheuristic Approaches for a Bi-Objective School Bus Scheduling Problem.
Chen, Xiaopan; Kong, Yunfeng; Dang, Lanxue; Hou, Yane; Ye, Xinyue
2015-01-01
As a class of hard combinatorial optimization problems, the school bus routing problem has received considerable attention in the last decades. For a multi-school system, given the bus trips for each school, the school bus scheduling problem aims at optimizing bus schedules to serve all the trips within the school time windows. In this paper, we propose two approaches for solving the bi-objective school bus scheduling problem: an exact method of mixed integer programming (MIP) and a metaheuristic method which combines simulated annealing with local search. We develop MIP formulations for homogenous and heterogeneous fleet problems respectively and solve the models by MIP solver CPLEX. The bus type-based formulation for heterogeneous fleet problem reduces the model complexity in terms of the number of decision variables and constraints. The metaheuristic method is a two-stage framework for minimizing the number of buses to be used as well as the total travel distance of buses. We evaluate the proposed MIP and the metaheuristic method on two benchmark datasets, showing that on both instances, our metaheuristic method significantly outperforms the respective state-of-the-art methods. PMID:26176764
NASA Astrophysics Data System (ADS)
Wang, Deyun; Grunder, Olivier; EL Moudni, Abdellah
2014-08-01
This paper considers an integrated lot sizing and scheduling problem for a production-distribution environment with arbitrary job volumes and distinct due dates considerations. In the problem, jobs are firstly batch processed on a batching machine at production stage and then delivered to a pre-specified customer at the subsequent delivery stage by a capacitated vehicle. Each job is associated with a distinct due date and a distinct volume, and has to be delivered to the customer before its due date, i.e. delay is not allowed. The processing time of a batch is a constant independent of the jobs it contains. In production, a constant set-up time as well as a constant set-up cost is required before the first job of this batch is processed. In delivery, a constant delivery time as well as a constant delivery cost is needed for each round-trip delivery between the factory and the customer. Moreover, it is supposed that a job that arrives at the customer before its due date will incur a customer inventory cost. The objective is to find a coordinated lot sizing and scheduling scheme such that the total cost is minimised while guaranteeing a certain customer service level. A mixed integer formulation is proposed for this problem, and then a genetic algorithm is developed to solve it. To evaluate the performance of the proposed genetic algorithm, a lower bound on the objective value is established. Computational experiments show that the proposed genetic algorithm performs well on randomly generated problem instances.
Morrow, Daniel; Raquel, Liza; Schriver, Angela; Redenbo, Seth; Rozovski, David; Weiss, Gillian
2008-09-01
Taking medication requires developing plans to accomplish the activity. This planning challenges older adults because of age-related cognitive limits and inadequate collaboration with health providers. The authors investigated whether an external aid (medtable) supports collaborative planning in the context of a simulated patient/provider task in which pairs of older adults worked together to create medication schedules. Experiment 1 compared pairs who used the medtable, blank paper (unstructured aid), or no aid to create schedules varying in complexity of medication constraints (number of medications and medication co-occurrence restrictions) and patient constraints (available times during the day to take medication). Both aids increased problem-solving accuracy and efficiency (time per unit accuracy) compared to the no-aid condition, primarily for more complex schedules. However, benefits were similar for the two aids. In Experiment 2, a redesigned medtable increased problem-solving accuracy and efficiency compared to blank paper. Both aids presumably supported problem solving by providing a jointly visible workspace for developing schedules. The medtable may be more effective because it externalizes constraints (relationships between medication and patient information), so that participants can more easily organize information. PMID:18808282
Minimizing conflicts: A heuristic repair method for constraint-satisfaction and scheduling problems
NASA Technical Reports Server (NTRS)
Minton, Steve; Johnston, Mark; Philips, Andrew; Laird, Phil
1992-01-01
This paper describes a simple heuristic approach to solving large-scale constraint satisfaction and scheduling problems. In this approach one starts with an inconsistent assignment for a set of variables and searches through the space of possible repairs. The search can be guided by a value-ordering heuristic, the min-conflicts heuristic, that attempts to minimize the number of constraint violations after each step. The heuristic can be used with a variety of different search strategies. We demonstrate empirically that on the n-queens problem, a technique based on this approach performs orders of magnitude better than traditional backtracking techniques. We also describe a scheduling application where the approach has been used successfully. A theoretical analysis is presented both to explain why this method works well on certain types of problems and to predict when it is likely to be most effective.
Genetic Algorithm to minimize flowtime in a no-wait flowshop scheduling problem
NASA Astrophysics Data System (ADS)
Chaudhry, Imran A.; Ahmed, Riaz; Munem Khan, Abdul
2014-07-01
No-wait flowshop is an important scheduling environment having application in many industries. This paper addresses a no-wait flowshop scheduling problem, where the objective function is to minimise total flowtime. A Genetic Algorithm (GA) optimization approach implemented in a spreadsheet environment is suggested to solve this important class of problem. The proposed algorithm employs a general purpose genetic algorithm which can be customised with ease to address any objective function without modifying the optimization routine. Performance of the proposed approach is compared with eight previously reported algorithms for two sets of benchmark problems. Experimental analysis shows that the performance of the suggested approach is comparable with earlier approaches in terms of quality of solution.
ERIC Educational Resources Information Center
Waters, Melissa B.; Lerman, Dorothea C.; Hovanetz, Alyson N.
2009-01-01
The separate and combined effects of visual schedules and extinction plus differential reinforcement of other behavior (DRO) were evaluated to decrease transition-related problem behavior of 2 children diagnosed with autism. Visual schedules alone were ineffective in reducing problem behavior when transitioning from preferred to nonpreferred…
NASA Astrophysics Data System (ADS)
Ramli, Razamin; Tein, Lim Huai
2016-08-01
A good work schedule can improve hospital operations by providing better coverage with appropriate staffing levels in managing nurse personnel. Hence, constructing the best nurse work schedule is the appropriate effort. In doing so, an improved selection operator in the Evolutionary Algorithm (EA) strategy for a nurse scheduling problem (NSP) is proposed. The smart and efficient scheduling procedures were considered. Computation of the performance of each potential solution or schedule was done through fitness evaluation. The best so far solution was obtained via special Maximax&Maximin (MM) parent selection operator embedded in the EA, which fulfilled all constraints considered in the NSP.
Enhancements of evolutionary algorithm for the complex requirements of a nurse scheduling problem
NASA Astrophysics Data System (ADS)
Tein, Lim Huai; Ramli, Razamin
2014-12-01
Over the years, nurse scheduling is a noticeable problem that is affected by the global nurse turnover crisis. The more nurses are unsatisfied with their working environment the more severe the condition or implication they tend to leave. Therefore, the current undesirable work schedule is partly due to that working condition. Basically, there is a lack of complimentary requirement between the head nurse's liability and the nurses' need. In particular, subject to highly nurse preferences issue, the sophisticated challenge of doing nurse scheduling is failure to stimulate tolerance behavior between both parties during shifts assignment in real working scenarios. Inevitably, the flexibility in shifts assignment is hard to achieve for the sake of satisfying nurse diverse requests with upholding imperative nurse ward coverage. Hence, Evolutionary Algorithm (EA) is proposed to cater for this complexity in a nurse scheduling problem (NSP). The restriction of EA is discussed and thus, enhancement on the EA operators is suggested so that the EA would have the characteristic of a flexible search. This paper consists of three types of constraints which are the hard, semi-hard and soft constraints that can be handled by the EA with enhanced parent selection and specialized mutation operators. These operators and EA as a whole contribute to the efficiency of constraint handling, fitness computation as well as flexibility in the search, which correspond to the employment of exploration and exploitation principles.
Duan, Qian-Qian; Yang, Gen-Ke; Pan, Chang-Chun
2014-01-01
A hybrid optimization algorithm combining finite state method (FSM) and genetic algorithm (GA) is proposed to solve the crude oil scheduling problem. The FSM and GA are combined to take the advantage of each method and compensate deficiencies of individual methods. In the proposed algorithm, the finite state method makes up for the weakness of GA which is poor at local searching ability. The heuristic returned by the FSM can guide the GA algorithm towards good solutions. The idea behind this is that we can generate promising substructure or partial solution by using FSM. Furthermore, the FSM can guarantee that the entire solution space is uniformly covered. Therefore, the combination of the two algorithms has better global performance than the existing GA or FSM which is operated individually. Finally, a real-life crude oil scheduling problem from the literature is used for conducting simulation. The experimental results validate that the proposed method outperforms the state-of-art GA method. PMID:24772031
A Novel Joint Problem of Routing, Scheduling, and Variable-Width Channel Allocation in WMNs
Liu, Wan-Yu; Chou, Chun-Hung
2014-01-01
This paper investigates a novel joint problem of routing, scheduling, and channel allocation for single-radio multichannel wireless mesh networks in which multiple channel widths can be adjusted dynamically through a new software technology so that more concurrent transmissions and suppressed overlapping channel interference can be achieved. Although the previous works have studied this joint problem, their linear programming models for the problem were not incorporated with some delicate constraints. As a result, this paper first constructs a linear programming model with more practical concerns and then proposes a simulated annealing approach with a novel encoding mechanism, in which the configurations of multiple time slots are devised to characterize the dynamic transmission process. Experimental results show that our approach can find the same or similar solutions as the optimal solutions for smaller-scale problems and can efficiently find good-quality solutions for a variety of larger-scale problems. PMID:24982990
NASA Astrophysics Data System (ADS)
Bai, Danyu; Zhang, Zhihai
2014-08-01
This article investigates the open-shop scheduling problem with the optimal criterion of minimising the sum of quadratic completion times. For this NP-hard problem, the asymptotic optimality of the shortest processing time block (SPTB) heuristic is proven in the sense of limit. Moreover, three different improvements, namely, the job-insert scheme, tabu search and genetic algorithm, are introduced to enhance the quality of the original solution generated by the SPTB heuristic. At the end of the article, a series of numerical experiments demonstrate the convergence of the heuristic, the performance of the improvements and the effectiveness of the quadratic objective.
NASA Astrophysics Data System (ADS)
Zhu, Li; He, Yongxiang; Xue, Haidong; Chen, Leichen
Traditional genetic algorithms (GA) displays a disadvantage of early-constringency in dealing with scheduling problem. To improve the crossover operators and mutation operators self-adaptively, this paper proposes a self-adaptive GA at the target of multitask scheduling optimization under limited resources. The experiment results show that the proposed algorithm outperforms the traditional GA in evolutive ability to deal with complex task scheduling optimization.
NASA Astrophysics Data System (ADS)
Jafari, Hamed; Salmasi, Nasser
2015-04-01
The nurse scheduling problem (NSP) has received a great amount of attention in recent years. In the NSP, the goal is to assign shifts to the nurses in order to satisfy the hospital's demand during the planning horizon by considering different objective functions. In this research, we focus on maximizing the nurses' preferences for working shifts and weekends off by considering several important factors such as hospital's policies, labor laws, governmental regulations, and the status of nurses at the end of the previous planning horizon in one of the largest hospitals in Iran i.e., Milad Hospital. Due to the shortage of available nurses, at first, the minimum total number of required nurses is determined. Then, a mathematical programming model is proposed to solve the problem optimally. Since the proposed research problem is NP-hard, a meta-heuristic algorithm based on simulated annealing (SA) is applied to heuristically solve the problem in a reasonable time. An initial feasible solution generator and several novel neighborhood structures are applied to enhance performance of the SA algorithm. Inspired from our observations in Milad hospital, random test problems are generated to evaluate the performance of the SA algorithm. The results of computational experiments indicate that the applied SA algorithm provides solutions with average percentage gap of 5.49 % compared to the upper bounds obtained from the mathematical model. Moreover, the applied SA algorithm provides significantly better solutions in a reasonable time than the schedules provided by the head nurses.
A time-shared machine repair problem with mixed spares under N-policy
NASA Astrophysics Data System (ADS)
Jain, Madhu; Shekhar, Chandra; Shukla, Shalini
2016-02-01
The present investigation deals with a machine repair problem consisting of cold and warm standby machines. The machines are subject to breakdown and are repaired by the permanent repairman operating under N-policy. There is provision of one additional removable repairman who is called upon when the work load of failed machines crosses a certain threshold level and is removed as soon as the work load again ceases to that level. Both repairmen recover the failed machines by following the time sharing concept which means that the repairmen share their repair job simultaneously among all the failed machines that have joined the system for repair. Markovian model has been developed by considering the queue dependent rates and solved analytically using the recursive technique. Various performance indices are derived which are further used to obtain the cost function. By taking illustration, numerical simulation and sensitivity analysis have been provided.
Minimizing the total completion time in a two-machine flowshop problem with time delays
NASA Astrophysics Data System (ADS)
Kais Msakni, Mohamed; Khallouli, Wael; Al-Salem, Mohamed; Ladhari, Talel
2016-07-01
This article proposes to solve the problem of minimizing the total completion time in a two-machine permutation flowshop environment in which time delays between the machines are considered. For this purpose, an enumeration algorithm based on the branch-and-bound framework is developed, which includes new lower and upper bounds as well as dominance rules. The computational study shows that problems with up to 40 jobs can be solved in a reasonable amount of time.
A Hybrid Cellular Genetic Algorithm for Multi-objective Crew Scheduling Problem
NASA Astrophysics Data System (ADS)
Jolai, Fariborz; Assadipour, Ghazal
Crew scheduling is one of the important problems of the airline industry. This problem aims to cover a number of flights by crew members, such that all the flights are covered. In a robust scheduling the assignment should be so that the total cost, delays, and unbalanced utilization are minimized. As the problem is NP-hard and the objectives are in conflict with each other, a multi-objective meta-heuristic called CellDE, which is a hybrid cellular genetic algorithm, is implemented as the optimization method. The proposed algorithm provides the decision maker with a set of non-dominated or Pareto-optimal solutions, and enables them to choose the best one according to their preferences. A set of problems of different sizes is generated and solved using the proposed algorithm. Evaluating the performance of the proposed algorithm, three metrics are suggested, and the diversity and the convergence of the achieved Pareto front are appraised. Finally a comparison is made between CellDE and PAES, another meta-heuristic algorithm. The results show the superiority of CellDE.
NASA Astrophysics Data System (ADS)
Guo, Peng; Cheng, Wenming; Wang, Yi
2014-10-01
The quay crane scheduling problem (QCSP) determines the handling sequence of tasks at ship bays by a set of cranes assigned to a container vessel such that the vessel's service time is minimized. A number of heuristics or meta-heuristics have been proposed to obtain the near-optimal solutions to overcome the NP-hardness of the problem. In this article, the idea of generalized extremal optimization (GEO) is adapted to solve the QCSP with respect to various interference constraints. The resulting GEO is termed the modified GEO. A randomized searching method for neighbouring task-to-QC assignments to an incumbent task-to-QC assignment is developed in executing the modified GEO. In addition, a unidirectional search decoding scheme is employed to transform a task-to-QC assignment to an active quay crane schedule. The effectiveness of the developed GEO is tested on a suite of benchmark problems introduced by K.H. Kim and Y.M. Park in 2004 (European Journal of Operational Research, Vol. 156, No. 3). Compared with other well-known existing approaches, the experiment results show that the proposed modified GEO is capable of obtaining the optimal or near-optimal solution in a reasonable time, especially for large-sized problems.
A Two-Stage Stochastic Mixed-Integer Programming Approach to the Smart House Scheduling Problem
NASA Astrophysics Data System (ADS)
Ozoe, Shunsuke; Tanaka, Yoichi; Fukushima, Masao
A “Smart House” is a highly energy-optimized house equipped with photovoltaic systems (PV systems), electric battery systems, fuel cell cogeneration systems (FC systems), electric vehicles (EVs) and so on. Smart houses are attracting much attention recently thanks to their enhanced ability to save energy by making full use of renewable energy and by achieving power grid stability despite an increased power draw for installed PV systems. Yet running a smart house's power system, with its multiple power sources and power storages, is no simple task. In this paper, we consider the problem of power scheduling for a smart house with a PV system, an FC system and an EV. We formulate the problem as a mixed integer programming problem, and then extend it to a stochastic programming problem involving recourse costs to cope with uncertain electricity demand, heat demand and PV power generation. Using our method, we seek to achieve the optimal power schedule running at the minimum expected operation cost. We present some results of numerical experiments with data on real-life demands and PV power generation to show the effectiveness of our method.
Solving a Production Scheduling Problem by Means of Two Biobjective Metaheuristic Procedures
NASA Astrophysics Data System (ADS)
Toncovich, Adrián; Oliveros Colay, María José; Moreno, José María; Corral, Jiménez; Corral, Rafael
2009-11-01
Production planning and scheduling problems emphasize the need for the availability of management tools that can help to assure proper service levels to customers, maintaining, at the same time, the production costs at acceptable levels and maximizing the utilization of the production facilities. In this case, a production scheduling problem that arises in the context of the activities of a company dedicated to the manufacturing of furniture for children and teenagers is addressed. Two bicriteria metaheuristic procedures are proposed to solve the sequencing problem in a production equipment that constitutes the bottleneck of the production process of the company. The production scheduling problem can be characterized as a general flow shop with sequence dependant setup times and additional inventory constraints. Two objectives are simultaneously taken into account when the quality of the candidate solutions is evaluated: the minimization of completion time of all jobs, or makespan, and the minimization of the total flow time of all jobs. Both procedures are based on a local search strategy that responds to the structure of the simulated annealing metaheuristic. In this case, both metaheuristic approaches generate a set of solutions that provides an approximation to the optimal Pareto front. In order to evaluate the performance of the proposed techniques a series of experiments was conducted. After analyzing the results, it can be said that the solutions provided by both approaches are adequate from the viewpoint of the quality as well as the computational effort involved in their generation. Nevertheless, a further refinement of the proposed procedures should be implemented with the aim of facilitating a quasi-automatic definition of the solution parameters.
Three-Stage Tabu Search for Solving Large-Scale Flow Shop Scheduling Problems
NASA Astrophysics Data System (ADS)
Xu, Yuedong; Tian, Yajie; Sannomiya, Nobuo
Tabu search is a meta-heuristic approach designed skillfully for finding a suboptimal solution of combinatorial optimization problems. In this paper the tabu search with three stages is proposed for solving large-scale flow shop scheduling problems. In order to obtain a better suboptimal solution in a short computation time, three different candidate lists are used to determine the incumbent solution in the respective search stages. The candidate lists are constructed by restricting the moving of each job. Test problems with four kinds of job data are examined. Based on analyzing the relationship between the candidate list and the suboptimal solution for each job data, a common parameter is given to construct the candidate list during the search process. Comparison of the computation result is made with the genetic algorithm and the basic tabu search, from which it is shown that the proposed tabu search outperforms two others.
NASA Astrophysics Data System (ADS)
Ohno, Akiyoshi; Nishi, Tatsushi; Inuiguchi, Masahiro; Takahashi, Satoru; Ueda, Kenji
In this paper, we propose a column generation for the train-set scheduling problem with regular maintenance constraints. The problem is to allocate the minimum train-set to the train operations required to operate a given train timetable. In the proposed method, a tight lower bound can be obtained from the continuous relaxation for Dantzig-Wolfe reformulation by column generation. The subproblem for the column generation is an elementary shortest path problem with resource constraints. A labeling algorithm is applied to solve the subproblem. In order to reduce the computational effort for solving subproblems, a new state space relaxation of the subproblem is developed in the labeling algorithm. An upper bound is computed by a heuristic algorithm. Computational results demonstrate the effectiveness of the proposed method.
A network flow model for short-term hydro-dominated hydrothermal scheduling problems
Franco, P.E.C. ); Carvalho, M.F. ); Soares, S. )
1994-05-01
This paper is concerned with the Short-Term Hydrothermal Scheduling (STHS) of hydro-dominated power systems. The problem's formulation includes the representation of operational constraints such as the hydraulic coupling between hydro plants in cascade and the transmission limits in the electric network. In order to allow the problem's decomposition into hydraulic and electric subproblems, a linear-quadratic penalty approach is applied to enforce the coupling between hydro and electric variables. As a result, the problem's natural network flow structure is fully exploited through special-purposed network flow algorithms. The technique has been implemented in FORTRAN in a SUN SPARCstation IPX and tested in a 440 KV subsystem of the main interconnected Brazilian power system.
An estimation of distribution algorithm (EDA) variant with QGA for Flowshop scheduling problem
NASA Astrophysics Data System (ADS)
Latif, Muhammad Shahid; Hong, Zhou; Ali, Amir
2014-04-01
In this research article, a hybrid approach is presented which based on well-known meta-heuristics algorithms. This study based on integration of Quantum Genetic Algorithm (QGA) and Estimation of Distribution Algorithm, EDA, (for simplicity we use Q-EDA) for flowshop scheduling, a well-known NP hard Problem, while focusing on the total flow time minimization criterion. A relatively new method has been adopted for the encoding of jobs sequence in flowshop known as angel rotations instead of random keys, so QGA become more efficient. Further, EDA has been integrated to update the population of QGA by making a probability model. This probabilistic model is built and used to generate new candidate solutions which comprised on best individuals, obtained after several repetitions of proposed (Q-EDA) approach. As both heuristics based on probabilistic characteristics, so exhibits excellent learning capability and have minimum chances of being trapped in local optima. The results obtained during this study are presented and compared with contemporary approaches in literature. The current hybrid Q-EDA has implemented on different benchmark problems. The experiments has showed better convergence and results. It is concluded that hybrid Q-EDA algorithm can generally produce better results while implemented for Flowshop Scheduling Problem (FSSP).
Insight and analysis problem solving in microbes to machines.
Clark, Kevin B
2015-11-01
A key feature for obtaining solutions to difficult problems, insight is oftentimes vaguely regarded as a special discontinuous intellectual process and/or a cognitive restructuring of problem representation or goal approach. However, this nearly century-old state of art devised by the Gestalt tradition to explain the non-analytical or non-trial-and-error, goal-seeking aptitude of primate mentality tends to neglect problem-solving capabilities of lower animal phyla, Kingdoms other than Animalia, and advancing smart computational technologies built from biological, artificial, and composite media. Attempting to provide an inclusive, precise definition of insight, two major criteria of insight, discontinuous processing and problem restructuring, are here reframed using terminology and statistical mechanical properties of computational complexity classes. Discontinuous processing becomes abrupt state transitions in algorithmic/heuristic outcomes or in types of algorithms/heuristics executed by agents using classical and/or quantum computational models. And problem restructuring becomes combinatorial reorganization of resources, problem-type substitution, and/or exchange of computational models. With insight bounded by computational complexity, humans, ciliated protozoa, and complex technological networks, for example, show insight when restructuring time requirements, combinatorial complexity, and problem type to solve polynomial and nondeterministic polynomial decision problems. Similar effects are expected from other problem types, supporting the idea that insight might be an epiphenomenon of analytical problem solving and consequently a larger information processing framework. Thus, this computational complexity definition of insight improves the power, external and internal validity, and reliability of operational parameters with which to classify, investigate, and produce the phenomenon for computational agents ranging from microbes to man-made devices. PMID
Sharpe, Louise; Walker, Michael; Coughlan, Maree-Jo; Enersen, Kirsten; Blaszczynski, Alex
2005-01-01
This study aimed to evaluate the effectiveness of three proposed modifications to the structural characteristics of electronic gaming machines as harm minimisation strategies for non-problem and probable problem gamblers. Structural changes included reducing the maximum bet size, reducing reel spin and removing large note acceptors. Behavioural patterns of play were observed in 779 participants attending clubs and hotels. Observations were conducted in the gaming venue during regular gaming sessions. Eight experimental machines were designed to represent every combination of the modifications. 210 participants played at least one modified and one unmodified machine. Following play, the South Oaks Gambling Screen (SOGS) was administered. More problem than non-problem gamblers used high denomination bill acceptors and bet over one-dollar per wager. Machines modified to accept the one-dollar maximum bet were played for less time and were associated with smaller losses, fewer individual wagers and lower levels of alcohol consumption and smoking. It was concluded that the reduction of maximum bet levels was the only modification likely to be effective as a harm minimization strategy for problem gamblers. PMID:16311879
NASA Astrophysics Data System (ADS)
Xu, Ye; Wang, Ling; Wang, Shengyao; Liu, Min
2014-09-01
In this article, an effective hybrid immune algorithm (HIA) is presented to solve the distributed permutation flow-shop scheduling problem (DPFSP). First, a decoding method is proposed to transfer a job permutation sequence to a feasible schedule considering both factory dispatching and job sequencing. Secondly, a local search with four search operators is presented based on the characteristics of the problem. Thirdly, a special crossover operator is designed for the DPFSP, and mutation and vaccination operators are also applied within the framework of the HIA to perform an immune search. The influence of parameter setting on the HIA is investigated based on the Taguchi method of design of experiment. Extensive numerical testing results based on 420 small-sized instances and 720 large-sized instances are provided. The effectiveness of the HIA is demonstrated by comparison with some existing heuristic algorithms and the variable neighbourhood descent methods. New best known solutions are obtained by the HIA for 17 out of 420 small-sized instances and 585 out of 720 large-sized instances.
Some multigrid algorithms for elliptic problems on data parallel machines
Bandy, V.A.; Dendy, J.E. Jr.; Spangenberg, W.H.
1998-01-01
Previously a semicoarsening multigrid algorithm suitable for use on data parallel architectures was investigated. Through the use of new software tools, the performance of this algorithm has been considerably improved. The method has also been extended to three space dimensions. The method performs well for strongly anisotropic problems and for problems with coefficients jumping by orders of magnitude across internal interfaces. The parallel efficiency of this method is analyzed, and its actual performance on the CM-5 is compared with its performance on the CRAY Y-MP and the Sparc-5. A standard coarsening multigrid algorithm is also considered, and they compare its performance on these three platforms as well.
On the integrability of the motion of 3D-Swinging Atwood machine and related problems
NASA Astrophysics Data System (ADS)
Elmandouh, A. A.
2016-03-01
In the present article, we study the problem of the motion of 3D- Swinging Atwood machine. A new integrable case for this problem is announced. We point out a new integrable case describing the motion of a heavy particle on a titled cone.
An Exploratory Study of Problem Gambling on Casino versus Non-Casino Electronic Gaming Machines
ERIC Educational Resources Information Center
Clarke, Dave; Pulford, Justin; Bellringer, Maria; Abbott, Max; Hodgins, David C.
2012-01-01
Electronic gaming machines (EGMs) have been frequently associated with problem gambling. Little research has compared the relative contribution of casino EGMs versus non-casino EGMs on current problem gambling, after controlling for demographic factors and gambling behaviour. Our exploratory study obtained data from questionnaires administered to…
Whitley, L. Darrell; Watson, Jean-Paul; Howe, Adele E.
2005-06-01
Over the last decade and a half, tabu search algorithms for machine scheduling have gained a near-mythical reputation by consistently equaling or establishing state-of-the-art performance levels on a range of academic and real-world problems. Yet, despite these successes, remarkably little research has been devoted to developing an understanding of why tabu search is so effective on this problem class. In this paper, we report results that provide significant progress in this direction. We consider Nowicki and Smutnicki's i-TSAB tabu search algorithm, which represents the current state-of-the-art for the makespan-minimization form of the classical jobshop scheduling problem. Via a series of controlled experiments, we identify those components of i-TSAB that enable it to achieve state-of-the-art performance levels. In doing so, we expose a number of misconceptions regarding the behavior and/or benefits of tabu search and other local search metaheuristics for the job-shop problem. Our results also serve to focus future research, by identifying those specific directions that are most likely to yield further improvements in performance.
NASA Astrophysics Data System (ADS)
Gu, Cunchang; Mu, Yundong
2013-03-01
In this paper, we consider a single machine on-line scheduling problem with the special chains precedence and delivery time. All jobs arrive over time. The chains chainsi arrive at time ri , it is known that the processing and delivery time of each job on the chain satisfy one special condition CD a forehand: if the job J(i)j is the predecessor of the job J(i)k on the chain chaini, then they satisfy p(i)j = p(i)k = p >= qj >= qk , i = 1,2, ---,n , where pj and qj denote the processing time and the delivery time of the job Jj respectively. Obviously, if the arrival jobs have no chains precedence, it shows that the length of the corresponding chain is 1. The objective is to minimize the time by which all jobs have been delivered. We provide an on-line algorithm with a competitive ratio of √2 , and the result is the best possible.
Analysis problems for sequential dynamical systems and communicating state machines
Barrett, C. L.; Hunt, H. B.; Marathe, M. V.; Ravi, S. S.; Rosenkrantz, D. J.; Stearns, R. E.
2001-01-01
A simple sequential dynamical system (SDS) is a triple (G, F, {pi}), where (i) G(V, E) is an undirected graph with n nodes with each node having a 1-bit state, (ii) F = {l_brace} f{sub 1},f{sub 2},...,f{sub n}{r_brace} is a set of local transition functions with f{sub i} denoting a Boolean function associated with node Vv{sub i} and (iii) {pi} is a fixed permutation of (i.e., a total order on) the nodes in V. A single SDS transition is obtained by updating the states of the nodes in V by evaluating the function associated with each of them in the order given by {pi}. Such a (finite) SDS is a mathematical abstraction of simulation systems [BMR99, BR99]. In this paper, we characterize the computational complexity of determining several phase space properties of SDSs. The properties considered are t-REACHABILITY ('Can a given SDS starting from configuration I reach configuration B in t or fewer transitions?'), REACHABILITY('Can a given SDS starting from configuration I ever reach configuration B?') and FIXED POINT REACHABILITY ('Can a given SDS starting from configuration I ever reach configuration in which it stays for ever?'). Our main result is a sharp dichotomy between classes of SDSs whose behavior is 'easy' to predict and those whose behavior is 'hard' to predict. Specifically, we show the following. (1) The t-REACHABILITY, REACHABILITY and the FIXED POINT REACHABILITY problems for SDSs are PSPACE-complete, even when restricted to graphs of bounded bandwidth (and hence of bounded pathwidth and treewidth) and when the function associated with each node is symmetric. The result holds even for regular graphs of constant degree where all the nodes compute the same symmetric Boolean function. (2) In contrast, the t-REACHABILITYm REACHABILITY and FIXED POINT REACHABILITY problems are solvable in polynomial time for SDSs when the Boolean function associated with each node is symmetric and monotone. Two important consequences of our results are the following: (i) The
A Mixed Integer Linear Program for Solving a Multiple Route Taxi Scheduling Problem
NASA Technical Reports Server (NTRS)
Montoya, Justin Vincent; Wood, Zachary Paul; Rathinam, Sivakumar; Malik, Waqar Ahmad
2010-01-01
Aircraft movements on taxiways at busy airports often create bottlenecks. This paper introduces a mixed integer linear program to solve a Multiple Route Aircraft Taxi Scheduling Problem. The outputs of the model are in the form of optimal taxi schedules, which include routing decisions for taxiing aircraft. The model extends an existing single route formulation to include routing decisions. An efficient comparison framework compares the multi-route formulation and the single route formulation. The multi-route model is exercised for east side airport surface traffic at Dallas/Fort Worth International Airport to determine if any arrival taxi time savings can be achieved by allowing arrivals to have two taxi routes: a route that crosses an active departure runway and a perimeter route that avoids the crossing. Results indicate that the multi-route formulation yields reduced arrival taxi times over the single route formulation only when a perimeter taxiway is used. In conditions where the departure aircraft are given an optimal and fixed takeoff sequence, accumulative arrival taxi time savings in the multi-route formulation can be as high as 3.6 hours more than the single route formulation. If the departure sequence is not optimal, the multi-route formulation results in less taxi time savings made over the single route formulation, but the average arrival taxi time is significantly decreased.
NASA Astrophysics Data System (ADS)
Gao, Qian
For both the conventional radio frequency and the comparably recent optical wireless communication systems, extensive effort from the academia had been made in improving the network spectrum efficiency and/or reducing the error rate. To achieve these goals, many fundamental challenges such as power efficient constellation design, nonlinear distortion mitigation, channel training design, network scheduling and etc. need to be properly addressed. In this dissertation, novel schemes are proposed accordingly to deal with specific problems falling in category of these challenges. Rigorous proofs and analyses are provided for each of our work to make a fair comparison with the corresponding peer works to clearly demonstrate the advantages. The first part of this dissertation considers a multi-carrier optical wireless system employing intensity modulation (IM) and direct detection (DD). A block-wise constellation design is presented, which treats the DC-bias that conventionally used solely for biasing purpose as an information basis. Our scheme, we term it MSM-JDCM, takes advantage of the compactness of sphere packing in a higher dimensional space, and in turn power efficient constellations are obtained by solving an advanced convex optimization problem. Besides the significant power gains, the MSM-JDCM has many other merits such as being capable of mitigating nonlinear distortion by including a peak-to-power ratio (PAPR) constraint, minimizing inter-symbol-interference (ISI) caused by frequency-selective fading with a novel precoder designed and embedded, and further reducing the bit-error-rate (BER) by combining with an optimized labeling scheme. The second part addresses several optimization problems in a multi-color visible light communication system, including power efficient constellation design, joint pre-equalizer and constellation design, and modeling of different structured channels with cross-talks. Our novel constellation design scheme, termed CSK-Advanced, is
NASA Astrophysics Data System (ADS)
Izah Anuar, Nurul; Saptari, Adi
2016-02-01
This paper addresses the types of particle representation (encoding) procedures in a population-based stochastic optimization technique in solving scheduling problems known in the job-shop manufacturing environment. It intends to evaluate and compare the performance of different particle representation procedures in Particle Swarm Optimization (PSO) in the case of solving Job-shop Scheduling Problems (JSP). Particle representation procedures refer to the mapping between the particle position in PSO and the scheduling solution in JSP. It is an important step to be carried out so that each particle in PSO can represent a schedule in JSP. Three procedures such as Operation and Particle Position Sequence (OPPS), random keys representation and random-key encoding scheme are used in this study. These procedures have been tested on FT06 and FT10 benchmark problems available in the OR-Library, where the objective function is to minimize the makespan by the use of MATLAB software. Based on the experimental results, it is discovered that OPPS gives the best performance in solving both benchmark problems. The contribution of this paper is the fact that it demonstrates to the practitioners involved in complex scheduling problems that different particle representation procedures can have significant effects on the performance of PSO in solving JSP.
A stochastic model for the cell formation problem considering machine reliability
NASA Astrophysics Data System (ADS)
Esmailnezhad, Bahman; Fattahi, Parviz; Kheirkhah, Amir Saman
2015-03-01
This paper presents a new mathematical model to solve cell formation problem in cellular manufacturing systems, where inter-arrival time, processing time, and machine breakdown time are probabilistic. The objective function maximizes the number of operations of each part with more arrival rate within one cell. Because a queue behind each machine; queuing theory is used to formulate the model. To solve the model, two metaheurstic algorithms such as modified particle swarm optimization and genetic algorithm are proposed. For the generation of initial solutions in these algorithms, a new heuristic method is developed, which always creates feasible solutions. Both metaheurstic algorithms are compared against global solutions obtained from Lingo software's branch and bound (B&B). Also, a statistical method will be used for comparison of solutions of two metaheurstic algorithms. The results of numerical examples indicate that considering the machine breakdown has significant effect on block structures of machine-part matrixes.
An extended abstract: A heuristic repair method for constraint-satisfaction and scheduling problems
NASA Technical Reports Server (NTRS)
Minton, Steven; Johnston, Mark D.; Philips, Andrew B.; Laird, Philip
1992-01-01
The work described in this paper was inspired by a surprisingly effective neural network developed for scheduling astronomical observations on the Hubble Space Telescope. Our heuristic constraint satisfaction problem (CSP) method was distilled from an analysis of the network. In the process of carrying out the analysis, we discovered that the effectiveness of the network has little to do with its connectionist implementation. Furthermore, the ideas employed in the network can be implemented very efficiently within a symbolic CSP framework. The symbolic implementation is extremely simple. It also has the advantage that several different search strategies can be employed, although we have found that hill-climbing methods are particularly well-suited for the applications that we have investigated. We begin the paper with a brief review of the neural network. Following this, we describe our symbolic method for heuristic repair.
Three hybridization models based on local search scheme for job shop scheduling problem
NASA Astrophysics Data System (ADS)
Balbi Fraga, Tatiana
2015-05-01
This work presents three different hybridization models based on the general schema of Local Search Heuristics, named Hybrid Successive Application, Hybrid Neighborhood, and Hybrid Improved Neighborhood. Despite similar approaches might have already been presented in the literature in other contexts, in this work these models are applied to analyzes the solution of the job shop scheduling problem, with the heuristics Taboo Search and Particle Swarm Optimization. Besides, we investigate some aspects that must be considered in order to achieve better solutions than those obtained by the original heuristics. The results demonstrate that the algorithms derived from these three hybrid models are more robust than the original algorithms and able to get better results than those found by the single Taboo Search.
NASA Astrophysics Data System (ADS)
Kumar, Vijay M.; Murthy, ANN; Chandrashekara, K.
2012-05-01
The production planning problem of flexible manufacturing system (FMS) concerns with decisions that have to be made before an FMS begins to produce parts according to a given production plan during an upcoming planning horizon. The main aspect of production planning deals with machine loading problem in which selection of a subset of jobs to be manufactured and assignment of their operations to the relevant machines are made. Such problems are not only combinatorial optimization problems, but also happen to be non-deterministic polynomial-time-hard, making it difficult to obtain satisfactory solutions using traditional optimization techniques. In this paper, an attempt has been made to address the machine loading problem with objectives of minimization of system unbalance and maximization of throughput simultaneously while satisfying the system constraints related to available machining time and tool slot designing and using a meta-hybrid heuristic technique based on genetic algorithm and particle swarm optimization. The results reported in this paper demonstrate the model efficiency and examine the performance of the system with respect to measures such as throughput and system utilization.
Decision theory for computing variable and value ordering decisions for scheduling problems
NASA Technical Reports Server (NTRS)
Linden, Theodore A.
1993-01-01
Heuristics that guide search are critical when solving large planning and scheduling problems, but most variable and value ordering heuristics are sensitive to only one feature of the search state. One wants to combine evidence from all features of the search state into a subjective probability that a value choice is best, but there has been no solid semantics for merging evidence when it is conceived in these terms. Instead, variable and value ordering decisions should be viewed as problems in decision theory. This led to two key insights: (1) The fundamental concept that allows heuristic evidence to be merged is the net incremental utility that will be achieved by assigning a value to a variable. Probability distributions about net incremental utility can merge evidence from the utility function, binary constraints, resource constraints, and other problem features. The subjective probability that a value is the best choice is then derived from probability distributions about net incremental utility. (2) The methods used for rumor control in Bayesian Networks are the primary way to prevent cycling in the computation of probable net incremental utility. These insights lead to semantically justifiable ways to compute heuristic variable and value ordering decisions that merge evidence from all available features of the search state.
Li, Shanlin; Li, Maoqin
2015-01-01
We consider an integrated production and distribution scheduling problem faced by a typical make-to-order manufacturer which relies on a third-party logistics (3PL) provider for finished product delivery to customers. In the beginning of a planning horizon, the manufacturer has received a set of orders to be processed on a single production line. Completed orders are delivered to customers by a finite number of vehicles provided by the 3PL company which follows a fixed daily or weekly shipping schedule such that the vehicles have fixed departure dates which are not part of the decisions. The problem is to find a feasible schedule that minimizes one of the following objective functions when processing times and weights are oppositely ordered: (1) the total weight of late orders and (2) the number of vehicles used subject to the condition that the total weight of late orders is minimum. We show that both problems are solvable in polynomial time. PMID:25785285
Li, Shanlin; Li, Maoqin
2015-01-01
We consider an integrated production and distribution scheduling problem faced by a typical make-to-order manufacturer which relies on a third-party logistics (3PL) provider for finished product delivery to customers. In the beginning of a planning horizon, the manufacturer has received a set of orders to be processed on a single production line. Completed orders are delivered to customers by a finite number of vehicles provided by the 3PL company which follows a fixed daily or weekly shipping schedule such that the vehicles have fixed departure dates which are not part of the decisions. The problem is to find a feasible schedule that minimizes one of the following objective functions when processing times and weights are oppositely ordered: (1) the total weight of late orders and (2) the number of vehicles used subject to the condition that the total weight of late orders is minimum. We show that both problems are solvable in polynomial time. PMID:25785285
Heuristic algorithms for solving of the tool routing problem for CNC cutting machines
NASA Astrophysics Data System (ADS)
Chentsov, P. A.; Petunin, A. A.; Sesekin, A. N.; Shipacheva, E. N.; Sholohov, A. E.
2015-11-01
The article is devoted to the problem of minimizing the path of the cutting tool to shape cutting machines began. This problem can be interpreted as a generalized traveling salesman problem. Earlier version of the dynamic programming method to solve this problem was developed. Unfortunately, this method allows to process an amount not exceeding thirty circuits. In this regard, the task of constructing quasi-optimal route becomes relevant. In this paper we propose options for quasi-optimal greedy algorithms. Comparison of the results of exact and approximate algorithms is given.
Duan, Qianqian; Yang, Genke; Xu, Guanglin; Pan, Changchun
2014-01-01
This paper is devoted to develop an approximation method for scheduling refinery crude oil operations by taking into consideration the demand uncertainty. In the stochastic model the demand uncertainty is modeled as random variables which follow a joint multivariate distribution with a specific correlation structure. Compared to deterministic models in existing works, the stochastic model can be more practical for optimizing crude oil operations. Using joint chance constraints, the demand uncertainty is treated by specifying proximity level on the satisfaction of product demands. However, the joint chance constraints usually hold strong nonlinearity and consequently, it is still hard to handle it directly. In this paper, an approximation method combines a relax-and-tight technique to approximately transform the joint chance constraints to a serial of parameterized linear constraints so that the complicated problem can be attacked iteratively. The basic idea behind this approach is to approximate, as much as possible, nonlinear constraints by a lot of easily handled linear constraints which will lead to a well balance between the problem complexity and tractability. Case studies are conducted to demonstrate the proposed methods. Results show that the operation cost can be reduced effectively compared with the case without considering the demand correlation. PMID:24757433
NASA Astrophysics Data System (ADS)
Motoyama, H.
1989-05-01
The present situations of superconducting electric power machines in the world and studied problems were investigated from viewpoint of the electric insulation. 50MVA generator (CRIE/Hitachi) or 120MVA generator (KWU/Siemens) where the dc superconducting technique was applied on field windings, are developed. As to Superconducting transformer, 220KVA transformer is trially manufactured and the conceptual design of 1,000MVA transformer is made by W.H. or Alstom. Future problems are the study of protecting method for the overvoltage to superconducting electric power machines and the study to prevent the quench for superconducting windings. The respective insulating characteristics of solid and liquid insulators become clear gradually under the cryogenic condition but a large part of insulating characteristics of composite insulator prepared by combination of both insulators are not clear, so that these problems must be clarified.
Non-clairvoyant Scheduling Games
NASA Astrophysics Data System (ADS)
Dürr, Christoph; Nguyen, Kim Thang
In a scheduling game, each player owns a job and chooses a machine to execute it. While the social cost is the maximal load over all machines (makespan), the cost (disutility) of each player is the completion time of its own job. In the game, players may follow selfish strategies to optimize their cost and therefore their behaviors do not necessarily lead the game to an equilibrium. Even in the case there is an equilibrium, its makespan might be much larger than the social optimum, and this inefficiency is measured by the price of anarchy - the worst ratio between the makespan of an equilibrium and the optimum. Coordination mechanisms aim to reduce the price of anarchy by designing scheduling policies that specify how jobs assigned to a same machine are to be scheduled. Typically these policies define the schedule according to the processing times as announced by the jobs. One could wonder if there are policies that do not require this knowledge, and still provide a good price of anarchy. This would make the processing times be private information and avoid the problem of truthfulness. In this paper we study these so-called non-clairvoyant policies. In particular, we study the RANDOM policy that schedules the jobs in a random order without preemption, and the EQUI policy that schedules the jobs in parallel using time-multiplexing, assigning each job an equal fraction of CPU time.
Linear-time algorithms for scheduling on parallel processors
Monma, C.L.
1982-01-01
Linear-time algorithms are presented for several problems of scheduling n equal-length tasks on m identical parallel processors subject to precedence constraints. This improves upon previous time bounds for the maximum lateness problem with treelike precedence constraints, the number-of-late-tasks problem without precedence constraints, and the one machine maximum lateness problem with general precedence constraints. 5 references.
NASA Astrophysics Data System (ADS)
Trunfio, Roberto
2015-06-01
In a recent article, Guo, Cheng and Wang proposed a randomized search algorithm, called modified generalized extremal optimization (MGEO), to solve the quay crane scheduling problem for container groups under the assumption that schedules are unidirectional. The authors claim that the proposed algorithm is capable of finding new best solutions with respect to a well-known set of benchmark instances taken from the literature. However, as shown in this note, there are some errors in their work that can be detected by analysing the Gantt charts of two solutions provided by MGEO. In addition, some comments on the method used to evaluate the schedule corresponding to a task-to-quay crane assignment and on the search scheme of the proposed algorithm are provided. Finally, to assess the effectiveness of the proposed algorithm, the computational experiments are repeated and additional computational experiments are provided.
Research on Production Scheduling System with Bottleneck Based on Multi-agent
NASA Astrophysics Data System (ADS)
Zhenqiang, Bao; Weiye, Wang; Peng, Wang; Pan, Quanke
Aimed at the imbalance problem of resource capacity in Production Scheduling System, this paper uses Production Scheduling System based on multi-agent which has been constructed, and combines the dynamic and autonomous of Agent; the bottleneck problem in the scheduling is solved dynamically. Firstly, this paper uses Bottleneck Resource Agent to find out the bottleneck resource in the production line, analyses the inherent mechanism of bottleneck, and describes the production scheduling process based on bottleneck resource. Bottleneck Decomposition Agent harmonizes the relationship of job's arrival time and transfer time in Bottleneck Resource Agent and Non-Bottleneck Resource Agents, therefore, the dynamic scheduling problem is simplified as the single machine scheduling of each resource which takes part in the scheduling. Finally, the dynamic real-time scheduling problem is effectively solved in Production Scheduling System.
A multi-objective scatter search for a bi-criteria no-wait flow shop scheduling problem
NASA Astrophysics Data System (ADS)
Rahimi-Vahed, A. R.; Javadi, B.; Rabbani, M.; Tavakkoli-Moghaddam, R.
2008-04-01
The flow shop problem as a typical manufacturing challenge has gained wide attention in academic fields. This article considers a bi-criteria no-wait flow shop scheduling problem (FSSP) in which weighted mean completion time and weighted mean tardiness are to be minimized simultaneously. Since a FSSP has been proved to be NP-hard in a strong sense, a new multi-objective scatter search (MOSS) is designed for finding the locally Pareto-optimal frontier of the problem. To prove the efficiency of the proposed algorithm, various test problems are solved and the reliability of the proposed algorithm, based on some comparison metrics, is compared with a distinguished multi-objective genetic algorithm (GA), i.e. SPEA-II. The computational results show that the proposed MOSS performs better than the above GA, especially for the large-sized problems.
Personalized Education; Solving a Group Formation and Scheduling Problem for Educational Content
ERIC Educational Resources Information Center
Bahargam, Sanaz; Erdos, Dóra; Bestavros, Azer; Terzi, Evimaria
2015-01-01
Whether teaching in a classroom or a Massive Online Open Course it is crucial to present the material in a way that benefits the audience as a whole. We identify two important tasks to solve towards this objective; (1) group students so that they can maximally benefit from peer interaction and (2) find an optimal schedule of the educational…
NASA Technical Reports Server (NTRS)
Gaspin, Christine
1989-01-01
How a neural network can work, compared to a hybrid system based on an operations research and artificial intelligence approach, is investigated through a mission scheduling problem. The characteristic features of each system are discussed.
A federated model for scheduling in wide-area systems
Weissman, J.B.; Grimshaw, A.S.
1996-12-31
In this paper a model for scheduling in wide-area systems is described. The model is federated and utilizes a collection of local site schedulers that control the use of their resources. The wide-area scheduler consults the local site schedulers to obtain candidate machine schedules. A set of issues and challenges inherent to wide-area scheduling are also described and the proposed model is shown to address many of these problems. A distributed algorithm for wide-area scheduling is presented and relies upon information made available about the resource needs of user jobs. The wide-area scheduler will be implemented in Legion, a wide-area computing system developed at the University of Virginia.
NASA Astrophysics Data System (ADS)
Wang, Hongfeng; Fu, Yaping; Huang, Min; Wang, Junwei
2016-03-01
The operation process design is one of the key issues in the manufacturing and service sectors. As a typical operation process, the scheduling with consideration of the deteriorating effect has been widely studied; however, the current literature only studied single function requirement and rarely considered the multiple function requirements which are critical for a real-world scheduling process. In this article, two function requirements are involved in the design of a scheduling process with consideration of the deteriorating effect and then formulated into two objectives of a mathematical programming model. A novel multiobjective evolutionary algorithm is proposed to solve this model with combination of three strategies, i.e. a multiple population scheme, a rule-based local search method and an elitist preserve strategy. To validate the proposed model and algorithm, a series of randomly-generated instances are tested and the experimental results indicate that the model is effective and the proposed algorithm can achieve the satisfactory performance which outperforms the other state-of-the-art multiobjective evolutionary algorithms, such as nondominated sorting genetic algorithm II and multiobjective evolutionary algorithm based on decomposition, on all the test instances.
Quay crane scheduling with dual cycling
NASA Astrophysics Data System (ADS)
Wang, Dandan; Li, Xiaoping
2015-10-01
In this article, the dual cycling quay crane scheduling problem (D-QCSP) with hatches is addressed to minimize the operation cycles of quay cranes. The problem is decomposed into two sub-problems: the intra-group stage (sequencing stacks within each hatch) and the inter-group stage (scheduling all hatches). A new stack sequencing method is constructed for stacks of each hatch, which is modelled as a two-machine non-permutation flow shop scheduling problem. By removing inner gaps using left-shifting, the adapted hatch scheduling sub-problem is modelled as a two-machine grouped flow shop scheduling problem, which contains more precise processing times. A composite heuristic is proposed for the D-QCSP. Based on the derived lower bound, the heuristic is compared with the best existing heuristics on a large number of instances. Experimental results illustrate that the proposal outperforms the existing methods on all instances and dual cycling needs many fewer quay crane operating cycles than single cycling.
NASA Astrophysics Data System (ADS)
Han, Yu-Yan; Gong, Dunwei; Sun, Xiaoyan
2015-07-01
A flow-shop scheduling problem with blocking has important applications in a variety of industrial systems but is underrepresented in the research literature. In this study, a novel discrete artificial bee colony (ABC) algorithm is presented to solve the above scheduling problem with a makespan criterion by incorporating the ABC with differential evolution (DE). The proposed algorithm (DE-ABC) contains three key operators. One is related to the employed bee operator (i.e. adopting mutation and crossover operators of discrete DE to generate solutions with good quality); the second is concerned with the onlooker bee operator, which modifies the selected solutions using insert or swap operators based on the self-adaptive strategy; and the last is for the local search, that is, the insert-neighbourhood-based local search with a small probability is adopted to improve the algorithm's capability in exploitation. The performance of the proposed DE-ABC algorithm is empirically evaluated by applying it to well-known benchmark problems. The experimental results show that the proposed algorithm is superior to the compared algorithms in minimizing the makespan criterion.
NASA Astrophysics Data System (ADS)
Basri, Azyanzuhaila Hasan; Zainuddin, Zaitul Marlizawati
2014-09-01
High efficiency of port operation is required to succeed in the competition between port container terminals. Berth Allocation and Quay Crane Scheduling are the most important part in container terminal operations. The integrated model is formulated as a MIP problem with the objective to minimize the sum of the dwell times, where a vessel's dwell time is measured between arrival and departure including both times waiting to be berthed and servicing time while berthed. The construction of suitable mathematical model is formulated by considering various practical constraints.
Agnetis, Alessandro; Coppi, Alberto; Corsini, Matteo; Dellino, Gabriella; Meloni, Carlo; Pranzo, Marco
2014-03-01
This research aims at supporting hospital management in making prompt Operating Room (OR) planning decisions, when either unpredicted events occur or alternative scenarios or configurations need to be rapidly evaluated. We design and test a planning tool enabling managers to efficiently analyse several alternatives to the current OR planning and scheduling. To this aim, we propose a decomposition approach. More specifically, we first focus on determining the Master Surgical Schedule (MSS) on a weekly basis, by assigning the different surgical disciplines to the available sessions. Next, we allocate surgeries to each session, focusing on elective patients only. Patients are selected from the waiting lists according to several parameters, including surgery duration, waiting time and priority class of the operations. We performed computational experiments to compare the performance of our decomposition approach with an (exact) integrated approach. The case study selected for our simulations is based on the characteristics of the operating theatre (OT) of a medium-size public Italian hospital. Scalability of the method is tested for different OT sizes. A pilot example is also proposed to highlight the usefulness of our approach for decision support. The proposed decomposition approach finds satisfactory solutions with significant savings in computation time. PMID:23783452
Learning to integrate reactivity and deliberation in uncertain planning and scheduling problems
NASA Technical Reports Server (NTRS)
Chien, Steve A.; Gervasio, Melinda T.; Dejong, Gerald F.
1992-01-01
This paper describes an approach to planning and scheduling in uncertain domains. In this approach, a system divides a task on a goal by goal basis into reactive and deliberative components. Initially, a task is handled entirely reactively. When failures occur, the system changes the reactive/deliverative goal division by moving goals into the deliberative component. Because our approach attempts to minimize the number of deliberative goals, we call our approach Minimal Deliberation (MD). Because MD allows goals to be treated reactively, it gains some of the advantages of reactive systems: computational efficiency, the ability to deal with noise and non-deterministic effects, and the ability to take advantage of unforseen opportunities. However, because MD can fall back upon deliberation, it can also provide some of the guarantees of classical planning, such as the ability to deal with complex goal interactions. This paper describes the Minimal Deliberation approach to integrating reactivity and deliberation and describe an ongoing application of the approach to an uncertain planning and scheduling domain.
NASA Astrophysics Data System (ADS)
Yusriski, R.; Sukoyo; Samadhi, T. M. A. A.; Halim, A. H.
2016-02-01
In the manufacturing industry, several identical parts can be processed in batches, and setup time is needed between two consecutive batches. Since the processing times of batches are not always fixed during a scheduling period due to learning and deterioration effects, this research deals with batch scheduling problems with simultaneous learning and deterioration effects. The objective is to minimize total actual flow time, defined as a time interval between the arrival of all parts at the shop and their common due date. The decision variables are the number of batches, integer batch sizes, and the sequence of the resulting batches. This research proposes a heuristic algorithm based on the Lagrange Relaxation. The effectiveness of the proposed algorithm is determined by comparing the resulting solutions of the algorithm to the respective optimal solution obtained from the enumeration method. Numerical experience results show that the average of difference among the solutions is 0.05%.
NASA Technical Reports Server (NTRS)
Phillips, K.
1976-01-01
A mathematical model for job scheduling in a specified context is presented. The model uses both linear programming and combinatorial methods. While designed with a view toward optimization of scheduling of facility and plant operations at the Deep Space Communications Complex, the context is sufficiently general to be widely applicable. The general scheduling problem including options for scheduling objectives is discussed and fundamental parameters identified. Mathematical algorithms for partitioning problems germane to scheduling are presented.
NASA Astrophysics Data System (ADS)
Cram, Ana Catalina
As worldwide environmental awareness grow, alternative sources of energy have become important to mitigate climate change. Biogas in particular reduces greenhouse gas emissions that contribute to global warming and has the potential of providing 25% of the annual demand for natural gas in the U.S. In 2011, 55,000 metric tons of methane emissions were reduced and 301 metric tons of carbon dioxide emissions were avoided through the use of biogas alone. Biogas is produced by anaerobic digestion through the fermentation of organic material. It is mainly composed of methane with a rage of 50 to 80% in its concentration. Carbon dioxide covers 20 to 50% and small amounts of hydrogen, carbon monoxide and nitrogen. The biogas production systems are anaerobic digestion facilities and the optimal operation of an anaerobic digester requires the scheduling of all batches from multiple feedstocks during a specific time horizon. The availability times, biomass quantities, biogas production rates and storage decay rates must all be taken into account for maximal biogas production to be achieved during the planning horizon. Little work has been done to optimize the scheduling of different types of feedstock in anaerobic digestion facilities to maximize the total biogas produced by these systems. Therefore, in the present thesis, a new genetic algorithm is developed with the main objective of obtaining the optimal sequence in which different feedstocks will be processed and the optimal time to allocate to each feedstock in the digester with the main objective of maximizing the production of biogas considering different types of feedstocks, arrival times and decay rates. Moreover, all batches need to be processed in the digester in a specified time with the restriction that only one batch can be processed at a time. The developed algorithm is applied to 3 different examples and a comparison with results obtained in previous studies is presented.
The Block Scheduling Handbook.
ERIC Educational Resources Information Center
Queen, J. Allen
Block scheduling encourages increased comprehensive immersion into subject matter, improved teacher-student relationships, and decreased disciplinary problems. While block scheduling may offer many advantages, moving to a block schedule from conventional scheduling can be a major adjustment for both students and teachers. This guide is intended to…
An efficient annealing in Boltzmann machine in Hopfield neural network
NASA Astrophysics Data System (ADS)
Kin, Teoh Yeong; Hasan, Suzanawati Abu; Bulot, Norhisam; Ismail, Mohammad Hafiz
2012-09-01
This paper proposes and implements Boltzmann machine in Hopfield neural network doing logic programming based on the energy minimization system. The temperature scheduling in Boltzmann machine enhancing the performance of doing logic programming in Hopfield neural network. The finest temperature is determined by observing the ratio of global solution and final hamming distance using computer simulations. The study shows that Boltzmann Machine model is more stable and competent in term of representing and solving difficult combinatory problems.
NASA Technical Reports Server (NTRS)
Rash, James
2014-01-01
NASA's space data-communications infrastructure-the Space Network and the Ground Network-provide scheduled (as well as some limited types of unscheduled) data-communications services to user spacecraft. The Space Network operates several orbiting geostationary platforms (the Tracking and Data Relay Satellite System (TDRSS)), each with its own servicedelivery antennas onboard. The Ground Network operates service-delivery antennas at ground stations located around the world. Together, these networks enable data transfer between user spacecraft and their mission control centers on Earth. Scheduling data-communications events for spacecraft that use the NASA communications infrastructure-the relay satellites and the ground stations-can be accomplished today with software having an operational heritage dating from the 1980s or earlier. An implementation of the scheduling methods and algorithms disclosed and formally specified herein will produce globally optimized schedules with not only optimized service delivery by the space data-communications infrastructure but also optimized satisfaction of all user requirements and prescribed constraints, including radio frequency interference (RFI) constraints. Evolutionary algorithms, a class of probabilistic strategies for searching large solution spaces, is the essential technology invoked and exploited in this disclosure. Also disclosed are secondary methods and algorithms for optimizing the execution efficiency of the schedule-generation algorithms themselves. The scheduling methods and algorithms as presented are adaptable to accommodate the complexity of scheduling the civilian and/or military data-communications infrastructure within the expected range of future users and space- or ground-based service-delivery assets. Finally, the problem itself, and the methods and algorithms, are generalized and specified formally. The generalized methods and algorithms are applicable to a very broad class of combinatorial
Hinselmann, Georg; Rosenbaum, Lars; Jahn, Andreas; Fechner, Nikolas; Ostermann, Claude; Zell, Andreas
2011-02-28
The goal of this study was to adapt a recently proposed linear large-scale support vector machine to large-scale binary cheminformatics classification problems and to assess its performance on various benchmarks using virtual screening performance measures. We extended the large-scale linear support vector machine library LIBLINEAR with state-of-the-art virtual high-throughput screening metrics to train classifiers on whole large and unbalanced data sets. The formulation of this linear support machine has an excellent performance if applied to high-dimensional sparse feature vectors. An additional advantage is the average linear complexity in the number of non-zero features of a prediction. Nevertheless, the approach assumes that a problem is linearly separable. Therefore, we conducted an extensive benchmarking to evaluate the performance on large-scale problems up to a size of 175000 samples. To examine the virtual screening performance, we determined the chemotype clusters using Feature Trees and integrated this information to compute weighted AUC-based performance measures and a leave-cluster-out cross-validation. We also considered the BEDROC score, a metric that was suggested to tackle the early enrichment problem. The performance on each problem was evaluated by a nested cross-validation and a nested leave-cluster-out cross-validation. We compared LIBLINEAR against a Naïve Bayes classifier, a random decision forest classifier, and a maximum similarity ranking approach. These reference approaches were outperformed in a direct comparison by LIBLINEAR. A comparison to literature results showed that the LIBLINEAR performance is competitive but without achieving results as good as the top-ranked nonlinear machines on these benchmarks. However, considering the overall convincing performance and computation time of the large-scale support vector machine, the approach provides an excellent alternative to established large-scale classification approaches. PMID
Automated Planning and Scheduling for Space Mission Operations
NASA Technical Reports Server (NTRS)
Chien, Steve; Jonsson, Ari; Knight, Russell
2005-01-01
Research Trends: a) Finite-capacity scheduling under more complex constraints and increased problem dimensionality (subcontracting, overtime, lot splitting, inventory, etc.) b) Integrated planning and scheduling. c) Mixed-initiative frameworks. d) Management of uncertainty (proactive and reactive). e) Autonomous agent architectures and distributed production management. e) Integration of machine learning capabilities. f) Wider scope of applications: 1) analysis of supplier/buyer protocols & tradeoffs; 2) integration of strategic & tactical decision-making; and 3) enterprise integration.
Integrating tactical and operational decisions in fixed job scheduling
NASA Astrophysics Data System (ADS)
Türsel Eliiyi, Deniz
2013-12-01
Two different problems are introduced in this article to handle capacity and scheduling decisions simultaneously in the fixed job scheduling framework. The combined fixed job scheduling (CFJS) problem integrates these decisions assuming fixed costs for the usage of identical parallel machines, whereas the working time determination (WTD) problem involves unit-time operating or rental costs. Mathematical models for both problems are presented along with the worst case time complexities. While an exact polynomial-time algorithm is proposed for the CFJS problem, a heuristic algorithm is developed for the WTD problem as it is shown to be strongly NP hard. Computational experiments are carried out for evaluating the performance of the algorithms. The results reveal that the solutions by the exact algorithm for the CFJS problem are much faster than a state-of-the-art commercial solver, particularly for large instances. For the WTD problem, the developed heuristic provides high-quality solutions in very short computation times.
Reinforcement learning in scheduling
NASA Technical Reports Server (NTRS)
Dietterich, Tom G.; Ok, Dokyeong; Zhang, Wei; Tadepalli, Prasad
1994-01-01
The goal of this research is to apply reinforcement learning methods to real-world problems like scheduling. In this preliminary paper, we show that learning to solve scheduling problems such as the Space Shuttle Payload Processing and the Automatic Guided Vehicle (AGV) scheduling can be usefully studied in the reinforcement learning framework. We discuss some of the special challenges posed by the scheduling domain to these methods and propose some possible solutions we plan to implement.
The home health care routing and scheduling problem with interdependent services.
Mankowska, Dorota Slawa; Meisel, Frank; Bierwirth, Christian
2014-03-01
This paper presents a model for the daily planning of health care services carried out at patients' homes by staff members of a home care company. The planning takes into account individual service requirements of the patients, individual qualifications of the staff and possible interdependencies between different service operations. Interdependencies of services can include, for example, a temporal separation of two services as is required if drugs have to be administered a certain time before providing a meal. Other services like handling a disabled patient may require two staff members working together at a patient's home. The time preferences of patients are included in terms of given time windows. In this paper, we propose a planning approach for the described problem, which can be used for optimizing economical and service oriented measures of performance. A mathematical model formulation is proposed together with a powerful heuristic based on a sophisticated solution representation. PMID:23780750
Integrated resource scheduling in a distributed scheduling environment
NASA Technical Reports Server (NTRS)
Zoch, David; Hall, Gardiner
1988-01-01
The Space Station era presents a highly-complex multi-mission planning and scheduling environment exercised over a highly distributed system. In order to automate the scheduling process, customers require a mechanism for communicating their scheduling requirements to NASA. A request language that a remotely-located customer can use to specify his scheduling requirements to a NASA scheduler, thus automating the customer-scheduler interface, is described. This notation, Flexible Envelope-Request Notation (FERN), allows the user to completely specify his scheduling requirements such as resource usage, temporal constraints, and scheduling preferences and options. The FERN also contains mechanisms for representing schedule and resource availability information, which are used in the inter-scheduler inconsistency resolution process. Additionally, a scheduler is described that can accept these requests, process them, generate schedules, and return schedule and resource availability information to the requester. The Request-Oriented Scheduling Engine (ROSE) was designed to function either as an independent scheduler or as a scheduling element in a network of schedulers. When used in a network of schedulers, each ROSE communicates schedule and resource usage information to other schedulers via the FERN notation, enabling inconsistencies to be resolved between schedulers. Individual ROSE schedules are created by viewing the problem as a constraint satisfaction problem with a heuristically guided search strategy.
Protocols for distributive scheduling
NASA Technical Reports Server (NTRS)
Richards, Stephen F.; Fox, Barry
1993-01-01
The increasing complexity of space operations and the inclusion of interorganizational and international groups in the planning and control of space missions lead to requirements for greater communication, coordination, and cooperation among mission schedulers. These schedulers must jointly allocate scarce shared resources among the various operational and mission oriented activities while adhering to all constraints. This scheduling environment is complicated by such factors as the presence of varying perspectives and conflicting objectives among the schedulers, the need for different schedulers to work in parallel, and limited communication among schedulers. Smooth interaction among schedulers requires the use of protocols that govern such issues as resource sharing, authority to update the schedule, and communication of updates. This paper addresses the development and characteristics of such protocols and their use in a distributed scheduling environment that incorporates computer-aided scheduling tools. An example problem is drawn from the domain of space shuttle mission planning.
Learning dominance relations in combinatorial search problems
NASA Technical Reports Server (NTRS)
Yu, Chee-Fen; Wah, Benjamin W.
1988-01-01
Dominance relations commonly are used to prune unnecessary nodes in search graphs, but they are problem-dependent and cannot be derived by a general procedure. The authors identify machine learning of dominance relations and the applicable learning mechanisms. A study of learning dominance relations using learning by experimentation is described. This system has been able to learn dominance relations for the 0/1-knapsack problem, an inventory problem, the reliability-by-replication problem, the two-machine flow shop problem, a number of single-machine scheduling problems, and a two-machine scheduling problem. It is considered that the same methodology can be extended to learn dominance relations in general.
Brown, Stephen L; Rodda, Simone; Phillips, James G
2004-12-01
Arousal-based theories of gambling suggest that excitement gained from gambling reinforces further gambling behavior. However, recent theories of emotion conceptualize mood as comprising both arousal and valence dimensions. Thus, excitement comprises arousal with positive valence. We examined self-reported changes in arousal and affective valence in 27 problem and 40 nonproblem gamblers playing electronic gaming machines (EGMs). Problem gamblers reported greater arousal increases after gambling and increases in negative valence if they lost. This accords poorly with an excitement-based explanation of problem gambling. PMID:15530730
Trajectory classification in circular restricted three-body problem using support vector machine
NASA Astrophysics Data System (ADS)
Li, Weipeng; Huang, Hai; Peng, Fujun
2015-07-01
In the circular restricted three-body problem (CR3BP), transit orbit is a class of orbit which can pass through the bottleneck region of the zero velocity curve and escapes from the vicinity of the primary or the secondary. This kind of orbit plays a very important role in the design of space exploration missions. A kind of low-energy interplanetary transfer, which is called Interplanetary Superhighway (IPS), can be realized by utilizing transit orbits. To use the transit orbit in actual mission design, a key issue is to find an algorithm which can separate the states corresponding to transit orbits from the states corresponding to other types of orbits rapidly. In fact, the distribution of transit orbit in the phase space has been investigated by numerical method, and a Fourier series approximation method has been introduced to describe the boundary of transit orbits. However, the Fourier series approximation method needs several hundred sets of Fourier series. The coefficients of these Fourier series are neither easy to be computed nor convenient to be stored, which makes the method can hardly be used in actual mission design. In this paper, the support vector machine (SVM) is used to classify the trajectories in the CR3BP. Using the Gaussian kernel, the 6-dimensional states in the CR3BP are mapped into an infinite-dimensional space, and the bound of the transit orbits is described by a hyperplane. A training data generation method is introduced, which reduces the size of training data by generating the states near the hyperplane. The numerical results show that the proposed algorithm gives the good correct rate of classification, and its computing speed is much faster than that of the Fourier series approximation method.
Rau, B.R.
1996-02-01
Modulo scheduling is a framework within which algorithms for software pipelining innermost loops may be defined. The framework specifies a set of constraints that must be met in order to achieve a legal modulo schedule. A wide variety of algorithms and heuristics can be defined within this framework. Little work has been done to evaluate and compare alternative algorithms and heuristics for modulo scheduling from the viewpoints of schedule quality as well as computational complexity. This, along with a vague and unfounded perception that modulo scheduling is computationally expensive as well as difficult to implement, have inhibited its corporation into product compilers. This paper presents iterative modulo scheduling, a practical algorithm that is capable of dealing with realistic machine models. The paper also characterizes the algorithm in terms of the quality of the generated schedules as well as the computational incurred.
Spike: Artificial intelligence scheduling for Hubble space telescope
NASA Technical Reports Server (NTRS)
Johnston, Mark; Miller, Glenn; Sponsler, Jeff; Vick, Shon; Jackson, Robert
1990-01-01
Efficient utilization of spacecraft resources is essential, but the accompanying scheduling problems are often computationally intractable and are difficult to approximate because of the presence of numerous interacting constraints. Artificial intelligence techniques were applied to the scheduling of the NASA/ESA Hubble Space Telescope (HST). This presents a particularly challenging problem since a yearlong observing program can contain some tens of thousands of exposures which are subject to a large number of scientific, operational, spacecraft, and environmental constraints. New techniques were developed for machine reasoning about scheduling constraints and goals, especially in cases where uncertainty is an important scheduling consideration and where resolving conflicts among conflicting preferences is essential. These technique were utilized in a set of workstation based scheduling tools (Spike) for HST. Graphical displays of activities, constraints, and schedules are an important feature of the system. High level scheduling strategies using both rule based and neural network approaches were developed. While the specific constraints implemented are those most relevant to HST, the framework developed is far more general and could easily handle other kinds of scheduling problems. The concept and implementation of the Spike system are described along with some experiments in adapting Spike to other spacecraft scheduling domains.
NASA Astrophysics Data System (ADS)
Haribara, Yoshitaka; Utsunomiya, Shoko; Yamamoto, Yoshihisa
An optical parametric oscillator network driven by a quantum measurement-feedback circuit, composed of optical homodyne detectors, analog-to-digital conversion devices and field programmable gate arrays (FPGA), is proposed and analysed as a scalable coherent Ising machine. The new scheme has an advantage that a large number of optical coupling paths, which is proportional to the square of a problem size in the worst case, can be replaced by a single quantum measurement-feedback circuit. There is additional advantage in the new scheme that a three body or higher order Ising interaction can be implemented in the FPGA digital circuit. Noise associated with the measurement-feedback process is governed by the standard quantum limit. Numerical simulation based on c-number coupled Langevin equations demonstrate a satisfying performance of the proposed Ising machine against the NP-hard MAX-CUT problems.
ERIC Educational Resources Information Center
Davis, Harold S.; Bechard, Joseph E.
A flexible schedule allows teachers to change group size, group composition, and class length according to the purpose of the lesson. This pamphlet presents various "master" schedules for flexible scheduling: (1) Simple block schedules, (2) back-to-back schedules, (3) interdisciplinary schedules, (4) school-wide block schedules, (5) open-lab…
NASA Technical Reports Server (NTRS)
Rash, James L.
2010-01-01
NASA's space data-communications infrastructure, the Space Network and the Ground Network, provide scheduled (as well as some limited types of unscheduled) data-communications services to user spacecraft via orbiting relay satellites and ground stations. An implementation of the methods and algorithms disclosed herein will be a system that produces globally optimized schedules with not only optimized service delivery by the space data-communications infrastructure but also optimized satisfaction of all user requirements and prescribed constraints, including radio frequency interference (RFI) constraints. Evolutionary search, a class of probabilistic strategies for searching large solution spaces, constitutes the essential technology in this disclosure. Also disclosed are methods and algorithms for optimizing the execution efficiency of the schedule-generation algorithm itself. The scheduling methods and algorithms as presented are adaptable to accommodate the complexity of scheduling the civilian and/or military data-communications infrastructure. Finally, the problem itself, and the methods and algorithms, are generalized and specified formally, with applicability to a very broad class of combinatorial optimization problems.
Parallel scheduling algorithms
Dekel, E.; Sahni, S.
1983-01-01
Parallel algorithms are given for scheduling problems such as scheduling to minimize the number of tardy jobs, job sequencing with deadlines, scheduling to minimize earliness and tardiness penalties, channel assignment, and minimizing the mean finish time. The shared memory model of parallel computers is used to obtain fast algorithms. 26 references.
A knowledge-based system for diagnosis of mastitis problems at the herd level. 2. Machine milking.
Hogeveen, H; van Vliet, J H; Noordhuizen-Stassen, E N; De Koning, C; Tepp, D M; Brand, A
1995-07-01
A knowledge-based system for the diagnosis of mastitis problems at the herd level must search for possible causes, including malfunctioning milking machines or incorrect milking technique. A knowledge-based system on general mechanisms of mastitis infection, using hierarchical conditional causal models, was extended. Model building entailed extensive cooperation between the knowledge engineer and a domain expert. The extended knowledge-based system contains 12 submodels underlying the overview models. Nine submodels were concerned with mastitis problems arising from machine milking. These models are briefly described. The knowledge-based system has been validated by other experts after which the models were adjusted slightly. The final knowledge-based system was validated to data collected at 17 commercial dairy farms with high SCC in the bulk milk. Reports containing the farm data were accompanied by recommendations made by a dairy farm advisor. This validation showed good agreement between the knowledge-based system and the dairy farm advisors. The described knowledge-based system is a good tool for dairy farm advisors to solve herd mastitis problems caused by a malfunctioning milking machine or incorrect milking technique. PMID:7593837
Job Shop Scheduling Focusing on Role of Buffer
NASA Astrophysics Data System (ADS)
Hino, Rei; Kusumi, Tetsuya; Yoo, Jae-Kyu; Shimizu, Yoshiaki
A scheduling problem is formulated in order to consistently manage each manufacturing resource, including machine tools, assembly robots, AGV, storehouses, material shelves, and so on. The manufacturing resources are classified into three types: producer, location, and mover. This paper focuses especially on the role of the buffer, and the differences among these types are analyzed. A unified scheduling formulation is derived from the analytical results based on the resource’s roles. Scheduling procedures based on dispatching rules are also proposed in order to numerically evaluate job shop-type production having finite buffer capacity. The influences of the capacity of bottle-necked production devices and the buffer on productivity are discussed.
Time sharing massively parallel machines. Draft
Gorda, B.; Wolski, R.
1995-03-01
As part of the Massively Parallel Computing Initiative (MPCI) at the Lawrence Livermore National Laboratory, the authors have developed a simple, effective and portable time sharing mechanism by scheduling gangs of processes on tightly coupled parallel machines. By time-sharing the resources, the system interleaves production and interactive jobs. Immediate priority is given to interactive use, maintaining good response time. Production jobs are scheduled during idle periods, making use of the otherwise unused resources. In this paper the authors discuss their experience with gang scheduling over the 3 year life-time of the project. In section 2, they motivate the project and discuss some of its details. Section 3.0 describes the general scheduling problem and how gang scheduling addresses it. In section 4.0, they describe the implementation. Section 8.0 presents results culled over the lifetime of the project. They conclude this paper with some observations and possible future directions.
Efficient static scheduling of loops on synchronous multiprocessors
Zaky, A.M.
1989-01-01
This dissertation investigates efficient compile-time scheduling techniques for exploiting parallelism on synchronous multiprocessors. Synchronous multiprocessors, e.g. Very Long Instruction Word (VLIW) machines, are very effective in utilizing unstructured fine-grained parallelism in programs. The effectiveness of such machines is crucially dependent on the static compile-time analysis and detection of potential parallelism. The first part of the dissertation focuses on scheduling sequential loops on multiprocessors with multiple identical processor units. The problem of determining the maximal initiation rate for the execution of a sequential loop with uniform dependence distances on a synchronous multiprocessor is addressed and cast as an eigenvalue problem in a path algebra. A low-order polynomial algorithm for the determination of the optimal loop initiation rate is developed, and a schedule that exploits fine-grained parallelism and achieves the optimal initiation rate is developed under an idealized unbounded processor model. Next, the concepts developed above are extended to deal with perfectly-nested loops with uniform dependences. A strategy is developed to identify both the loop level and fine-grained expression level parallelism in nested loops, and to efficiently schedule such loops on synchronous multiprocessors. Loop scheduling techniques such as Do-Across, Wavefront scheduling, and fine-grained scheduling techniques such as loop unfolding are shown to be derivable within the presented framework.
Investigating the Human Computer Interaction Problems with Automated Teller Machine Navigation Menus
ERIC Educational Resources Information Center
Curran, Kevin; King, David
2008-01-01
Purpose: The automated teller machine (ATM) has become an integral part of our society. However, using the ATM can often be a frustrating experience as people frequently reinsert cards to conduct multiple transactions. This has led to the research question of whether ATM menus are designed in an optimal manner. This paper aims to address the…
Richards, E P; Walter, C
1989-01-01
This article is the first in a series about the dilemmas posed by products-liability law. The authors point out that manufacturers of complex medical devices may have an unrealistic expectation about user expertise, the lack of which often causes injury to patients, thereby leading to litigation over the devices. The learned-intermediary problem arises when engineers and lawyers disagree over whether a device is designed for a sophisticated rather than unsophisticated user. The authors discuss why the law assumes that anesthesia machines are like lawnmowers, which are usually operated by untrained users, rather than airplanes. However, they argue that anesthesia machines, being complicated to understand as well as dangerous, are much more like airplanes. They conclude that medical-device manufacturers must develop strategies to assure that legally risky devices are restricted to competent users; otherwise, innovation will suffer and legal costs will destroy the competitiveness of the US medical device manufacturers. PMID:18238310
ERIC Educational Resources Information Center
Karabinus, Robert A.; Boris, Richard
This document describes the development and implementation of a computer based registration and scheduling information system at Northern Illinois University. Because of personnel shortages, the University sought and received help from commercial computer firms. Implementation of scheduling, registration, and billing systems was accomplished in a…
Engineering solution to the problem of ingot solidification in slab continuous-casting machines
NASA Astrophysics Data System (ADS)
Shichkov, A. N.; Bykasova, E. N.; Bashirov, N. G.; Klochai, V. V.; Bystrov, L. G.
1995-03-01
An engineering solution to ingot solidification and the regularities of growth of an ingot envelope thickness and the coordinate of the end of slab solidification directly on slab continuous-casting machines (SCCM) are given, and ingot solidification conditions are determined. Examples of calculation of the envelope thickness and the coordinate of the end of solidification are provided for slab continuous-casting machines utilized at the Cherepovetsk integrated metallurgical complex (CherMC) and at the cast-and-iron works of the Aisenhüttenstadt Joint-Stock Company. A graphical algorithm for determining the cooling capacity of the secondary cooling zone is presented, and a nomogram for calibration of the cooling capacity of forced secondary cooling against the major and minor radii of an SCCM is developed.
Scheduling Jobs with Genetic Algorithms
NASA Astrophysics Data System (ADS)
Ferrolho, António; Crisóstomo, Manuel
Most scheduling problems are NP-hard, the time required to solve the problem optimally increases exponentially with the size of the problem. Scheduling problems have important applications, and a number of heuristic algorithms have been proposed to determine relatively good solutions in polynomial time. Recently, genetic algorithms (GA) are successfully used to solve scheduling problems, as shown by the growing numbers of papers. GA are known as one of the most efficient algorithms for solving scheduling problems. But, when a GA is applied to scheduling problems various crossovers and mutations operators can be applicable. This paper presents and examines a new concept of genetic operators for scheduling problems. A software tool called hybrid and flexible genetic algorithm (HybFlexGA) was developed to examine the performance of various crossover and mutation operators by computing simulations of job scheduling problems.
Approximation Schemes for Scheduling with Availability Constraints
NASA Astrophysics Data System (ADS)
Fu, Bin; Huo, Yumei; Zhao, Hairong
We investigate the problems of scheduling n weighted jobs to m identical machines with availability constraints. We consider two different models of availability constraints: the preventive model where the unavailability is due to preventive machine maintenance, and the fixed job model where the unavailability is due to a priori assignment of some of the n jobs to certain machines at certain times. Both models have applications such as turnaround scheduling or overlay computing. In both models, the objective is to minimize the total weighted completion time. We assume that m is a constant, and the jobs are non-resumable. For the preventive model, it has been shown that there is no approximation algorithm if all machines have unavailable intervals even when w i = p i for all jobs. In this paper, we assume there is one machine permanently available and the processing time of each job is equal to its weight for all jobs. We develop the first PTAS when there are constant number of unavailable intervals. One main feature of our algorithm is that the classification of large and small jobs is with respect to each individual interval, thus not fixed. This classification allows us (1) to enumerate the assignments of large jobs efficiently; (2) and to move small jobs around without increasing the objective value too much, and thus derive our PTAS. Then we show that there is no FPTAS in this case unless P = NP.
Automated telescope scheduling
NASA Technical Reports Server (NTRS)
Johnston, Mark D.
1988-01-01
With the ever increasing level of automation of astronomical telescopes the benefits and feasibility of automated planning and scheduling are becoming more apparent. Improved efficiency and increased overall telescope utilization are the most obvious goals. Automated scheduling at some level has been done for several satellite observatories, but the requirements on these systems were much less stringent than on modern ground or satellite observatories. The scheduling problem is particularly acute for Hubble Space Telescope: virtually all observations must be planned in excruciating detail weeks to months in advance. Space Telescope Science Institute has recently made significant progress on the scheduling problem by exploiting state-of-the-art artificial intelligence software technology. What is especially interesting is that this effort has already yielded software that is well suited to scheduling groundbased telescopes, including the problem of optimizing the coordinated scheduling of more than one telescope.
Automated telescope scheduling
NASA Astrophysics Data System (ADS)
Johnston, Mark D.
1988-08-01
With the ever increasing level of automation of astronomical telescopes the benefits and feasibility of automated planning and scheduling are becoming more apparent. Improved efficiency and increased overall telescope utilization are the most obvious goals. Automated scheduling at some level has been done for several satellite observatories, but the requirements on these systems were much less stringent than on modern ground or satellite observatories. The scheduling problem is particularly acute for Hubble Space Telescope: virtually all observations must be planned in excruciating detail weeks to months in advance. Space Telescope Science Institute has recently made significant progress on the scheduling problem by exploiting state-of-the-art artificial intelligence software technology. What is especially interesting is that this effort has already yielded software that is well suited to scheduling groundbased telescopes, including the problem of optimizing the coordinated scheduling of more than one telescope.
The preview control problem with application to man-machine system analysis
NASA Technical Reports Server (NTRS)
Tomizuka, M.; Whitney, D. E.
1973-01-01
The preview control problem is formulated in a general form and its solution is obtained. The analytical tool used is discrete stochastic optimal control theory. Aiming the application to manual control situations with preview, time delay, observation noise, motor noise, etc. were included in formulating the problem. Manual preview control experiments were performed to qualitatively check the validity of the model, and it was found that the mechanism of the manual control problem was explained by the developed model.
Scheduling constrained tools using heuristic techniques
NASA Astrophysics Data System (ADS)
Maram, Venkataramana; Rahman, Syariza Abdul; Maram, Sandhya Rani
2015-12-01
One of the main challenge to the current manufacturing production planning is to provide schedules of operations to maximize resource utilization to yield highest overall productivity. This is achieved by scheduling available resources to activities. There can be many different real time scenarios with different combination of input resources to produce parts. In this paper, the problem is simplified to single machine with individual process times and due dates to represent the real world scheduling problem. The main objective function is to minimize the total tardiness or late jobs. Nearest greedy method of assignment problem algorithm is used to find the initial solution followed by Simulated Annealing (SA) algorithm for the improvement part. Simulated Annealing is one of the meta-heuristic techniques in solving combinatorial optimization problem. The general purpose Microsoft Visual C++ is used to developed algorithm for finding the best solution. The proposed hybrid approach able to generate best schedule in 7th and optimal in 170th iteration with tardiness 8 and 7 hours respectively.
NASA Astrophysics Data System (ADS)
Effati, Meysam; Thill, Jean-Claude; Shabani, Shahin
2015-04-01
The contention of this paper is that many social science research problems are too "wicked" to be suitably studied using conventional statistical and regression-based methods of data analysis. This paper argues that an integrated geospatial approach based on methods of machine learning is well suited to this purpose. Recognizing the intrinsic wickedness of traffic safety issues, such approach is used to unravel the complexity of traffic crash severity on highway corridors as an example of such problems. The support vector machine (SVM) and coactive neuro-fuzzy inference system (CANFIS) algorithms are tested as inferential engines to predict crash severity and uncover spatial and non-spatial factors that systematically relate to crash severity, while a sensitivity analysis is conducted to determine the relative influence of crash severity factors. Different specifications of the two methods are implemented, trained, and evaluated against crash events recorded over a 4-year period on a regional highway corridor in Northern Iran. Overall, the SVM model outperforms CANFIS by a notable margin. The combined use of spatial analysis and artificial intelligence is effective at identifying leading factors of crash severity, while explicitly accounting for spatial dependence and spatial heterogeneity effects. Thanks to the demonstrated effectiveness of a sensitivity analysis, this approach produces comprehensive results that are consistent with existing traffic safety theories and supports the prioritization of effective safety measures that are geographically targeted and behaviorally sound on regional highway corridors.
Effects of problem structure and dialogue type on the performance of the man/machine interface
Palko, J.B.
1986-01-01
This research project is intended to provide guidance for software designers who must decide on a dialogue style for interactive problem solving support for new or infrequent users. In addition, it examines the relative performance of these dialogues in environments with a varying amount of problem structure. The three dialogue styles that are recommended by various writers in the field of information systems are: (1) menu driven, (2) question/response, and (3) form filling. It is concluded that if an interface is properly designed it probably does not matter which dialogue style is used in some problem settings. It is also clear that adding problem parameters reduces the ability of the problem solver to determine the best solution.
NASA Technical Reports Server (NTRS)
Messing, Fredric
1993-01-01
This paper provides an analytical formulation to predict scheduling success for a class of problems frequently referred to as activity scheduling. Space Network communications scheduling is an example of activity scheduling. The principal assumption is that the activity start times are randomly distributed over the available time in the time line. The formulation makes it possible to estimate how much of the demand can be scheduled as a function of the demand, number of resources, activity duration, and activity flexibility. The paper includes computed results for a variety of resource and demand conditions. The results demonstrate that even with highly flexible activities, it is difficult to schedule demand greater than 60 percent of resources without the use of optimization and conflict resolution capabilities in the scheduling system.
Automated Scheduling Via Artificial Intelligence
NASA Technical Reports Server (NTRS)
Biefeld, Eric W.; Cooper, Lynne P.
1991-01-01
Artificial-intelligence software that automates scheduling developed in Operations Mission Planner (OMP) research project. Software used in both generation of new schedules and modification of existing schedules in view of changes in tasks and/or available resources. Approach based on iterative refinement. Although project focused upon scheduling of operations of scientific instruments and other equipment aboard spacecraft, also applicable to such terrestrial problems as scheduling production in factory.
Research on schedulers for astronomical observatories
NASA Astrophysics Data System (ADS)
Colome, Josep; Colomer, Pau; Guàrdia, Josep; Ribas, Ignasi; Campreciós, Jordi; Coiffard, Thierry; Gesa, Lluis; Martínez, Francesc; Rodler, Florian
2012-09-01
The main task of a scheduler applied to astronomical observatories is the time optimization of the facility and the maximization of the scientific return. Scheduling of astronomical observations is an example of the classical task allocation problem known as the job-shop problem (JSP), where N ideal tasks are assigned to M identical resources, while minimizing the total execution time. A problem of higher complexity, called the Flexible-JSP (FJSP), arises when the tasks can be executed by different resources, i.e. by different telescopes, and it focuses on determining a routing policy (i.e., which machine to assign for each operation) other than the traditional scheduling decisions (i.e., to determine the starting time of each operation). In most cases there is no single best approach to solve the planning system and, therefore, various mathematical algorithms (Genetic Algorithms, Ant Colony Optimization algorithms, Multi-Objective Evolutionary algorithms, etc.) are usually considered to adapt the application to the system configuration and task execution constraints. The scheduling time-cycle is also an important ingredient to determine the best approach. A shortterm scheduler, for instance, has to find a good solution with the minimum computation time, providing the system with the capability to adapt the selected task to varying execution constraints (i.e., environment conditions). We present in this contribution an analysis of the task allocation problem and the solutions currently in use at different astronomical facilities. We also describe the schedulers for three different projects (CTA, CARMENES and TJO) where the conclusions of this analysis are applied to develop a suitable routine.
SPINN: a straightforward machine learning solution to the pulsar candidate selection problem
NASA Astrophysics Data System (ADS)
Morello, V.; Barr, E. D.; Bailes, M.; Flynn, C. M.; Keane, E. F.; van Straten, W.
2014-09-01
We describe SPINN (Straightforward Pulsar Identification using Neural Networks), a high-performance machine learning solution developed to process increasingly large data outputs from pulsar surveys. SPINN has been cross-validated on candidates from the southern High Time Resolution Universe (HTRU) survey and shown to identify every known pulsar found in the survey data while maintaining a false positive rate of 0.64 per cent. Furthermore, it ranks 99 per cent of pulsars among the top 0.11 per cent of candidates, and 95 per cent among the top 0.01 per cent. In conjunction with the PEASOUP pipeline, it has already discovered four new pulsars in a re-processing of the intermediate Galactic latitude area of HTRU, three of which have spin periods shorter than 5 ms. SPINN's ability to reduce the amount of candidates to visually inspect by up to four orders of magnitude makes it a very promising tool for future large-scale pulsar surveys. In an effort to provide a common testing ground for pulsar candidate selection tools and stimulate interest in their development, we also make publicly available the set of candidates on which SPINN was cross-validated.
Engineering-Based Problem Solving in the Middle School: Design and Construction with Simple Machines
ERIC Educational Resources Information Center
English, Lyn D.; Hudson, Peter; Dawes, Les
2013-01-01
Incorporating engineering concepts into middle school curriculum is seen as an effective way to improve students' problem-solving skills. A selection of findings is reported from a science, technology, engineering and mathematics (STEM)-based unit in which students in the second year (grade 8) of a three-year longitudinal study explored…
MONTO: A Machine-Readable Ontology for Teaching Word Problems in Mathematics
ERIC Educational Resources Information Center
Lalingkar, Aparna; Ramnathan, Chandrashekar; Ramani, Srinivasan
2015-01-01
The Indian National Curriculum Framework has as one of its objectives the development of mathematical thinking and problem solving ability. However, recent studies conducted in Indian metros have expressed concern about students' mathematics learning. Except in some private coaching academies, regular classroom teaching does not include problem…
NASA Astrophysics Data System (ADS)
Chen, Xili; Hao, Xinchang; Lin, Hao Wen; Murata, Tomohiro
The applications of composite dispatching rules for multi objective dynamic scheduling have been widely studied in literature. In general, a composite dispatching rule is a combination of several elementary dispatching rules, which is designed to optimize multiple objectives of interest under a certain scheduling environment. The relative importance of elementary dispatching rules is modeled by weight factors. A critical issue for implementation of composite dispatching rule is that the inappropriate weight values may result in poor performance. This paper presents an offline scheduling knowledge acquisition method based on reinforcement learning using simulation technique. The scheduling knowledge is applied to adjust the appropriate weight values of elementary dispatching rules in composite manner with respect to work in process fluctuation of machines during online scheduling. Implementation of the proposed method in a two objectives dynamic job shop scheduling problem is demonstrated and the results are satisfactory.
Completable scheduling: An integrated approach to planning and scheduling
NASA Technical Reports Server (NTRS)
Gervasio, Melinda T.; Dejong, Gerald F.
1992-01-01
The planning problem has traditionally been treated separately from the scheduling problem. However, as more realistic domains are tackled, it becomes evident that the problem of deciding on an ordered set of tasks to achieve a set of goals cannot be treated independently of the problem of actually allocating resources to the tasks. Doing so would result in losing the robustness and flexibility needed to deal with imperfectly modeled domains. Completable scheduling is an approach which integrates the two problems by allowing an a priori planning module to defer particular planning decisions, and consequently the associated scheduling decisions, until execution time. This allows a completable scheduling system to maximize plan flexibility by allowing runtime information to be taken into consideration when making planning and scheduling decision. Furthermore, through the criteria of achievability placed on deferred decision, a completable scheduling system is able to retain much of the goal-directedness and guarantees of achievement afforded by a priori planning. The completable scheduling approach is further enhanced by the use of contingent explanation-based learning, which enables a completable scheduling system to learn general completable plans from example and improve its performance through experience. Initial experimental results show that completable scheduling outperforms classical scheduling as well as pure reactive scheduling in a simple scheduling domain.
Planning and scheduling research at NASA Ames Research Center
NASA Technical Reports Server (NTRS)
Friedland, Peter
1990-01-01
Planning and scheduling is the area of artificial intelligence research that focuses on the determination of a series of operations to achieve some set of (possibly) interacting goals and the placement of those operations in a timeline that allows them to be accomplished given available resources. Work in this area at the NASA Ames Research Center ranging from basic research in constrain-based reasoning and machine learning, to the development of efficient scheduling tools, to the application of such tools to complex agency problems is described.
Adaptive Parallel Job Scheduling with Flexible CoScheduling
Frachtenberg, Eitan; Feitelson, Dror; Petrini, Fabrizio; Fernandez, Juan
2005-11-01
Abstract—Many scientific and high-performance computing applications consist of multiple processes running on different processors that communicate frequently. Because of their synchronization needs, these applications can suffer severe performance penalties if their processes are not all coscheduled to run together. Two common approaches to coscheduling jobs are batch scheduling, wherein nodes are dedicated for the duration of the run, and gang scheduling, wherein time slicing is coordinated across processors. Both work well when jobs are load-balanced and make use of the entire parallel machine. However, these conditions are rarely met and most realistic workloads consequently suffer from both internal and external fragmentation, in which resources and processors are left idle because jobs cannot be packed with perfect efficiency. This situation leads to reduced utilization and suboptimal performance. Flexible CoScheduling (FCS) addresses this problem by monitoring each job’s computation granularity and communication pattern and scheduling jobs based on their synchronization and load-balancing requirements. In particular, jobs that do not require stringent synchronization are identified, and are not coscheduled; instead, these processes are used to reduce fragmentation. FCS has been fully implemented on top of the STORM resource manager on a 256-processor Alpha cluster and compared to batch, gang, and implicit coscheduling algorithms. This paper describes in detail the implementation of FCS and its performance evaluation with a variety of workloads, including large-scale benchmarks, scientific applications, and dynamic workloads. The experimental results show that FCS saturates at higher loads than other algorithms (up to 54 percent higher in some cases), and displays lower response times and slowdown than the other algorithms in nearly all scenarios.
Multiagent scheduling method with earliness and tardiness objectives in flexible job shops.
Wu, Zuobao; Weng, Michael X
2005-04-01
Flexible job-shop scheduling problems are an important extension of the classical job-shop scheduling problems and present additional complexity. Such problems are mainly due to the existence of a considerable amount of overlapping capacities with modern machines. Classical scheduling methods are generally incapable of addressing such capacity overlapping. We propose a multiagent scheduling method with job earliness and tardiness objectives in a flexible job-shop environment. The earliness and tardiness objectives are consistent with the just-in-time production philosophy which has attracted significant attention in both industry and academic community. A new job-routing and sequencing mechanism is proposed. In this mechanism, two kinds of jobs are defined to distinguish jobs with one operation left from jobs with more than one operation left. Different criteria are proposed to route these two kinds of jobs. Job sequencing enables to hold a job that may be completed too early. Two heuristic algorithms for job sequencing are developed to deal with these two kinds of jobs. The computational experiments show that the proposed multiagent scheduling method significantly outperforms the existing scheduling methods in the literature. In addition, the proposed method is quite fast. In fact, the simulation time to find a complete schedule with over 2000 jobs on ten machines is less than 1.5 min. PMID:15828657
Wayne F. Boyer; Gurdeep S. Hura
2005-09-01
The Problem of obtaining an optimal matching and scheduling of interdependent tasks in distributed heterogeneous computing (DHC) environments is well known to be an NP-hard problem. In a DHC system, task execution time is dependent on the machine to which it is assigned and task precedence constraints are represented by a directed acyclic graph. Recent research in evolutionary techniques has shown that genetic algorithms usually obtain more efficient schedules that other known algorithms. We propose a non-evolutionary random scheduling (RS) algorithm for efficient matching and scheduling of inter-dependent tasks in a DHC system. RS is a succession of randomized task orderings and a heuristic mapping from task order to schedule. Randomized task ordering is effectively a topological sort where the outcome may be any possible task order for which the task precedent constraints are maintained. A detailed comparison to existing evolutionary techniques (GA and PSGA) shows the proposed algorithm is less complex than evolutionary techniques, computes schedules in less time, requires less memory and fewer tuning parameters. Simulation results show that the average schedules produced by RS are approximately as efficient as PSGA schedules for all cases studied and clearly more efficient than PSGA for certain cases. The standard formulation for the scheduling problem addressed in this paper is Rm|prec|Cmax.,
NASA Technical Reports Server (NTRS)
Tanner, Steve; Hughes, Angi; Byrd, Jim
1987-01-01
Resupply scheduling for the Space Station presents some formidable logistics problems. One of the most basic problems is assigning supplies to a series of shuttle resupply missions. A prototype logistics expert system which constructs resupply schedules was developed. This prototype is able to reconstruct feasible resupply plans. In addition, analysts can use the system to evaluate the impact of adding, deleting or modifying launches, cargo space, experiments, etc.
State-based scheduling: An architecture for telescope observation scheduling
NASA Technical Reports Server (NTRS)
Muscettola, Nicola; Smith, Stephen F.
1989-01-01
The applicability of constraint-based scheduling, a methodology previously developed and validated in the domain of factory scheduling, is extended to problem domains that require attendance to a wider range of state-dependent constraints. The problem of constructing and maintaining a short-term observation schedule for the Hubble Space Telescope (HST), which typifies this type of domain is the focus of interest. The nature of the constraints encountered in the HST domain is examined, system requirements are discussed with respect to utilization of a constraint-based scheduling methodology in such domains, and a general framework for state-based scheduling is presented.
... Labor & birth > Scheduling a c-section Scheduling a c-section E-mail to a friend Please fill ... develop before she’s born. Why can scheduling a c-section for non-medical reasons be a problem? ...
Sun, Yan; Lang, Maoxiang; Wang, Danzhu
2016-01-01
The transportation of hazardous materials is always accompanied by considerable risk that will impact public and environment security. As an efficient and reliable transportation organization, a multimodal service should participate in the transportation of hazardous materials. In this study, we focus on transporting hazardous materials through the multimodal service network and explore the hazardous materials multimodal routing problem from the operational level of network planning. To formulate this problem more practicably, minimizing the total generalized costs of transporting the hazardous materials and the social risk along the planned routes are set as the optimization objectives. Meanwhile, the following formulation characteristics will be comprehensively modelled: (1) specific customer demands; (2) multiple hazardous material flows; (3) capacitated schedule-based rail service and uncapacitated time-flexible road service; and (4) environmental risk constraint. A bi-objective mixed integer nonlinear programming model is first built to formulate the routing problem that combines the formulation characteristics above. Then linear reformations are developed to linearize and improve the initial model so that it can be effectively solved by exact solution algorithms on standard mathematical programming software. By utilizing the normalized weighted sum method, we can generate the Pareto solutions to the bi-objective optimization problem for a specific case. Finally, a large-scale empirical case study from the Beijing-Tianjin-Hebei Region in China is presented to demonstrate the feasibility of the proposed methods in dealing with the practical problem. Various scenarios are also discussed in the case study. PMID:27483294
Sun, Yan; Lang, Maoxiang; Wang, Danzhu
2016-01-01
The transportation of hazardous materials is always accompanied by considerable risk that will impact public and environment security. As an efficient and reliable transportation organization, a multimodal service should participate in the transportation of hazardous materials. In this study, we focus on transporting hazardous materials through the multimodal service network and explore the hazardous materials multimodal routing problem from the operational level of network planning. To formulate this problem more practicably, minimizing the total generalized costs of transporting the hazardous materials and the social risk along the planned routes are set as the optimization objectives. Meanwhile, the following formulation characteristics will be comprehensively modelled: (1) specific customer demands; (2) multiple hazardous material flows; (3) capacitated schedule-based rail service and uncapacitated time-flexible road service; and (4) environmental risk constraint. A bi-objective mixed integer nonlinear programming model is first built to formulate the routing problem that combines the formulation characteristics above. Then linear reformations are developed to linearize and improve the initial model so that it can be effectively solved by exact solution algorithms on standard mathematical programming software. By utilizing the normalized weighted sum method, we can generate the Pareto solutions to the bi-objective optimization problem for a specific case. Finally, a large-scale empirical case study from the Beijing–Tianjin–Hebei Region in China is presented to demonstrate the feasibility of the proposed methods in dealing with the practical problem. Various scenarios are also discussed in the case study. PMID:27483294
Conway, L.
1984-01-01
To meet the challenge of certain critical problems in defense, the Defense Advanced Research Projects Agency (DARPA) is initiating an important new program in strategic computing. By seizing an opportunity to leverage recent advances in artificial intelligence, computer science, and microelectronics, the agency plans to create a new generation of machine-intelligence technology.
Minimizing makespan on parallel machines with batch arrivals
NASA Astrophysics Data System (ADS)
Chung, Tsui-Ping; Liao, Ching-Jong; Lin, Chien-Hung
2012-04-01
Most studies in the scheduling literature assume that jobs arrive at time zero, while some studies assume that jobs arrive individually at non-zero times. However, both assumptions may not be valid in practice because jobs usually arrive in batches. In this article, a scheduling model for an identical parallel machine problem with batch arrivals is formulated. Because of the NP-hardness of the problem, a heuristic based on a simplified version of lexicographical search is proposed. To verify the heuristic, two lower bounding schemes are developed, where one lower bound is tight, and the list scheduling heuristic is compared. Extensive computational experiments demonstrate that the proposed heuristic is quite efficient in obtaining near optimal solution with an average error of less than 1.58%. The percentage improvement (from the lower bound) of the heuristic solution on the solution by the list scheduling is as large as 31.68.
NASA Astrophysics Data System (ADS)
Kalsom Yusof, Umi; Nor Akmal Khalid, Mohd
2015-05-01
Semiconductor industries need to constantly adjust to the rapid pace of change in the market. Most manufactured products usually have a very short life cycle. These scenarios imply the need to improve the efficiency of capacity planning, an important aspect of the machine allocation plan known for its complexity. Various studies have been performed to balance productivity and flexibility in the flexible manufacturing system (FMS). Many approaches have been developed by the researchers to determine the suitable balance between exploration (global improvement) and exploitation (local improvement). However, not much work has been focused on the domain of machine allocation problem that considers the effects of machine breakdowns. This paper develops a model to minimize the effect of machine breakdowns, thus increasing the productivity. The objectives are to minimize system unbalance and makespan as well as increase throughput while satisfying the technological constraints such as machine time availability. To examine the effectiveness of the proposed model, results for throughput, system unbalance and makespan on real industrial datasets were performed with applications of intelligence techniques, that is, a hybrid of genetic algorithm and harmony search. The result aims to obtain a feasible solution to the domain problem.
Cakar, Tarik; Koker, Rasit
2015-01-01
A particle swarm optimization algorithm (PSO) has been used to solve the single machine total weighted tardiness problem (SMTWT) with unequal release date. To find the best solutions three different solution approaches have been used. To prepare subhybrid solution system, genetic algorithms (GA) and simulated annealing (SA) have been used. In the subhybrid system (GA and SA), GA obtains a solution in any stage, that solution is taken by SA and used as an initial solution. When SA finds better solution than this solution, it stops working and gives this solution to GA again. After GA finishes working the obtained solution is given to PSO. PSO searches for better solution than this solution. Later it again sends the obtained solution to GA. Three different solution systems worked together. Neurohybrid system uses PSO as the main optimizer and SA and GA have been used as local search tools. For each stage, local optimizers are used to perform exploitation to the best particle. In addition to local search tools, neurodominance rule (NDR) has been used to improve performance of last solution of hybrid-PSO system. NDR checked sequential jobs according to total weighted tardiness factor. All system is named as neurohybrid-PSO solution system. PMID:26221134
Cakar, Tarik; Koker, Rasit
2015-01-01
A particle swarm optimization algorithm (PSO) has been used to solve the single machine total weighted tardiness problem (SMTWT) with unequal release date. To find the best solutions three different solution approaches have been used. To prepare subhybrid solution system, genetic algorithms (GA) and simulated annealing (SA) have been used. In the subhybrid system (GA and SA), GA obtains a solution in any stage, that solution is taken by SA and used as an initial solution. When SA finds better solution than this solution, it stops working and gives this solution to GA again. After GA finishes working the obtained solution is given to PSO. PSO searches for better solution than this solution. Later it again sends the obtained solution to GA. Three different solution systems worked together. Neurohybrid system uses PSO as the main optimizer and SA and GA have been used as local search tools. For each stage, local optimizers are used to perform exploitation to the best particle. In addition to local search tools, neurodominance rule (NDR) has been used to improve performance of last solution of hybrid-PSO system. NDR checked sequential jobs according to total weighted tardiness factor. All system is named as neurohybrid-PSO solution system. PMID:26221134
NASA Astrophysics Data System (ADS)
Pei, Jun; Liu, Xinbao; Pardalos, Panos M.; Fan, Wenjuan; Wang, Ling; Yang, Shanlin
2016-03-01
Motivated by applications in manufacturing industry, we consider a supply chain scheduling problem, where each job is characterised by non-identical sizes, different release times and unequal processing times. The objective is to minimise the makespan by making batching and sequencing decisions. The problem is formalised as a mixed integer programming model and proved to be strongly NP-hard. Some structural properties are presented for both the general case and a special case. Based on these properties, a lower bound is derived, and a novel two-phase heuristic (TP-H) is developed to solve the problem, which guarantees to obtain a worst case performance ratio of ?. Computational experiments with a set of different sizes of random instances are conducted to evaluate the proposed approach TP-H, which is superior to another two heuristics proposed in the literature. Furthermore, the experimental results indicate that TP-H can effectively and efficiently solve large-size problems in a reasonable time.
Genetic algorithm based system for patient scheduling in highly constrained situations.
Podgorelec, V; Kokol, P
1997-12-01
In medicine and health care there are a lot of situations when patients have to be scheduled on different devices and/or with different physicians or therapists. It may concern preventive examinations, laboratory tests or convalescent therapies, therefore we are always looking for an optimal schedule that would result in finishing all the activities scheduled as soon as possible, with the least patient waiting time and maximum device utilization. Since patient scheduling is a highly complex problem, it is impossible to make a qualitative schedule by hand or even with exact heuristic methods. Therefore we developed a powerful automated scheduling method for highly constrained situations based on genetic algorithms and machine learning. In this paper we present the method, together with the whole process of schedule generation, the important parameters to direct the evolution and how the algorithm is guaranteed to produce only feasible solutions, not breaking any of the required constraints. We applied the described method to a problem of scheduling patients with different therapy needs to a limited number of therapeutic devices, but the algorithm can be easily modified for use in similar situations. The results are quite encouraging and since all the solutions are feasible, the method can be easily incorporated into an interactive user interface, which can be of major importance when scheduling patients, and human resources in general, is considered. PMID:9555628
NASA Astrophysics Data System (ADS)
Paksi, A. B. N.; Ma'ruf, A.
2016-02-01
In general, both machines and human resources are needed for processing a job on production floor. However, most classical scheduling problems have ignored the possible constraint caused by availability of workers and have considered only machines as a limited resource. In addition, along with production technology development, routing flexibility appears as a consequence of high product variety and medium demand for each product. Routing flexibility is caused by capability of machines that offers more than one machining process. This paper presents a method to address scheduling problem constrained by both machines and workers, considering routing flexibility. Scheduling in a Dual-Resource Constrained shop is categorized as NP-hard problem that needs long computational time. Meta-heuristic approach, based on Genetic Algorithm, is used due to its practical implementation in industry. Developed Genetic Algorithm uses indirect chromosome representative and procedure to transform chromosome into Gantt chart. Genetic operators, namely selection, elitism, crossover, and mutation are developed to search the best fitness value until steady state condition is achieved. A case study in a manufacturing SME is used to minimize tardiness as objective function. The algorithm has shown 25.6% reduction of tardiness, equal to 43.5 hours.
Tricker, Christopher; Rock, Adam J; Clark, Gavin I
2016-06-01
In order to enhance our understanding of the nature of poker-machine problem-gambling, a community sample of 37 poker-machine gamblers (M age = 32 years, M PGSI = 5; PGSI = Problem Gambling Severity Index) were assessed for urge to gamble (responses on a visual analogue scale) and altered state of consciousness (assessed by the Altered State of Awareness dimension of the Phenomenology of Consciousness Inventory) at baseline, after a neutral cue, and after a gambling cue. It was found that (a) problem-gambling severity (PGSI score) predicted increase in urge (from neutral cue to gambling cue, controlling for baseline; sr (2) = .19, p = .006) and increase in altered state of consciousness (from neutral cue to gambling cue, controlling for baseline; sr (2) = .57, p < .001), and (b) increase in altered state of consciousness (from neutral cue to gambling cue) mediated the relationship between problem-gambling severity and increase in urge (from neutral cue to gambling cue; κ(2) = .40, 99 % CI [.08, .71]). These findings suggest that cue-reactive altered state of consciousness is an important component of cue-reactive urge in poker-machine problem-gamblers. PMID:26026986
Artificial intelligence approaches to astronomical observation scheduling
NASA Technical Reports Server (NTRS)
Johnston, Mark D.; Miller, Glenn
1988-01-01
Automated scheduling will play an increasing role in future ground- and space-based observatory operations. Due to the complexity of the problem, artificial intelligence technology currently offers the greatest potential for the development of scheduling tools with sufficient power and flexibility to handle realistic scheduling situations. Summarized here are the main features of the observatory scheduling problem, how artificial intelligence (AI) techniques can be applied, and recent progress in AI scheduling for Hubble Space Telescope.
A Genetic Algorithm for Flow Shop Scheduling with Assembly Operations to Minimize Makespan
NASA Astrophysics Data System (ADS)
Bhongade, A. S.; Khodke, P. M.
2014-04-01
Manufacturing systems, in which, several parts are processed through machining workstations and later assembled to form final products, is common. Though scheduling of such problems are solved using heuristics, available solution approaches can provide solution for only moderate sized problems due to large computation time required. In this work, scheduling approach is developed for such flow-shop manufacturing system having machining workstations followed by assembly workstations. The initial schedule is generated using Disjunctive method and genetic algorithm (GA) is applied further for generating schedule for large sized problems. GA is found to give near optimal solution based on the deviation of makespan from lower bound. The lower bound of makespan of such problem is estimated and percent deviation of makespan from lower bounds is used as a performance measure to evaluate the schedules. Computational experiments are conducted on problems developed using fractional factorial orthogonal array, varying the number of parts per product, number of products, and number of workstations (ranging upto 1,520 number of operations). A statistical analysis indicated the significance of all the three factors considered. It is concluded that GA method can obtain optimal makespan.
NASA Technical Reports Server (NTRS)
Borse, John E.; Owens, Christopher C.
1992-01-01
Our research focuses on the problem of recovering from perturbations in large-scale schedules, specifically on the ability of a human-machine partnership to dynamically modify an airline schedule in response to unanticipated disruptions. This task is characterized by massive interdependencies and a large space of possible actions. Our approach is to apply the following: qualitative, knowledge-intensive techniques relying on a memory of stereotypical failures and appropriate recoveries; and quantitative techniques drawn from the Operations Research community's work on scheduling. Our main scientific challenge is to represent schedules, failures, and repairs so as to make both sets of techniques applicable to the same data. This paper outlines ongoing research in which we are cooperating with United Airlines to develop our understanding of the scientific issues underlying the practicalities of dynamic, real-time schedule repair.
Flexible Job Shop Scheduling with Multi-level Job Structures
NASA Astrophysics Data System (ADS)
Jang, Yang-Ja; Kim, Ki-Dong; Jang, Seong-Yong; Park, Jinwoo
This paper deals with a scheduling problem in a flexible job shop with multi-level job structures where end products are assembled from sub-assemblies or manufactured components. For such shops MRP (Material Requirement Planning) logic is frequently used to synchronize and pace the production activities for the required parts. However, in MRP, the planning of operational-level activities is left to short term scheduling. So, we need a good scheduling algorithm to generate feasible schedules taking into account shop floor characteristics and multi-level job structures used in MRP. In this paper, we present a GA (Genetic Algorithm) solution for this complex scheduling problem based on a new gene to reflect the machine assignment, operation sequences and the levels of the operations relative to final assembly operation. The relative operation level is the control parameter that paces the completion timing of the components belonging to the same branch in the multi-level job hierarchy. We compare the genetic algorithm with several dispatching rules in terms of total tardiness and the genetic algorithm shows outstanding performance for about forty modified standard job-shop problem instances.
NASA Astrophysics Data System (ADS)
Setiawan, A.; Wangsaputra, R.; Martawirya, Y. Y.; Halim, A. H.
2016-02-01
This paper deals with Flexible Manufacturing System (FMS) production rescheduling due to unavailability of cutting tools caused either of cutting tool failure or life time limit. The FMS consists of parallel identical machines integrated with an automatic material handling system and it runs fully automatically. Each machine has a same cutting tool configuration that consists of different geometrical cutting tool types on each tool magazine. The job usually takes two stages. Each stage has sequential operations allocated to machines considering the cutting tool life. In the real situation, the cutting tool can fail before the cutting tool life is reached. The objective in this paper is to develop a dynamic scheduling algorithm when a cutting tool is broken during unmanned and a rescheduling needed. The algorithm consists of four steps. The first step is generating initial schedule, the second step is determination the cutting tool failure time, the third step is determination of system status at cutting tool failure time and the fourth step is the rescheduling for unfinished jobs. The approaches to solve the problem are complete-reactive scheduling and robust-proactive scheduling. The new schedules result differences starting time and completion time of each operations from the initial schedule.
Zhang, Weizhe; Bai, Enci; He, Hui; Cheng, Albert M.K.
2015-01-01
Reducing energy consumption is becoming very important in order to keep battery life and lower overall operational costs for heterogeneous real-time multiprocessor systems. In this paper, we first formulate this as a combinatorial optimization problem. Then, a successful meta-heuristic, called Shuffled Frog Leaping Algorithm (SFLA) is proposed to reduce the energy consumption. Precocity remission and local optimal avoidance techniques are proposed to avoid the precocity and improve the solution quality. Convergence acceleration significantly reduces the search time. Experimental results show that the SFLA-based energy-aware meta-heuristic uses 30% less energy than the Ant Colony Optimization (ACO) algorithm, and 60% less energy than the Genetic Algorithm (GA) algorithm. Remarkably, the running time of the SFLA-based meta-heuristic is 20 and 200 times less than ACO and GA, respectively, for finding the optimal solution. PMID:26110406
Zhang, Weizhe; Bai, Enci; He, Hui; Cheng, Albert M K
2015-01-01
Reducing energy consumption is becoming very important in order to keep battery life and lower overall operational costs for heterogeneous real-time multiprocessor systems. In this paper, we first formulate this as a combinatorial optimization problem. Then, a successful meta-heuristic, called Shuffled Frog Leaping Algorithm (SFLA) is proposed to reduce the energy consumption. Precocity remission and local optimal avoidance techniques are proposed to avoid the precocity and improve the solution quality. Convergence acceleration significantly reduces the search time. Experimental results show that the SFLA-based energy-aware meta-heuristic uses 30% less energy than the Ant Colony Optimization (ACO) algorithm, and 60% less energy than the Genetic Algorithm (GA) algorithm. Remarkably, the running time of the SFLA-based meta-heuristic is 20 and 200 times less than ACO and GA, respectively, for finding the optimal solution. PMID:26110406
DTS: Building custom, intelligent schedulers
NASA Technical Reports Server (NTRS)
Hansson, Othar; Mayer, Andrew
1994-01-01
DTS is a decision-theoretic scheduler, built on top of a flexible toolkit -- this paper focuses on how the toolkit might be reused in future NASA mission schedulers. The toolkit includes a user-customizable scheduling interface, and a 'Just-For-You' optimization engine. The customizable interface is built on two metaphors: objects and dynamic graphs. Objects help to structure problem specifications and related data, while dynamic graphs simplify the specification of graphical schedule editors (such as Gantt charts). The interface can be used with any 'back-end' scheduler, through dynamically-loaded code, interprocess communication, or a shared database. The 'Just-For-You' optimization engine includes user-specific utility functions, automatically compiled heuristic evaluations, and a postprocessing facility for enforcing scheduling policies. The optimization engine is based on BPS, the Bayesian Problem-Solver (1,2), which introduced a similar approach to solving single-agent and adversarial graph search problems.
GAMMA: a high performance dataflow database machine
DeWitt, D.J.; Gerber, R.H.; Graefe, G.; Heytens, M.L.; Kumar, K.B.; Muralikrishna, M.
1986-03-01
In this paper, the design, implementation techniques, and initial performance evaluation of Gamma are presented. Gamma is a new relational database machine that exploits dataflow query processing techniques. Ganma is a fully operational prototype consisting of 20 VAX 11/750 computers. The design of Gamma is based on an earlier multiprocessor database machine prototype (DIRECT) and several years of subsequent research on the problems raised by the DIRECT prototype. In addition to demonstrating that parallelism can really be made to work in a database machine context, the Gamma prototype shows how parallelism can be controlled with minimal control overhead through a combination of the use of algorithms based on hashing and the pipelining of data between processes. Except for 2 messages to initiate each operator of a query tree and 1 message when the operator terminates, the execution of a query is entirely self-scheduling. 52 refs., 12 figs.
Kleck, W
1982-04-01
Structuring a schedule - whether by Critical Path Method (CPM) or Precedence Charting System (PCS) - involves estimating the duration of one or more activities and arranging them in the most logical sequence. Given the start date, the completion date is relatively simple to determine. What is then so complicated about the process. It is complicated by the people involved - the people who make the schedules and the people who attempt to follow them. Schedules are an essential part of project management and construction contract administration. Much of the material available pertains to the mechanics of schedules, the types of logic networks, the ways that data can be generated and presented. This paper sheds light on other facets of the subject - the statistical and philosophical fundamentals involved in scheduling.
McCormick, Jessica; Delfabbro, Paul; Denson, Linley A
2012-12-01
The aim of this study was to conduct an empirical investigation of the validity of Jacobs' (in J Gambl Behav 2:15-31, 1986) general theory of addictions in relation to gambling problems associated with electronic gaming machines (EGM). Regular EGM gamblers (n = 190) completed a series of standardised measures relating to psychological and physiological vulnerability, substance use, dissociative experiences, early childhood trauma and abuse and problem gambling (the Problem Gambling Severity Index). Statistical analysis using structural equation modelling revealed clear relationships between childhood trauma and life stressors and psychological vulnerability, dissociative-like experiences and problem gambling. These findings confirm and extend a previous model validated by Gupta and Derevensky (in J Gambl Stud 14: 17-49, 1998) using an adolescent population. The significance of these findings are discussed for existing pathway models of problem gambling, for Jacobs' theory, and for clinicians engaged in assessment and intervention. PMID:22116713
NASA Astrophysics Data System (ADS)
Ondrášek, J.
The paper deals with the issues of self-excited vibration. Such vibration occurs in those systems in which an internal source continuously exists from which the system draws power to maintain or even increase the amplitude of vibration when this take-off is controlled by the oscillating motion of the system itself. Thus, this energy source considerably influences the dynamic and stability features of the system. An example of such a system is a machine tool during machining as well when a part of the energy of the cutting process in chip machining can be changed into energy that will be vibrating the machine as a whole. The vibrations are then shown with a strong machined surface waviness and are usually accompanied by noise. In general, it can be determined the range of cutting conditions in which, when applying them, no chatter will occur, i.e. the cutting process is stable. One of the ways of such a determining is a stability lobe diagram e.g. which expresses the dependence of chip thickness on the speed of a workpiece. The subject of this paper is a possible procedure and approach when creating stability lobe diagrams.
NASA Technical Reports Server (NTRS)
Zweben, Monte
1991-01-01
The GERRY scheduling system developed by NASA Ames with assistance from the Lockheed Space Operations Company, and the Lockheed Artificial Intelligence Center, uses a method called constraint-based iterative repair. Using this technique, one encodes both hard rules and preference criteria into data structures called constraints. GERRY repeatedly attempts to improve schedules by seeking repairs for violated constraints. The system provides a general scheduling framework which is being tested on two NASA applications. The larger of the two is the Space Shuttle Ground Processing problem which entails the scheduling of all the inspection, repair, and maintenance tasks required to prepare the orbiter for flight. The other application involves power allocation for the NASA Ames wind tunnels. Here the system will be used to schedule wind tunnel tests with the goal of minimizing power costs. In this paper, we describe the GERRY system and its application to the Space Shuttle problem. We also speculate as to how the system would be used for manufacturing, transportation, and military problems.
NASA Technical Reports Server (NTRS)
Zweben, Monte
1993-01-01
The GERRY scheduling system developed by NASA Ames with assistance from the Lockheed Space Operations Company, and the Lockheed Artificial Intelligence Center, uses a method called constraint-based iterative repair. Using this technique, one encodes both hard rules and preference criteria into data structures called constraints. GERRY repeatedly attempts to improve schedules by seeking repairs for violated constraints. The system provides a general scheduling framework which is being tested on two NASA applications. The larger of the two is the Space Shuttle Ground Processing problem which entails the scheduling of all the inspection, repair, and maintenance tasks required to prepare the orbiter for flight. The other application involves power allocation for the NASA Ames wind tunnels. Here the system will be used to schedule wind tunnel tests with the goal of minimizing power costs. In this paper, we describe the GERRY system and its application to the Space Shuttle problem. We also speculate as to how the system would be used for manufacturing, transportation, and military problems.
NASA Technical Reports Server (NTRS)
Zweben, Monte
1991-01-01
The GERRY scheduling system developed by NASA Ames with assistance from the Lockheed Space Operations Company, and the Lockheed Artificial Intelligence Center, uses a method called constraint based iterative repair. Using this technique, one encodes both hard rules and preference criteria into data structures called constraints. GERRY repeatedly attempts to improve schedules by seeking repairs for violated constraints. The system provides a general scheduling framework which is being tested on two NASA applications. The larger of the two is the Space Shuttle Ground Processing problem which entails the scheduling of all inspection, repair, and maintenance tasks required to prepare the orbiter for flight. The other application involves power allocations for the NASA Ames wind tunnels. Here the system will be used to schedule wind tunnel tests with the goal of minimizing power costs. In this paper, we describe the GERRY system and its applications to the Space Shuttle problem. We also speculate as to how the system would be used for manufacturing, transportation, and military problems.
Enhanced Memetic Algorithm for Task Scheduling
NASA Astrophysics Data System (ADS)
Padmavathi, S.; Shalinie, S. Mercy; Someshwar, B. C.; Sasikumar, T.
Scheduling tasks onto the processors of a parallel system is a crucial part of program parallelization. Due to the NP-hardness of the task scheduling problem, scheduling algorithms are based on heuristics that try to produce good rather than optimal schedules. This paper proposes a Memetic algorithm with Tabu search and Simulated Annealing as local search for solving Task scheduling problem considering communication contention. This problem consists of finding a schedule for a general task graph to be executed on a cluster of workstations and hence the schedule length can be minimized. Our approach combines local search (by self experience) and global search (by neighboring experience) possessing high search efficiency. The proposed approach is compared with existing list scheduling heuristics. The numerical results clearly indicate that our proposed approach produces solutions which are closer to optimality and/or better quality than the existing list scheduling heuristics.
Scheduling: A guide for program managers
NASA Technical Reports Server (NTRS)
1994-01-01
The following topics are discussed concerning scheduling: (1) milestone scheduling; (2) network scheduling; (3) program evaluation and review technique; (4) critical path method; (5) developing a network; (6) converting an ugly duckling to a swan; (7) network scheduling problem; (8) (9) network scheduling when resources are limited; (10) multi-program considerations; (11) influence on program performance; (12) line-of-balance technique; (13) time management; (14) recapitulization; and (15) analysis.
Amirghasemi, Mehrdad; Zamani, Reza
2014-01-01
This paper presents an effective procedure for solving the job shop problem. Synergistically combining small and large neighborhood schemes, the procedure consists of four components, namely (i) a construction method for generating semi-active schedules by a forward-backward mechanism, (ii) a local search for manipulating a small neighborhood structure guided by a tabu list, (iii) a feedback-based mechanism for perturbing the solutions generated, and (iv) a very large-neighborhood local search guided by a forward-backward shifting bottleneck method. The combination of shifting bottleneck mechanism and tabu list is used as a means of the manipulation of neighborhood structures, and the perturbation mechanism employed diversifies the search. A feedback mechanism, called repeat-check, detects consequent repeats and ignites a perturbation when the total number of consecutive repeats for two identical makespan values reaches a given threshold. The results of extensive computational experiments on the benchmark instances indicate that the combination of these four components is synergetic, in the sense that they collectively make the procedure fast and robust. PMID:24808999
Optimization-based manufacturing scheduling with multiple resources and setup requirements
NASA Astrophysics Data System (ADS)
Chen, Dong; Luh, Peter B.; Thakur, Lakshman S.; Moreno, Jack, Jr.
1998-10-01
The increasing demand for on-time delivery and low price forces manufacturer to seek effective schedules to improve coordination of multiple resources and to reduce product internal costs associated with labor, setup and inventory. This study describes the design and implementation of a scheduling system for J. M. Product Inc. whose manufacturing is characterized by the need to simultaneously consider machines and operators while an operator may attend several operations at the same time, and the presence of machines requiring significant setup times. The scheduling problem with these characteristics are typical for many manufacturers, very difficult to be handled, and have not been adequately addressed in the literature. In this study, both machine and operators are modeled as resources with finite capacities to obtain efficient coordination between them, and an operator's time can be shared by several operations at the same time to make full use of the operator. Setups are explicitly modeled following our previous work, with additional penalties on excessive setups to reduce setup costs and avoid possible scraps. An integer formulation with a separable structure is developed to maximize on-time delivery of products, low inventory and small number of setups. Within the Lagrangian relaxation framework, the problem is decomposed into individual subproblems that are effectively solved by using dynamic programming with additional penalties embedded in state transitions. Heuristics is then developed to obtain a feasible schedule following on our previous work with new mechanism to satisfy operator capacity constraints. The method has been implemented using the object-oriented programming language C++ with a user-friendly interface, and numerical testing shows that the method generates high quality schedules in a timely fashion. Through simultaneous consideration of machines and operators, machines and operators are well coordinated to facilitate the smooth flow of
User requirements for a patient scheduling system
NASA Technical Reports Server (NTRS)
Zimmerman, W.
1979-01-01
A rehabilitation institute's needs and wants from a scheduling system were established by (1) studying the existing scheduling system and the variables that affect patient scheduling, (2) conducting a human-factors study to establish the human interfaces that affect patients' meeting prescribed therapy schedules, and (3) developing and administering a questionnaire to the staff which pertains to the various interface problems in order to identify staff requirements to minimize scheduling problems and other factors that may limit the effectiveness of any new scheduling system.
Reactive Scheduling in Multipurpose Batch Plants
NASA Astrophysics Data System (ADS)
Narayani, A.; Shaik, Munawar A.
2010-10-01
Scheduling is an important operation in process industries for improving resource utilization resulting in direct economic benefits. It has a two-fold objective of fulfilling customer orders within the specified time as well as maximizing the plant profit. Unexpected disturbances such as machine breakdown, arrival of rush orders and cancellation of orders affect the schedule of the plant. Reactive scheduling is generation of a new schedule which has minimum deviation from the original schedule in spite of the occurrence of unexpected events in the plant operation. Recently, Shaik & Floudas (2009) proposed a novel unified model for short-term scheduling of multipurpose batch plants using unit-specific event-based continuous time representation. In this paper, we extend the model of Shaik & Floudas (2009) to handle reactive scheduling.
Scheduling with genetic algorithms
NASA Technical Reports Server (NTRS)
Fennel, Theron R.; Underbrink, A. J., Jr.; Williams, George P. W., Jr.
1994-01-01
In many domains, scheduling a sequence of jobs is an important function contributing to the overall efficiency of the operation. At Boeing, we develop schedules for many different domains, including assembly of military and commercial aircraft, weapons systems, and space vehicles. Boeing is under contract to develop scheduling systems for the Space Station Payload Planning System (PPS) and Payload Operations and Integration Center (POIC). These applications require that we respect certain sequencing restrictions among the jobs to be scheduled while at the same time assigning resources to the jobs. We call this general problem scheduling and resource allocation. Genetic algorithms (GA's) offer a search method that uses a population of solutions and benefits from intrinsic parallelism to search the problem space rapidly, producing near-optimal solutions. Good intermediate solutions are probabalistically recombined to produce better offspring (based upon some application specific measure of solution fitness, e.g., minimum flowtime, or schedule completeness). Also, at any point in the search, any intermediate solution can be accepted as a final solution; allowing the search to proceed longer usually produces a better solution while terminating the search at virtually any time may yield an acceptable solution. Many processes are constrained by restrictions of sequence among the individual jobs. For a specific job, other jobs must be completed beforehand. While there are obviously many other constraints on processes, it is these on which we focussed for this research: how to allocate crews to jobs while satisfying job precedence requirements and personnel, and tooling and fixture (or, more generally, resource) requirements.
Development of Watch Schedule Using Rules Approach
NASA Astrophysics Data System (ADS)
Jurkevicius, Darius; Vasilecas, Olegas
The software for schedule creation and optimization solves a difficult, important and practical problem. The proposed solution is an online employee portal where administrator users can create and manage watch schedules and employee requests. Each employee can login with his/her own account and see his/her assignments, manage requests, etc. Employees set as administrators can perform the employee scheduling online, manage requests, etc. This scheduling software allows users not only to see the initial and optimized watch schedule in a simple and understandable form, but also to create special rules and criteria and input their business. The system using rules automatically will generate watch schedule.
Intelligent perturbation algorithms for space scheduling optimization
NASA Technical Reports Server (NTRS)
Kurtzman, Clifford R.
1991-01-01
Intelligent perturbation algorithms for space scheduling optimization are presented in the form of the viewgraphs. The following subject areas are covered: optimization of planning, scheduling, and manifesting; searching a discrete configuration space; heuristic algorithms used for optimization; use of heuristic methods on a sample scheduling problem; intelligent perturbation algorithms are iterative refinement techniques; properties of a good iterative search operator; dispatching examples of intelligent perturbation algorithm and perturbation operator attributes; scheduling implementations using intelligent perturbation algorithms; major advances in scheduling capabilities; the prototype ISF (industrial Space Facility) experiment scheduler; optimized schedule (max revenue); multi-variable optimization; Space Station design reference mission scheduling; ISF-TDRSS command scheduling demonstration; and example task - communications check.
JIGSAW: Preference-directed, co-operative scheduling
NASA Technical Reports Server (NTRS)
Linden, Theodore A.; Gaw, David
1992-01-01
Techniques that enable humans and machines to cooperate in the solution of complex scheduling problems have evolved out of work on the daily allocation and scheduling of Tactical Air Force resources. A generalized, formal model of these applied techniques is being developed. It is called JIGSAW by analogy with the multi-agent, constructive process used when solving jigsaw puzzles. JIGSAW begins from this analogy and extends it by propagating local preferences into global statistics that dynamically influence the value and variable ordering decisions. The statistical projections also apply to abstract resources and time periods--allowing more opportunities to find a successful variable ordering by reserving abstract resources and deferring the choice of a specific resource or time period.
Scheduler Design Criteria: Requirements and Considerations
NASA Technical Reports Server (NTRS)
Lee, Hanbong
2016-01-01
This presentation covers fundamental requirements and considerations for developing schedulers in airport operations. We first introduce performance and functional requirements for airport surface schedulers. Among various optimization problems in airport operations, we focus on airport surface scheduling problem, including runway and taxiway operations. We then describe a basic methodology for airport surface scheduling such as node-link network model and scheduling algorithms previously developed. Next, we explain how to design a mathematical formulation in more details, which consists of objectives, decision variables, and constraints. Lastly, we review other considerations, including optimization tools, computational performance, and performance metrics for evaluation.
Decomposability and scalability in space-based observatory scheduling
NASA Technical Reports Server (NTRS)
Muscettola, Nicola; Smith, Stephen F.
1992-01-01
In this paper, we discuss issues of problem and model decomposition within the HSTS scheduling framework. HSTS was developed and originally applied in the context of the Hubble Space Telescope (HST) scheduling problem, motivated by the limitations of the current solution and, more generally, the insufficiency of classical planning and scheduling approaches in this problem context. We first summarize the salient architectural characteristics of HSTS and their relationship to previous scheduling and AI planning research. Then, we describe some key problem decomposition techniques supported by HSTS and underlying our integrated planning and scheduling approach, and we discuss the leverage they provide in solving space-based observatory scheduling problems.
Oakes, J; Pols, R; Battersby, M; Lawn, S; Pulvirenti, M; Smith, D
2012-09-01
This study aimed to develop an empirically based description of relapse in Electronic Gaming Machine problem gambling. In this paper the authors describe part one of a two part, linked relapse process: the 'push' towards relapse. In this two-part process, factors interact sequentially and simultaneously within the problem gambler to produce a series of mental and behavioural events that ends with relapse when the 'push' overcomes 'pull' (part one); or as described in part two, continued abstinence when 'pull' overcomes 'push'. In the second paper, the authors describe how interacting factors 'pull' the problem gambler away from relapse. This study used four focus groups comprising thirty participants who were gamblers, gamblers' significant others, therapists and counsellors. The groups were recorded, recordings were then transcribed and analysed using thematic, textual analysis. With the large number of variables considered to be related to relapse in problem gamblers, five key factors emerged that 'push' the gambler towards relapse. These were urge, erroneous cognitions about the outcomes of gambling, negative affect, dysfunctional relationships and environmental gambling triggers. Two theories emerged: (1) each relapse episode comprised a sequence of mental and behavioural events, which evolves over time and was modified by factors that 'push' this sequence towards relapse and (2) a number of gamblers develop an altered state of consciousness during relapse described as the 'zone' which prolongs the relapse. PMID:21901457
ASTER Scheduling Prioritization Function
NASA Technical Reports Server (NTRS)
Cohen, Ron
1996-01-01
ASTER schedules are generated by an automated scheduling system. This scheduler will generate psuedo-optimal schedules based on a priority scheme. This priority scheme is controlled by the Science Team.
ERIC Educational Resources Information Center
Pactor, Paul
1970-01-01
The U.S. Department of Labor has projected a 106 percent increase in the demand for office machine operators over the next 10 years. Machines with a high frequency of use include printing calculators, 10-key adding machines, and key punch machines. The 12th grade is the logical time for teaching business machines. (CH)
Scheduling Software for Complex Scenarios
NASA Technical Reports Server (NTRS)
2006-01-01
Preparing a vehicle and its payload for a single launch is a complex process that involves thousands of operations. Because the equipment and facilities required to carry out these operations are extremely expensive and limited in number, optimal assignment and efficient use are critically important. Overlapping missions that compete for the same resources, ground rules, safety requirements, and the unique needs of processing vehicles and payloads destined for space impose numerous constraints that, when combined, require advanced scheduling. Traditional scheduling systems use simple algorithms and criteria when selecting activities and assigning resources and times to each activity. Schedules generated by these simple decision rules are, however, frequently far from optimal. To resolve mission-critical scheduling issues and predict possible problem areas, NASA historically relied upon expert human schedulers who used their judgment and experience to determine where things should happen, whether they will happen on time, and whether the requested resources are truly necessary.
Oakes, J; Pols, R; Battersby, M; Lawn, S; Pulvirenti, M; Smith, D
2012-09-01
This study aimed to develop an empirically based description of relapse in Electronic Gaming Machine (EGM) problem gambling (PG) by describing the processes and factors that 'pull' the problem gambler away from relapse contrasted with the 'push' towards relapse. These conceptualisations describe two opposing, interacting emotional processes occurring within the problem gambler during any relapse episode. Each relapse episode comprises a complex set of psychological and social behaviours where many factors interact sequentially and simultaneously within the problem gambler to produce a series of mental and behaviour events that end (1) with relapse where 'push' overcomes 'pull' or (2) continued abstinence where 'pull' overcomes 'push'. Four focus groups comprising thirty participants who were EGM problem gamblers, gamblers' significant others, therapists and counsellors described their experiences and understanding of relapse. The groups were recorded, recordings were then transcribed and analysed using thematic textual analysis. It was established that vigilance, motivation to commit to change, positive social support, cognitive strategies such as remembering past gambling harms or distraction techniques to avoid thinking about gambling to enable gamblers to manage the urge to gamble and urge extinction were key factors that protected against relapse. Three complementary theories emerged from the analysis. Firstly, a process of reappraisal of personal gambling behaviour pulls the gambler away from relapse. This results in a commitment to change that develops over time and affects but is independent of each episode of relapse. Secondly, relapse may be halted by interacting factors that 'pull' the problem gambler away from the sequence of mental and behavioural events, which follow the triggering of the urge and cognitions to gamble. Thirdly, urge extinction and apparent 'cure' is possible for EGM gambling. This study provides a qualitative, empirical model for
Xiao, Feng; Shen, Hong-Bin
2015-11-23
The α-helical transmembrane proteins constitute 25% of the entire human proteome space and are difficult targets in high-resolution wet-lab structural studies, calling for accurate computational predictors. We present a novel sequence-based method called MemBrain-Rasa to predict relative solvent accessibility surface area (rASA) from primary sequences. MemBrain-Rasa features by an ensemble prediction protocol composed of a statistical machine-learning engine, which is trained in the sequential feature space, and a segment template similarity-based engine, which is constructed with solved structures and sequence alignment. We locally constructed a comprehensive database of residue relative solvent accessibility surface area from the solved protein 3D structures in the PDB database. It is searched against for segment templates that are expected to be structurally similar to the query sequence's segments. The segment template-based prediction is then fused with the support vector regression outputs using knowledge rules. Our experiments show that pure machine learning output cannot cover the entire rASA solution space and will have a serious prediction preference problem due to the relatively small size of membrane protein structures that can be used as the training samples. The template-based engine solves this problem very well, resulting in significant improvement of the prediction performance. MemBrain-Rasa achieves a Pearson correlation coefficient of 0.733 and mean absolute error of 13.593 on the benchmark dataset, which are 26.4% and 26.1% better than existing predictors. MemBrain-Rasa represents a new progress in structure modeling of α-helical transmembrane proteins. MemBrain-Rasa is available at www.csbio.sjtu.edu.cn/bioinf/MemBrain/. PMID:26455366
Satellite mission scheduling algorithm based on GA
NASA Astrophysics Data System (ADS)
Sun, Baolin; Mao, Lifei; Wang, Wenxiang; Xie, Xing; Qin, Qianqing
2007-11-01
The Satellite Mission Scheduling problem (SMS) involves scheduling tasks to be performed by a satellite, where new task requests can arrive at any time, non-deterministically, and must be scheduled in real-time. This paper describes a new Satellite Mission Scheduling problem based on Genetic Algorithm (SMSGA). In this paper, it investigates algorithmic approaches for determining an optimal or near-optimal sequence of tasks, allocated to a satellite payload over time, with dynamic tasking considerations. The simulation results show that the proposed approach is effective and efficient in applications to the real problems.
A survey of planning and scheduling research at the NASA Ames Research Center
NASA Technical Reports Server (NTRS)
Zweben, Monte
1988-01-01
NASA Ames Research Center has a diverse program in planning and scheduling. This paper highlights some of our research projects as well as some of our applications. Topics addressed include machine learning techniques, action representations and constraint-based scheduling systems. The applications discussed are planetary rovers, Hubble Space Telescope scheduling, and Pioneer Venus orbit scheduling.
A survey of planning and scheduling research at the NASA Ames Research Center
NASA Technical Reports Server (NTRS)
Zweben, Monte
1989-01-01
NASA Ames Research Center has a diverse program in planning and scheduling. Some research projects as well as some applications are highlighted. Topics addressed include machine learning techniques, action representations and constraint-based scheduling systems. The applications discussed are planetary rovers, Hubble Space Telescope scheduling, and Pioneer Venus orbit scheduling.
Longest jobs first algorithm in solving job shop scheduling using adaptive genetic algorithm (GA)
NASA Astrophysics Data System (ADS)
Alizadeh Sahzabi, Vahid; Karimi, Iman; Alizadeh Sahzabi, Navid; Mamaani Barnaghi, Peiman
2011-12-01
In this paper, genetic algorithm was used to solve job shop scheduling problems. One example discussed in JSSP (Job Shop Scheduling Problem) and I described how we can solve such these problems by genetic algorithm. The goal in JSSP is to gain the shortest process time. Furthermore I proposed a method to obtain best performance on performing all jobs in shortest time. The method mainly, is according to Genetic algorithm (GA) and crossing over between parents always follows the rule which the longest process is at the first in the job queue. In the other word chromosomes is suggested to sorts based on the longest processes to shortest i.e. "longest job first" says firstly look which machine contains most processing time during its performing all its jobs and that is the bottleneck. Secondly, start sort those jobs which are belonging to that specific machine descending. Based on the achieved results," longest jobs first" is the optimized status in job shop scheduling problems. In our results the accuracy would grow up to 94.7% for total processing time and the method improved 4% the accuracy of performing all jobs in the presented example.
Longest jobs first algorithm in solving job shop scheduling using adaptive genetic algorithm (GA)
NASA Astrophysics Data System (ADS)
Alizadeh Sahzabi, Vahid; Karimi, Iman; Alizadeh Sahzabi, Navid; Mamaani Barnaghi, Peiman
2012-01-01
In this paper, genetic algorithm was used to solve job shop scheduling problems. One example discussed in JSSP (Job Shop Scheduling Problem) and I described how we can solve such these problems by genetic algorithm. The goal in JSSP is to gain the shortest process time. Furthermore I proposed a method to obtain best performance on performing all jobs in shortest time. The method mainly, is according to Genetic algorithm (GA) and crossing over between parents always follows the rule which the longest process is at the first in the job queue. In the other word chromosomes is suggested to sorts based on the longest processes to shortest i.e. "longest job first" says firstly look which machine contains most processing time during its performing all its jobs and that is the bottleneck. Secondly, start sort those jobs which are belonging to that specific machine descending. Based on the achieved results," longest jobs first" is the optimized status in job shop scheduling problems. In our results the accuracy would grow up to 94.7% for total processing time and the method improved 4% the accuracy of performing all jobs in the presented example.
Machine Shop Grinding Machines.
ERIC Educational Resources Information Center
Dunn, James
This curriculum manual is one in a series of machine shop curriculum manuals intended for use in full-time secondary and postsecondary classes, as well as part-time adult classes. The curriculum can also be adapted to open-entry, open-exit programs. Its purpose is to equip students with basic knowledge and skills that will enable them to enter the…
Bridging the Gap Between Planning and Scheduling
NASA Technical Reports Server (NTRS)
Smith, David E.; Frank, Jeremy; Jonsson, Ari K.; Norvig, Peter (Technical Monitor)
2000-01-01
Planning research in Artificial Intelligence (AI) has often focused on problems where there are cascading levels of action choice and complex interactions between actions. In contrast. Scheduling research has focused on much larger problems where there is little action choice, but the resulting ordering problem is hard. In this paper, we give an overview of M planning and scheduling techniques, focusing on their similarities, differences, and limitations. We also argue that many difficult practical problems lie somewhere between planning and scheduling, and that neither area has the right set of tools for solving these vexing problems.
NASA Technical Reports Server (NTRS)
Chien, Steve A.; Tran, Daniel Q.; Rabideau, Gregg R.; Schaffer, Steven R.
2011-01-01
Software has been designed to schedule remote sensing with the Earth Observing One spacecraft. The software attempts to satisfy as many observation requests as possible considering each against spacecraft operation constraints such as data volume, thermal, pointing maneuvers, and others. More complex constraints such as temperature are approximated to enable efficient reasoning while keeping the spacecraft within safe limits. Other constraints are checked using an external software library. For example, an attitude control library is used to determine the feasibility of maneuvering between pairs of observations. This innovation can deal with a wide range of spacecraft constraints and solve large scale scheduling problems like hundreds of observations and thousands of combinations of observation sequences.
Dawn Usage, Scheduling, and Governance Model
Louis, S
2009-11-02
This document describes Dawn use, scheduling, and governance concerns. Users started running full-machine science runs in early April 2009 during the initial open shakedown period. Scheduling Dawn while in the Open Computing Facility (OCF) was controlled and coordinated via phone calls, emails, and a small number of controlled banks. With Dawn moving to the Secure Computing Facility (SCF) in fall of 2009, a more detailed scheduling and governance model is required. The three major objectives are: (1) Ensure Dawn resources are allocated on a program priority-driven basis; (2) Utilize Dawn resources on the job mixes for which they were intended; and (3) Minimize idle cycles through use of partitions, banks and proper job mix. The SCF workload for Dawn will be inherently different than Purple or BG/L, and therefore needs a different approach. Dawn's primary function is to permit adequate access for tri-lab code development in preparation for Sequoia, and in particular for weapons multi-physics codes in support of UQ. A second purpose is to provide time allocations for large-scale science runs and for UQ suite calculations to advance SSP program priorities. This proposed governance model will be the basis for initial time allocation of Dawn computing resources for the science and UQ workloads that merit priority on this class of resource, either because they cannot be reasonably attempted on any other resources due to size of problem, or because of the unavailability of sizable allocations on other ASC capability or capacity platforms. This proposed model intends to make the most effective use of Dawn as possible, but without being overly constrained by more formal proposal processes such as those now used for Purple CCCs.
Progress in Documentation: Machine Translation and Machine-Aided Translation.
ERIC Educational Resources Information Center
Hutchins, W. J.
1978-01-01
Discusses the prospects for fully automatic machine translation of good quality. Sections include history and background, operational and experimental machine translation systems of recent years, descriptions of interactive systems and machine-assisted translation, and a general survey of present problems and future possibilities. (VT)
Isomap based supporting vector machine
NASA Astrophysics Data System (ADS)
Liang, W. N.
2015-12-01
This research presents a new isomap based supporting vector machine method. Isomap is a dimension reduction method which is able to analyze nonlinear relationship of data on manifolds. Accordingly, support vector machine is established on the isomap manifold to classify given and predict unknown data. A case study of the isomap based supporting vector machine for environmental planning problems is conducted.
Future aircraft networks and schedules
NASA Astrophysics Data System (ADS)
Shu, Yan
2011-07-01
Because of the importance of air transportation scheduling, the emergence of small aircraft and the vision of future fuel-efficient aircraft, this thesis has focused on the study of aircraft scheduling and network design involving multiple types of aircraft and flight services. It develops models and solution algorithms for the schedule design problem and analyzes the computational results. First, based on the current development of small aircraft and on-demand flight services, this thesis expands a business model for integrating on-demand flight services with the traditional scheduled flight services. This thesis proposes a three-step approach to the design of aircraft schedules and networks from scratch under the model. In the first step, both a frequency assignment model for scheduled flights that incorporates a passenger path choice model and a frequency assignment model for on-demand flights that incorporates a passenger mode choice model are created. In the second step, a rough fleet assignment model that determines a set of flight legs, each of which is assigned an aircraft type and a rough departure time is constructed. In the third step, a timetable model that determines an exact departure time for each flight leg is developed. Based on the models proposed in the three steps, this thesis creates schedule design instances that involve almost all the major airports and markets in the United States. The instances of the frequency assignment model created in this thesis are large-scale non-convex mixed-integer programming problems, and this dissertation develops an overall network structure and proposes iterative algorithms for solving these instances. The instances of both the rough fleet assignment model and the timetable model created in this thesis are large-scale mixed-integer programming problems, and this dissertation develops subproblem schemes for solving these instances. Based on these solution algorithms, this dissertation also presents
Parallel job scheduling policies to improve fairness : a case study.
Leung, Vitus Joseph; Sabin, Gerald; Sadayappan, Ponnuswamy
2008-02-01
Balancing fairness, user performance, and system performance is a critical concern when developing and installing parallel schedulers. Sandia uses a customized scheduler to manage many of their parallel machines. A primary function of the scheduler is to ensure that the machines have good utilization and that users are treated in a 'fair' manner. A separate compute process allocator (CPA) ensures that the jobs on the machines are not too fragmented in order to maximize throughput. Until recently, there has been no established technique to measure the fairness of parallel job schedulers. This paper introduces a 'hybrid' fairness metric that is similar to recently proposed metrics. The metric uses the Sandia version of a 'fairshare' queuing priority as the basis for fairness. The hybrid fairness metric is used to evaluate a Sandia workload. Using these results, multiple scheduling strategies are introduced to improve performance while satisfying user and system performance constraints.
A System for Automatically Generating Scheduling Heuristics
NASA Technical Reports Server (NTRS)
Morris, Robert
1996-01-01
The goal of this research is to improve the performance of automated schedulers by designing and implementing an algorithm by automatically generating heuristics by selecting a schedule. The particular application selected by applying this method solves the problem of scheduling telescope observations, and is called the Associate Principal Astronomer. The input to the APA scheduler is a set of observation requests submitted by one or more astronomers. Each observation request specifies an observation program as well as scheduling constraints and preferences associated with the program. The scheduler employs greedy heuristic search to synthesize a schedule that satisfies all hard constraints of the domain and achieves a good score with respect to soft constraints expressed as an objective function established by an astronomer-user.
Scheduling Reconsidered (Again!)
ERIC Educational Resources Information Center
Hentschke, Guilbert C.; Fowler, William J.
1974-01-01
Computer technicians bring to school scheduling a certain naivete regarding the operation of schools. School administrators play a fundamental role of informing technicians about education scheduling needs. (Author)
Intelligent perturbation algorithms to space scheduling optimization
NASA Technical Reports Server (NTRS)
Kurtzman, Clifford R.
1991-01-01
The limited availability and high cost of crew time and scarce resources make optimization of space operations critical. Advances in computer technology coupled with new iterative search techniques permit the near optimization of complex scheduling problems that were previously considered computationally intractable. Described here is a class of search techniques called Intelligent Perturbation Algorithms. Several scheduling systems which use these algorithms to optimize the scheduling of space crew, payload, and resource operations are also discussed.
Scheduling of an aircraft fleet
NASA Technical Reports Server (NTRS)
Paltrinieri, Massimo; Momigliano, Alberto; Torquati, Franco
1992-01-01
Scheduling is the task of assigning resources to operations. When the resources are mobile vehicles, they describe routes through the served stations. To emphasize such aspect, this problem is usually referred to as the routing problem. In particular, if vehicles are aircraft and stations are airports, the problem is known as aircraft routing. This paper describes the solution to such a problem developed in OMAR (Operative Management of Aircraft Routing), a system implemented by Bull HN for Alitalia. In our approach, aircraft routing is viewed as a Constraint Satisfaction Problem. The solving strategy combines network consistency and tree search techniques.
Optimal randomized scheduling by replacement
Saias, I.
1996-05-01
In the replacement scheduling problem, a system is composed of n processors drawn from a pool of p. The processors can become faulty while in operation and faulty processors never recover. A report is issued whenever a fault occurs. This report states only the existence of a fault but does not indicate its location. Based on this report, the scheduler can reconfigure the system and choose another set of n processors. The system operates satisfactorily as long as, upon report of a fault, the scheduler chooses n non-faulty processors. We provide a randomized protocol maximizing the expected number of faults the system can sustain before the occurrence of a crash. The optimality of the protocol is established by considering a closely related dual optimization problem. The game-theoretic technical difficulties that we solve in this paper are very general and encountered whenever proving the optimality of a randomized algorithm in parallel and distributed computation.
Linux Kernel Co-Scheduling For Bulk Synchronous Parallel Applications
Jones, Terry R
2011-01-01
This paper describes a kernel scheduling algorithm that is based on co-scheduling principles and that is intended for parallel applications running on 1000 cores or more where inter-node scalability is key. Experimental results for a Linux implementation on a Cray XT5 machine are presented.1 The results indicate that Linux is a suitable operating system for this new scheduling scheme, and that this design provides a dramatic improvement in scaling performance for synchronizing collective operations at scale.
NASA Technical Reports Server (NTRS)
Stiefel, M. L.
1983-01-01
The functions and performance characteristics of data base machines (DBM), including machines currently being studied in research laboratories and those currently offered on a commerical basis are discussed. The cost/benefit considerations that must be recognized in selecting a DBM are discussed, as well as the future outlook for such machines.
Sort-Mid tasks scheduling algorithm in grid computing.
Reda, Naglaa M; Tawfik, A; Marzok, Mohamed A; Khamis, Soheir M
2015-11-01
Scheduling tasks on heterogeneous resources distributed over a grid computing system is an NP-complete problem. The main aim for several researchers is to develop variant scheduling algorithms for achieving optimality, and they have shown a good performance for tasks scheduling regarding resources selection. However, using of the full power of resources is still a challenge. In this paper, a new heuristic algorithm called Sort-Mid is proposed. It aims to maximizing the utilization and minimizing the makespan. The new strategy of Sort-Mid algorithm is to find appropriate resources. The base step is to get the average value via sorting list of completion time of each task. Then, the maximum average is obtained. Finally, the task has the maximum average is allocated to the machine that has the minimum completion time. The allocated task is deleted and then, these steps are repeated until all tasks are allocated. Experimental tests show that the proposed algorithm outperforms almost other algorithms in terms of resources utilization and makespan. PMID:26644937
Dypas: A dynamic payload scheduler for shuttle missions
NASA Technical Reports Server (NTRS)
Davis, Stephen
1988-01-01
Decision and analysis systems have had broad and very practical application areas in the human decision making process. These software systems range from the help sections in simple accounting packages, to the more complex computer configuration programs. Dypas is a decision and analysis system that aids prelaunch shutlle scheduling, and has added functionality to aid the rescheduling done in flight. Dypas is written in Common Lisp on a Symbolics Lisp machine. Dypas differs from other scheduling programs in that it can draw its knowledge from different rule bases and apply them to different rule interpretation schemes. The system has been coded with Flavors, an object oriented extension to Common Lisp on the Symbolics hardware. This allows implementation of objects (experiments) to better match the problem definition, and allows a more coherent solution space to be developed. Dypas was originally developed to test a programmer's aptitude toward Common Lisp and the Symbolics software environment. Since then the system has grown into a large software effort with several programmers and researchers thrown into the effort. Dypas is currently using two expert systems and three inferencing procedures to generate a many object schedule. The paper will review the abilities of Dypas and comment on its functionality.
NASA Technical Reports Server (NTRS)
Smith, Greg
2003-01-01
Schedule risk assessments determine the likelihood of finishing on time. Each task in a schedule has a varying degree of probability of being finished on time. A schedule risk assessment quantifies these probabilities by assigning values to each task. This viewgraph presentation contains a flow chart for conducting a schedule risk assessment, and profiles applicable several methods of data analysis.
Metalworking and machining fluids
Erdemir, Ali; Sykora, Frank; Dorbeck, Mark
2010-10-12
Improved boron-based metal working and machining fluids. Boric acid and boron-based additives that, when mixed with certain carrier fluids, such as water, cellulose and/or cellulose derivatives, polyhydric alcohol, polyalkylene glycol, polyvinyl alcohol, starch, dextrin, in solid and/or solvated forms result in improved metalworking and machining of metallic work pieces. Fluids manufactured with boric acid or boron-based additives effectively reduce friction, prevent galling and severe wear problems on cutting and forming tools.
A scheduling algorithm for Spacelab telescope observations
NASA Technical Reports Server (NTRS)
Grone, B.
1982-01-01
An algorithm is developed for sequencing and scheduling of observations of stellar targets by equipment on Spacelab. The method is a general one. The scheduling problem is defined and examined. The method developed for its solution is documented. Suggestions for further development and implementation of this method are made.
Scheduling Sanity in the Elementary Schools.
ERIC Educational Resources Information Center
Curatilo, Joseph S.
1983-01-01
Use of the music block schedule for elementary band, chorus, and orchestra programs has many advantages. For example, it eliminates the need to pull students from class and abates classroom teacher and music teacher friction. Solutions to other problems often encountered in scheduling are also discussed. (RM)
Husain, O A; Grainger, J M; Sims, J
1978-01-01
Further development of an individual staining machine is to be strongly encouraged but meanwhile, using bulk stainers, frequent changing of wash fluids and staining solutions, particularly leading up to and following the haematoxylin pot, is essential to reduce the risk of cross contamination. Certain smears, such as from semen or from serous fluids where malignancy is suspected or known, must be stained on separate racks. In some laboratories it is the rule not to stain semen or serous fluids in bulk staining machines at all and this may have to become the rule everywhere until we are provided with safe individual slide stainers. PMID:75214
Optimizing Operating Room Scheduling.
Levine, Wilton C; Dunn, Peter F
2015-12-01
This article reviews the management of an operating room (OR) schedule and use of the schedule to add value to an organization. We review the methodology of an OR block schedule, daily OR schedule management, and post anesthesia care unit patient flow. We discuss the importance of a well-managed OR schedule to ensure smooth patient care, not only in the OR, but throughout the entire hospital. PMID:26610624
Production scheduling and rescheduling with genetic algorithms.
Bierwirth, C; Mattfeld, D C
1999-01-01
A general model for job shop scheduling is described which applies to static, dynamic and non-deterministic production environments. Next, a Genetic Algorithm is presented which solves the job shop scheduling problem. This algorithm is tested in a dynamic environment under different workload situations. Thereby, a highly efficient decoding procedure is proposed which strongly improves the quality of schedules. Finally, this technique is tested for scheduling and rescheduling in a non-deterministic environment. It is shown by experiment that conventional methods of production control are clearly outperformed at reasonable run-time costs. PMID:10199993
Perspex machine: VII. The universal perspex machine
NASA Astrophysics Data System (ADS)
Anderson, James A. D. W.
2006-01-01
The perspex machine arose from the unification of projective geometry with the Turing machine. It uses a total arithmetic, called transreal arithmetic, that contains real arithmetic and allows division by zero. Transreal arithmetic is redefined here. The new arithmetic has both a positive and a negative infinity which lie at the extremes of the number line, and a number nullity that lies off the number line. We prove that nullity, 0/0, is a number. Hence a number may have one of four signs: negative, zero, positive, or nullity. It is, therefore, impossible to encode the sign of a number in one bit, as floating-point arithmetic attempts to do, resulting in the difficulty of having both positive and negative zeros and NaNs. Transrational arithmetic is consistent with Cantor arithmetic. In an extension to real arithmetic, the product of zero, an infinity, or nullity with its reciprocal is nullity, not unity. This avoids the usual contradictions that follow from allowing division by zero. Transreal arithmetic has a fixed algebraic structure and does not admit options as IEEE, floating-point arithmetic does. Most significantly, nullity has a simple semantics that is related to zero. Zero means "no value" and nullity means "no information." We argue that nullity is as useful to a manufactured computer as zero is to a human computer. The perspex machine is intended to offer one solution to the mind-body problem by showing how the computable aspects of mind and, perhaps, the whole of mind relates to the geometrical aspects of body and, perhaps, the whole of body. We review some of Turing's writings and show that he held the view that his machine has spatial properties. In particular, that it has the property of being a 7D lattice of compact spaces. Thus, we read Turing as believing that his machine relates computation to geometrical bodies. We simplify the perspex machine by substituting an augmented Euclidean geometry for projective geometry. This leads to a general
Flexible shift scheduling of physicians.
Brunner, Jens O; Bard, Jonathan F; Kolisch, Rainer
2009-09-01
This research addresses a shift scheduling problem in which physicians at a German university hospital are assigned to demand periods over a planning horizon that can extend up to several weeks. When performing the scheduling it is necessary to take into account a variety of legal and institutional constraints that are imposed by a national labor agreement, which governs all physicians in German university hospitals. Currently, most medical departments develop their staff schedules manually at great cost and time. To solve the problem, a new modeling approach is developed that requires shifts to be generated implicitly. Rather than beginning with a predetermined number of shift types and start times, shifts are allowed to start at every pre-defined period in the planning horizon and extend up to 13 h with an hour-long break included. The objective is to find an assignment such that the total hours that have to be paid out as overtime are minimal under the restrictions given by the labor agreement. The problem is formulated as a mixed-integer program and solved with CPLEX. During the solution process individual lines-of-work are constructed for each physician. Using data from an anesthesia department, computational results indicate that high quality schedules can be obtained much more quickly than by current practice. PMID:19739361
21 CFR 1310.16 - Exemptions for certain scheduled listed chemical products.
Code of Federal Regulations, 2011 CFR
2011-04-01
... 21 Food and Drugs 9 2011-04-01 2011-04-01 false Exemptions for certain scheduled listed chemical... RECORDS AND REPORTS OF LISTED CHEMICALS AND CERTAIN MACHINES § 1310.16 Exemptions for certain scheduled listed chemical products. (a) Upon the application of a manufacturer of a scheduled listed...
Dedicated heterogeneous node scheduling including backfill scheduling
Wood, Robert R.; Eckert, Philip D.; Hommes, Gregg
2006-07-25
A method and system for job backfill scheduling dedicated heterogeneous nodes in a multi-node computing environment. Heterogeneous nodes are grouped into homogeneous node sub-pools. For each sub-pool, a free node schedule (FNS) is created so that the number of to chart the free nodes over time. For each prioritized job, using the FNS of sub-pools having nodes useable by a particular job, to determine the earliest time range (ETR) capable of running the job. Once determined for a particular job, scheduling the job to run in that ETR. If the ETR determined for a lower priority job (LPJ) has a start time earlier than a higher priority job (HPJ), then the LPJ is scheduled in that ETR if it would not disturb the anticipated start times of any HPJ previously scheduled for a future time. Thus, efficient utilization and throughput of such computing environments may be increased by utilizing resources otherwise remaining idle.
NASA Astrophysics Data System (ADS)
Vincent, Graylan
2003-01-01
A CNC mill was flown aboard NASA's KC-135 ``Weightless Wonder'' microgravity research aircraft to investigate the effect of gravity on the machining process and to demonstrate the feasibility and functionality of a CNC mill in a weightless environment, such as aboard the International Space Station. The experiment hypothesis was that the surface roughness of milling cuts made in microgravity would be of higher quality than cuts made in a gravitational environment due to increased chip removal. The technical problems associated with microgravity machining (such as the chip removal and collection process), and the engineering solutions to these problems were also evaluated in this experiment.
Batch Scheduling a Fresh Approach
NASA Technical Reports Server (NTRS)
Cardo, Nicholas P.; Woodrow, Thomas (Technical Monitor)
1994-01-01
The Network Queueing System (NQS) was designed to schedule jobs based on limits within queues. As systems obtain more memory, the number of queues increased to take advantage of the added memory resource. The problem now becomes too many queues. Having a large number of queues provides users with the capability to gain an unfair advantage over other users by tailoring their job to fit in an empty queue. Additionally, the large number of queues becomes confusing to the user community. The High Speed Processors group at the Numerical Aerodynamics Simulation (NAS) Facility at NASA Ames Research Center developed a new approach to batch job scheduling. This new method reduces the number of queues required by eliminating the need for queues based on resource limits. The scheduler examines each request for necessary resources before initiating the job. Also additional user limits at the complex level were added to provide a fairness to all users. Additional tools which include user job reordering are under development to work with the new scheduler. This paper discusses the objectives, design and implementation results of this new scheduler
Machine Learning and Radiology
Wang, Shijun; Summers, Ronald M.
2012-01-01
In this paper, we give a short introduction to machine learning and survey its applications in radiology. We focused on six categories of applications in radiology: medical image segmentation, registration, computer aided detection and diagnosis, brain function or activity analysis and neurological disease diagnosis from fMR images, content-based image retrieval systems for CT or MRI images, and text analysis of radiology reports using natural language processing (NLP) and natural language understanding (NLU). This survey shows that machine learning plays a key role in many radiology applications. Machine learning identifies complex patterns automatically and helps radiologists make intelligent decisions on radiology data such as conventional radiographs, CT, MRI, and PET images and radiology reports. In many applications, the performance of machine learning-based automatic detection and diagnosis systems has shown to be comparable to that of a well-trained and experienced radiologist. Technology development in machine learning and radiology will benefit from each other in the long run. Key contributions and common characteristics of machine learning techniques in radiology are discussed. We also discuss the problem of translating machine learning applications to the radiology clinical setting, including advantages and potential barriers. PMID:22465077
A task scheduler framework for self-powered wireless sensors.
Nordman, Mikael M
2003-10-01
The cost and inconvenience of cabling is a factor limiting widespread use of intelligent sensors. Recent developments in short-range, low-power radio seem to provide an opening to this problem, making development of wireless sensors feasible. However, for these sensors the energy availability is a main concern. The common solution is either to use a battery or to harvest ambient energy. The benefit of harvested ambient energy is that the energy feeder can be considered as lasting a lifetime, thus it saves the user from concerns related to energy management. The problem is, however, the unpredictability and unsteady behavior of ambient energy sources. This becomes a main concern for sensors that run multiple tasks at different priorities. This paper proposes a new scheduler framework that enables the reliable assignment of task priorities and scheduling in sensors powered by ambient energy. The framework being based on environment parameters, virtual queues, and a state machine with transition conditions, dynamically manages task execution according to priorities. The framework is assessed in a test system powered by a solar panel. The results show the functionality of the framework and how task execution reliably is handled without violating the priority scheme that has been assigned to it. PMID:14582879
A new mathematical programming model for scheduling flexible manufacturing systems
MacCarthy, B.; Liu, J.
1994-12-31
Flexibility is now a major consideration in the design of many manufacturing systems. Flexible manufacturing systems (FMS) have been developed in the last two decades. The principal elements of an FMS are (1) computer controlled machine tools, (2) a transport system and (3) a host computer system. Such systems may combine high flexibility with high productivity and may allow unsupervised production. However, in order to achieve these benefits, the control system must be capable of exercising intelligent supervisory management. Scheduling is at the heart of the control system and is still a major problem area. This paper describes a new mathematical programming model for a wide class of FMS scheduling problems based on a new classification scheme. A global optimization approach is adopted based on a mixed-integer linear programming model. Many important aspects of operational FMS, omitted form earlier models, are included. Key elements of model structure are highlighted. Computational experience with a comprehensive set of designed experiments is described. The applications of the model are noted and the development of effective heuristic procedures based on the model is highlighted.
Performance evaluation of vector-machine architectures
Tang, Ju-ho.
1989-01-01
Vector machines are well known for their high-peak performance, but the delivered performance varies greatly over different workloads and depends strongly on compiler optimizations. Recently it has been claimed that several horizontal superscalar architectures, e.g., VLIW and polycyclic architectures, provide a more balanced performance across a wider range of scientific workloads than do vector machines. The purpose of this research is to study the performance of register-register vector processors, such as Cray supercomputers, as a function of their architectural features, scheduling schemes, compiler optimization capabilities, and program parameters. The results of this study also provide a base for comparing vector machines with horizontal superscalar machines. An evaluation methodology, based on timing parameters, bottle-necks, and run time bounds, is developed. Cray-1 performance is degraded by the multiple memory loads of index-misaligned vectors and the inability of the Cray Fortran Compiler (CFT) to produce code that hits all the chain slot times. The impact of chaining and two instruction scheduling schemes on one-memory-port vector supercomputers, illustrated by the Cray-1 and Cray-2, is studied. The lack of instruction chaining on the Cray-2 requires a different instruction scheduling scheme from that of the Cray-1. Situations are characterized in which simple vector scheduling can generate code that fully utilizes one functional unit for machines with chaining. Even without chaining, polycyclic scheduling guarantees full utilization of one functional unit, after an initial transient, for loops with acyclic dependence graphs.
Immunization Schedules for Adults
... ACIP Vaccination Recommendations Why Immunize? Vaccines: The Basics Immunization Schedules for Adults in Easy-to-read Formats ... previous immunizations. View or Print a Schedule Recommended Immunizations for Adults (19 Years and Older) by Age ...
Childhood Immunization Schedule
... Recommendations Why Immunize? Vaccines: The Basics Instant Childhood Immunization Schedule Recommend on Facebook Tweet Share Compartir Get ... date. See Disclaimer for additional details. Based on Immunization Schedule for Children 0 through 6 Years of ...
A Test Scheduling Algorithm Based on Two-Stage GA
NASA Astrophysics Data System (ADS)
Yu, Y.; Peng, X. Y.; Peng, Y.
2006-10-01
In this paper, we present a new algorithm to co-optimize the core wrapper design and the SOC test scheduling. The SOC test scheduling problem is first formulated into a twodimension floorplan problem and a sequence pair architecture is used to represent it. Then we propose a two-stage GA (Genetic Algorithm) to solve the SOC test scheduling problem. Experiments on ITC'02 benchmark show that our algorithm can effectively reduce test time so as to decrease SOC test cost.
Deep Space Network Scheduling Using Evolutionary Computational Methods
NASA Technical Reports Server (NTRS)
Guillaume, Alexandre; Lee, Seugnwon; Wang, Yeou-Fang; Terrile, Richard J.
2007-01-01
The paper presents the specific approach taken to formulate the problem in terms of gene encoding, fitness function, and genetic operations. The genome is encoded such that a subset of the scheduling constraints is automatically satisfied. Several fitness functions are formulated to emphasize different aspects of the scheduling problem. The optimal solutions of the different fitness functions demonstrate the trade-off of the scheduling problem and provide insight into a conflict resolution process.
NASA Technical Reports Server (NTRS)
Smith, Greg
2003-01-01
Schedule Risk Assessment needs to determine the probability of finishing on or before a given point in time. Task in a schedule should reflect the "most likely" duration for each task. IN reality, each task is different and has a varying degree of probability of finishing within or after the duration specified. Schedule risk assessment attempt to quantify these probabilities by assigning values to each task. Bridges the gap between CPM scheduling and the project's need to know the likelihood of "when".
Daily Modular Scheduling Practice at Pahranagat Valley High School. Report.
ERIC Educational Resources Information Center
Anderson, David Neil
The main topic discussed is a daily modular scheduling system initiated for the small enrollment at Pahranagat Valley High School in Alamo, Nevada, with specific reference to types of instruction, schedule procedures, and conflict problems. An evaluation of the scheduling system is also included. The report is written in dissertation format, which…
Ritson, D. )
1989-05-01
This talk examines methods available to minimize, but never entirely eliminate, degradation of machine performance caused by terrain following. Breaking of planar machine symmetry for engineering convenience and/or monetary savings must be balanced against small performance degradation, and can only be decided on a case-by-case basis. 5 refs.
El-Refaie, Ayman Mohamed Fawzi; Reddy, Patel Bhageerath
2012-07-17
An interior permanent magnet electric machine is disclosed. The interior permanent magnet electric machine comprises a rotor comprising a plurality of radially placed magnets each having a proximal end and a distal end, wherein each magnet comprises a plurality of magnetic segments and at least one magnetic segment towards the distal end comprises a high resistivity magnetic material.
Bhatia, Swapnil; LaBoda, Craig; Yanez, Vanessa; Haddock-Angelli, Traci; Densmore, Douglas
2016-08-19
We define a new inversion-based machine called a permuton of n genetic elements, which allows the n elements to be rearranged in any of the n·(n - 1)·(n - 2)···2 = n! distinct orderings. We present two design algorithms for architecting such a machine. We define a notion of a feasible design and use the framework to discuss the feasibility of the permuton architectures. We have implemented our design algorithms in a freely usable web-accessible software for exploration of these machines. Permutation machines could be used as memory elements or state machines and explicitly illustrate a rational approach to designing biological systems. PMID:27383067
Block Scheduling. Research Brief
ERIC Educational Resources Information Center
Muir, Mike
2003-01-01
What are the effects of block scheduling? Results of transitioning from traditional to block scheduling are mixed. Some studies indicate no change in achievement results, nor change in teachers' opinions about instructional strategies. Other studies show that block scheduling doesn't work well for Advanced Placement or Music courses, that "hard to…
Job Scheduling in a Heterogeneous Grid Environment
NASA Technical Reports Server (NTRS)
Shan, Hong-Zhang; Smith, Warren; Oliker, Leonid; Biswas, Rupak
2004-01-01
Computational grids have the potential for solving large-scale scientific problems using heterogeneous and geographically distributed resources. However, a number of major technical hurdles must be overcome before this potential can be realized. One problem that is critical to effective utilization of computational grids is the efficient scheduling of jobs. This work addresses this problem by describing and evaluating a grid scheduling architecture and three job migration algorithms. The architecture is scalable and does not assume control of local site resources. The job migration policies use the availability and performance of computer systems, the network bandwidth available between systems, and the volume of input and output data associated with each job. An extensive performance comparison is presented using real workloads from leading computational centers. The results, based on several key metrics, demonstrate that the performance of our distributed migration algorithms is significantly greater than that of a local scheduling framework and comparable to a non-scalable global scheduling approach.
Job scheduling in a heterogenous grid environment
Oliker, Leonid; Biswas, Rupak; Shan, Hongzhang; Smith, Warren
2004-02-11
Computational grids have the potential for solving large-scale scientific problems using heterogeneous and geographically distributed resources. However, a number of major technical hurdles must be overcome before this potential can be realized. One problem that is critical to effective utilization of computational grids is the efficient scheduling of jobs. This work addresses this problem by describing and evaluating a grid scheduling architecture and three job migration algorithms. The architecture is scalable and does not assume control of local site resources. The job migration policies use the availability and performance of computer systems, the network bandwidth available between systems, and the volume of input and output data associated with each job. An extensive performance comparison is presented using real workloads from leading computational centers. The results, based on several key metrics, demonstrate that the performance of our distributed migration algorithms is significantly greater than that of a local scheduling framework and comparable to a non-scalable global scheduling approach.
Parallel-Batch Scheduling and Transportation Coordination with Waiting Time Constraint
Gong, Hua; Chen, Daheng; Xu, Ke
2014-01-01
This paper addresses a parallel-batch scheduling problem that incorporates transportation of raw materials or semifinished products before processing with waiting time constraint. The orders located at the different suppliers are transported by some vehicles to a manufacturing facility for further processing. One vehicle can load only one order in one shipment. Each order arriving at the facility must be processed in the limited waiting time. The orders are processed in batches on a parallel-batch machine, where a batch contains several orders and the processing time of the batch is the largest processing time of the orders in it. The goal is to find a schedule to minimize the sum of the total flow time and the production cost. We prove that the general problem is NP-hard in the strong sense. We also demonstrate that the problem with equal processing times on the machine is NP-hard. Furthermore, a dynamic programming algorithm in pseudopolynomial time is provided to prove its ordinarily NP-hardness. An optimal algorithm in polynomial time is presented to solve a special case with equal processing times and equal transportation times for each order. PMID:24883385
Multi-Objective Scheduling for the Cluster II Constellation
NASA Technical Reports Server (NTRS)
Johnston, Mark D.; Giuliano, Mark
2011-01-01
This paper describes the application of the MUSE multiobjecctive scheduling framework to the Cluster II WBD scheduling domain. Cluster II is an ESA four-spacecraft constellation designed to study the plasma environment of the Earth and it's magnetosphere. One of the instruments on each of the four spacecraft is the Wide Band Data (WBD) plasma wave experiment. We have applied the MUSE evolutionary algorithm to the scheduling problem represented by this instrument, and the result has been adopted and utilized by the WBD schedulers for nearly a year. This paper describes the WBD scheduling problem, its representation in MUSE, and some of the visualization elements that provide insight into objective value tradeoffs.
49 CFR 214.533 - Schedule of repairs subject to availability of parts.
Code of Federal Regulations, 2013 CFR
2013-10-01
... Maintenance Machines and Hi-Rail Vehicles § 214.533 Schedule of repairs subject to availability of parts. (a... maintenance machine or a hi-rail vehicle by the end of the next business day following the report of the... maintenance machine or hi-rail vehicle within seven calendar days after receiving the necessary part....
49 CFR 214.533 - Schedule of repairs subject to availability of parts.
Code of Federal Regulations, 2010 CFR
2010-10-01
... Maintenance Machines and Hi-Rail Vehicles § 214.533 Schedule of repairs subject to availability of parts. (a... maintenance machine or a hi-rail vehicle by the end of the next business day following the report of the... maintenance machine or hi-rail vehicle within seven calendar days after receiving the necessary part....
49 CFR 214.533 - Schedule of repairs subject to availability of parts.
Code of Federal Regulations, 2011 CFR
2011-10-01
... Maintenance Machines and Hi-Rail Vehicles § 214.533 Schedule of repairs subject to availability of parts. (a... maintenance machine or a hi-rail vehicle by the end of the next business day following the report of the... maintenance machine or hi-rail vehicle within seven calendar days after receiving the necessary part....
49 CFR 214.533 - Schedule of repairs subject to availability of parts.
Code of Federal Regulations, 2012 CFR
2012-10-01
... Maintenance Machines and Hi-Rail Vehicles § 214.533 Schedule of repairs subject to availability of parts. (a... maintenance machine or a hi-rail vehicle by the end of the next business day following the report of the... maintenance machine or hi-rail vehicle within seven calendar days after receiving the necessary part....
49 CFR 214.533 - Schedule of repairs subject to availability of parts.
Code of Federal Regulations, 2014 CFR
2014-10-01
... Maintenance Machines and Hi-Rail Vehicles § 214.533 Schedule of repairs subject to availability of parts. (a... maintenance machine or a hi-rail vehicle by the end of the next business day following the report of the... maintenance machine or hi-rail vehicle within seven calendar days after receiving the necessary part....
NASA Technical Reports Server (NTRS)
Clement, Bradley; Johnston, Mark; Wax, Allan; Chouinard, Caroline
2008-01-01
The DSN (Deep Space Network) Scheduling Engine targets all space missions that use DSN services. It allows clients to issue scheduling, conflict identification, conflict resolution, and status requests in XML over a Java Message Service interface. The scheduling requests may include new requirements that represent a set of tracks to be scheduled under some constraints. This program uses a heuristic local search to schedule a variety of schedule requirements, and is being infused into the Service Scheduling Assembly, a mixed-initiative scheduling application. The engine resolves conflicting schedules of resource allocation according to a range of existing and possible requirement specifications, including optional antennas; start of track and track duration ranges; periodic tracks; locks on track start, duration, and allocated antenna; MSPA (multiple spacecraft per aperture); arraying/VLBI (very long baseline interferometry)/delta DOR (differential one-way ranging); continuous tracks; segmented tracks; gap-to-track ratio; and override or block-out of requirements. The scheduling models now include conflict identification for SOA(start of activity), BOT (beginning of track), RFI (radio frequency interference), and equipment constraints. This software will search through all possible allocations while providing a best-effort solution at any time. The engine reschedules to accommodate individual emergency tracks in 0.2 second, and emergency antenna downtime in 0.2 second. The software handles doubling of one mission's track requests over one week (to 42 total) in 2.7 seconds. Further tests will be performed in the context of actual schedules.
NASA Technical Reports Server (NTRS)
Richards, Stephen F.
1991-01-01
Although computerized operations have significant gains realized in many areas, one area, scheduling, has enjoyed few benefits from automation. The traditional methods of industrial engineering and operations research have not proven robust enough to handle the complexities associated with the scheduling of realistic problems. To address this need, NASA has developed the computer-aided scheduling system (COMPASS), a sophisticated, interactive scheduling tool that is in wide-spread use within NASA and the contractor community. Therefore, COMPASS provides no explicit support for the large class of problems in which several people, perhaps at various locations, build separate schedules that share a common pool of resources. This research examines the issue of distributing scheduling, as applied to application domains characterized by the partial ordering of tasks, limited resources, and time restrictions. The focus of this research is on identifying issues related to distributed scheduling, locating applicable problem domains within NASA, and suggesting areas for ongoing research. The issues that this research identifies are goals, rescheduling requirements, database support, the need for communication and coordination among individual schedulers, the potential for expert system support for scheduling, and the possibility of integrating artificially intelligent schedulers into a network of human schedulers.
NASA Technical Reports Server (NTRS)
Adair, Jerry R.
1994-01-01
This paper is a consolidated report on ten major planning and scheduling systems that have been developed by the National Aeronautics and Space Administration (NASA). A description of each system, its components, and how it could be potentially used in private industry is provided in this paper. The planning and scheduling technology represented by the systems ranges from activity based scheduling employing artificial intelligence (AI) techniques to constraint based, iterative repair scheduling. The space related application domains in which the systems have been deployed vary from Space Shuttle monitoring during launch countdown to long term Hubble Space Telescope (HST) scheduling. This paper also describes any correlation that may exist between the work done on different planning and scheduling systems. Finally, this paper documents the lessons learned from the work and research performed in planning and scheduling technology and describes the areas where future work will be conducted.
Introduction to Exploring Machines
ERIC Educational Resources Information Center
Early Childhood Today, 2006
2006-01-01
Young children are fascinated by how things "work." They are at a stage of development where they want to experiment with the many ways to use an object or take things apart and put them back together. In the process of exploring tools and machines, children use the scientific method and problem-solving skills. They observe how things work, wonder…
ERIC Educational Resources Information Center
Sedwal, Mona; Kamat, Sangeeta
2008-01-01
The Scheduled Castes (SCs, also known as Dalits) and Scheduled Tribes (STs, also known as Adivasis) are among the most socially and educationally disadvantaged groups in India. This paper examines issues concerning school access and equity for Scheduled Caste and Scheduled Tribe communities and also highlights their unique problems, which may…
Experiments with a decision-theoretic scheduler
NASA Technical Reports Server (NTRS)
Hansson, Othar; Holt, Gerhard; Mayer, Andrew
1992-01-01
This paper describes DTS, a decision-theoretic scheduler designed to employ state-of-the-art probabilistic inference technology to speed the search for efficient solutions to constraint-satisfaction problems. Our approach involves assessing the performance of heuristic control strategies that are normally hard-coded into scheduling systems, and using probabilistic inference to aggregate this information in light of features of a given problem. BPS, the Bayesian Problem-Solver, introduced a similar approach to solving single-agent and adversarial graph search problems, yielding orders-of-magnitude improvement over traditional techniques. Initial efforts suggest that similar improvements will be realizable when applied to typical constraint-satisfaction scheduling problems.
Integrated online job-shop scheduling system
NASA Astrophysics Data System (ADS)
Zhao, Xing; Chen, Kuan H.; Luh, Peter B.; Chiueh, T. D.; Chang, ShihChang; Thakur, Lakshman S.
1999-11-01
The rapid development of information technology and e- commerce requires fast response form scheduling systems. Based on the Lagrangian relaxation approach for job shop scheduling, this paper present an integrated system that will generate schedules quickly. The Lagrangian relaxation approach is an iterative optimization process, where dynamic programming is solved in each iteration. Since dynamic programming is computational expensive especially for large problems, this paper develops the simplified dynamic programming, which will cut the computation time of each iteration by one order. Furthermore, a digital circuit to be embedded in PC is designed to implement the iterative optimization algorithm, leading to another order of speed improvement. The resulting integrated scheduling system consists of the hardware for optimization and the related software. It is estimated that two orders of magnitude gain in speed can be obtained, which will make on-line scheduling for practical job shops possible.
Parrott, G.A.
1985-05-07
A haulage system for a mining machine comprises a mining machine mounted on and/or guided by a conveyor and reciprocable with respect thereto, the conveyor being provided with a rack having plural rows of teeth of identical pitch, with the teeth of one row staggered with respect to an adjacent row(s), and the machine being provided with at least one power driven haulage sprocket comprising plural sets of peripherally arranged teeth of identical pitch, one set being angularly staggered with respect to an adjacent set(s), whereby one set is engageable with each row of teeth of the rack. The invention also includes a mining machine provided with such a power driven haulage sprocket, and a rack as above described and provided with end fittings for securing in articulated manner to an adjacent rack.
Schedule-Organizer Computer Program
NASA Technical Reports Server (NTRS)
Collazo, Fernando F.
1990-01-01
Schedule Organizer provides simple method for generating distribution lists. Contains readers' names for each task schedule defined by input files. Schedule Organizer (SO), Schedule Tracker (ST) (COSMIC program MSC-21526), and Schedule Report Generator (SRG) (COSMIC program MSC-21527) computer programs manipulating data-base files in ways advantageous in scheduling. Written in PL/1 and DEC Command Language (DCL).
NASA Technical Reports Server (NTRS)
1983-01-01
Castle Industries, Inc. is a small machine shop manufacturing replacement plumbing repair parts, such as faucet, tub and ballcock seats. Therese Castley, president of Castle decided to introduce Monel because it offered a chance to improve competitiveness and expand the product line. Before expanding, Castley sought NERAC assistance on Monel technology. NERAC (New England Research Application Center) provided an information package which proved very helpful. The NASA database was included in NERAC's search and yielded a wealth of information on machining Monel.
Space communications scheduler: A rule-based approach to adaptive deadline scheduling
NASA Technical Reports Server (NTRS)
Straguzzi, Nicholas
1990-01-01
Job scheduling is a deceptively complex subfield of computer science. The highly combinatorial nature of the problem, which is NP-complete in nearly all cases, requires a scheduling program to intelligently transverse an immense search tree to create the best possible schedule in a minimal amount of time. In addition, the program must continually make adjustments to the initial schedule when faced with last-minute user requests, cancellations, unexpected device failures, quests, cancellations, unexpected device failures, etc. A good scheduler must be quick, flexible, and efficient, even at the expense of generating slightly less-than-optimal schedules. The Space Communication Scheduler (SCS) is an intelligent rule-based scheduling system. SCS is an adaptive deadline scheduler which allocates modular communications resources to meet an ordered set of user-specified job requests on board the NASA Space Station. SCS uses pattern matching techniques to detect potential conflicts through algorithmic and heuristic means. As a result, the system generates and maintains high density schedules without relying heavily on backtracking or blind search techniques. SCS is suitable for many common real-world applications.
Distributed job scheduling in MetaCentrum
NASA Astrophysics Data System (ADS)
Tóth, Šimon; Ruda, Miroslav
2015-05-01
MetaCentrum - The Czech National Grid provides access to various resources across the Czech Republic. The utilized resource management and scheduling system is based on a heavily modified version of the Torque Batch System. This open source resource manager is maintained in a local fork and was extended to facilitate the requirements of such a large installation. This paper provides an overview of unique features deployed in MetaCentrum. Notably, we describe our distributed setup that encompasses several standalone independent servers while still maintaining full cooperative scheduling across the grid. We also present the benefits of our virtualized infrastructure that enables our schedulers to dynamically request ondemand virtual machines, that are then used to facilitate the varied requirements of users in our system, as well as enabling support for user requested virtual clusters that can be further interconnected using a private VLAN.
A scheduling model for astronomy
NASA Astrophysics Data System (ADS)
Solar, M.; Michelon, P.; Avarias, J.; Garces, M.
2016-04-01
Astronomical scheduling problem has several external conditions that change dynamically at any time during observations, like weather condition (humidity, temperature, wind speed, opacity, etc.), and target visibility conditions (target over the horizon, Sun/Moon blocking the target). Therefore, a dynamic re-scheduling is needed. An astronomical project will be scheduled as one or more Scheduling Blocks (SBs) as an atomic unit of astronomical observations. We propose a mixed integer linear programming (MILP) solution to select the best SBs, favors SBs with high scientific values, and thus maximizing the quantity of completed observation projects. The data content of Atacama Large Millimeter/Submillimeter Array (ALMA) projects of cycle 0 and cycle 1 were analyzed, and a synthetic set of tests of the real instances was created. Two configurations, one of 5000 SBs in a 3 months season and another 10,000 SBs a 6 months season were created. These instances were evaluated with excellent results. Through the testing it is showed that the MILP proposal has optimal solutions.
NASA Astrophysics Data System (ADS)
Muscettola, Nicola; Smith, Steven S.
1996-09-01
This final report summarizes research performed under NASA contract NCC 2-531 toward generalization of constraint-based scheduling theories and techniques for application to space telescope observation scheduling problems. Our work into theories and techniques for solution of this class of problems has led to the development of the Heuristic Scheduling Testbed System (HSTS), a software system for integrated planning and scheduling. Within HSTS, planning and scheduling are treated as two complementary aspects of the more general process of constructing a feasible set of behaviors of a target system. We have validated the HSTS approach by applying it to the generation of observation schedules for the Hubble Space Telescope. This report summarizes the HSTS framework and its application to the Hubble Space Telescope domain. First, the HSTS software architecture is described, indicating (1) how the structure and dynamics of a system is modeled in HSTS, (2) how schedules are represented at multiple levels of abstraction, and (3) the problem solving machinery that is provided. Next, the specific scheduler developed within this software architecture for detailed management of Hubble Space Telescope operations is presented. Finally, experimental performance results are given that confirm the utility and practicality of the approach.
NASA Technical Reports Server (NTRS)
Muscettola, Nicola; Smith, Steven S.
1996-01-01
This final report summarizes research performed under NASA contract NCC 2-531 toward generalization of constraint-based scheduling theories and techniques for application to space telescope observation scheduling problems. Our work into theories and techniques for solution of this class of problems has led to the development of the Heuristic Scheduling Testbed System (HSTS), a software system for integrated planning and scheduling. Within HSTS, planning and scheduling are treated as two complementary aspects of the more general process of constructing a feasible set of behaviors of a target system. We have validated the HSTS approach by applying it to the generation of observation schedules for the Hubble Space Telescope. This report summarizes the HSTS framework and its application to the Hubble Space Telescope domain. First, the HSTS software architecture is described, indicating (1) how the structure and dynamics of a system is modeled in HSTS, (2) how schedules are represented at multiple levels of abstraction, and (3) the problem solving machinery that is provided. Next, the specific scheduler developed within this software architecture for detailed management of Hubble Space Telescope operations is presented. Finally, experimental performance results are given that confirm the utility and practicality of the approach.
Evolutionary driver scheduling with relief chains.
Kwan, R S; Kwan, A S; Wren, A
2001-01-01
Public transport driver scheduling problems are well known to be NP-hard. Although some mathematically based methods are being used in the transport industry, there is room for improvement. A hybrid approach incorporating a genetic algorithm (GA) is presented. The role of the GA is to derive a small selection of good shifts to seed a greedy schedule construction heuristic. A group of shifts called a relief chain is identified and recorded. The relief chain is then inherited by the offspring and used by the GA for schedule construction. The new approach has been tested using real-life data sets, some of which represent very large problem instances. The results are generally better than those compiled by experienced schedulers and are comparable to solutions found by integer linear programming (ILP). In some cases, solutions were obtained when the ILP failed within practical computational limits. PMID:11709104
Holonic manufacturing scheduling: architecture, cooperation mechanism, and implementation
NASA Astrophysics Data System (ADS)
Gou, Ling; Luh, Peter B.; Kyoya, Yuji
1997-12-01
A holonic manufacturing system (HMS) is a manufacturing system where key elements, such as machines, cells, factories, parts, products, persons, teams, etc., are modeled as 'holons' having autonomous and cooperative properties. The distributed decision authority, the cooperative nature, and the integration of physical and informational aspects of system elements or holons make the HMS a new manufacturing paradigm, with great potential for meeting today's agile manufacturing challenges. Critical issues to be investigated include how to define holons for a given problem context, what should be the appropriate system architecture, and how to design effective cooperation mechanisms among holons for overall system performance. In this paper, holonic scheduling is developed for a factory consisting of multiple cells. Holons are identified, and their relationships are delineated through a novel modeling of the interactions among cells. The cooperation mechanisms among holons are established based on the 'Lagrangian relaxation technique' of mathematical optimization, and cooperation across cells is performed without accessing individual cells' local information nor intruding on their decision authority. The holonic system developed also possesses structural recursivity, and this property, combined with the integrability of individual holons, enables high system extendibility. Numerical testing shows that the method can generate high quality schedules in a timely fashion, and has comparable computational requirements as compared to a single-level Lagrangian relaxation method. The method thus provides a theoretical foundation for guiding the cooperation among holons, leading to globally near-optimal performance.
NASA Technical Reports Server (NTRS)
Halbfinger, Eliezer M.; Smith, Barry D.
1991-01-01
The Air Force Space Command schedules telemetry, tracking and control activities across the Air Force Satellite Control network. The Range Scheduling Aid (RSA) is a rapid prototype combining a user-friendly, portable, graphical interface with a sophisticated object-oriented database. The RSA has been a rapid prototyping effort whose purpose is to elucidate and define suitable technology for enhancing the performance of the range schedulers. Designing a system to assist schedulers in their task and using their current techniques as well as enhancements enabled by an electronic environment, has created a continuously developing model that will serve as a standard for future range scheduling systems. The RSA system is easy to use, easily ported between platforms, fast, and provides a set of tools for the scheduler that substantially increases his productivity.
Quay crane scheduling for an indented berth
NASA Astrophysics Data System (ADS)
Lee, Der-Horng; Chen, Jiang Hang; Cao, Jin Xin
2011-09-01
This article explores the quay crane scheduling problem at an indented berth. The indented berth is known as an innovative implementation in the container terminals to tackle the challenge from the emergence of more and more mega-containerships. A mixed integer programming model by considering the non-crossing and safety distance constraints is formulated. A Tabu search heuristic is developed to solve the proposed problem. The computational results from this research indicate that the designed Tabu search is an effective method to handle the quay crane scheduling problem at an indented berth.
Job schedul in Grid batch farms
NASA Astrophysics Data System (ADS)
Gellrich, Andreas
2014-06-01
We present here a study for a scheduler which cooperates with the queueing system TORQUE and is tailored to the needs of a HEP-dominated large Grid site with around 10000 jobs slots. Triggered by severe scaling problems of MAUI, a scheduler, referred to as MYSCHED, was developed and put into operation. We discuss conceptional aspects as well as experiences after almost two years of running.
Intelligent scheduling support for the US Coast Guard
Darby-Dowman, K.; Lucas, C.; Mitra, G.; Fink, R.; Kingsley, L.; Smith, J.W.
1992-12-31
This paper will discuss a joint effort by the U.S. Coast Guard Research & Development Center, Idaho National Engineering Laboratory and Brunel University to provide the necessary tools to increase the human scheduler`s capability to handle the scheduling process more efficiently and effectively. Automating the scheduling process required a system that could think independently of the scheduler, that is, the systems needed its own control mechanism and knowledge base. Further, automated schedule generation became a design requirement and sophisticated algorithms were formulated to solve a complex combinatorial problem. In short, the resulting design can be viewed as a hybrid knowledge-based mathematical programming application system. This document contains an overview of the integrated system, a discrete optimization model for scheduling, and schedule diagnosis and analysis.
Hagopian, Louis P; Boelter, Eric W; Jarmolowicz, David P
2011-01-01
This paper extends the Tiger, Hanley, and Bruzek (2008) review of functional communication training (FCT) by reviewing the published literature on reinforcement schedule thinning following FCT. As noted by Tiger et al. and others, schedule thinning may be necessary when the newly acquired communication response occurs excessively, to the extent that reinforcing it consistently is not practical in the natural environment. We provide a review of this literature including a discussion of each of the more commonly used schedule arrangements used for this purpose, outcomes obtained, a description of methods for progressing toward the terminal schedule, and a description of supplemental treatment components aimed at maintaining low levels of problem behavior during schedule thinning. Recommendations for schedule thinning are then provided. Finally, conceptual issues related to the reemergence of problem behavior during schedule thinning and areas for future research are discussed. PMID:22532899
Debugging Fortran on a shared memory machine
Allen, T.R.; Padua, D.A.
1987-01-01
Debugging on a parallel processor is more difficult than debugging on a serial machine because errors in a parallel program may introduce nondeterminism. The approach to parallel debugging presented here attempts to reduce the problem of debugging on a parallel machine to that of debugging on a serial machine by automatically detecting nondeterminism. 20 refs., 6 figs.
Brown coal preparation machines
Bleckmann, H.; Sitte, W.; Kellerwessel, H.
1981-05-01
Lignite usually requires comminuting and screening before being used as a fuel in power plants. Reduction machines normally used for coarse crushing bituminous coal, such as jaw crushers, roll crushers, and impact crushers, are not generally suitable for lignite as they require a brittle feed and large grain size. In contrast to these requirements, lignite can be easily compressed and has a small grain size. Therefore, special crusher types have been developed for the coarse reduction of lignite. These machines resemble roll crushers but subject the feed to shearing and tearing forces rather than to compressive stress. It is often necessary to screen the lignite to remove the undersize or to limit the maximum particle size before the next comminution process. Screening the lignite is a particularly difficult operation due to the high water content and the presence of clay minerals which tend to clog the screening machines. These problems can be overcome with multi-roll sizers.
Abdullahi, Mohammed; Ngadi, Md Asri
2016-01-01
Cloud computing has attracted significant attention from research community because of rapid migration rate of Information Technology services to its domain. Advances in virtualization technology has made cloud computing very popular as a result of easier deployment of application services. Tasks are submitted to cloud datacenters to be processed on pay as you go fashion. Task scheduling is one the significant research challenges in cloud computing environment. The current formulation of task scheduling problems has been shown to be NP-complete, hence finding the exact solution especially for large problem sizes is intractable. The heterogeneous and dynamic feature of cloud resources makes optimum task scheduling non-trivial. Therefore, efficient task scheduling algorithms are required for optimum resource utilization. Symbiotic Organisms Search (SOS) has been shown to perform competitively with Particle Swarm Optimization (PSO). The aim of this study is to optimize task scheduling in cloud computing environment based on a proposed Simulated Annealing (SA) based SOS (SASOS) in order to improve the convergence rate and quality of solution of SOS. The SOS algorithm has a strong global exploration capability and uses fewer parameters. The systematic reasoning ability of SA is employed to find better solutions on local solution regions, hence, adding exploration ability to SOS. Also, a fitness function is proposed which takes into account the utilization level of virtual machines (VMs) which reduced makespan and degree of imbalance among VMs. CloudSim toolkit was used to evaluate the efficiency of the proposed method using both synthetic and standard workload. Results of simulation showed that hybrid SOS performs better than SOS in terms of convergence speed, response time, degree of imbalance, and makespan. PMID:27348127
Abdullahi, Mohammed; Ngadi, Md Asri
2016-01-01
Cloud computing has attracted significant attention from research community because of rapid migration rate of Information Technology services to its domain. Advances in virtualization technology has made cloud computing very popular as a result of easier deployment of application services. Tasks are submitted to cloud datacenters to be processed on pay as you go fashion. Task scheduling is one the significant research challenges in cloud computing environment. The current formulation of task scheduling problems has been shown to be NP-complete, hence finding the exact solution especially for large problem sizes is intractable. The heterogeneous and dynamic feature of cloud resources makes optimum task scheduling non-trivial. Therefore, efficient task scheduling algorithms are required for optimum resource utilization. Symbiotic Organisms Search (SOS) has been shown to perform competitively with Particle Swarm Optimization (PSO). The aim of this study is to optimize task scheduling in cloud computing environment based on a proposed Simulated Annealing (SA) based SOS (SASOS) in order to improve the convergence rate and quality of solution of SOS. The SOS algorithm has a strong global exploration capability and uses fewer parameters. The systematic reasoning ability of SA is employed to find better solutions on local solution regions, hence, adding exploration ability to SOS. Also, a fitness function is proposed which takes into account the utilization level of virtual machines (VMs) which reduced makespan and degree of imbalance among VMs. CloudSim toolkit was used to evaluate the efficiency of the proposed method using both synthetic and standard workload. Results of simulation showed that hybrid SOS performs better than SOS in terms of convergence speed, response time, degree of imbalance, and makespan. PMID:27348127
Linux Kernel Co-Scheduling and Bulk Synchronous Parallelism
Jones, Terry R
2012-01-01
This paper describes a kernel scheduling algorithm that is based on coscheduling principles and that is intended for parallel applications running on 1000 cores or more. Experimental results for a Linux implementation on a Cray XT5 machine are presented. The results indicate that Linux is a suitable operating system for this new scheduling scheme, and that this design provides a dramatic improvement in scaling performance for synchronizing collective operations at scale.
Empirical results on scheduling and dynamic backtracking
NASA Technical Reports Server (NTRS)
Boddy, Mark S.; Goldman, Robert P.
1994-01-01
At the Honeywell Technology Center (HTC), we have been working on a scheduling problem related to commercial avionics. This application is large, complex, and hard to solve. To be a little more concrete: 'large' means almost 20,000 activities, 'complex' means several activity types, periodic behavior, and assorted types of temporal constraints, and 'hard to solve' means that we have been unable to eliminate backtracking through the use of search heuristics. At this point, we can generate solutions, where solutions exist, or report failure and sometimes why the system failed. To the best of our knowledge, this is among the largest and most complex scheduling problems to have been solved as a constraint satisfaction problem, at least that has appeared in the published literature. This abstract is a preliminary report on what we have done and how. In the next section, we present our approach to treating scheduling as a constraint satisfaction problem. The following sections present the application in more detail and describe how we solve scheduling problems in the application domain. The implemented system makes use of Ginsberg's Dynamic Backtracking algorithm, with some minor extensions to improve its utility for scheduling. We describe those extensions and the performance of the resulting system. The paper concludes with some general remarks, open questions and plans for future work.
Achieving spectrum conservation for the minimum-span and minimum-order frequency assignment problems
NASA Technical Reports Server (NTRS)
Heyward, Ann O.
1992-01-01
Effective and efficient solutions of frequency assignment problems assumes increasing importance as the radiofrequency spectrum experiences ever increasing utilization by diverse communications services, requiring that the most efficient use of this resource be achieved. The research presented explores a general approach to the frequency assignment problem, in which such problems are categorized by the appropriate spectrum conserving objective function, and are each treated as an N-job, M-machine scheduling problem appropriate for the objective. Results obtained and presented illustrate that such an approach presents an effective means of achieving spectrum conserving frequency assignments for communications systems in a variety of environments.
Mixed-Integer Formulations for Constellation Scheduling
NASA Astrophysics Data System (ADS)
Valicka, C.; Hart, W.; Rintoul, M.
Remote sensing systems have expanded the set of capabilities available for and critical to national security. Cooperating, high-fidelity sensing systems and growing mission applications have exponentially increased the set of potential schedules. A definitive lack of advanced tools places an increased burden on operators, as planning and scheduling remain largely manual tasks. This is particularly true in time-critical planning activities where operators aim to accomplish a large number of missions through optimal utilization of single or multiple sensor systems. Automated scheduling through identification and comparison of alternative schedules remains a challenging problem applicable across all remote sensing systems. Previous approaches focused on a subset of sensor missions and do not consider ad-hoc tasking. We have begun development of a robust framework that leverages the Pyomo optimization modeling language for the design of a tool to assist sensor operators planning under the constraints of multiple concurrent missions and uncertainty. Our scheduling models have been formulated to address the stochastic nature of ad-hoc tasks inserted under a variety of scenarios. Operator experience is being leveraged to select appropriate model objectives. Successful development of the framework will include iterative development of high-fidelity mission models that consider and expose various schedule performance metrics. Creating this tool will aid time-critical scheduling by increasing planning efficiency, clarifying the value of alternative modalities uniquely provided by multi-sensor systems, and by presenting both sets of organized information to operators. Such a tool will help operators more quickly and fully utilize sensing systems, a high interest objective within the current remote sensing operations community. Preliminary results for mixed-integer programming formulations of a sensor scheduling problem will be presented. Assumptions regarding sensor geometry
Optimal radiotherapy dose schedules under parametric uncertainty
NASA Astrophysics Data System (ADS)
Badri, Hamidreza; Watanabe, Yoichi; Leder, Kevin
2016-01-01
We consider the effects of parameter uncertainty on the optimal radiation schedule in the context of the linear-quadratic model. Our interest arises from the observation that if inter-patient variability in normal and tumor tissue radiosensitivity or sparing factor of the organs-at-risk (OAR) are not accounted for during radiation scheduling, the performance of the therapy may be strongly degraded or the OAR may receive a substantially larger dose than the allowable threshold. This paper proposes a stochastic radiation scheduling concept to incorporate inter-patient variability into the scheduling optimization problem. Our method is based on a probabilistic approach, where the model parameters are given by a set of random variables. Our probabilistic formulation ensures that our constraints are satisfied with a given probability, and that our objective function achieves a desired level with a stated probability. We used a variable transformation to reduce the resulting optimization problem to two dimensions. We showed that the optimal solution lies on the boundary of the feasible region and we implemented a branch and bound algorithm to find the global optimal solution. We demonstrated how the configuration of optimal schedules in the presence of uncertainty compares to optimal schedules in the absence of uncertainty (conventional schedule). We observed that in order to protect against the possibility of the model parameters falling into a region where the conventional schedule is no longer feasible, it is required to avoid extremal solutions, i.e. a single large dose or very large total dose delivered over a long period. Finally, we performed numerical experiments in the setting of head and neck tumors including several normal tissues to reveal the effect of parameter uncertainty on optimal schedules and to evaluate the sensitivity of the solutions to the choice of key model parameters.
Support vector machine in machine condition monitoring and fault diagnosis
NASA Astrophysics Data System (ADS)
Widodo, Achmad; Yang, Bo-Suk
2007-08-01
Recently, the issue of machine condition monitoring and fault diagnosis as a part of maintenance system became global due to the potential advantages to be gained from reduced maintenance costs, improved productivity and increased machine availability. This paper presents a survey of machine condition monitoring and fault diagnosis using support vector machine (SVM). It attempts to summarize and review the recent research and developments of SVM in machine condition monitoring and diagnosis. Numerous methods have been developed based on intelligent systems such as artificial neural network, fuzzy expert system, condition-based reasoning, random forest, etc. However, the use of SVM for machine condition monitoring and fault diagnosis is still rare. SVM has excellent performance in generalization so it can produce high accuracy in classification for machine condition monitoring and diagnosis. Until 2006, the use of SVM in machine condition monitoring and fault diagnosis is tending to develop towards expertise orientation and problem-oriented domain. Finally, the ability to continually change and obtain a novel idea for machine condition monitoring and fault diagnosis using SVM will be future works.
Nicholson, Daniel J
2013-12-01
The machine conception of the organism (MCO) is one of the most pervasive notions in modern biology. However, it has not yet received much attention by philosophers of biology. The MCO has its origins in Cartesian natural philosophy, and it is based on the metaphorical redescription of the organism as a machine. In this paper I argue that although organisms and machines resemble each other in some basic respects, they are actually very different kinds of systems. I submit that the most significant difference between organisms and machines is that the former are intrinsically purposive whereas the latter are extrinsically purposive. Using this distinction as a starting point, I discuss a wide range of dissimilarities between organisms and machines that collectively lay bare the inadequacy of the MCO as a general theory of living systems. To account for the MCO's prevalence in biology, I distinguish between its theoretical, heuristic, and rhetorical functions. I explain why the MCO is valuable when it is employed heuristically but not theoretically, and finally I illustrate the serious problems that arise from the rhetorical appeal to the MCO. PMID:23810470
SOFIA's Choice: Automating the Scheduling of Airborne Observations
NASA Technical Reports Server (NTRS)
Frank, Jeremy; Norvig, Peter (Technical Monitor)
1999-01-01
This paper describes the problem of scheduling observations for an airborne telescope. Given a set of prioritized observations to choose from, and a wide range of complex constraints governing legitimate choices and orderings, how can we efficiently and effectively create a valid flight plan which supports high priority observations? This problem is quite different from scheduling problems which are routinely solved automatically in industry. For instance, the problem requires making choices which lead to other choices later, and contains many interacting complex constraints over both discrete and continuous variables. Furthermore, new types of constraints may be added as the fundamental problem changes. As a result of these features, this problem cannot be solved by traditional scheduling techniques. The problem resembles other problems in NASA and industry, from observation scheduling for rovers and other science instruments to vehicle routing. The remainder of the paper is organized as follows. In 2 we describe the observatory in order to provide some background. In 3 we describe the problem of scheduling a single flight. In 4 we compare flight planning and other scheduling problems and argue that traditional techniques are not sufficient to solve this problem. We also mention similar complex scheduling problems which may benefit from efforts to solve this problem. In 5 we describe an approach for solving this problem based on research into a similar problem, that of scheduling observations for a space-borne probe. In 6 we discuss extensions of the flight planning problem as well as other problems which are similar to flight planning. In 7 we conclude and discuss future work.
Secure Autonomous Automated Scheduling (SAAS). Rev. 1.1
NASA Technical Reports Server (NTRS)
Walke, Jon G.; Dikeman, Larry; Sage, Stephen P.; Miller, Eric M.
2010-01-01
This report describes network-centric operations, where a virtual mission operations center autonomously receives sensor triggers, and schedules space and ground assets using Internet-based technologies and service-oriented architectures. For proof-of-concept purposes, sensor triggers are received from the United States Geological Survey (USGS) to determine targets for space-based sensors. The Surrey Satellite Technology Limited (SSTL) Disaster Monitoring Constellation satellite, the UK-DMC, is used as the space-based sensor. The UK-DMC's availability is determined via machine-to-machine communications using SSTL's mission planning system. Access to/from the UK-DMC for tasking and sensor data is via SSTL's and Universal Space Network's (USN) ground assets. The availability and scheduling of USN's assets can also be performed autonomously via machine-to-machine communications. All communication, both on the ground and between ground and space, uses open Internet standards
NASA Technical Reports Server (NTRS)
1995-01-01
The Orbotron is a tri-axle exercise machine patterned after a NASA training simulator for astronaut orientation in the microgravity of space. It has three orbiting rings corresponding to roll, pitch and yaw. The user is in the middle of the inner ring with the stomach remaining in the center of all axes, eliminating dizziness. Human power starts the rings spinning, unlike the NASA air-powered system. Marketed by Fantasy Factory (formerly Orbotron, Inc.), the machine can improve aerobic capacity, strength and endurance in five to seven minute workouts.
Characterization of robotics parallel algorithms and mapping onto a reconfigurable SIMD machine
NASA Technical Reports Server (NTRS)
Lee, C. S. G.; Lin, C. T.
1989-01-01
The kinematics, dynamics, Jacobian, and their corresponding inverse computations are six essential problems in the control of robot manipulators. Efficient parallel algorithms for these computations are discussed and analyzed. Their characteristics are identified and a scheme on the mapping of these algorithms to a reconfigurable parallel architecture is presented. Based on the characteristics including type of parallelism, degree of parallelism, uniformity of the operations, fundamental operations, data dependencies, and communication requirement, it is shown that most of the algorithms for robotic computations possess highly regular properties and some common structures, especially the linear recursive structure. Moreover, they are well-suited to be implemented on a single-instruction-stream multiple-data-stream (SIMD) computer with reconfigurable interconnection network. The model of a reconfigurable dual network SIMD machine with internal direct feedback is introduced. A systematic procedure internal direct feedback is introduced. A systematic procedure to map these computations to the proposed machine is presented. A new scheduling problem for SIMD machines is investigated and a heuristic algorithm, called neighborhood scheduling, that reorders the processing sequence of subtasks to reduce the communication time is described. Mapping results of a benchmark algorithm are illustrated and discussed.
Schedule Matters: Understanding the Relationship between Schedule Delays and Costs on Overruns
NASA Technical Reports Server (NTRS)
Majerowicz, Walt; Shinn, Stephen A.
2016-01-01
This paper examines the relationship between schedule delays and cost overruns on complex projects. It is generally accepted by many project practitioners that cost overruns are directly related to schedule delays. But what does "directly related to" actually mean? Some reasons or root causes for schedule delays and associated cost overruns are obvious, if only in hindsight. For example, unrealistic estimates, supply chain difficulties, insufficient schedule margin, technical problems, scope changes, or the occurrence of risk events can negatively impact schedule performance. Other factors driving schedule delays and cost overruns may be less obvious and more difficult to quantify. Examples of these less obvious factors include project complexity, flawed estimating assumptions, over-optimism, political factors, "black swan" events, or even poor leadership and communication. Indeed, is it even possible the schedule itself could be a source of delay and subsequent cost overrun? Through literature review, surveys of project practitioners, and the authors' own experience on NASA programs and projects, the authors will categorize and examine the various factors affecting the relationship between project schedule delays and cost growth. The authors will also propose some ideas for organizations to consider to help create an awareness of the factors which could cause or influence schedule delays and associated cost growth on complex projects.
Transportation Baseline Schedule
Fawcett, Ricky Lee; John, Mark Earl
2000-01-01
The “1999 National Transportation Program - Transportation Baseline Report” presents data that form a baseline to enable analysis and planning for future Department of Energy (DOE) Environmental Management (EM) waste/material transportation. The companion “1999 Transportation ‘Barriers’ Analysis” analyzes the data and identifies existing and potential problems that may prevent or delay transportation activities based on the data presented. The “1999 Transportation Baseline Schedule” (this report) uses the same data to provide an overview of the transportation activities of DOE EM waste/materials. This report can be used to identify areas where stakeholder interface is needed, and to communicate to stakeholders the quantity/schedule of shipments going through their area. Potential bottlenecks in the transportation system can be identified; the number of packages needed, and the capacity needed at receiving facilities can be planned. This report offers a visualization of baseline DOE EM transportation activities for the 11 major sites and the “Geologic Repository Disposal” site (GRD).
Hybrid Scheduling Model for Independent Grid Tasks
Shanthini, J.; Kalaikumaran, T.; Karthik, S.
2015-01-01
Grid computing facilitates the resource sharing through the administrative domains which are geographically distributed. Scheduling in a distributed heterogeneous environment is intrinsically very hard because of the heterogeneous nature of resource collection. Makespan and tardiness are two different measures of scheduling, and many of the previous researches concentrated much on reduction of makespan, which measures the machine utilization. In this paper, we propose a hybrid scheduling algorithm for scheduling independent grid tasks with the objective of reducing total weighted tardiness of grid tasks. Tardiness is to measure the due date performance, which has a direct impact on cost for executing the jobs. In this paper we propose BG_ATC algorithm which is a combination of best gap (BG) search and Apparent Tardiness Cost (ATC) indexing algorithm. Furthermore, we implemented these two algorithms in two different phases of the scheduling process. In addition to that, the comparison was made on results with various benchmark algorithms and the experimental results show that our algorithm outperforms the benchmark algorithms. PMID:26543897
Integrated scheduling and resource management. [for Space Station Information System
NASA Technical Reports Server (NTRS)
Ward, M. T.
1987-01-01
This paper examines the problem of integrated scheduling during the Space Station era. Scheduling for Space Station entails coordinating the support of many distributed users who are sharing common resources and pursuing individual and sometimes conflicting objectives. This paper compares the scheduling integration problems of current missions with those anticipated for the Space Station era. It examines the facilities and the proposed operations environment for Space Station. It concludes that the pattern of interdependecies among the users and facilities, which are the source of the integration problem is well structured, allowing a dividing of the larger problem into smaller problems. It proposes an architecture to support integrated scheduling by scheduling efficiently at local facilities as a function of dependencies with other facilities of the program. A prototype is described that is being developed to demonstrate this integration concept.
NASA Technical Reports Server (NTRS)
Logan, J. R.; Pulvermacher, M. K.
1991-01-01
Range Scheduling Aid (RSA) is presented in the form of the viewgraphs. The following subject areas are covered: satellite control network; current and new approaches to range scheduling; MITRE tasking; RSA features; RSA display; constraint based analytic capability; RSA architecture; and RSA benefits.
ERIC Educational Resources Information Center
Haley, Marjorie
A discussion of block scheduling for second language instruction looks at the advantages and disadvantages and offers some suggestions for classroom management and course organization. It is argued that block scheduling may offer a potential solution to large classes, insufficient time for labs, too little individualized instruction; few…
ERIC Educational Resources Information Center
Queen, J. Allen
2000-01-01
Successful block scheduling depends on provision of initial and ongoing instructional training. Teaching strategies should vary and include cooperative learning, the case method, the socratic seminar, synectics, concept attainment, the inquiry method, and simulations. Recommendations for maximizing block scheduling are outlined. (Contains 52…
ERIC Educational Resources Information Center
Ubben, Gerald C.
1976-01-01
Achieving flexibility without losing student accountability is a challenge that faces every school. With a fluid block schedule, as described here, accountability is maintained without inhibiting flexibility. An additional advantage is that three levels of schedule decision making take some of the pressure off the principal. (Editor)
ERIC Educational Resources Information Center
Williamson, Ronald
2010-01-01
Driven by stable or declining financial resources many school districts are considering the costs and benefits of a seven-period day. While there is limited evidence that any particular scheduling model has a greater impact on student learning than any other, it is clear that the school schedule is a tool that can significantly impact teacher…
Trimester Schedule. Research Brief
ERIC Educational Resources Information Center
Education Partnerships, Inc., 2012
2012-01-01
Why do a trimester schedule? With the advent of block scheduling, many high schools conducted research on utilizing that plan in a trimester format. There appeared to be three issues that most schools faced: (1) How to provide substantive instructional time that was not fragmented?; (2) How does the school climate contribute positively to…
Anaesthesia Machine: Checklist, Hazards, Scavenging
Goneppanavar, Umesh; Prabhu, Manjunath
2013-01-01
From a simple pneumatic device of the early 20th century, the anaesthesia machine has evolved to incorporate various mechanical, electrical and electronic components to be more appropriately called anaesthesia workstation. Modern machines have overcome many drawbacks associated with the older machines. However, addition of several mechanical, electronic and electric components has contributed to recurrence of some of the older problems such as leak or obstruction attributable to newer gadgets and development of newer problems. No single checklist can satisfactorily test the integrity and safety of all existing anaesthesia machines due to their complex nature as well as variations in design among manufacturers. Human factors have contributed to greater complications than machine faults. Therefore, better understanding of the basics of anaesthesia machine and checking each component of the machine for proper functioning prior to use is essential to minimise these hazards. Clear documentation of regular and appropriate servicing of the anaesthesia machine, its components and their satisfactory functioning following servicing and repair is also equally important. Trace anaesthetic gases polluting the theatre atmosphere can have several adverse effects on the health of theatre personnel. Therefore, safe disposal of these gases away from the workplace with efficiently functioning scavenging system is necessary. Other ways of minimising atmospheric pollution such as gas delivery equipment with negligible leaks, low flow anaesthesia, minimal leak around the airway equipment (facemask, tracheal tube, laryngeal mask airway, etc.) more than 15 air changes/hour and total intravenous anaesthesia should also be considered. PMID:24249887
NASA Technical Reports Server (NTRS)
Wang, Yeou-Fang; Baldwin, John
2007-01-01
TIGRAS is client-side software, which provides tracking-station equipment planning, allocation, and scheduling services to the DSMS (Deep Space Mission System). TIGRAS provides functions for schedulers to coordinate the DSN (Deep Space Network) antenna usage time and to resolve the resource usage conflicts among tracking passes, antenna calibrations, maintenance, and system testing activities. TIGRAS provides a fully integrated multi-pane graphical user interface for all scheduling operations. This is a great improvement over the legacy VAX VMS command line user interface. TIGRAS has the capability to handle all DSN resource scheduling aspects from long-range to real time. TIGRAS assists NASA mission operations for DSN tracking of station equipment resource request processes from long-range load forecasts (ten years or longer), to midrange, short-range, and real-time (less than one week) emergency tracking plan changes. TIGRAS can be operated by NASA mission operations worldwide to make schedule requests for the DSN station equipment.
SOFIA's Choice: Scheduling Observations for an Airborne Observatory
NASA Technical Reports Server (NTRS)
Frank, Jeremy; Kurklu, Elif; Koga, Dennis (Technical Monitor)
2002-01-01
We describe the problem of scheduling observations for an airborne observatory. The problem is more complex than traditional scheduling problems in that it incorporates complex constraints relating the feasibility of an astronomical observation to the position and time of a mobile observatory, as well as traditional temporal constraints and optimization criteria. We describe the problem, its proposed solution and the empirical validation of that solution.
ERIC Educational Resources Information Center
Fendrich, Jean
2002-01-01
Collectors everywhere know that local antique shops and flea markets are treasure troves just waiting to be plundered. Science teachers might take a hint from these hobbyists, for the next community yard sale might be a repository of old, quirky items that are just the things to get students thinking about simple machines. By introducing some…
Power system scheduling with fuzzy reserve requirements
Guan, X.; Luh, P.B.; Prasannan, B.
1996-05-01
The modeling of constraints is an important issue in power system scheduling. Constraints can be generally classified into two categories: (1) physical limits and (2) operating limits. A schedule violating physical limit or constraint would not be acceptable. An operating limit, however, is often imposed to enhance system security but does not represent a physical bound. This kind of soft limits can be temporarily violated a little bit if necessary, but not too much. These constraints are therefore fuzzy in nature, and crisp treatment of them may lead to over conservative solutions. In this paper, a fuzzy optimization-based method is developed to solve power system scheduling problem with fuzzy reserve requirements. The problem is first converted to a crisp and separable optimization problem. Lagrange multipliers are then used to relax system-wide constraints and decompose the problem into a number of unit-wise subproblems and a membership subproblem. These subproblems can be efficiently solved, and the multipliers are updated by using a subgradient method. Heuristics are then used to construct a feasible schedule based on subproblem solutions. Numerical testing results show that near optimal schedules are obtained, and the method can provide a good balance between reducing costs and satisfying reserve requirements.
Mission Operations Planning and Scheduling System (MOPSS)
NASA Technical Reports Server (NTRS)
Wood, Terri; Hempel, Paul
2011-01-01
MOPSS is a generic framework that can be configured on the fly to support a wide range of planning and scheduling applications. It is currently used to support seven missions at Goddard Space Flight Center (GSFC) in roles that include science planning, mission planning, and real-time control. Prior to MOPSS, each spacecraft project built its own planning and scheduling capability to plan satellite activities and communications and to create the commands to be uplinked to the spacecraft. This approach required creating a data repository for storing planning and scheduling information, building user interfaces to display data, generating needed scheduling algorithms, and implementing customized external interfaces. Complex scheduling problems that involved reacting to multiple variable situations were analyzed manually. Operators then used the results to add commands to the schedule. Each architecture was unique to specific satellite requirements. MOPSS is an expert system that automates mission operations and frees the flight operations team to concentrate on critical activities. It is easily reconfigured by the flight operations team as the mission evolves. The heart of the system is a custom object-oriented data layer mapped onto an Oracle relational database. The combination of these two technologies allows a user or system engineer to capture any type of scheduling or planning data in the system's generic data storage via a GUI.
Schedule-Aware Workflow Management Systems
NASA Astrophysics Data System (ADS)
Mans, Ronny S.; Russell, Nick C.; van der Aalst, Wil M. P.; Moleman, Arnold J.; Bakker, Piet J. M.
Contemporary workflow management systems offer work-items to users through specific work-lists. Users select the work-items they will perform without having a specific schedule in mind. However, in many environments work needs to be scheduled and performed at particular times. For example, in hospitals many work-items are linked to appointments, e.g., a doctor cannot perform surgery without reserving an operating theater and making sure that the patient is present. One of the problems when applying workflow technology in such domains is the lack of calendar-based scheduling support. In this paper, we present an approach that supports the seamless integration of unscheduled (flow) and scheduled (schedule) tasks. Using CPN Tools we have developed a specification and simulation model for schedule-aware workflow management systems. Based on this a system has been realized that uses YAWL, Microsoft Exchange Server 2007, Outlook, and a dedicated scheduling service. The approach is illustrated using a real-life case study at the AMC hospital in the Netherlands. In addition, we elaborate on the experiences obtained when developing and implementing a system of this scale using formal techniques.
NASA Astrophysics Data System (ADS)
Prasetyaningsih, E.; Suprayogi; Samadhi, TMAA; Halim, AH
2016-02-01
This paper studies production and delivery batch scheduling problems for a single- supplier-to-a-single-manufacturer case, with multiple capacitated vehicles wherein different holding costs between in-process and completed parts are allowed. In the problem, the parts of a single item are first batched,then the resulting batches are processed on a single machine. All completed batches are transported in a number of deliveries in order to be received at a common due date. The objective is to find the integrated schedule of production and delivery batches so as to satisfy its due date and to minimize the total cost of associated in-process parts inventory, completed parts inventory and delivery. It should be noted that both holding costs constitute a derivation of the so-called actual flow time, and the delivery cost is proportional to the required number of deliveries. The problem can be formulated as an integer non-linier programming and it is solved optimally by Lingo 11.0 software. Numerical experiences show that there are two patterns of batch sizes affected by the ratio of holding costs of in-process and completed parts. It can be used by practitioners to solve the realistic integrated production and delivery batch scheduling problem.
Job Scheduling with Efficient Resource Monitoring in Cloud Datacenter
Loganathan, Shyamala; Mukherjee, Saswati
2015-01-01
Cloud computing is an on-demand computing model, which uses virtualization technology to provide cloud resources to users in the form of virtual machines through internet. Being an adaptable technology, cloud computing is an excellent alternative for organizations for forming their own private cloud. Since the resources are limited in these private clouds maximizing the utilization of resources and giving the guaranteed service for the user are the ultimate goal. For that, efficient scheduling is needed. This research reports on an efficient data structure for resource management and resource scheduling technique in a private cloud environment and discusses a cloud model. The proposed scheduling algorithm considers the types of jobs and the resource availability in its scheduling decision. Finally, we conducted simulations using CloudSim and compared our algorithm with other existing methods, like V-MCT and priority scheduling algorithms. PMID:26473166
Scheduling periodic jobs using imprecise results
NASA Technical Reports Server (NTRS)
Chung, Jen-Yao; Liu, Jane W. S.; Lin, Kwei-Jay
1987-01-01
One approach to avoid timing faults in hard, real-time systems is to make available intermediate, imprecise results produced by real-time processes. When a result of the desired quality cannot be produced in time, an imprecise result of acceptable quality produced before the deadline can be used. The problem of scheduling periodic jobs to meet deadlines on a system that provides the necessary programming language primitives and run-time support for processes to return imprecise results is discussed. Since the scheduler may choose to terminate a task before it is completed, causing it to produce an acceptable but imprecise result, the amount of processor time assigned to any task in a valid schedule can be less than the amount of time required to complete the task. A meaningful formulation of the scheduling problem must take into account the overall quality of the results. Depending on the different types of undesirable effects caused by errors, jobs are classified as type N or type C. For type N jobs, the effects of errors in results produced in different periods are not cumulative. A reasonable performance measure is the average error over all jobs. Three heuristic algorithms that lead to feasible schedules with small average errors are described. For type C jobs, the undesirable effects of errors produced in different periods are cumulative. Schedulability criteria of type C jobs are discussed.
Chandra mission scheduling on-orbit experience
NASA Astrophysics Data System (ADS)
Bucher, Sabina; Williams, Brent; Pendexter, Misty; Balke, David
2008-07-01
Scheduling observatory time to maximize both day-to-day science target integration time and the lifetime of the observatory is a formidable challenge. Furthermore, it is not a static problem. Of course, every schedule brings a new set of observations, but the boundaries of the problem change as well. As spacecraft ages, its capabilities may degrade. As in-flight experience grows, capabilities may expand. As observing programs are completed, the needs and expectations of the science community may evolve. Changes such as these impact the rules by which a mission scheduled. In eight years on orbit, the Chandra X-Ray Observatory Mission Planning process has adapted to meet the challenge of maximizing day-to-day and mission lifetime science return, despite a consistently evolving set of scheduling constraints. The success of the planning team has been achieved, not through the use of complex algorithms and optimization routines, but through processes and home grown tools that help individuals make smart short term and long term Mission Planning decisions. This paper walks through the processes and tools used to plan and produce mission schedules for the Chandra X-Ray Observatory. Nominal planning and scheduling, target of opportunity response, and recovery from on-board autonomous safing actions are all addressed. Evolution of tools and processes, best practices, and lessons learned are highlighted along the way.
Technology for planning and scheduling under complex constraints
NASA Astrophysics Data System (ADS)
Alguire, Karen M.; Pedro Gomes, Carla O.
1997-02-01
Within the context of law enforcement, several problems fall into the category of planning and scheduling under constraints. Examples include resource and personnel scheduling, and court scheduling. In the case of court scheduling, a schedule must be generated considering available resources, e.g., court rooms and personnel. Additionally, there are constraints on individual court cases, e.g., temporal and spatial, and between different cases, e.g., precedence. Finally, there are overall objectives that the schedule should satisfy such as timely processing of cases and optimal use of court facilities. Manually generating a schedule that satisfies all of the constraints is a very time consuming task. As the number of court cases and constraints increases, this becomes increasingly harder to handle without the assistance of automatic scheduling techniques. This paper describes artificial intelligence (AI) technology that has been used to develop several high performance scheduling applications including a military transportation scheduler, a military in-theater airlift scheduler, and a nuclear power plant outage scheduler. We discuss possible law enforcement applications where we feel the same technology could provide long-term benefits to law enforcement agencies and their operations personnel.
NASA Technical Reports Server (NTRS)
1981-01-01
Manufacturer of the Model 2210 copying machine was looking for a plastic valve bushing material that could be produced by a low-cost injection molding process to replace the unsuitable valve bushing they were using. NERAC conducted a computer search of the NASA database and was able to supply Nashua Corporation with several technical reports in their area of interest. Information aided the company's development of a urethane valve bushing which solved the problem and created a dramatic reduction in unit cost.
A Synthesized Heuristic Task Scheduling Algorithm
Dai, Yanyan; Zhang, Xiangli
2014-01-01
Aiming at the static task scheduling problems in heterogeneous environment, a heuristic task scheduling algorithm named HCPPEFT is proposed. In task prioritizing phase, there are three levels of priority in the algorithm to choose task. First, the critical tasks have the highest priority, secondly the tasks with longer path to exit task will be selected, and then algorithm will choose tasks with less predecessors to schedule. In resource selection phase, the algorithm is selected task duplication to reduce the interresource communication cost, besides forecasting the impact of an assignment for all children of the current task permits better decisions to be made in selecting resources. The algorithm proposed is compared with STDH, PEFT, and HEFT algorithms through randomly generated graphs and sets of task graphs. The experimental results show that the new algorithm can achieve better scheduling performance. PMID:25254244
Monitoring Building Systems for Schedule Compliance
Jensen, Andrew M.; Belew, Shan T.
2013-02-19
As Pacific Northwest National Laboratory (PNNL) initiated a Core Business Hours program, it became a challenge to ensure that the hundreds of systems campus wide were operating within their programmed schedules. Therefore, a collaborative exchange between PNNL operations and PNNL researchers developing the Decision Support for Operations and Maintenance (DSOM) software package was initiated to create a tool to solve this problem. This new DSOM tool verifies systems are operating within scheduled operation times by polling Building Automation and Control Network (BACnet) identifiers of systems’ on/off or command statuses. The tool records the time spent in operation state (ON) and totalizes each system over a rolling 7-day period, highlighting systems that are running over the scheduled hours. This snapshot view allows building management to look quickly at the entire campus to ensure that systems are not operating beyond their scheduled hours.
NASA Schedule Management Handbook
NASA Technical Reports Server (NTRS)
2011-01-01
The purpose of schedule management is to provide the framework for time-phasing, resource planning, coordination, and communicating the necessary tasks within a work effort. The intent is to improve schedule management by providing recommended concepts, processes, and techniques used within the Agency and private industry. The intended function of this handbook is two-fold: first, to provide guidance for meeting the scheduling requirements contained in NPR 7120.5, NASA Space Flight Program and Project Management Requirements, NPR 7120.7, NASA Information Technology and Institutional Infrastructure Program and Project Requirements, NPR 7120.8, NASA Research and Technology Program and Project Management Requirements, and NPD 1000.5, Policy for NASA Acquisition. The second function is to describe the schedule management approach and the recommended best practices for carrying out this project control function. With regards to the above project management requirements documents, it should be noted that those space flight projects previously established and approved under the guidance of prior versions of NPR 7120.5 will continue to comply with those requirements until project completion has been achieved. This handbook will be updated as needed, to enhance efficient and effective schedule management across the Agency. It is acknowledged that most, if not all, external organizations participating in NASA programs/projects will have their own internal schedule management documents. Issues that arise from conflicting schedule guidance will be resolved on a case by case basis as contracts and partnering relationships are established. It is also acknowledged and understood that all projects are not the same and may require different levels of schedule visibility, scrutiny and control. Project type, value, and complexity are factors that typically dictate which schedule management practices should be employed.
Temporal planning for transportation planning and scheduling
NASA Technical Reports Server (NTRS)
Frederking, Robert E.; Muscettola, Nicola
1992-01-01
In this paper we describe preliminary work done in the CORTES project, applying the Heuristic Scheduling Testbed System (HSTS) to a transportation planning and scheduling domain. First, we describe in more detail the transportation problems that we are addressing. We then describe the fundamental characteristics of HSTS and we concentrate on the representation of multiple capacity resources. We continue with a more detailed description of the transportation planning problem that we have initially addressed in HSTS and of its solution. Finally we describe future directions for our research.
Introducing Machine Learning Concepts with WEKA.
Smith, Tony C; Frank, Eibe
2016-01-01
This chapter presents an introduction to data mining with machine learning. It gives an overview of various types of machine learning, along with some examples. It explains how to download, install, and run the WEKA data mining toolkit on a simple data set, then proceeds to explain how one might approach a bioinformatics problem. Finally, it includes a brief summary of machine learning algorithms for other types of data mining problems, and provides suggestions about where to find additional information. PMID:27008023
Drilling Machines: Vocational Machine Shop.
ERIC Educational Resources Information Center
Thomas, John C.
The lessons and supportive information in this field tested instructional block provide a guide for teachers in developing a machine shop course of study in drilling. The document is comprised of operation sheets, information sheets, and transparency masters for 23 lessons. Each lesson plan includes a performance objective, material and tools,…
Decision-theoretic control of EUVE telescope scheduling
NASA Technical Reports Server (NTRS)
Hansson, Othar; Mayer, Andrew
1993-01-01
This paper describes a decision theoretic scheduler (DTS) designed to employ state-of-the-art probabilistic inference technology to speed the search for efficient solutions to constraint-satisfaction problems. Our approach involves assessing the performance of heuristic control strategies that are normally hard-coded into scheduling systems and using probabilistic inference to aggregate this information in light of the features of a given problem. The Bayesian Problem-Solver (BPS) introduced a similar approach to solving single agent and adversarial graph search patterns yielding orders-of-magnitude improvement over traditional techniques. Initial efforts suggest that similar improvements will be realizable when applied to typical constraint-satisfaction scheduling problems.
Ada and cyclic runtime scheduling
NASA Technical Reports Server (NTRS)
Hood, Philip E.
1986-01-01
An important issue that must be faced while introducing Ada into the real time world is efficient and prodictable runtime behavior. One of the most effective methods employed during the traditional design of a real time system is the cyclic executive. The role cyclic scheduling might play in an Ada application in terms of currently available implementations and in terms of implementations that might be developed especially to support real time system development is examined. The cyclic executive solves many of the problems faced by real time designers, resulting in a system for which it is relatively easy to achieve approporiate timing behavior. Unfortunately a cyclic executive carries with it a very high maintenance penalty over the lifetime of the software that is schedules. Additionally, these cyclic systems tend to be quite fragil when any aspect of the system changes. The findings are presented of an ongoing SofTech investigation into Ada methods for real time system development. The topics covered include a description of the costs involved in using cyclic schedulers, the sources of these costs, and measures for future systems to avoid these costs without giving up the runtime performance of a cyclic system.
Universal Memcomputing Machines.
Traversa, Fabio Lorenzo; Di Ventra, Massimiliano
2015-11-01
We introduce the notion of universal memcomputing machines (UMMs): a class of brain-inspired general-purpose computing machines based on systems with memory, whereby processing and storing of information occur on the same physical location. We analytically prove that the memory properties of UMMs endow them with universal computing power (they are Turing-complete), intrinsic parallelism, functional polymorphism, and information overhead, namely, their collective states can support exponential data compression directly in memory. We also demonstrate that a UMM has the same computational power as a nondeterministic Turing machine, namely, it can solve nondeterministic polynomial (NP)-complete problems in polynomial time. However, by virtue of its information overhead, a UMM needs only an amount of memory cells (memprocessors) that grows polynomially with the problem size. As an example, we provide the polynomial-time solution of the subset-sum problem and a simple hardware implementation of the same. Even though these results do not prove the statement NP = P within the Turing paradigm, the practical realization of these UMMs would represent a paradigm shift from the present von Neumann architectures, bringing us closer to brain-like neural computation. PMID:25667360
Intelligent scheduling support for the US Coast Guard
Darby-Dowman, K.; Lucas, C.; Mitra, G. ); Fink, R. ); Kingsley, L.; Smith, J.W. )
1992-01-01
This paper will discuss a joint effort by the U.S. Coast Guard Research Development Center, Idaho National Engineering Laboratory and Brunel University to provide the necessary tools to increase the human scheduler's capability to handle the scheduling process more efficiently and effectively. Automating the scheduling process required a system that could think independently of the scheduler, that is, the systems needed its own control mechanism and knowledge base. Further, automated schedule generation became a design requirement and sophisticated algorithms were formulated to solve a complex combinatorial problem. In short, the resulting design can be viewed as a hybrid knowledge-based mathematical programming application system. This document contains an overview of the integrated system, a discrete optimization model for scheduling, and schedule diagnosis and analysis.
Medlin, John B.
1976-05-25
A charging machine for loading fuel slugs into the process tubes of a nuclear reactor includes a tubular housing connected to the process tube, a charging trough connected to the other end of the tubular housing, a device for loading the charging trough with a group of fuel slugs, means for equalizing the coolant pressure in the charging trough with the pressure in the process tubes, means for pushing the group of fuel slugs into the process tube and a latch and a seal engaging the last object in the group of fuel slugs to prevent the fuel slugs from being ejected from the process tube when the pusher is removed and to prevent pressure liquid from entering the charging machine.
NASA Technical Reports Server (NTRS)
Globus, Al; Saini, Subhash (Technical Monitor)
1998-01-01
Fullerenes possess remarkable properties and many investigators have examined the mechanical, electronic and other characteristics of carbon SP2 systems in some detail. In addition, C-60 can be functionalized with many classes of molecular fragments and we may expect the caps of carbon nanotubes to have a similar chemistry. Finally, carbon nanotubes have been attached to t he end of scanning probe microscope (Spill) tips. Spills can be manipulated with sub-angstrom accuracy. Together, these investigations suggest that complex molecular machines made of fullerenes may someday be created and manipulated with very high accuracy. We have studied some such systems computationally (primarily functionalized carbon nanotube gears and computer components). If such machines can be combined appropriately, a class of materials may be created that can sense their environment, calculate a response, and act. The implications of such hypothetical materials are substantial.
NASA Technical Reports Server (NTRS)
Globus, Al; Saini, Subhash
1998-01-01
Recent computational efforts at NASA Ames Research Center and computation and experiment elsewhere suggest that a nanotechnology of machine phase functionalized fullerenes may be synthetically accessible and of great interest. We have computationally demonstrated that molecular gears fashioned from (14,0) single-walled carbon nanotubes and benzyne teeth should operate well at 50-100 gigahertz. Preliminary results suggest that these gears can be cooled by a helium atmosphere and a laser motor can power fullerene gears if a positive and negative charge have been added to form a dipole. In addition, we have unproven concepts based on experimental and computational evidence for support structures, computer control, a system architecture, a variety of components, and manufacture. Combining fullerene machines with the remarkable mechanical properties of carbon nanotubes, there is some reason to believe that a focused effort to develop fullerene nanotechnology could yield materials with tremendous properties.
Owen, Whitney H.
1980-01-01
A polyphase rotary induction machine for use as a motor or generator utilizing a single rotor assembly having two series connected sets of rotor windings, a first stator winding disposed around the first rotor winding and means for controlling the current induced in one set of the rotor windings compared to the current induced in the other set of the rotor windings. The rotor windings may be wound rotor windings or squirrel cage windings.
Scheduling Future Water Supply Investments Under Uncertainty
NASA Astrophysics Data System (ADS)
Huskova, I.; Matrosov, E. S.; Harou, J. J.; Kasprzyk, J. R.; Reed, P. M.
2014-12-01
Uncertain hydrological impacts of climate change, population growth and institutional changes pose a major challenge to planning of water supply systems. Planners seek optimal portfolios of supply and demand management schemes but also when to activate assets whilst considering many system goals and plausible futures. Incorporation of scheduling into the planning under uncertainty problem strongly increases its complexity. We investigate some approaches to scheduling with many-objective heuristic search. We apply a multi-scenario many-objective scheduling approach to the Thames River basin water supply system planning problem in the UK. Decisions include which new supply and demand schemes to implement, at what capacity and when. The impact of different system uncertainties on scheme implementation schedules are explored, i.e. how the choice of future scenarios affects the search process and its outcomes. The activation of schemes is influenced by the occurrence of extreme hydrological events in the ensemble of plausible scenarios and other factors. The approach and results are compared with a previous study where only the portfolio problem is addressed (without scheduling).
Parallel Batch Scheduling of Deteriorating Jobs with Release Dates and Rejection
Zou, Juan; Miao, Cuixia
2014-01-01
We consider the unbounded parallel batch scheduling with deterioration, release dates, and rejection. Each job is either accepted and processed on a single batching machine, or rejected by paying penalties. The processing time of a job is a simple linear increasing function of its starting time. The objective is to minimize the sum of the makespan of the accepted jobs and the total penalty of the rejected jobs. First, we show that the problem is NP-hard in the ordinary sense. Then, we present two pseudopolynomial time algorithms and a fully polynomial-time approximation scheme to solve this problem. Furthermore, we provide an optimal O(nlogn) time algorithm for the case where jobs have identical release dates. PMID:25143969
A comparison of dense-to-lean and fixed lean schedules of alternative reinforcement and extinction.
Hagopian, Louis P; Toole, Lisa M; Long, Ethan S; Bowman, Lynn G; Lieving, Gregory A
2004-01-01
Behavior-reduction interventions typically employ dense schedules of alternative reinforcement in conjunction with operant extinction for problem behavior. After problem behavior is reduced in the initial treatment stages, schedule thinning is routinely conducted to make the intervention more practical in natural environments. In the current investigation, two methods for thinning alternative reinforcement schedules were compared for 3 clients who exhibited severe problem behavior. In the dense-to-lean (DTL) condition, reinforcement was delivered on relatively dense schedules (using noncontingent reinforcement for 1 participant and functional communication training for 2 participants), followed by systematic schedule thinning to progressively leaner schedules. During the fixed lean (FL) condition, reinforcement was delivered on lean schedules (equivalent to the terminal schedule of the DTL condition). The FL condition produced a quicker attainment of individual treatment goals for 2 of the 3 participants. The results are discussed in terms of the potential utility of using relatively lean schedules at treatment outset. PMID:15529889
ERIC Educational Resources Information Center
Blakesley, James F.
1982-01-01
College scheduling system and operating procedure objectives influence efficient use of resources. Theoretical and practical considerations are outlined, including a step-by-step assessment of an institution's enrollment capacity and discussion of centralized or reassignable classrooms. (Author/MSE)
... Navigation Bar Home Current Issue Past Issues Childhood Vaccine Schedule Past Issues / Spring 2008 Table of Contents ... please turn Javascript on. When to Vaccinate What Vaccine Why Birth (or any age if not previously ...
NASA Technical Reports Server (NTRS)
Smith, Grego
2004-01-01
Schedule Risk Assessment (SRA) determines the probability of finishing on or before a given point in time. This viewgraph presentation introduces the prerequisites, probability distribution curves, special conditions, calculations, and results analysis for SRA.
Distributed network scheduling
NASA Technical Reports Server (NTRS)
Clement, Bradley J.; Schaffer, Steven R.
2005-01-01
We investigate missions where communications resources are limited, requiring autonomous planning and execution. Unlike typical networks, spacecraft networks are also suited to automated planning and scheduling because many communications can be planned in advance.
Initial Hardware Development Schedule
NASA Technical Reports Server (NTRS)
Culpepper, William X.
1991-01-01
The hardware development schedule for the Common Lunar Lander's (CLLs) tracking system is presented. Among the topics covered are the following: historical perspective, solution options, industry contacts, and the rationale for selection.
Cardoso, Goncalo; Stadler, Michael; Bozchalui, Mohammed C.; Sharma, Ratnesh; Marnay, Chris; Barbosa-Povoa, Ana; Ferrao, Paulo
2013-12-06
The large scale penetration of electric vehicles (EVs) will introduce technical challenges to the distribution grid, but also carries the potential for vehicle-to-grid services. Namely, if available in large enough numbers, EVs can be used as a distributed energy resource (DER) and their presence can influence optimal DER investment and scheduling decisions in microgrids. In this work, a novel EV fleet aggregator model is introduced in a stochastic formulation of DER-CAM [1], an optimization tool used to address DER investment and scheduling problems. This is used to assess the impact of EV interconnections on optimal DER solutions considering uncertainty in EV driving schedules. Optimization results indicate that EVs can have a significant impact on DER investments, particularly if considering short payback periods. Furthermore, results suggest that uncertainty in driving schedules carries little significance to total energy costs, which is corroborated by results obtained using the stochastic formulation of the problem.
Chen, B; Hickling, T; Krnjajic, M; Hanley, W; Clark, G; Nitao, J; Knapp, D; Hiller, L; Mugge, M
2007-01-09
In this project, the basic problem is to automatically separate test samples into one of two categories: clean or corrupt. This type of classification problem is known as a two-class classification problem or detection problem. In what follows, we refer to clean examples as negative examples and corrupt examples as positive examples. In a detection problem, a classifier decision on any one sample can be grouped into one of four decision categories: true negative, true positive, false negative and false positive. These four categories are illustrated by Table 1. True negatives and true positives are cases where the classifier has made the correct decision. False positives are cases where the classifier decides positive when the true nature of the sample was negative, and false negatives are cases where the classifier decides negative when the sample was actually positive. To evaluate the performance of a classifier, we run the classifier on all the samples of a data set and then count all the instances of true negatives, true positives, false negatives, and false positives. All of the performance metrics in this report are then formed from a combination of these four basic decision categories.
Improving Resource Selection and Scheduling Using Predictions. Chapter 1
NASA Technical Reports Server (NTRS)
Smith, Warren
2003-01-01
The introduction of computational grids has resulted in several new problems in the area of scheduling that can be addressed using predictions. The first problem is selecting where to run an application on the many resources available in a grid. Our approach to help address this problem is to provide predictions of when an application would start to execute if submitted to specific scheduled computer systems. The second problem is gaining simultaneous access to multiple computer systems so that distributed applications can be executed. We help address this problem by investigating how to support advance reservations in local scheduling systems. Our approaches to both of these problems are based on predictions for the execution time of applications on space- shared parallel computers. As a side effect of this work, we also discuss how predictions of application run times can be used to improve scheduling performance.
Performance Analysis of Selective Breeding Algorithm on One Dimensional Bin Packing Problems
NASA Astrophysics Data System (ADS)
Sriramya, P.; Parvathavarthini, B.
2012-12-01
The bin packing optimization problem packs a set of objects into a set of bins so that the amount of wasted space is minimized. The bin packing problem has many important applications. The objective is to find a feasible assignment of all weights to bins that minimizes the total number of bins used. The bin packing problem models several practical problems in such diverse areas as industrial control, computer systems, machine scheduling, VLSI chip layout and etc. Selective breeding algorithm (SBA) is an iterative procedure which borrows the ideas of artificial selection and breeding process. By simulating artificial evolution in this way SBA algorithm can easily solve complex problems. One dimensional bin packing benchmark problems are taken for evaluating the performance of the SBA. The computational results of SBA algorithm show optimal solution for the tested benchmark problems. The proposed SBA algorithm is a good problem-solving technique for one dimensional bin packing problems.
49 CFR 214.531 - Schedule of repairs; general.
Code of Federal Regulations, 2011 CFR
2011-10-01
... Hi-Rail Vehicles § 214.531 Schedule of repairs; general. Except as provided in §§ 214.527(c)(5), 214.529, and 214.533, an on-track roadway maintenance machine or hi-rail vehicle that does not meet all... or hi-rail vehicle shall be placed out of on-track service....
49 CFR 214.531 - Schedule of repairs; general.
Code of Federal Regulations, 2014 CFR
2014-10-01
... Hi-Rail Vehicles § 214.531 Schedule of repairs; general. Except as provided in §§ 214.527(c)(5), 214.529, and 214.533, an on-track roadway maintenance machine or hi-rail vehicle that does not meet all... or hi-rail vehicle shall be placed out of on-track service....
49 CFR 214.531 - Schedule of repairs; general.
Code of Federal Regulations, 2010 CFR
2010-10-01
... Hi-Rail Vehicles § 214.531 Schedule of repairs; general. Except as provided in §§ 214.527(c)(5), 214.529, and 214.533, an on-track roadway maintenance machine or hi-rail vehicle that does not meet all... or hi-rail vehicle shall be placed out of on-track service....
49 CFR 214.531 - Schedule of repairs; general.
Code of Federal Regulations, 2012 CFR
2012-10-01
... Hi-Rail Vehicles § 214.531 Schedule of repairs; general. Except as provided in §§ 214.527(c)(5), 214.529, and 214.533, an on-track roadway maintenance machine or hi-rail vehicle that does not meet all... or hi-rail vehicle shall be placed out of on-track service....
Optimal scheduling of logistical support for an emergency roadway repair work schedule
NASA Astrophysics Data System (ADS)
Yan, S.; Lin, C. K.; Chen, S. Y.
2012-09-01
The completion of every disaster rescue task performed by repair work teams relies on the in-time supply of materials to the rescue workers. Up to now, logistical support planning for emergency repair work in Taiwan has been done manually, which is neither effective nor efficient. To remedy the problem, this study presents a logistical support scheduling model for the given emergency repair work schedule. The objective is to minimize the short-term operating cost subject to time constraints and other related operating constraints. This model is formulated as an integer multiple-commodity network flow problem which is characterized as NP-hard. A heuristic algorithm, based on the problem decomposition and variable fixing techniques, is also proposed to efficiently solve this problem. Computational tests are performed using data from Taiwan's 1999 Chi-Chi earthquake. The results show that the model and the solution algorithm would be useful for the logistical support scheduling.
S-CHART - SCHEDULING CHART PROGRAM
NASA Technical Reports Server (NTRS)
Klinkner, E. R.
1994-01-01
The Scheduling Chart Program (S-Chart) uses simple menu choices to produce high quality Gantt type scheduling and production time-line charts with minimal data entry requirements. The software produces high quality charts (20, 40, and 80 day options) in 10-20 minutes including start up, data entry, and printing. Reprints take less than one minute. Some of the more significant features are as follows: speed and ease of use with menu driven selections from start to finish, storage and catalogs of all data sets for rapid access and manipulation, compact program size to permit storage of dozens of data sets on the program disc, and creation of ASCII text files of graphs for rapid printing. The final product of the plot routine is an extended ASCII character file which shows the events scheduled and actual work dates. This file may be printed from the program, from DOS, or imported to a word processor for customizing. S-Chart is not intended to take the place of commercial scheduling/project management software. However, it is a simple, somewhat limited Gantt chart plotter and scheduling tracking program with an emphasis on straightforward data entry. There is no complex relationship building between the scheduled events. Dates are entered as dates, not as algebraic functions. All the files of S-Chart, except for the graph picture and the stand-alone executable, were written in 100% dBASE III compatible code. The executable code was created with CLIPPER and the graph picture files are written in ASCII code. S-Chart was implemented on an IBM PC series machine under DOS and requires 215k bytes of memory. The program was developed in 1988.
Scheduling Operations for Massive Heterogeneous Clusters
NASA Technical Reports Server (NTRS)
Humphrey, John; Spagnoli, Kyle
2013-01-01
High-performance computing (HPC) programming has become increasingly difficult with the advent of hybrid supercomputers consisting of multicore CPUs and accelerator boards such as the GPU. Manual tuning of software to achieve high performance on this type of machine has been performed by programmers. This is needlessly difficult and prone to being invalidated by new hardware, new software, or changes in the underlying code. A system was developed for task-based representation of programs, which when coupled with a scheduler and runtime system, allows for many benefits, including higher performance and utilization of computational resources, easier programming and porting, and adaptations of code during runtime. The system consists of a method of representing computer algorithms as a series of data-dependent tasks. The series forms a graph, which can be scheduled for execution on many nodes of a supercomputer efficiently by a computer algorithm. The schedule is executed by a dispatch component, which is tailored to understand all of the hardware types that may be available within the system. The scheduler is informed by a cluster mapping tool, which generates a topology of available resources and their strengths and communication costs. Software is decoupled from its hardware, which aids in porting to future architectures. A computer algorithm schedules all operations, which for systems of high complexity (i.e., most NASA codes), cannot be performed optimally by a human. The system aids in reducing repetitive code, such as communication code, and aids in the reduction of redundant code across projects. It adds new features to code automatically, such as recovering from a lost node or the ability to modify the code while running. In this project, the innovators at the time of this reporting intend to develop two distinct technologies that build upon each other and both of which serve as building blocks for more efficient HPC usage. First is the scheduling and dynamic
Reinforcement schedule thinning following treatment with functional communication training.
Hanley, G P; Iwata, B A; Thompson, R H
2001-01-01
We evaluated four methods for increasing the practicality of functional communication training (FCT) by decreasing the frequency of reinforcement for alternative behavior. Three participants whose problem behaviors were maintained by positive reinforcement were treated successfully with FCT in which reinforcement for alternative behavior was initially delivered on fixed-ratio (FR) 1 schedules. One participant was then exposed to increasing delays to reinforcement under FR 1, a graduated fixed-interval (FI) schedule, and a graduated multiple-schedule arrangement in which signaled periods of reinforcement and extinction were alternated. Results showed that (a) increasing delays resulted in extinction of the alternative behavior, (b) the FI schedule produced undesirably high rates of the alternative behavior, and (c) the multiple schedule resulted in moderate and stable levels of the alternative behavior as the duration of the extinction component was increased. The other 2 participants were exposed to graduated mixed-schedule (unsignaled alternation between reinforcement and extinction components) and multiple-schedule (signaled alternation between reinforcement and extinction components) arrangements in which the durations of the reinforcement and extinction components were modified. Results obtained for these 2 participants indicated that the use of discriminative stimuli in the multiple schedule facilitated reinforcement schedule thinning. Upon completion of treatment, problem behavior remained low (or at zero), whereas alternative behavior was maintained as well as differentiated during a multiple-schedule arrangement consisting of a 4-min extinction period followed by a 1-min reinforcement period. PMID:11317985
Scheduling in the Face of Uncertain Resource Consumption and Utility
NASA Technical Reports Server (NTRS)
Frank, Jeremy; Dearden, Richard
2003-01-01
We discuss the problem of scheduling tasks that consume uncertain amounts of a resource with known capacity and where the tasks have uncertain utility. In these circumstances, we would like to find schedules that exceed a lower bound on the expected utility when executed. We show that the problems are NP- complete, and present some results that characterize the behavior of some simple heuristics over a variety of problem classes.
Features of plastics edge cutting machining
NASA Astrophysics Data System (ADS)
Handozhko, A. V.; Shcherbakov, A. N.; Zaharov, L. A.; Gavrilenko, T. V.
2016-04-01
This article describes the features of pieces from thermoplastic materials in the form of electrical insulators cut by a disk edge tool. The problems in question are possible defects arising during machining and technological conditions that reduce their quantity. The necessity of required machining conditions matching substantiated in accordance with a specific grade of the material which is treated. Equipment and machining attachments, developed for experimental studies, determine the rational conditions of plastic electrical insulators machining. As a result of experiments the dependences of cut face quality parameters of plastics are obtained by machining conditions. The obtained results allowed us to make valid conclusions and recommendations.
Steps Toward Optimal Competitive Scheduling
NASA Technical Reports Server (NTRS)
Frank, Jeremy; Crawford, James; Khatib, Lina; Brafman, Ronen
2006-01-01
This paper is concerned with the problem of allocating a unit capacity resource to multiple users within a pre-defined time period. The resource is indivisible, so that at most one user can use it at each time instance. However, different users may use it at different times. The users have independent, se@sh preferences for when and for how long they are allocated this resource. Thus, they value different resource access durations differently, and they value different time slots differently. We seek an optimal allocation schedule for this resource. This problem arises in many institutional settings where, e.g., different departments, agencies, or personal, compete for a single resource. We are particularly motivated by the problem of scheduling NASA's Deep Space Satellite Network (DSN) among different users within NASA. Access to DSN is needed for transmitting data from various space missions to Earth. Each mission has different needs for DSN time, depending on satellite and planetary orbits. Typically, the DSN is over-subscribed, in that not all missions will be allocated as much time as they want. This leads to various inefficiencies - missions spend much time and resource lobbying for their time, often exaggerating their needs. NASA, on the other hand, would like to make optimal use of this resource, ensuring that the good for NASA is maximized. This raises the thorny problem of how to measure the utility to NASA of each allocation. In the typical case, it is difficult for the central agency, NASA in our case, to assess the value of each interval to each user - this is really only known to the users who understand their needs. Thus, our problem is more precisely formulated as follows: find an allocation schedule for the resource that maximizes the sum of users preferences, when the preference values are private information of the users. We bypass this problem by making the assumptions that one can assign money to customers. This assumption is reasonable; a
A space station onboard scheduling assistant
NASA Technical Reports Server (NTRS)
Brindle, A. F.; Anderson, B. H.
1988-01-01
One of the goals for the Space Station is to achieve greater autonomy, and have less reliance on ground commanding than previous space missions. This means that the crew will have to take an active role in scheduling and rescheduling their activities onboard, perhaps working from preliminary schedules generated on the ground. Scheduling is a time intensive task, whether performed manually or automatically, so the best approach to solving onboard scheduling problems may involve crew members working with an interactive software scheduling package. A project is described which investigates a system that uses knowledge based techniques for the rescheduling of experiments within the Materials Technology Laboratory of the Space Station. Particular attention is paid to: (1) methods for rapid response rescheduling to accommodate unplanned changes in resource availability, (2) the nature of the interface to the crew, (3) the representation of the many types of data within the knowledge base, and (4) the possibility of applying rule-based and constraint-based reasoning methods to onboard activity scheduling.
Deo, Rahul C
2015-11-17
Spurred by advances in processing power, memory, storage, and an unprecedented wealth of data, computers are being asked to tackle increasingly complex learning tasks, often with astonishing success. Computers have now mastered a popular variant of poker, learned the laws of physics from experimental data, and become experts in video games - tasks that would have been deemed impossible not too long ago. In parallel, the number of companies centered on applying complex data analysis to varying industries has exploded, and it is thus unsurprising that some analytic companies are turning attention to problems in health care. The purpose of this review is to explore what problems in medicine might benefit from such learning approaches and use examples from the literature to introduce basic concepts in machine learning. It is important to note that seemingly large enough medical data sets and adequate learning algorithms have been available for many decades, and yet, although there are thousands of papers applying machine learning algorithms to medical data, very few have contributed meaningfully to clinical care. This lack of impact stands in stark contrast to the enormous relevance of machine learning to many other industries. Thus, part of my effort will be to identify what obstacles there may be to changing the practice of medicine through statistical learning approaches, and discuss how these might be overcome. PMID:26572668
Rohwein, G.J.; Lancaster, K.T.; Lawson, R.N.
1986-06-01
TEMPO is a transformer powered megavolt pulse generator with an output pulse of 100 ns duration. The machine was designed for burst mode operation at pulse repetition rates up to 10 Hz with minimum pulse-to-pulse voltage variations. To meet the requirement for pulse duration a nd a 20-..omega.. output impedance within reasonable size constraints, the pulse forming transmission line was designed as two parallel water-insulated, strip-type Blumleins. Stray capacitance and electric fields along the edges of the line elements were controlled by lining the tank with plastic sheet.
De Bock, Hendrik Pieter Jacobus; Alexander, James Pellegrino; El-Refaie, Ayman Mohamed Fawzi; Gerstler, William Dwight; Shah, Manoj Ramprasad; Shen, Xiaochun
2016-06-21
An apparatus, such as an electrical machine, is provided. The apparatus can include a rotor defining a rotor bore and a conduit disposed in and extending axially along the rotor bore. The conduit can have an annular conduit body defining a plurality of orifices disposed axially along the conduit and extending through the conduit body. The rotor can have an inner wall that at least partially defines the rotor bore. The orifices can extend through the conduit body along respective orifice directions, and the rotor and conduit can be configured to provide a line of sight along the orifice direction from the respective orifices to the inner wall.
The Ames-Lockheed orbiter processing scheduling system
NASA Technical Reports Server (NTRS)
Zweben, Monte; Gargan, Robert
1991-01-01
A general purpose scheduling system and its application to Space Shuttle Orbiter Processing at the Kennedy Space Center (KSC) are described. Orbiter processing entails all the inspection, testing, repair, and maintenance necessary to prepare the Shuttle for launch and takes place within the Orbiter Processing Facility (OPF) at KSC, the Vehicle Assembly Building (VAB), and on the launch pad. The problems are extremely combinatoric in that there are thousands of tasks, resources, and other temporal considerations that must be coordinated. Researchers are building a scheduling tool that they hope will be an integral part of automating the planning and scheduling process at KSC. The scheduling engine is domain independent and is also being applied to Space Shuttle cargo processing problems as well as wind tunnel scheduling problems.
Checking the anaesthetic machine.
Wicker, Paul; Smith, Brian
2006-12-01
Making sure that anaesthetic equipment is working correctly is an essential part of the anaesthetic practitioner's role. During induction of anaesthesia the patient is at one of the most vulnerable points in his or her perioperative care. This is the point at which equipment error may put the patient at high risk of harm, for example, through compromising the airway, causing circulatory problems, preventing satisfactory oxygenation or even causing death. Many writers have drawn comparisons between anaesthesia and aviation, with the suggestion that practitioners should check the anaesthetic machines using a 'cockpit drill' (Ranasinghe 2000). The purpose of this detailed check is to ensure the machine is safe to use. The careful attention to the check is a reflection of good practice which the practitioner's codes of professional practice demand (HPC 2004, NMC 2004). This article discusses the importance of following the anaesthetic checklist to the recommended standards for both the practitioner and the patient. PMID:17193997
NASA Astrophysics Data System (ADS)
Freund, Richard F.; Braun, Tracy D.; Kussow, Matthew; Godfrey, Michael; Koyama, Terry
2001-07-01
SPANR (Schedule, Plan, Assess Networked Resources) is (i) a pre-run, off-line planning and (ii) a runtime, just-in-time scheduling mechanism. It is designed to support primarily commercial applications in that it optimizes throughput rather than individual jobs (unless they have highest priority). Thus it is a tool for a commercial production manager to maximize total work. First the SPANR Planner is presented showing the ability to do predictive 'what-if' planning. It can answer such questions as, (i) what is the overall effect of acquiring new hardware or (ii) what would be the effect of a different scheduler. The ability of the SPANR Planner to formulate in advance tree-trimming strategies is useful in several commercial applications, such as electronic design or pharmaceutical simulations. The SPANR Planner is demonstrated using a variety of benchmarks. The SPANR Runtime Scheduler (RS) is briefly presented. The SPANR RS can provide benefit for several commercial applications, such as airframe design and financial applications. Finally a design is shown whereby SPANR can provide scheduling advice to most resource management systems.
Snyder, L.L.
1980-02-19
A diametrically compact tunneling machine for boring tunnels is disclosed. The machine includes a tubular support frame having a hollow piston mounted therein which is movable from a retracted position in the support frame to an extended position. A drive shaft is rotatably mounted in the hollow piston and carries a cutter head at one end. The hollow piston is restrained against rotational movement relative to the support frame and the drive shaft is constrained against longitudinal movement relative to the hollow piston. A plurality of radially extendible feet project from the support frame to the tunnel wall to grip the tunnel wall during a tunneling operation wherein the hollow piston is driven forwardly so that the cutter head works on the tunnel face. When the hollow piston is fully extended, a plurality of extendible support feet, which are fixed to the rearward and forward ends of the hollow piston, are extended, the radially extendible feet are retracted and the support frame is shifted forwardly by the piston so that a further tunneling operation may be initiated.
Schedule-Tracker Computer Program
NASA Technical Reports Server (NTRS)
Collazo, Fernando F.
1990-01-01
Schedule Tracker provides effective method for tracking tasks "past due" and/or "near term". Generates reports for each responsible staff member having one or more assigned tasks falling within two listed categories. Schedule Organizer (SO) (COSMIC program MSC-21525), Schedule Tracker (ST), and Schedule Report Generator (SRG) (COSMIC program MSC-21527) computer programs manipulating data-base files in ways advantageous in scheduling. Written in PL/1 and DEC Command Language (DCL).
NASA Astrophysics Data System (ADS)
Ekberg, Peter; Stiblert, Lars; Mattsson, Lars
2014-05-01
The manufacturing of flat panel displays requires a number of photomasks for the placement of pixel patterns and supporting transistor arrays. For large area photomasks, dedicated ultra-precision writers have been developed for the production of these chromium patterns on glass or quartz plates. The dimensional tolerances in X and Y for absolute pattern placement on these plates, with areas measured in square meters, are in the range of 200-300 nm (3σ). To verify these photomasks, 2D ultra-precision coordinate measurement machines are used having even tighter tolerance requirements. This paper will present how the world standard metrology tool used for verifying large masks, the Micronic Mydata MMS15000, is calibrated without any other references than the wavelength of the interferometers in an extremely well-controlled temperature environment. This process is called self-calibration and is the only way to calibrate the metrology tool, as no square-meter-sized large area 2D traceable artifact is available. The only parameter that cannot be found using self-calibration is the absolute length scale. To make the MMS15000 traceable, a 1D reference rod, calibrated at a national metrology lab, is used. The reference plates used in the calibration of the MMS15000 may have sizes up to 1 m2 and a weight of 50 kg. Therefore, standard methods for self-calibration on a small scale with exact placements cannot be used in the large area case. A new, more general method had to be developed for the purpose of calibrating the MMS15000. Using this method, it is possible to calibrate the measurement tool down to an uncertainty level of <90 nm (3σ) over an area of (0.8 × 0.8) m2. The method used, which is based on the concept of iteration, does not introduce any more noise than the random noise introduced by the measurements, resulting in the lowest possible noise level that can be achieved by any self-calibration method.
A software tool for dataflow graph scheduling
NASA Technical Reports Server (NTRS)
Jones, Robert L., III
1994-01-01
A graph-theoretic design process and software tool is presented for selecting a multiprocessing scheduling solution for a class of computational problems. The problems of interest are those that can be described using a dataflow graph and are intended to be executed repetitively on multiple processors. The dataflow paradigm is very useful in exposing the parallelism inherent in algorithms. It provides a graphical and mathematical model which describes a partial ordering of algorithm tasks based on data precedence.
Long term fuel scheduling linear programming
Asgarpoor, S. . Dept. of Electrical Engineering); Gul, N. )
1992-01-01
This paper presents an application of linear programming (LP) revised simplex method in order to solve the fuel scheduling problem. A regression method is applied to determine the polynomial cost curves, and a separable programming technique is used to linearize the objective function and the constraints for LP application. Results based on sample data obtained from Omaha Public Power District (OPPD) are presented to demonstrate the LP application to this problem.
Parallel machines: Parallel machine languages
Iannucci, R.A. )
1990-01-01
This book presents a framework for understanding the tradeoffs between the conventional view and the dataflow view with the objective of discovering the critical hardware structures which must be present in any scalable, general-purpose parallel computer to effectively tolerate latency and synchronization costs. The author presents an approach to scalable general purpose parallel computation. Linguistic Concerns, Compiling Issues, Intermediate Language Issues, and hardware/technological constraints are presented as a combined approach to architectural Develoement. This book presents the notion of a parallel machine language.
APGEN Scheduling: 15 Years of Experience in Planning Automation
NASA Technical Reports Server (NTRS)
Maldague, Pierre F.; Wissler, Steve; Lenda, Matthew; Finnerty, Daniel
2014-01-01
In this paper, we discuss the scheduling capability of APGEN (Activity Plan Generator), a multi-mission planning application that is part of the NASA AMMOS (Advanced Multi- Mission Operations System), and how APGEN scheduling evolved over its applications to specific Space Missions. Our analysis identifies two major reasons for the successful application of APGEN scheduling to real problems: an expressive DSL (Domain-Specific Language) for formulating scheduling algorithms, and a well-defined process for enlisting the help of auxiliary modeling tools in providing high-fidelity, system-level simulations of the combined spacecraft and ground support system.
A harmonized immunization schedule for Canada: A call to action
MacDonald, NE; Bortolussi, R
2011-01-01
In Canada, the National Advisory Committee on Immunization systematically reviews the evidence for the effectiveness and safety of new and old vaccines, and sets a ‘minimum’ recommended schedule. However, in contrast to other industrialized countries where single, harmonized countrywide immunization schedules are de rigeur, Canada has a confusing system, with each province and territory defining its own schedule – and none are the same. The time has come to rectify this decades-old patient equity and safety problem. The Canadian Paediatric Society calls for a harmonized schedule to improve the health and safety of Canadian children and youth. PMID:22211070
Bishop, Christopher M.
2013-01-01
Several decades of research in the field of machine learning have resulted in a multitude of different algorithms for solving a broad range of problems. To tackle a new application, a researcher typically tries to map their problem onto one of these existing methods, often influenced by their familiarity with specific algorithms and by the availability of corresponding software implementations. In this study, we describe an alternative methodology for applying machine learning, in which a bespoke solution is formulated for each new application. The solution is expressed through a compact modelling language, and the corresponding custom machine learning code is then generated automatically. This model-based approach offers several major advantages, including the opportunity to create highly tailored models for specific scenarios, as well as rapid prototyping and comparison of a range of alternative models. Furthermore, newcomers to the field of machine learning do not have to learn about the huge range of traditional methods, but instead can focus their attention on understanding a single modelling environment. In this study, we show how probabilistic graphical models, coupled with efficient inference algorithms, provide a very flexible foundation for model-based machine learning, and we outline a large-scale commercial application of this framework involving tens of millions of users. We also describe the concept of probabilistic programming as a powerful software environment for model-based machine learning, and we discuss a specific probabilistic programming language called Infer.NET, which has been widely used in practical applications. PMID:23277612
Gaussian processes for machine learning.
Seeger, Matthias
2004-04-01
Gaussian processes (GPs) are natural generalisations of multivariate Gaussian random variables to infinite (countably or continuous) index sets. GPs have been applied in a large number of fields to a diverse range of ends, and very many deep theoretical analyses of various properties are available. This paper gives an introduction to Gaussian processes on a fairly elementary level with special emphasis on characteristics relevant in machine learning. It draws explicit connections to branches such as spline smoothing models and support vector machines in which similar ideas have been investigated. Gaussian process models are routinely used to solve hard machine learning problems. They are attractive because of their flexible non-parametric nature and computational simplicity. Treated within a Bayesian framework, very powerful statistical methods can be implemented which offer valid estimates of uncertainties in our predictions and generic model selection procedures cast as nonlinear optimization problems. Their main drawback of heavy computational scaling has recently been alleviated by the introduction of generic sparse approximations.13,78,31 The mathematical literature on GPs is large and often uses deep concepts which are not required to fully understand most machine learning applications. In this tutorial paper, we aim to present characteristics of GPs relevant to machine learning and to show up precise connections to other "kernel machines" popular in the community. Our focus is on a simple presentation, but references to more detailed sources are provided. PMID:15112367
Bishop, Christopher M
2013-02-13
Several decades of research in the field of machine learning have resulted in a multitude of different algorithms for solving a broad range of problems. To tackle a new application, a researcher typically tries to map their problem onto one of these existing methods, often influenced by their familiarity with specific algorithms and by the availability of corresponding software implementations. In this study, we describe an alternative methodology for applying machine learning, in which a bespoke solution is formulated for each new application. The solution is expressed through a compact modelling language, and the corresponding custom machine learning code is then generated automatically. This model-based approach offers several major advantages, including the opportunity to create highly tailored models for specific scenarios, as well as rapid prototyping and comparison of a range of alternative models. Furthermore, newcomers to the field of machine learning do not have to learn about the huge range of traditional methods, but instead can focus their attention on understanding a single modelling environment. In this study, we show how probabilistic graphical models, coupled with efficient inference algorithms, provide a very flexible foundation for model-based machine learning, and we outline a large-scale commercial application of this framework involving tens of millions of users. We also describe the concept of probabilistic programming as a powerful software environment for model-based machine learning, and we discuss a specific probabilistic programming language called Infer.NET, which has been widely used in practical applications. PMID:23277612
Distributed project scheduling at NASA: Requirements for manual protocols and computer-based support
NASA Technical Reports Server (NTRS)
Richards, Stephen F.
1992-01-01
The increasing complexity of space operations and the inclusion of interorganizational and international groups in the planning and control of space missions lead to requirements for greater communication, coordination, and cooperation among mission schedulers. These schedulers must jointly allocate scarce shared resources among the various operational and mission oriented activities while adhering to all constraints. This scheduling environment is complicated by such factors as the presence of varying perspectives and conflicting objectives among the schedulers, the need for different schedulers to work in parallel, and limited communication among schedulers. Smooth interaction among schedulers requires the use of protocols that govern such issues as resource sharing, authority to update the schedule, and communication of updates. This paper addresses the development and characteristics of such protocols and their use in a distributed scheduling environment that incorporates computer-aided scheduling tools. An example problem is drawn from the domain of Space Shuttle mission planning.
RSM 1.0 - A RESUPPLY SCHEDULER USING INTEGER OPTIMIZATION
NASA Technical Reports Server (NTRS)
Viterna, L. A.
1994-01-01
RSM, Resupply Scheduling Modeler, is a fully menu-driven program that uses integer programming techniques to determine an optimum schedule for replacing components on or before the end of a fixed replacement period. Although written to analyze the electrical power system on the Space Station Freedom, RSM is quite general and can be used to model the resupply of almost any system subject to user-defined resource constraints. RSM is based on a specific form of the general linear programming problem in which all variables in the objective function and all variables in the constraints are integers. While more computationally intensive, integer programming was required for accuracy when modeling systems with small quantities of components. Input values for component life cane be real numbers, RSM converts them to integers by dividing the lifetime by the period duration, then reducing the result to the next lowest integer. For each component, there is a set of constraints that insure that it is replaced before its lifetime expires. RSM includes user-defined constraints such as transportation mass and volume limits, as well as component life, available repair crew time and assembly sequences. A weighting factor allows the program to minimize factors such as cost. The program then performs an iterative analysis, which is displayed during the processing. A message gives the first period in which resources are being exceeded on each iteration. If the scheduling problem is unfeasible, the final message will also indicate the first period in which resources were exceeded. RSM is written in APL2 for IBM PC series computers and compatibles. A stand-alone executable version of RSM is provided; however, this is a "packed" version of RSM which can only utilize the memory within the 640K DOS limit. This executable requires at least 640K of memory and DOS 3.1 or higher. Source code for an APL2/PC workspace version is also provided. This version of RSM can make full use of any
NASA Astrophysics Data System (ADS)
Wainger, Lisa A.
The new schedule for Space Shuttle missions and expendable launch vehicles (ELV's) calls for a 7-month delay in sending up the Hubble Space Telescope. NASA was forced to put off launching the telescope until February 1990 to keep the Magellan and Galileo missions within their narrow launch windows. The first post-Challenger shuttle launch is now scheduled for late this month. Discovery's most recent delays were due to a hydrogen leak discovered July 29 that has still not been corrected and an engine valve malfunction during an August 4 test fire.
Intelligent retail logistics scheduling
Rowe, J.; Jewers, K.; Codd, A.; Alcock, A.
1996-12-31
The Supply Chain Integrated Ordering Network (SCION) Depot Bookings system automates the planning and scheduling of perishable and non-perishable commodities and the vehicles that carry them into J. Sainsbury depots. This is a strategic initiative, enabling the business to make the key move from weekly to daily ordering. The system is mission critical, managing the inwards flow of commodities from suppliers into J. Sainsbury`s depots. The system leverages Al techniques to provide a business solution that meets challenging functional and performance needs. The SCION Depot Bookings system is operational providing schedules for 22 depots across the UK.
Multiple daily fractionation schedules
Peschel, R.E.; Fischer, J.J.
1982-10-01
Although conventional fractionation schedules have been satisfactory for the treatment of some tumors, there is reason to believe that the results of radiation therapy could be improved in some cases by appropriate alterations in treatment schedules. The pharmacological characteristics of some of the electron affinic radiation sensitizers have provided added incentive to investigate newer fractionation schemes, particularly ones which deliver the majority of the radiation dose in short periods of time. This editorial discusses three papers describing preliminary clinical studies using multi-daily fractionated (MDF) radiation therapy. Two of these studies also make use of the radiation sensitizer misonidazole. (KRM)
Heuristics for scheduling Earth observing satellites
NASA Astrophysics Data System (ADS)
Wolfe, William J.; Sorensen, Stephen E.
1999-09-01
This paper describes several methods for assigning tasks to Earth Observing Systems Satellites (EOS). We present empirical results for three heuristics, called: Priority Dispatch (PD), Look Ahead (LA), and Genetic Algorithm (GA). These heuristics progress from simple to complex, from less accurate to more accurate, and from fast to slow. We present empirical results as applied to the Window-Constrained Packing problem (WCP). The WCP is a simplified version of the EOS scheduling problem. We discuss the problem of having more than one optimization criteria. We will also discuss the relationship between the WCP and the more traditional Knapsack and Weighted Early/Tardy problems.
Automatic Generation of Heuristics for Scheduling
NASA Technical Reports Server (NTRS)
Morris, Robert A.; Bresina, John L.; Rodgers, Stuart M.
1997-01-01
This paper presents a technique, called GenH, that automatically generates search heuristics for scheduling problems. The impetus for developing this technique is the growing consensus that heuristics encode advice that is, at best, useful in solving most, or typical, problem instances, and, at worst, useful in solving only a narrowly defined set of instances. In either case, heuristic problem solvers, to be broadly applicable, should have a means of automatically adjusting to the idiosyncrasies of each problem instance. GenH generates a search heuristic for a given problem instance by hill-climbing in the space of possible multi-attribute heuristics, where the evaluation of a candidate heuristic is based on the quality of the solution found under its guidance. We present empirical results obtained by applying GenH to the real world problem of telescope observation scheduling. These results demonstrate that GenH is a simple and effective way of improving the performance of an heuristic scheduler.
Advanced coordinate measuring machine at Sandia National Laboratories/California
Pilkey, R.D.; Klevgard, P.A.
1993-03-01
Sandia National Laboratories/California has acquired a new Moore M-48V CNC five-axis universal coordinate measuring machine (CMM). Site preparation, acceptance testing, and initial performance results are discussed. Unique features of the machine include a ceramic ram and vacuum evacuated laser pathways (VELPS). The implementation of a VELPS system on the machine imposed certain design requirements and entailed certain start-up problems. The machine`s projected capabilities, workload, and research possibilities are outlined.
Improving Energy Efficiency in CNC Machining
NASA Astrophysics Data System (ADS)
Pavanaskar, Sushrut S.
We present our work on analyzing and improving the energy efficiency of multi-axis CNC milling process. Due to the differences in energy consumption behavior, we treat 3- and 5-axis CNC machines separately in our work. For 3-axis CNC machines, we first propose an energy model that estimates the energy requirement for machining a component on a specified 3-axis CNC milling machine. Our model makes machine-specific predictions of energy requirements while also considering the geometric aspects of the machining toolpath. Our model - and the associated software tool - facilitate direct comparison of various alternative toolpath strategies based on their energy-consumption performance. Further, we identify key factors in toolpath planning that affect energy consumption in CNC machining. We then use this knowledge to propose and demonstrate a novel toolpath planning strategy that may be used to generate new toolpaths that are inherently energy-efficient, inspired by research on digital micrography -- a form of computational art. For 5-axis CNC machines, the process planning problem consists of several sub-problems that researchers have traditionally solved separately to obtain an approximate solution. After illustrating the need to solve all sub-problems simultaneously for a truly optimal solution, we propose a unified formulation based on configuration space theory. We apply our formulation to solve a problem variant that retains key characteristics of the full problem but has lower dimensionality, allowing visualization in 2D. Given the complexity of the full 5-axis toolpath planning problem, our unified formulation represents an important step towards obtaining a truly optimal solution. With this work on the two types of CNC machines, we demonstrate that without changing the current infrastructure or business practices, machine-specific, geometry-based, customized toolpath planning can save energy in CNC machining.
Autonomous scheduling technology for Earth orbital missions
NASA Technical Reports Server (NTRS)
Srivastava, S.
1982-01-01
The development of a dynamic autonomous system (DYASS) of resources for the mission support of near-Earth NASA spacecraft is discussed and the current NASA space data system is described from a functional perspective. The future (late 80's and early 90's) NASA space data system is discussed. The DYASS concept, the autonomous process control, and the NASA space data system are introduced. Scheduling and related disciplines are surveyed. DYASS as a scheduling problem is also discussed. Artificial intelligence and knowledge representation is considered as well as the NUDGE system and the I-Space system.
Planning, scheduling, and control for automatic telescopes
NASA Technical Reports Server (NTRS)
Drummond, Mark; Swanson, Keith; Philips, Andy; Levinson, Rich; Bresina, John
1992-01-01
This paper presents an argument for the appropriateness of Entropy Reduction Engine (ERE) technology to the planning, scheduling, and control components of Automatic Photoelectric Telescope (APT) management. The paper is organized as follows. In the next section, we give a brief summary of the planning and scheduling requirements for APTs. Following this, in section 3, we give an ERE project precis, couched primarily in terms of project objectives. Section 4 gives a sketch of the match-up between problem and technology, and section 5 outlines where we want to go with this work.
The finite element machine: An experiment in parallel processing
NASA Technical Reports Server (NTRS)
Storaasli, O. O.; Peebles, S. W.; Crockett, T. W.; Knott, J. D.; Adams, L.
1982-01-01
The finite element machine is a prototype computer designed to support parallel solutions to structural analysis problems. The hardware architecture and support software for the machine, initial solution algorithms and test applications, and preliminary results are described.
Scheduling Real-Time Mixed-Criticality Jobs
NASA Astrophysics Data System (ADS)
Baruah, Sanjoy K.; Bonifaci, Vincenzo; D'Angelo, Gianlorenzo; Li, Haohan; Marchetti-Spaccamela, Alberto; Megow, Nicole; Stougie, Leen
Many safety-critical embedded systems are subject to certification requirements; some systems may be required to meet multiple sets of certification requirements, from different certification authorities. Certification requirements in such "mixed-criticality" systems give rise to interesting scheduling problems, that cannot be satisfactorily addressed using techniques from conventional scheduling theory. In this paper, we study a formal model for representing such mixed-criticality workloads. We demonstrate first the intractability of determining whether a system specified in this model can be scheduled to meet all its certification requirements, even for systems subject to two sets of certification requirements. Then we quantify, via the metric of processor speedup factor, the effectiveness of two techniques, reservation-based scheduling and priority-based scheduling, that are widely used in scheduling such mixed-criticality systems, showing that the latter of the two is superior to the former. We also show that the speedup factors are tight for these two techniques.
Machine speech and speaking about machines
Nye, A.
1996-12-31
Current philosophy of language prides itself on scientific status. It boasts of being no longer contaminated with queer mental entities or idealist essences. It theorizes language as programmable variants of formal semantic systems, reimaginable either as the properly epiphenomenal machine functions of computer science or the properly material neural networks of physiology. Whether or not such models properly capture the physical workings of a living human brain is a question that scientists will have to answer. I, as a philosopher, come at the problem from another direction. Does contemporary philosophical semantics, in its dominant truth-theoretic and related versions, capture actual living human thought as it is experienced, or does it instead reflect, regardless of (perhaps dubious) scientific credentials, pathology of thought, a pathology with a disturbing social history.
An Evaluation of Parallel Job Scheduling for ASCI Blue-Pacific
Franke, H.; Jann, J.; Moreira, J.; Pattnaik, P.; Jette, M.
1999-11-09
In this paper we analyze the behavior of a gang-scheduling strategy that we are developing for the ASCI Blue-Pacific machines. Using actual job logs for one of the ASCI machines we generate a statistical model of the current workload with hyper Erlang distributions. We then vary the parameters of those distributions to generate various workloads, representative of different operating points of the machine. Through simulation we obtain performance parameters for three different scheduling strategies: (i) first-come first-serve, (ii) gang-scheduling, and (iii) backfilling. Our results show that backfilling, can be very effective for the common operating points in the 60-70% utilization range. However, for higher utilization rates, time-sharing techniques such as gang-scheduling offer much better performance.
Stacked Extreme Learning Machines.
Zhou, Hongming; Huang, Guang-Bin; Lin, Zhiping; Wang, Han; Soh, Yeng Chai
2015-09-01
Extreme learning machine (ELM) has recently attracted many researchers' interest due to its very fast learning speed, good generalization ability, and ease of implementation. It provides a unified solution that can be used directly to solve regression, binary, and multiclass classification problems. In this paper, we propose a stacked ELMs (S-ELMs) that is specially designed for solving large and complex data problems. The S-ELMs divides a single large ELM network into multiple stacked small ELMs which are serially connected. The S-ELMs can approximate a very large ELM network with small memory requirement. To further improve the testing accuracy on big data problems, the ELM autoencoder can be implemented during each iteration of the S-ELMs algorithm. The simulation results show that the S-ELMs even with random hidden nodes can achieve similar testing accuracy to support vector machine (SVM) while having low memory requirements. With the help of ELM autoencoder, the S-ELMs can achieve much better testing accuracy than SVM and slightly better accuracy than deep belief network (DBN) with much faster training speed. PMID:25361517
Entering the operative correction machining processes CNC
NASA Astrophysics Data System (ADS)
Nekrasov, R. Yu; Starikov, A. I.; Lasukov, A. A.
2015-09-01
The article describes the solution to the problem of compensation of errors occurring during machining on CNC machines. We propose a method of mathematical modeling of processes diagnostics and control of technological equipment. The results of the diagnosis of the CNC machine, as well as the mathematical model describing the dependence of the positioning error of the executive bodies of operating component of cutting force PZ, in the range of movement OX.
CMS multicore scheduling strategy
Perez-Calero Yzquierdo, Antonio; Hernandez, Jose; Holzman, Burt; Majewski, Krista; McCrea, Alison
2014-01-01
In the next years, processor architectures based on much larger numbers of cores will be most likely the model to continue 'Moore's Law' style throughput gains. This not only results in many more jobs in parallel running the LHC Run 1 era monolithic applications, but also the memory requirements of these processes push the workernode architectures to the limit. One solution is parallelizing the application itself, through forking and memory sharing or through threaded frameworks. CMS is following all of these approaches and has a comprehensive strategy to schedule multicore jobs on the GRID based on the glideinWMS submission infrastructure. The main component of the scheduling strategy, a pilot-based model with dynamic partitioning of resources that allows the transition to multicore or whole-node scheduling without disallowing the use of single-core jobs, is described. This contribution also presents the experiences made with the proposed multicore scheduling schema and gives an outlook of further developments working towards the restart of the LHC in 2015.
CMS multicore scheduling strategy
NASA Astrophysics Data System (ADS)
Pérez-Calero Yzquierdo, Antonio; Hernández, Jose; Holzman, Burt; Majewski, Krista; McCrea, Alison; Cms Collaboration
2014-06-01
In the next years, processor architectures based on much larger numbers of cores will be most likely the model to continue "Moore's Law" style throughput gains. This not only results in many more jobs in parallel running the LHC Run 1 era monolithic applications, but also the memory requirements of these processes push the workernode architectures to the limit. One solution is parallelizing the application itself, through forking and memory sharing or through threaded frameworks. CMS is following all of these approaches and has a comprehensive strategy to schedule multicore jobs on the GRID based on the glideinWMS submission infrastructure. The main component of the scheduling strategy, a pilot-based model with dynamic partitioning of resources that allows the transition to multicore or whole-node scheduling without disallowing the use of single-core jobs, is described. This contribution also presents the experiences made with the proposed multicore scheduling schema and gives an outlook of further developments working towards the restart of the LHC in 2015.
ERIC Educational Resources Information Center
Blai, Boris
Many creative or flexible work scheduling options are becoming available to the many working parents, students, handicapped persons, elderly individuals, and others who are either unable or unwilling to work a customary 40-hour work week. These options may be broadly categorized as either restructured or reduced work time options. The three main…
ERIC Educational Resources Information Center
Purdue Univ., Lafayette, IN. Educational Research Center.
This 116-item interview schedule designed for parents who failed to respond to the Questionnaire for Parents, is individually administered to the mother of the child of elementary school age. It consists of scales measuring 14 parent variables plus a section devoted to demographic variables: (1) parent's achievement aspirations for the child, (2)…
ERIC Educational Resources Information Center
Emerson, Eric; Howard, Denise
1992-01-01
The phenomena of the induction and entrainment of adjunctive behaviors was investigated in 8 people (ages 5-51) with severe or profound mental retardation who exhibited stereotypic behaviors. Seven of the eight demonstrated evidence of schedule-induced stereotypic behavior, whereas five also showed evidence of the entrainment of these behaviors by…
Spanning Tree Calculations on D-Wave 2 Machines
NASA Astrophysics Data System (ADS)
Novotny, M. A.; Hobl, Q. L.; Hall, J. S.; Michielsen, K.
2016-02-01
Calculations on D-Wave machines are presented, both for the 500-qubit and the 1000-qubit machines. Results are presented for spanning trees on the available K4,4 Chimera graphs of both machines. Comparing trees of approximately the same size, the frequency of finding the ground state for the 1000-qubit machine is significantly improved over the 500- qubit older generation machine. Spanning trees are difficult problems for solution by adiabatic quantum computers, so the enhanced frequency of finding the ground state for newer machine generations and larger machines is encouraging for this immature technology.
Planning and Scheduling for Fleets of Earth Observing Satellites
NASA Technical Reports Server (NTRS)
Frank, Jeremy; Jonsson, Ari; Morris, Robert; Smith, David E.; Norvig, Peter (Technical Monitor)
2001-01-01
We address the problem of scheduling observations for a collection of earth observing satellites. This scheduling task is a difficult optimization problem, potentially involving many satellites, hundreds of requests, constraints on when and how to service each request, and resources such as instruments, recording devices, transmitters, and ground stations. High-fidelity models are required to ensure the validity of schedules; at the same time, the size and complexity of the problem makes it unlikely that systematic optimization search methods will be able to solve them in a reasonable time. This paper presents a constraint-based approach to solving the Earth Observing Satellites (EOS) scheduling problem, and proposes a stochastic heuristic search method for solving it.
Constrained Task Assignment and Scheduling On Networks of Arbitrary Topology
NASA Astrophysics Data System (ADS)
Jackson, Justin Patrick
This dissertation develops a framework to address centralized and distributed constrained task assignment and task scheduling problems. This framework is used to prove properties of these problems that can be exploited, develop effective solution algorithms, and to prove important properties such as correctness, completeness and optimality. The centralized task assignment and task scheduling problem treated here is expressed as a vehicle routing problem with the goal of optimizing mission time subject to mission constraints on task precedence and agent capability. The algorithm developed to solve this problem is able to coordinate vehicle (agent) timing for task completion. This class of problems is NP-hard and analytical guarantees on solution quality are often unavailable. This dissertation develops a technique for determining solution quality that can be used on a large class of problems and does not rely on traditional analytical guarantees. For distributed problems several agents must communicate to collectively solve a distributed task assignment and task scheduling problem. The distributed task assignment and task scheduling algorithms developed here allow for the optimization of constrained military missions in situations where the communication network may be incomplete and only locally known. Two problems are developed. The distributed task assignment problem incorporates communication constraints that must be satisfied; this is the Communication-Constrained Distributed Assignment Problem. A novel distributed assignment algorithm, the Stochastic Bidding Algorithm, solves this problem. The algorithm is correct, probabilistically complete, and has linear average-case time complexity. The distributed task scheduling problem addressed here is to minimize mission time subject to arbitrary predicate mission constraints; this is the Minimum-time Arbitrarily-constrained Distributed Scheduling Problem. The Optimal Distributed Non-sequential Backtracking Algorithm
Electrical spectrum analysis of operating Hydro Electric machines
NASA Astrophysics Data System (ADS)
Timperley, J. E.
1981-12-01
The electrical spectrum analysis of the operation of five pumped storage machines is discussed. It was found that machines without electrical problems produced little radio noise, although all machines produced some noise. Severe problems produced severe radio noise. If stator deterioration increases, the noise level increases. Similar machines produce similar electrical spectrum signatures. The general source of discharges can be located. A likelihood of failure can be calculated from spectrum analysis.
Use of evolutionary algorithms for telescope scheduling
NASA Astrophysics Data System (ADS)
Grim, Ruud; Jansen, Mischa; Baan, Arno; van Hemert, Jano; de Wolf, Hans
2002-07-01
LOFAR, a new radio telescope, will be designed to observe with up to 8 independent beams, thus allowing several simultaneous observations. Scheduling of multiple observations parallel in time, each having their own constraints, requires a more intelligent and flexible scheduling function then operated before. In support of the LOFAR radio telescope project, and in co-operation with Leiden University, Fokker Space has started a study to investigate the suitability of the use of evolutionary algorithms applied to complex scheduling problems. After a positive familiarization phase, we now examine the potential use of evolutionary algorithms via a demonstration project. Results of the familiarization phase, and the first results of the demonstration project are presented in this paper.
Scheduling Earth Observing Satellites with Evolutionary Algorithms
NASA Technical Reports Server (NTRS)
Globus, Al; Crawford, James; Lohn, Jason; Pryor, Anna
2003-01-01
We hypothesize that evolutionary algorithms can effectively schedule coordinated fleets of Earth observing satellites. The constraints are complex and the bottlenecks are not well understood, a condition where evolutionary algorithms are often effective. This is, in part, because evolutionary algorithms require only that one can represent solutions, modify solutions, and evaluate solution fitness. To test the hypothesis we have developed a representative set of problems, produced optimization software (in Java) to solve them, and run experiments comparing techniques. This paper presents initial results of a comparison of several evolutionary and other optimization techniques; namely the genetic algorithm, simulated annealing, squeaky wheel optimization, and stochastic hill climbing. We also compare separate satellite vs. integrated scheduling of a two satellite constellation. While the results are not definitive, tests to date suggest that simulated annealing is the best search technique and integrated scheduling is superior.
Scheduling IT staff at a bank: a mathematical programming approach.
Labidi, M; Mrad, M; Gharbi, A; Louly, M A
2014-01-01
We address a real-world optimization problem: the scheduling of a Bank Information Technologies (IT) staff. This problem can be defined as the process of constructing optimized work schedules for staff. In a general sense, it requires the allocation of suitably qualified staff to specific shifts to meet the demands for services of an organization while observing workplace regulations and attempting to satisfy individual work preferences. A monthly shift schedule is prepared to determine the shift duties of each staff considering shift coverage requirements, seniority-based workload rules, and staff work preferences. Due to the large number of conflicting constraints, a multiobjective programming model has been proposed to automate the schedule generation process. The suggested mathematical model has been implemented using Lingo software. The results indicate that high quality solutions can be obtained within a few seconds compared to the manually prepared schedules. PMID:24772032
Scheduling IT Staff at a Bank: A Mathematical Programming Approach
Labidi, M.; Mrad, M.; Gharbi, A.; Louly, M. A.
2014-01-01
We address a real-world optimization problem: the scheduling of a Bank Information Technologies (IT) staff. This problem can be defined as the process of constructing optimized work schedules for staff. In a general sense, it requires the allocation of suitably qualified staff to specific shifts to meet the demands for services of an organization while observing workplace regulations and attempting to satisfy individual work preferences. A monthly shift schedule is prepared to determine the shift duties of each staff considering shift coverage requirements, seniority-based workload rules, and staff work preferences. Due to the large number of conflicting constraints, a multiobjective programming model has been proposed to automate the schedule generation process. The suggested mathematical model has been implemented using Lingo software. The results indicate that high quality solutions can be obtained within a few seconds compared to the manually prepared schedules. PMID:24772032
Concurrent reinforcement schedules: behavior change and maintenance without extinction.
Hoch, Hannah; McComas, Jennifer J; Thompson, Andrea L; Paone, Debra
2002-01-01
We evaluated the effects of concurrent schedules of reinforcement on negatively reinforced problem behavior and task completion with 3 children with autism. Results indicated that problem behavior occurred at high levels and relatively few tasks were completed when problem behavior produced a break (from tasks) and task completion produced either no consequence or a break. By contrast, problem behavior was eliminated and tasks were completed when problem behavior produced a break and task completion produced a break with access to preferred activities. Treatment gains were maintained without the use of extinction when the response requirement was increased and the schedule of reinforcement was thinned. PMID:12102135
Scheduling in the Face of Uncertain Resource Consumption and Utility
NASA Technical Reports Server (NTRS)
Koga, Dennis (Technical Monitor); Frank, Jeremy; Dearden, Richard
2003-01-01
We discuss the problem of scheduling tasks that consume a resource with known capacity and where the tasks have varying utility. We consider problems in which the resource consumption and utility of each activity is described by probability distributions. In these circumstances, we would like to find schedules that exceed a lower bound on the expected utility when executed. We first show that while some of these problems are NP-complete, others are only NP-Hard. We then describe various heuristic search algorithms to solve these problems and their drawbacks. Finally, we present empirical results that characterize the behavior of these heuristics over a variety of problem classes.
Scheduling prioritized patients in emergency department laboratories.
Azadeh, A; Hosseinabadi Farahani, M; Torabzadeh, S; Baghersad, M
2014-11-01
This research focuses on scheduling patients in emergency department laboratories according to the priority of patients' treatments, determined by the triage factor. The objective is to minimize the total waiting time of patients in the emergency department laboratories with emphasis on patients with severe conditions. The problem is formulated as a flexible open shop scheduling problem and a mixed integer linear programming model is proposed. A genetic algorithm (GA) is developed for solving the problem. Then, the response surface methodology is applied for tuning the GA parameters. The algorithm is tested on a set of real data from an emergency department. Simulation results show that the proposed algorithm can significantly improve the efficiency of the emergency department by reducing the total waiting time of prioritized patients. PMID:25214024
High performance techniques for space mission scheduling
NASA Technical Reports Server (NTRS)
Smith, Stephen F.
1994-01-01
In this paper, we summarize current research at Carnegie Mellon University aimed at development of high performance techniques and tools for space mission scheduling. Similar to prior research in opportunistic scheduling, our approach assumes the use of dynamic analysis of problem constraints as a basis for heuristic focusing of problem solving search. This methodology, however, is grounded in representational assumptions more akin to those adopted in recent temporal planning research, and in a problem solving framework which similarly emphasizes constraint posting in an explicitly maintained solution constraint network. These more general representational assumptions are necessitated by the predominance of state-dependent constraints in space mission planning domains, and the consequent need to integrate resource allocation and plan synthesis processes. First, we review the space mission problems we have considered to date and indicate the results obtained in these application domains. Next, we summarize recent work in constraint posting scheduling procedures, which offer the promise of better future solutions to this class of problems.
From human-machine interaction to human-machine cooperation.
Hoc, J M
2000-07-01
Since the 1960s, the rapid growth of information systems has led to the wide development of research on human-computer interaction (HCI) that aims at the designing of human-computer interfaces presenting ergonomic properties, such as friendliness, usability, transparency, etc. Various work situations have been covered--clerical work, computer programming, design, etc. However, they were mainly static in the sense that the user fully controls the computer. More recently, public and private organizations have engaged themselves in the enterprise of managing more and more complex and coupled systems by the means of automation. Modern machines not only process information, but also act on dynamic situations as humans have done in the past, managing stock exchange, industrial plants, aircraft, etc. These dynamic situations are not fully controlled and are affected by uncertain factors. Hence, degrees of freedom must be maintained to allow the humans and the machine to adapt to unforeseen contingencies. A human-machine cooperation (HMC) approach is necessary to address the new stakes introduced by this trend. This paper describes the possible improvement of HCI by HMC, the need for a new conception of function allocation between humans and machines, and the main problems encountered within the new forms of human-machine relationship. It proposes a conceptual framework to study HMC from a cognitive point of view in highly dynamic situations like aircraft piloting or air-traffic control, and concludes on the design of 'cooperative' machines. PMID:10929820
Scheduling techniques in the Request Oriented Scheduling Engine (ROSE)
NASA Technical Reports Server (NTRS)
Zoch, David R.
1991-01-01
Scheduling techniques in the ROSE are presented in the form of the viewgraphs. The following subject areas are covered: agenda; ROSE summary and history; NCC-ROSE task goals; accomplishments; ROSE timeline manager; scheduling concerns; current and ROSE approaches; initial scheduling; BFSSE overview and example; and summary.
Morton, D.P.
1994-01-01
Handling uncertainty in natural inflow is an important part of a hydroelectric scheduling model. In a stochastic programming formulation, natural inflow may be modeled as a random vector with known distribution, but the size of the resulting mathematical program can be formidable. Decomposition-based algorithms take advantage of special structure and provide an attractive approach to such problems. We develop an enhanced Benders decomposition algorithm for solving multistage stochastic linear programs. The enhancements include warm start basis selection, preliminary cut generation, the multicut procedure, and decision tree traversing strategies. Computational results are presented for a collection of stochastic hydroelectric scheduling problems. Stochastic programming, Hydroelectric scheduling, Large-scale Systems.
The role of artificial intelligence techniques in scheduling systems
NASA Technical Reports Server (NTRS)
Geoffroy, Amy L.; Britt, Daniel L.; Gohring, John R.
1990-01-01
Artificial Intelligence (AI) techniques provide good solutions for many of the problems which are characteristic of scheduling applications. However, scheduling is a large, complex heterogeneous problem. Different applications will require different solutions. Any individual application will require the use of a variety of techniques, including both AI and conventional software methods. The operational context of the scheduling system will also play a large role in design considerations. The key is to identify those places where a specific AI technique is in fact the preferable solution, and to integrate that technique into the overall architecture.
NASA Astrophysics Data System (ADS)
Rowe, Robert
2002-05-01
The training of musicians begins by teaching basic musical concepts, a collection of knowledge commonly known as musicianship. Computer programs designed to implement musical skills (e.g., to make sense of what they hear, perform music expressively, or compose convincing pieces) can similarly benefit from access to a fundamental level of musicianship. Recent research in music cognition, artificial intelligence, and music theory has produced a repertoire of techniques that can make the behavior of computer programs more musical. Many of these were presented in a recently published book/CD-ROM entitled Machine Musicianship. For use in interactive music systems, we are interested in those which are fast enough to run in real time and that need only make reference to the material as it appears in sequence. This talk will review several applications that are able to identify the tonal center of musical material during performance. Beyond this specific task, the design of real-time algorithmic listening through the concurrent operation of several connected analyzers is examined. The presentation includes discussion of a library of C++ objects that can be combined to perform interactive listening and a demonstration of their capability.
McShea, Daniel W
2013-12-01
Wants, preferences, and cares are physical things or events, not ideas or propositions, and therefore no chain of pure logic can conclude with a want, preference, or care. It follows that no pure-logic machine will ever want, prefer, or care. And its behavior will never be driven in the way that deliberate human behavior is driven, in other words, it will not be motivated or goal directed. Therefore, if we want to simulate human-style interactions with the world, we will need to first understand the physical structure of goal-directed systems. I argue that all such systems share a common nested structure, consisting of a smaller entity that moves within and is driven by a larger field that contains it. In such systems, the smaller contained entity is directed by the field, but also moves to some degree independently of it, allowing the entity to deviate and return, to show the plasticity and persistence that is characteristic of goal direction. If all this is right, then human want-driven behavior probably involves a behavior-generating mechanism that is contained within a neural field of some kind. In principle, for goal directedness generally, the containment can be virtual, raising the possibility that want-driven behavior could be simulated in standard computational systems. But there are also reasons to believe that goal-direction works better when containment is also physical, suggesting that a new kind of hardware may be necessary. PMID:23792091
An assessment of the connection machine
NASA Technical Reports Server (NTRS)
Schreiber, Robert
1990-01-01
The CM-2 is an example of a connection machine. The strengths and problems of this implementation are considered as well as important issues in the architecture and programming environment of connection machines in general. These are contrasted to the same issues in Multiple Instruction/Multiple Data (MIMD) microprocessors and multicomputers.
Planning and Scheduling for Environmental Sensor Networks
NASA Astrophysics Data System (ADS)
Frank, J. D.
2005-12-01
Environmental Sensor Networks are a new way of monitoring the environment. They comprise autonomous sensor nodes in the environment that record real-time data, which is retrieved, analyzed, integrated with other data sets (e.g. satellite images, GIS, process models) and ultimately lead to scientific discoveries. Sensor networks must operate within time and resource constraints. Sensors have limited onboard memory, energy, computational power, communications windows and communications bandwidth. The value of data will depend on when, where and how it was collected, how detailed the data is, how long it takes to integrate the data, and how important the data was to the original scientific question. Planning and scheduling of sensor networks is necessary for effective, safe operations in the face of these constraints. For example, power bus limitations may preclude sensors from simultaneously collecting data and communicating without damaging the sensor; planners and schedulers can ensure these operations are ordered so that they do not happen simultaneously. Planning and scheduling can also ensure best use of the sensor network to maximize the value of collected science data. For example, if data is best recorded using a particular camera angle but it is costly in time and energy to achieve this, planners and schedulers can search for times when time and energy are available to achieve the optimal camera angle. Planning and scheduling can handle uncertainty in the problem specification; planners can be re-run when new information is made available, or can generate plans that include contingencies. For example, if bad weather may prevent the collection of data, a contingent plan can check lighting conditions and turn off data collection to save resources if lighting is not ideal. Both mobile and immobile sensors can benefit from planning and scheduling. For example, data collection on otherwise passive sensors can be halted to preserve limited power and memory
Basic Mathematics Machine Calculator Course.
ERIC Educational Resources Information Center
Windsor Public Schools, CT.
This series of four text-workbooks was designed for tenth grade mathematics students who have exhibited lack of problem-solving skills. Electric desk calculators are to be used with the text. In the first five chapters of the series, students learn how to use the machine while reviewing basic operations with whole numbers, decimals, fractions, and…
ERIC Educational Resources Information Center
Snider, Robert C.
1992-01-01
Since the 1960s, difficulty of developing a technology of instruction in public schools has proved insurmountable; results have been spotty, machines have come and gone, and classroom practices remain largely unchanged. Public clamor for reform has provided neither direction nor purpose. Technology will ultimately prevail; the problem is educating…
Minimizing metastatic risk in radiotherapy fractionation schedules
NASA Astrophysics Data System (ADS)
Badri, Hamidreza; Ramakrishnan, Jagdish; Leder, Kevin
2015-11-01
Metastasis is the process by which cells from a primary tumor disperse and form new tumors at distant anatomical locations. The treatment and prevention of metastatic cancer remains an extremely challenging problem. This work introduces a novel biologically motivated objective function to the radiation optimization community that takes into account metastatic risk instead of the status of the primary tumor. In this work, we consider the problem of developing fractionated irradiation schedules that minimize production of metastatic cancer cells while keeping normal tissue damage below an acceptable level. A dynamic programming framework is utilized to determine the optimal fractionation scheme. We evaluated our approach on a breast cancer case using the heart and the lung as organs-at-risk (OAR). For small tumor α /β values, hypo-fractionated schedules were optimal, which is consistent with standard models. However, for relatively larger α /β values, we found the type of schedule depended on various parameters such as the time when metastatic risk was evaluated, the α /β values of the OARs, and the normal tissue sparing factors. Interestingly, in contrast to standard models, hypo-fractionated and semi-hypo-fractionated schedules (large initial doses with doses tapering off with time) were suggested even with large tumor α/β values. Numerical results indicate the potential for significant reduction in metastatic risk.