ERIC Educational Resources Information Center
Cooper, Melanie M.; Cox, Charles T., Jr.; Nammouz, Minory; Case, Edward; Stevens, Ronald
2008-01-01
Improving students' problem-solving skills is a major goal for most science educators. While a large body of research on problem solving exists, assessment of meaningful problem solving is very difficult, particularly for courses with large numbers of students in which one-on-one interactions are not feasible. We have used a suite of software…
Penders, Bart; Vos, Rein; Horstman, Klasien
2009-11-01
Solving complex problems in large-scale research programmes requires cooperation and division of labour. Simultaneously, large-scale problem solving also gives rise to unintended side effects. Based upon 5 years of researching two large-scale nutrigenomic research programmes, we argue that problems are fragmented in order to be solved. These sub-problems are given priority for practical reasons and in the process of solving them, various changes are introduced in each sub-problem. Combined with additional diversity as a result of interdisciplinarity, this makes reassembling the original and overall goal of the research programme less likely. In the case of nutrigenomics and health, this produces a diversification of health. As a result, the public health goal of contemporary nutrition science is not reached in the large-scale research programmes we studied. Large-scale research programmes are very successful in producing scientific publications and new knowledge; however, in reaching their political goals they often are less successful.
The Cyclic Nature of Problem Solving: An Emergent Multidimensional Problem-Solving Framework
ERIC Educational Resources Information Center
Carlson, Marilyn P.; Bloom, Irene
2005-01-01
This paper describes the problem-solving behaviors of 12 mathematicians as they completed four mathematical tasks. The emergent problem-solving framework draws on the large body of research, as grounded by and modified in response to our close observations of these mathematicians. The resulting "Multidimensional Problem-Solving Framework" has four…
NASA Astrophysics Data System (ADS)
Noor-E-Alam, Md.; Doucette, John
2015-08-01
Grid-based location problems (GBLPs) can be used to solve location problems in business, engineering, resource exploitation, and even in the field of medical sciences. To solve these decision problems, an integer linear programming (ILP) model is designed and developed to provide the optimal solution for GBLPs considering fixed cost criteria. Preliminary results show that the ILP model is efficient in solving small to moderate-sized problems. However, this ILP model becomes intractable in solving large-scale instances. Therefore, a decomposition heuristic is proposed to solve these large-scale GBLPs, which demonstrates significant reduction of solution runtimes. To benchmark the proposed heuristic, results are compared with the exact solution via ILP. The experimental results show that the proposed method significantly outperforms the exact method in runtime with minimal (and in most cases, no) loss of optimality.
The solution of large multi-dimensional Poisson problems
NASA Technical Reports Server (NTRS)
Stone, H. S.
1974-01-01
The Buneman algorithm for solving Poisson problems can be adapted to solve large Poisson problems on computers with a rotating drum memory so that the computation is done with very little time lost due to rotational latency of the drum.
Using Coaching to Improve the Teaching of Problem Solving to Year 8 Students in Mathematics
ERIC Educational Resources Information Center
Kargas, Christine Anestis; Stephens, Max
2014-01-01
This study investigated how to improve the teaching of problem solving in a large Melbourne secondary school. Coaching was used to support and equip five teachers, some with limited experiences in teaching problem solving, with knowledge and strategies to build up students' problem solving and reasoning skills. The results showed increased…
Xia, Youshen; Sun, Changyin; Zheng, Wei Xing
2012-05-01
There is growing interest in solving linear L1 estimation problems for sparsity of the solution and robustness against non-Gaussian noise. This paper proposes a discrete-time neural network which can calculate large linear L1 estimation problems fast. The proposed neural network has a fixed computational step length and is proved to be globally convergent to an optimal solution. Then, the proposed neural network is efficiently applied to image restoration. Numerical results show that the proposed neural network is not only efficient in solving degenerate problems resulting from the nonunique solutions of the linear L1 estimation problems but also needs much less computational time than the related algorithms in solving both linear L1 estimation and image restoration problems.
The Role of the Goal in Solving Hard Computational Problems: Do People Really Optimize?
ERIC Educational Resources Information Center
Carruthers, Sarah; Stege, Ulrike; Masson, Michael E. J.
2018-01-01
The role that the mental, or internal, representation plays when people are solving hard computational problems has largely been overlooked to date, despite the reality that this internal representation drives problem solving. In this work we investigate how performance on versions of two hard computational problems differs based on what internal…
Problem Solving and the Development of Expertise in Management.
ERIC Educational Resources Information Center
Lash, Fredrick B.
This study investigated novice and expert problem solving behavior in management to examine the role of domain specific knowledge on problem solving processes. Forty-one middle level marketing managers in a large petrochemical organization provided think aloud protocols in response to two hypothetical management scenarios. Protocol analysis…
Parallel Algorithm Solves Coupled Differential Equations
NASA Technical Reports Server (NTRS)
Hayashi, A.
1987-01-01
Numerical methods adapted to concurrent processing. Algorithm solves set of coupled partial differential equations by numerical integration. Adapted to run on hypercube computer, algorithm separates problem into smaller problems solved concurrently. Increase in computing speed with concurrent processing over that achievable with conventional sequential processing appreciable, especially for large problems.
Planning meals: Problem-solving on a real data-base
ERIC Educational Resources Information Center
Byrne, Richard
1977-01-01
Planning the menu for a dinner party, which involves problem-solving with a large body of knowledge, is used to study the daily operation of human memory. Verbal protocol analysis, a technique devised to investigate formal problem-solving, is examined theoretically and adapted for analysis of this task. (Author/MV)
Internet computer coaches for introductory physics problem solving
NASA Astrophysics Data System (ADS)
Xu Ryan, Qing
The ability to solve problems in a variety of contexts is becoming increasingly important in our rapidly changing technological society. Problem-solving is a complex process that is important for everyday life and crucial for learning physics. Although there is a great deal of effort to improve student problem solving skills throughout the educational system, national studies have shown that the majority of students emerge from such courses having made little progress toward developing good problem-solving skills. The Physics Education Research Group at the University of Minnesota has been developing Internet computer coaches to help students become more expert-like problem solvers. During the Fall 2011 and Spring 2013 semesters, the coaches were introduced into large sections (200+ students) of the calculus based introductory mechanics course at the University of Minnesota. This dissertation, will address the research background of the project, including the pedagogical design of the coaches and the assessment of problem solving. The methodological framework of conducting experiments will be explained. The data collected from the large-scale experimental studies will be discussed from the following aspects: the usage and usability of these coaches; the usefulness perceived by students; and the usefulness measured by final exam and problem solving rubric. It will also address the implications drawn from this study, including using this data to direct future coach design and difficulties in conducting authentic assessment of problem-solving.
NASA Astrophysics Data System (ADS)
Li, Yuzhong
Using GA solve the winner determination problem (WDP) with large bids and items, run under different distribution, because the search space is large, constraint complex and it may easy to produce infeasible solution, would affect the efficiency and quality of algorithm. This paper present improved MKGA, including three operator: preprocessing, insert bid and exchange recombination, and use Monkey-king elite preservation strategy. Experimental results show that improved MKGA is better than SGA in population size and computation. The problem that traditional branch and bound algorithm hard to solve, improved MKGA can solve and achieve better effect.
Neural Network Solves "Traveling-Salesman" Problem
NASA Technical Reports Server (NTRS)
Thakoor, Anilkumar P.; Moopenn, Alexander W.
1990-01-01
Experimental electronic neural network solves "traveling-salesman" problem. Plans round trip of minimum distance among N cities, visiting every city once and only once (without backtracking). This problem is paradigm of many problems of global optimization (e.g., routing or allocation of resources) occuring in industry, business, and government. Applied to large number of cities (or resources), circuits of this kind expected to solve problem faster and more cheaply.
Solving LP Relaxations of Large-Scale Precedence Constrained Problems
NASA Astrophysics Data System (ADS)
Bienstock, Daniel; Zuckerberg, Mark
We describe new algorithms for solving linear programming relaxations of very large precedence constrained production scheduling problems. We present theory that motivates a new set of algorithmic ideas that can be employed on a wide range of problems; on data sets arising in the mining industry our algorithms prove effective on problems with many millions of variables and constraints, obtaining provably optimal solutions in a few minutes of computation.
NASA Technical Reports Server (NTRS)
1977-01-01
The large area crop inventory experiment is being developed to predict crop production through satellite photographs. This experiment demonstrates how space age technology can contribute to solving practical problems of agriculture management.
Fast, Nonlinear, Fully Probabilistic Inversion of Large Geophysical Problems
NASA Astrophysics Data System (ADS)
Curtis, A.; Shahraeeni, M.; Trampert, J.; Meier, U.; Cho, G.
2010-12-01
Almost all Geophysical inverse problems are in reality nonlinear. Fully nonlinear inversion including non-approximated physics, and solving for probability distribution functions (pdf’s) that describe the solution uncertainty, generally requires sampling-based Monte-Carlo style methods that are computationally intractable in most large problems. In order to solve such problems, physical relationships are usually linearized leading to efficiently-solved, (possibly iterated) linear inverse problems. However, it is well known that linearization can lead to erroneous solutions, and in particular to overly optimistic uncertainty estimates. What is needed across many Geophysical disciplines is a method to invert large inverse problems (or potentially tens of thousands of small inverse problems) fully probabilistically and without linearization. This talk shows how very large nonlinear inverse problems can be solved fully probabilistically and incorporating any available prior information using mixture density networks (driven by neural network banks), provided the problem can be decomposed into many small inverse problems. In this talk I will explain the methodology, compare multi-dimensional pdf inversion results to full Monte Carlo solutions, and illustrate the method with two applications: first, inverting surface wave group and phase velocities for a fully-probabilistic global tomography model of the Earth’s crust and mantle, and second inverting industrial 3D seismic data for petrophysical properties throughout and around a subsurface hydrocarbon reservoir. The latter problem is typically decomposed into 104 to 105 individual inverse problems, each solved fully probabilistically and without linearization. The results in both cases are sufficiently close to the Monte Carlo solution to exhibit realistic uncertainty, multimodality and bias. This provides far greater confidence in the results, and in decisions made on their basis.
ERIC Educational Resources Information Center
Anderson, William L.; Mitchell, Steven M.; Osgood, Marcy P.
2008-01-01
For the past 3 yr, faculty at the University of New Mexico, Department of Biochemistry and Molecular Biology have been using interactive online Problem-Based Learning (PBL) case discussions in our large-enrollment classes. We have developed an illustrative tracking method to monitor student use of problem-solving strategies to provide targeted…
Learning Analysis of K-12 Students' Online Problem Solving: A Three-Stage Assessment Approach
ERIC Educational Resources Information Center
Hu, Yiling; Wu, Bian; Gu, Xiaoqing
2017-01-01
Problem solving is considered a fundamental human skill. However, large-scale assessment of problem solving in K-12 education remains a challenging task. Researchers have argued for the development of an enhanced assessment approach through joint effort from multiple disciplines. In this study, a three-stage approach based on an evidence-centered…
Gong, Pinghua; Zhang, Changshui; Lu, Zhaosong; Huang, Jianhua Z; Ye, Jieping
2013-01-01
Non-convex sparsity-inducing penalties have recently received considerable attentions in sparse learning. Recent theoretical investigations have demonstrated their superiority over the convex counterparts in several sparse learning settings. However, solving the non-convex optimization problems associated with non-convex penalties remains a big challenge. A commonly used approach is the Multi-Stage (MS) convex relaxation (or DC programming), which relaxes the original non-convex problem to a sequence of convex problems. This approach is usually not very practical for large-scale problems because its computational cost is a multiple of solving a single convex problem. In this paper, we propose a General Iterative Shrinkage and Thresholding (GIST) algorithm to solve the nonconvex optimization problem for a large class of non-convex penalties. The GIST algorithm iteratively solves a proximal operator problem, which in turn has a closed-form solution for many commonly used penalties. At each outer iteration of the algorithm, we use a line search initialized by the Barzilai-Borwein (BB) rule that allows finding an appropriate step size quickly. The paper also presents a detailed convergence analysis of the GIST algorithm. The efficiency of the proposed algorithm is demonstrated by extensive experiments on large-scale data sets.
A novel heuristic algorithm for capacitated vehicle routing problem
NASA Astrophysics Data System (ADS)
Kır, Sena; Yazgan, Harun Reşit; Tüncel, Emre
2017-09-01
The vehicle routing problem with the capacity constraints was considered in this paper. It is quite difficult to achieve an optimal solution with traditional optimization methods by reason of the high computational complexity for large-scale problems. Consequently, new heuristic or metaheuristic approaches have been developed to solve this problem. In this paper, we constructed a new heuristic algorithm based on the tabu search and adaptive large neighborhood search (ALNS) with several specifically designed operators and features to solve the capacitated vehicle routing problem (CVRP). The effectiveness of the proposed algorithm was illustrated on the benchmark problems. The algorithm provides a better performance on large-scaled instances and gained advantage in terms of CPU time. In addition, we solved a real-life CVRP using the proposed algorithm and found the encouraging results by comparison with the current situation that the company is in.
An electromagnetism-like metaheuristic for open-shop problems with no buffer
NASA Astrophysics Data System (ADS)
Naderi, Bahman; Najafi, Esmaeil; Yazdani, Mehdi
2012-12-01
This paper considers open-shop scheduling with no intermediate buffer to minimize total tardiness. This problem occurs in many production settings, in the plastic molding, chemical, and food processing industries. The paper mathematically formulates the problem by a mixed integer linear program. The problem can be optimally solved by the model. The paper also develops a novel metaheuristic based on an electromagnetism algorithm to solve the large-sized problems. The paper conducts two computational experiments. The first includes small-sized instances by which the mathematical model and general performance of the proposed metaheuristic are evaluated. The second evaluates the metaheuristic for its performance to solve some large-sized instances. The results show that the model and algorithm are effective to deal with the problem.
Reasoning by analogy as an aid to heuristic theorem proving.
NASA Technical Reports Server (NTRS)
Kling, R. E.
1972-01-01
When heuristic problem-solving programs are faced with large data bases that contain numbers of facts far in excess of those needed to solve any particular problem, their performance rapidly deteriorates. In this paper, the correspondence between a new unsolved problem and a previously solved analogous problem is computed and invoked to tailor large data bases to manageable sizes. This paper outlines the design of an algorithm for generating and exploiting analogies between theorems posed to a resolution-logic system. These algorithms are believed to be the first computationally feasible development of reasoning by analogy to be applied to heuristic theorem proving.
Klegeris, Andis; Bahniwal, Manpreet; Hurren, Heather
2013-01-01
Problem-based learning (PBL) was originally introduced in medical education programs as a form of small-group learning, but its use has now spread to large undergraduate classrooms in various other disciplines. Introduction of new teaching techniques, including PBL-based methods, needs to be justified by demonstrating the benefits of such techniques over classical teaching styles. Previously, we demonstrated that introduction of tutor-less PBL in a large third-year biochemistry undergraduate class increased student satisfaction and attendance. The current study assessed the generic problem-solving abilities of students from the same class at the beginning and end of the term, and compared student scores with similar data obtained in three classes not using PBL. Two generic problem-solving tests of equal difficulty were administered such that students took different tests at the beginning and the end of the term. Blinded marking showed a statistically significant 13% increase in the test scores of the biochemistry students exposed to PBL, while no trend toward significant change in scores was observed in any of the control groups not using PBL. Our study is among the first to demonstrate that use of tutor-less PBL in a large classroom leads to statistically significant improvement in generic problem-solving skills of students. PMID:23463230
Finite difference and Runge-Kutta methods for solving vibration problems
NASA Astrophysics Data System (ADS)
Lintang Renganis Radityani, Scolastika; Mungkasi, Sudi
2017-11-01
The vibration of a storey building can be modelled into a system of second order ordinary differential equations. If the number of floors of a building is large, then the result is a large scale system of second order ordinary differential equations. The large scale system is difficult to solve, and if it can be solved, the solution may not be accurate. Therefore, in this paper, we seek for accurate methods for solving vibration problems. We compare the performance of numerical finite difference and Runge-Kutta methods for solving large scale systems of second order ordinary differential equations. The finite difference methods include the forward and central differences. The Runge-Kutta methods include the Euler and Heun methods. Our research results show that the central finite difference and the Heun methods produce more accurate solutions than the forward finite difference and the Euler methods do.
Predicting protein structures with a multiplayer online game.
Cooper, Seth; Khatib, Firas; Treuille, Adrien; Barbero, Janos; Lee, Jeehyung; Beenen, Michael; Leaver-Fay, Andrew; Baker, David; Popović, Zoran; Players, Foldit
2010-08-05
People exert large amounts of problem-solving effort playing computer games. Simple image- and text-recognition tasks have been successfully 'crowd-sourced' through games, but it is not clear if more complex scientific problems can be solved with human-directed computing. Protein structure prediction is one such problem: locating the biologically relevant native conformation of a protein is a formidable computational challenge given the very large size of the search space. Here we describe Foldit, a multiplayer online game that engages non-scientists in solving hard prediction problems. Foldit players interact with protein structures using direct manipulation tools and user-friendly versions of algorithms from the Rosetta structure prediction methodology, while they compete and collaborate to optimize the computed energy. We show that top-ranked Foldit players excel at solving challenging structure refinement problems in which substantial backbone rearrangements are necessary to achieve the burial of hydrophobic residues. Players working collaboratively develop a rich assortment of new strategies and algorithms; unlike computational approaches, they explore not only the conformational space but also the space of possible search strategies. The integration of human visual problem-solving and strategy development capabilities with traditional computational algorithms through interactive multiplayer games is a powerful new approach to solving computationally-limited scientific problems.
Fast Combinatorial Algorithm for the Solution of Linearly Constrained Least Squares Problems
Van Benthem, Mark H.; Keenan, Michael R.
2008-11-11
A fast combinatorial algorithm can significantly reduce the computational burden when solving general equality and inequality constrained least squares problems with large numbers of observation vectors. The combinatorial algorithm provides a mathematically rigorous solution and operates at great speed by reorganizing the calculations to take advantage of the combinatorial nature of the problems to be solved. The combinatorial algorithm exploits the structure that exists in large-scale problems in order to minimize the number of arithmetic operations required to obtain a solution.
New displacement-based methods for optimal truss topology design
NASA Technical Reports Server (NTRS)
Bendsoe, Martin P.; Ben-Tal, Aharon; Haftka, Raphael T.
1991-01-01
Two alternate methods for maximum stiffness truss topology design are presented. The ground structure approach is used, and the problem is formulated in terms of displacements and bar areas. This large, nonconvex optimization problem can be solved by a simultaneous analysis and design approach. Alternatively, an equivalent, unconstrained, and convex problem in the displacements only can be formulated, and this problem can be solved by a nonsmooth, steepest descent algorithm. In both methods, the explicit solving of the equilibrium equations and the assembly of the global stiffness matrix are circumvented. A large number of examples have been studied, showing the attractive features of topology design as well as exposing interesting features of optimal topologies.
ERIC Educational Resources Information Center
Sonnleitner, Philipp; Brunner, Martin; Keller, Ulrich; Martin, Romain
2014-01-01
Whereas the assessment of complex problem solving (CPS) has received increasing attention in the context of international large-scale assessments, its fairness in regard to students' cultural background has gone largely unexplored. On the basis of a student sample of 9th-graders (N = 299), including a representative number of immigrant students (N…
Solving Fuzzy Optimization Problem Using Hybrid Ls-Sa Method
NASA Astrophysics Data System (ADS)
Vasant, Pandian
2011-06-01
Fuzzy optimization problem has been one of the most and prominent topics inside the broad area of computational intelligent. It's especially relevant in the filed of fuzzy non-linear programming. It's application as well as practical realization can been seen in all the real world problems. In this paper a large scale non-linear fuzzy programming problem has been solved by hybrid optimization techniques of Line Search (LS), Simulated Annealing (SA) and Pattern Search (PS). As industrial production planning problem with cubic objective function, 8 decision variables and 29 constraints has been solved successfully using LS-SA-PS hybrid optimization techniques. The computational results for the objective function respect to vagueness factor and level of satisfaction has been provided in the form of 2D and 3D plots. The outcome is very promising and strongly suggests that the hybrid LS-SA-PS algorithm is very efficient and productive in solving the large scale non-linear fuzzy programming problem.
Performance of Grey Wolf Optimizer on large scale problems
NASA Astrophysics Data System (ADS)
Gupta, Shubham; Deep, Kusum
2017-01-01
For solving nonlinear continuous problems of optimization numerous nature inspired optimization techniques are being proposed in literature which can be implemented to solve real life problems wherein the conventional techniques cannot be applied. Grey Wolf Optimizer is one of such technique which is gaining popularity since the last two years. The objective of this paper is to investigate the performance of Grey Wolf Optimization Algorithm on large scale optimization problems. The Algorithm is implemented on 5 common scalable problems appearing in literature namely Sphere, Rosenbrock, Rastrigin, Ackley and Griewank Functions. The dimensions of these problems are varied from 50 to 1000. The results indicate that Grey Wolf Optimizer is a powerful nature inspired Optimization Algorithm for large scale problems, except Rosenbrock which is a unimodal function.
Personal and parental problem drinking: effects on problem-solving performance and self-appraisal.
Slavkin, S L; Heimberg, R G; Winning, C D; McCaffrey, R J
1992-01-01
This study examined the problem-solving performances and self-appraisals of problem-solving ability of college-age subjects with and without parental history of problem drinking. Contrary to our predictions, children of problem drinkers (COPDs) were rated as somewhat more effective in their problem-solving skills than non-COPDs, undermining prevailing assumptions about offspring from alcoholic households. While this difference was not large and was qualified by other variables, subjects' own alcohol abuse did exert a detrimental effect on problem-solving performance, regardless of parental history of problem drinking. However, a different pattern was evident for problem-solving self-appraisals. Alcohol-abusing non-COPDs saw themselves as effective problem-solvers while alcohol-abusing COPDs appraised themselves as poor problem-solvers. In addition, the self-appraisals of alcohol-abusing COPDs were consistent with objective ratings of solution effectiveness (i.e., they were both negative) while alcohol-abusing non-COPDs were overly positive in their appraisals, opposing the judgments of trained raters. This finding suggests that the relationship between personal alcohol abuse and self-appraised problem-solving abilities may differ as a function of parental history of problem drinking. Limitations on the generalizability of findings are addressed.
Efficient ICCG on a shared memory multiprocessor
NASA Technical Reports Server (NTRS)
Hammond, Steven W.; Schreiber, Robert
1989-01-01
Different approaches are discussed for exploiting parallelism in the ICCG (Incomplete Cholesky Conjugate Gradient) method for solving large sparse symmetric positive definite systems of equations on a shared memory parallel computer. Techniques for efficiently solving triangular systems and computing sparse matrix-vector products are explored. Three methods for scheduling the tasks in solving triangular systems are implemented on the Sequent Balance 21000. Sample problems that are representative of a large class of problems solved using iterative methods are used. We show that a static analysis to determine data dependences in the triangular solve can greatly improve its parallel efficiency. We also show that ignoring symmetry and storing the whole matrix can reduce solution time substantially.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moryakov, A. V., E-mail: sailor@orc.ru
2016-12-15
An algorithm for solving the linear Cauchy problem for large systems of ordinary differential equations is presented. The algorithm for systems of first-order differential equations is implemented in the EDELWEISS code with the possibility of parallel computations on supercomputers employing the MPI (Message Passing Interface) standard for the data exchange between parallel processes. The solution is represented by a series of orthogonal polynomials on the interval [0, 1]. The algorithm is characterized by simplicity and the possibility to solve nonlinear problems with a correction of the operator in accordance with the solution obtained in the previous iterative process.
The application of artificial intelligence techniques to large distributed networks
NASA Technical Reports Server (NTRS)
Dubyah, R.; Smith, T. R.; Star, J. L.
1985-01-01
Data accessibility and transfer of information, including the land resources information system pilot, are structured as large computer information networks. These pilot efforts include the reduction of the difficulty to find and use data, reducing processing costs, and minimize incompatibility between data sources. Artificial Intelligence (AI) techniques were suggested to achieve these goals. The applicability of certain AI techniques are explored in the context of distributed problem solving systems and the pilot land data system (PLDS). The topics discussed include: PLDS and its data processing requirements, expert systems and PLDS, distributed problem solving systems, AI problem solving paradigms, query processing, and distributed data bases.
Analysis of the Efficacy of an Intervention to Improve Parent-Adolescent Problem Solving
Semeniuk, Yulia Yuriyivna; Brown, Roger L.; Riesch, Susan K.
2016-01-01
We conducted a two-group longitudinal partially nested randomized controlled trial to examine whether young adolescent youth-parent dyads participating in Mission Possible: Parents and Kids Who Listen, in contrast to a comparison group, would demonstrate improved problem solving skill. The intervention is based on the Circumplex Model and Social Problem Solving Theory. The Circumplex Model posits that families who are balanced, that is characterized by high cohesion and flexibility and open communication, function best. Social Problem Solving Theory informs the process and skills of problem solving. The Conditional Latent Growth Modeling analysis revealed no statistically significant differences in problem solving among the final sample of 127 dyads in the intervention and comparison groups. Analyses of effect sizes indicated large magnitude group effects for selected scales for youth and dyads portraying a potential for efficacy and identifying for whom the intervention may be efficacious if study limitations and lessons learned were addressed. PMID:26936844
ERIC Educational Resources Information Center
Sousa, Fernando Cardoso; Monteiro, Ileana Pardal; Pellissier, René
2014-01-01
This article presents the development of a small-world network using an adapted version of the large-group problem-solving method "Future Search." Two management classes in a higher education setting were selected and required to plan a project. The students completed a survey focused on the frequency of communications before and after…
Problem Solving in Electricity.
ERIC Educational Resources Information Center
Caillot, Michel; Chalouhi, Elias
Two studies were conducted to describe how students perform direct current (D-C) circuit problems. It was hypothesized that problem solving in the electricity domain depends largely on good visual processing of the circuit diagram and that this processing depends on the ability to recognize when two or more electrical components are in series or…
Engineering management of large scale systems
NASA Technical Reports Server (NTRS)
Sanders, Serita; Gill, Tepper L.; Paul, Arthur S.
1989-01-01
The organization of high technology and engineering problem solving, has given rise to an emerging concept. Reasoning principles for integrating traditional engineering problem solving with system theory, management sciences, behavioral decision theory, and planning and design approaches can be incorporated into a methodological approach to solving problems with a long range perspective. Long range planning has a great potential to improve productivity by using a systematic and organized approach. Thus, efficiency and cost effectiveness are the driving forces in promoting the organization of engineering problems. Aspects of systems engineering that provide an understanding of management of large scale systems are broadly covered here. Due to the focus and application of research, other significant factors (e.g., human behavior, decision making, etc.) are not emphasized but are considered.
Solving Large-Scale Inverse Magnetostatic Problems using the Adjoint Method
Bruckner, Florian; Abert, Claas; Wautischer, Gregor; Huber, Christian; Vogler, Christoph; Hinze, Michael; Suess, Dieter
2017-01-01
An efficient algorithm for the reconstruction of the magnetization state within magnetic components is presented. The occurring inverse magnetostatic problem is solved by means of an adjoint approach, based on the Fredkin-Koehler method for the solution of the forward problem. Due to the use of hybrid FEM-BEM coupling combined with matrix compression techniques the resulting algorithm is well suited for large-scale problems. Furthermore the reconstruction of the magnetization state within a permanent magnet as well as an optimal design application are demonstrated. PMID:28098851
Internationalisation and Economic Growth: The Portuguese Case
ERIC Educational Resources Information Center
da Costa, Renato J. Lopes; António, Nélson J. Santos; Miguel, Maria Isabel
2017-01-01
Historically, a policy of enforcement in internationalisation processes is still seen by many as an approach to solve certain economic crises. However, Portugal's solution for this problem is part of a greater problem, namely trying to solve a European problem that has recently worsened and is largely uncontrolled. This paper aims to contribute,…
NASA Technical Reports Server (NTRS)
Vanderplaats, Garrett; Townsend, James C. (Technical Monitor)
2002-01-01
The purpose of this research under the NASA Small Business Innovative Research program was to develop algorithms and associated software to solve very large nonlinear, constrained optimization tasks. Key issues included efficiency, reliability, memory, and gradient calculation requirements. This report describes the general optimization problem, ten candidate methods, and detailed evaluations of four candidates. The algorithm chosen for final development is a modern recreation of a 1960s external penalty function method that uses very limited computer memory and computational time. Although of lower efficiency, the new method can solve problems orders of magnitude larger than current methods. The resulting BIGDOT software has been demonstrated on problems with 50,000 variables and about 50,000 active constraints. For unconstrained optimization, it has solved a problem in excess of 135,000 variables. The method includes a technique for solving discrete variable problems that finds a "good" design, although a theoretical optimum cannot be guaranteed. It is very scalable in that the number of function and gradient evaluations does not change significantly with increased problem size. Test cases are provided to demonstrate the efficiency and reliability of the methods and software.
Temperament and problem solving in a population of adolescent guide dogs.
Bray, Emily E; Sammel, Mary D; Seyfarth, Robert M; Serpell, James A; Cheney, Dorothy L
2017-09-01
It is often assumed that measures of temperament within individuals are more correlated to one another than to measures of problem solving. However, the exact relationship between temperament and problem-solving tasks remains unclear because large-scale studies have typically focused on each independently. To explore this relationship, we tested 119 prospective adolescent guide dogs on a battery of 11 temperament and problem-solving tasks. We then summarized the data using both confirmatory factor analysis and exploratory principal components analysis. Results of confirmatory analysis revealed that a priori separation of tests as measuring either temperament or problem solving led to weak results, poor model fit, some construct validity, and no predictive validity. In contrast, results of exploratory analysis were best summarized by principal components that mixed temperament and problem-solving traits. These components had both construct and predictive validity (i.e., association with success in the guide dog training program). We conclude that there is complex interplay between tasks of "temperament" and "problem solving" and that the study of both together will be more informative than approaches that consider either in isolation.
NASA Technical Reports Server (NTRS)
Bless, Robert R.
1991-01-01
A time-domain finite element method is developed for optimal control problems. The theory derived is general enough to handle a large class of problems including optimal control problems that are continuous in the states and controls, problems with discontinuities in the states and/or system equations, problems with control inequality constraints, problems with state inequality constraints, or problems involving any combination of the above. The theory is developed in such a way that no numerical quadrature is necessary regardless of the degree of nonlinearity in the equations. Also, the same shape functions may be employed for every problem because all strong boundary conditions are transformed into natural or weak boundary conditions. In addition, the resulting nonlinear algebraic equations are very sparse. Use of sparse matrix solvers allows for the rapid and accurate solution of very difficult optimization problems. The formulation is applied to launch-vehicle trajectory optimization problems, and results show that real-time optimal guidance is realizable with this method. Finally, a general problem solving environment is created for solving a large class of optimal control problems. The algorithm uses both FORTRAN and a symbolic computation program to solve problems with a minimum of user interaction. The use of symbolic computation eliminates the need for user-written subroutines which greatly reduces the setup time for solving problems.
Fitzpatrick, Stephanie L; Hill-Briggs, Felicia
2015-10-01
Identification of patients with poor chronic disease self-management skills can facilitate treatment planning, determine effectiveness of interventions, and reduce disease complications. This paper describes the use of a Rasch model, the Rating Scale Model, to examine psychometric properties of the 50-item Health Problem-Solving Scale (HPSS) among 320 African American patients with high risk for cardiovascular disease. Items on the positive/effective HPSS subscales targeted patients at low, moderate, and high levels of positive/effective problem solving, whereas items on the negative/ineffective problem solving subscales mostly targeted those at moderate or high levels of ineffective problem solving. Validity was examined by correlating factor scores on the measure with clinical and behavioral measures. Items on the HPSS show promise in the ability to assess health-related problem solving among high risk patients. However, further revisions of the scale are needed to increase its usability and validity with large, diverse patient populations in the future.
Analysis of the Efficacy of an Intervention to Improve Parent-Adolescent Problem Solving.
Semeniuk, Yulia Yuriyivna; Brown, Roger L; Riesch, Susan K
2016-07-01
We conducted a two-group longitudinal partially nested randomized controlled trial to examine whether young adolescent youth-parent dyads participating in Mission Possible: Parents and Kids Who Listen, in contrast to a comparison group, would demonstrate improved problem-solving skill. The intervention is based on the Circumplex Model and Social Problem-Solving Theory. The Circumplex Model posits that families who are balanced, that is characterized by high cohesion and flexibility and open communication, function best. Social Problem-Solving Theory informs the process and skills of problem solving. The Conditional Latent Growth Modeling analysis revealed no statistically significant differences in problem solving among the final sample of 127 dyads in the intervention and comparison groups. Analyses of effect sizes indicated large magnitude group effects for selected scales for youth and dyads portraying a potential for efficacy and identifying for whom the intervention may be efficacious if study limitations and lessons learned were addressed. © The Author(s) 2016.
Innovative problem solving by wild spotted hyenas
Benson-Amram, Sarah; Holekamp, Kay E.
2012-01-01
Innovative animals are those able to solve novel problems or invent novel solutions to existing problems. Despite the important ecological and evolutionary consequences of innovation, we still know very little about the traits that vary among individuals within a species to make them more or less innovative. Here we examine innovative problem solving by spotted hyenas (Crocuta crocuta) in their natural habitat, and demonstrate for the first time in a non-human animal that those individuals exhibiting a greater diversity of initial exploratory behaviours are more successful problem solvers. Additionally, as in earlier work, we found that neophobia was a critical inhibitor of problem-solving success. Interestingly, although juveniles and adults were equally successful in solving the problem, juveniles were significantly more diverse in their initial exploratory behaviours, more persistent and less neophobic than were adults. We found no significant effects of social rank or sex on success, the diversity of initial exploratory behaviours, behavioural persistence or neophobia. Our results suggest that the diversity of initial exploratory behaviours, akin to some measures of human creativity, is an important, but largely overlooked, determinant of problem-solving success in non-human animals. PMID:22874748
The Association of DRD2 with Insight Problem Solving.
Zhang, Shun; Zhang, Jinghuan
2016-01-01
Although the insight phenomenon has attracted great attention from psychologists, it is still largely unknown whether its variation in well-functioning human adults has a genetic basis. Several lines of evidence suggest that genes involved in dopamine (DA) transmission might be potential candidates. The present study explored for the first time the association of dopamine D2 receptor gene ( DRD2 ) with insight problem solving. Fifteen single-nucleotide polymorphisms (SNPs) covering DRD2 were genotyped in 425 unrelated healthy Chinese undergraduates, and were further tested for association with insight problem solving. Both single SNP and haplotype analysis revealed several associations of DRD2 SNPs and haplotypes with insight problem solving. In conclusion, the present study provides the first evidence for the involvement of DRD2 in insight problem solving, future studies are necessary to validate these findings.
The Association of DRD2 with Insight Problem Solving
Zhang, Shun; Zhang, Jinghuan
2016-01-01
Although the insight phenomenon has attracted great attention from psychologists, it is still largely unknown whether its variation in well-functioning human adults has a genetic basis. Several lines of evidence suggest that genes involved in dopamine (DA) transmission might be potential candidates. The present study explored for the first time the association of dopamine D2 receptor gene (DRD2) with insight problem solving. Fifteen single-nucleotide polymorphisms (SNPs) covering DRD2 were genotyped in 425 unrelated healthy Chinese undergraduates, and were further tested for association with insight problem solving. Both single SNP and haplotype analysis revealed several associations of DRD2 SNPs and haplotypes with insight problem solving. In conclusion, the present study provides the first evidence for the involvement of DRD2 in insight problem solving, future studies are necessary to validate these findings. PMID:27933030
Pattern-set generation algorithm for the one-dimensional multiple stock sizes cutting stock problem
NASA Astrophysics Data System (ADS)
Cui, Yaodong; Cui, Yi-Ping; Zhao, Zhigang
2015-09-01
A pattern-set generation algorithm (PSG) for the one-dimensional multiple stock sizes cutting stock problem (1DMSSCSP) is presented. The solution process contains two stages. In the first stage, the PSG solves the residual problems repeatedly to generate the patterns in the pattern set, where each residual problem is solved by the column-generation approach, and each pattern is generated by solving a single large object placement problem. In the second stage, the integer linear programming model of the 1DMSSCSP is solved using a commercial solver, where only the patterns in the pattern set are considered. The computational results of benchmark instances indicate that the PSG outperforms existing heuristic algorithms and rivals the exact algorithm in solution quality.
NASA Astrophysics Data System (ADS)
Mashood, K. K.; Singh, Vijay A.
2013-09-01
Research suggests that problem-solving skills are transferable across domains. This claim, however, needs further empirical substantiation. We suggest correlation studies as a methodology for making preliminary inferences about transfer. The correlation of the physics performance of students with their performance in chemistry and mathematics in highly competitive problem-solving examinations was studied using a massive database. The sample sizes ranged from hundreds to a few hundred thousand. Encouraged by the presence of significant correlations, we interviewed 20 students to explore the pedagogic potential of physics in imparting transferable problem-solving skills. We report strategies and practices relevant to physics employed by these students which foster transfer.
NASA Astrophysics Data System (ADS)
Navas, Pedro; Sanavia, Lorenzo; López-Querol, Susana; Yu, Rena C.
2017-12-01
Solving dynamic problems for fluid saturated porous media at large deformation regime is an interesting but complex issue. An implicit time integration scheme is herein developed within the framework of the u-w (solid displacement-relative fluid displacement) formulation for the Biot's equations. In particular, liquid water saturated porous media is considered and the linearization of the linear momentum equations taking into account all the inertia terms for both solid and fluid phases is for the first time presented. The spatial discretization is carried out through a meshfree method, in which the shape functions are based on the principle of local maximum entropy LME. The current methodology is firstly validated with the dynamic consolidation of a soil column and the plastic shear band formulation of a square domain loaded by a rigid footing. The feasibility of this new numerical approach for solving large deformation dynamic problems is finally demonstrated through the application to an embankment problem subjected to an earthquake.
NASA Astrophysics Data System (ADS)
Maries, Alexandru; Singh, Chandralekha
2018-01-01
An appropriate diagram is a required element of a solution building process in physics problem solving and it can transform a given problem into a representation that is easier to exploit for solving the problem. A major focus while helping introductory physics students learn problem solving is to help them appreciate that drawing diagrams facilitates problem solving. We conducted an investigation in which two different interventions were implemented during recitation quizzes throughout the semester in a large enrolment, algebra-based introductory physics course. Students were either (1) asked to solve problems in which the diagrams were drawn for them or (2) explicitly told to draw a diagram. A comparison group was not given any instruction regarding diagrams. We developed a rubric to score the problem solving performance of students in different intervention groups. We investigated two problems involving electric field and electric force and found that students who drew productive diagrams were more successful problem solvers and that a higher level of relevant detail in a student’s diagram corresponded to a better score. We also conducted think-aloud interviews with nine students who were at the time taking an equivalent introductory algebra-based physics course in order to gain insight into how drawing diagrams affects the problem solving process. These interviews supported some of the interpretations of the quantitative results. We end by discussing instructional implications of the findings.
Ramirez, Gerardo; Chang, Hyesang; Maloney, Erin A; Levine, Susan C; Beilock, Sian L
2016-01-01
Even at young ages, children self-report experiencing math anxiety, which negatively relates to their math achievement. Leveraging a large dataset of first and second grade students' math achievement scores, math problem solving strategies, and math attitudes, we explored the possibility that children's math anxiety (i.e., a fear or apprehension about math) negatively relates to their use of more advanced problem solving strategies, which in turn relates to their math achievement. Our results confirm our hypothesis and, moreover, demonstrate that the relation between math anxiety and math problem solving strategies is strongest in children with the highest working memory capacity. Ironically, children who have the highest cognitive capacity avoid using advanced problem solving strategies when they are high in math anxiety and, as a result, underperform in math compared with their lower working memory peers. Copyright © 2015 Elsevier Inc. All rights reserved.
Structural design using equilibrium programming formulations
NASA Technical Reports Server (NTRS)
Scotti, Stephen J.
1995-01-01
Solutions to increasingly larger structural optimization problems are desired. However, computational resources are strained to meet this need. New methods will be required to solve increasingly larger problems. The present approaches to solving large-scale problems involve approximations for the constraints of structural optimization problems and/or decomposition of the problem into multiple subproblems that can be solved in parallel. An area of game theory, equilibrium programming (also known as noncooperative game theory), can be used to unify these existing approaches from a theoretical point of view (considering the existence and optimality of solutions), and be used as a framework for the development of new methods for solving large-scale optimization problems. Equilibrium programming theory is described, and existing design techniques such as fully stressed design and constraint approximations are shown to fit within its framework. Two new structural design formulations are also derived. The first new formulation is another approximation technique which is a general updating scheme for the sensitivity derivatives of design constraints. The second new formulation uses a substructure-based decomposition of the structure for analysis and sensitivity calculations. Significant computational benefits of the new formulations compared with a conventional method are demonstrated.
Vectorial finite elements for solving the radiative transfer equation
NASA Astrophysics Data System (ADS)
Badri, M. A.; Jolivet, P.; Rousseau, B.; Le Corre, S.; Digonnet, H.; Favennec, Y.
2018-06-01
The discrete ordinate method coupled with the finite element method is often used for the spatio-angular discretization of the radiative transfer equation. In this paper we attempt to improve upon such a discretization technique. Instead of using standard finite elements, we reformulate the radiative transfer equation using vectorial finite elements. In comparison to standard finite elements, this reformulation yields faster timings for the linear system assemblies, as well as for the solution phase when using scattering media. The proposed vectorial finite element discretization for solving the radiative transfer equation is cross-validated against a benchmark problem available in literature. In addition, we have used the method of manufactured solutions to verify the order of accuracy for our discretization technique within different absorbing, scattering, and emitting media. For solving large problems of radiation on parallel computers, the vectorial finite element method is parallelized using domain decomposition. The proposed domain decomposition method scales on large number of processes, and its performance is unaffected by the changes in optical thickness of the medium. Our parallel solver is used to solve a large scale radiative transfer problem of the Kelvin-cell radiation.
NASA Technical Reports Server (NTRS)
Morrell, R. A.; Odoherty, R. J.; Ramsey, H. R.; Reynolds, C. C.; Willoughby, J. K.; Working, R. D.
1975-01-01
Data and analyses related to a variety of algorithms for solving typical large-scale scheduling and resource allocation problems are presented. The capabilities and deficiencies of various alternative problem solving strategies are discussed from the viewpoint of computer system design.
Interaction Network Estimation: Predicting Problem-Solving Diversity in Interactive Environments
ERIC Educational Resources Information Center
Eagle, Michael; Hicks, Drew; Barnes, Tiffany
2015-01-01
Intelligent tutoring systems and computer aided learning environments aimed at developing problem solving produce large amounts of transactional data which make it a challenge for both researchers and educators to understand how students work within the environment. Researchers have modeled student-tutor interactions using complex networks in…
Solving Large Problems with a Small Working Memory
ERIC Educational Resources Information Center
Pizlo, Zygmunt; Stefanov, Emil
2013-01-01
We describe an important elaboration of our multiscale/multiresolution model for solving the Traveling Salesman Problem (TSP). Our previous model emulated the non-uniform distribution of receptors on the human retina and the shifts of visual attention. This model produced near-optimal solutions of TSP in linear time by performing hierarchical…
Problem Solving: Physics Modeling-Based Interactive Engagement
ERIC Educational Resources Information Center
Ornek, Funda
2009-01-01
The purpose of this study was to investigate how modeling-based instruction combined with an interactive-engagement teaching approach promotes students' problem solving abilities. I focused on students in a calculus-based introductory physics course, based on the matter and interactions curriculum of Chabay & Sherwood (2002) at a large state…
Integrating Study Skills and Problem Solving into Remedial Mathematics
ERIC Educational Resources Information Center
Cornick, Jonathan; Guy, G. Michael; Beckford, Ian
2015-01-01
Students at a large urban community college enrolled in seven classes of an experimental remedial algebra programme, which integrated study skills instruction and collaborative problem solving. A control group of seven classes was taught in a traditional lecture format without study skills instruction. Student performance in the course was…
Eskin, M; Savk, E; Uslu, M; Küçükaydoğan, N
2014-11-01
Psoriasis is a chronic dermatosis which may cause significant impairment of the patient's quality of life. The purpose of this study was to investigate the social problem-solving skills, perceived stress, negative life events, depression and life satisfaction in psoriasis patients. Data were gathered by means of questionnaires and clinical evaluations from 51 psoriatic patients and 51 matched healthy controls. Average disease duration was 16.47 years and average Psoriasis Area and Severity Index score was 3.67. Compared with the controls, the patients displayed lower social problem-solving skills. They displayed higher negative problem orientation and impulsive-careless problem-solving style scores than the controls. Patients tended also to show more avoidant problem-solving style and lower life satisfaction than controls. There was no difference between psoriatic patients and controls in terms of depression, perceived stress and negative life events. Higher social problem-solving skills were associated with lower depression, perceived stress and fewer numbers of negative life events but higher level of life satisfaction. The patient group largely included mild and moderate psoriatic cases. The findings of the study suggest that problem-solving training or therapy may be a suitable option for alleviating levels of psychological distress in patients suffering from psoriasis. © 2014 European Academy of Dermatology and Venereology.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chin, George
1999-01-11
A workshop on collaborative problem-solving environments (CPSEs) was held June 29 through July 1, 1999, in San Diego, California. The workshop was sponsored by the U.S. Department of Energy and the High Performance Network Applications Team of the Large Scale Networking Working Group. The workshop brought together researchers and developers from industry, academia, and government to identify, define, and discuss future directions in collaboration and problem-solving technologies in support of scientific research.
NASA Astrophysics Data System (ADS)
Maries, Alexandru; Singh, Chandralekha
2018-06-01
Drawing appropriate diagrams is a useful problem solving heuristic that can transform a problem into a representation that is easier to exploit for solving it. One major focus while helping introductory physics students learn effective problem solving is to help them understand that drawing diagrams can facilitate problem solution. We conducted an investigation in which two different interventions were implemented during recitation quizzes in a large enrollment algebra-based introductory physics course. Students were either (i) asked to solve problems in which the diagrams were drawn for them or (ii) explicitly told to draw a diagram. A comparison group was not given any instruction regarding diagrams. We developed rubrics to score the problem solving performance of students in different intervention groups and investigated ten problems. We found that students who were provided diagrams never performed better and actually performed worse than the other students on three problems, one involving standing sound waves in a tube (discussed elsewhere) and two problems in electricity which we focus on here. These two problems were the only problems in electricity that involved considerations of initial and final conditions, which may partly account for why students provided with diagrams performed significantly worse than students who were not provided with diagrams. In order to explore potential reasons for this finding, we conducted interviews with students and found that some students provided with diagrams may have spent less time on the conceptual analysis and planning stage of the problem solving process. In particular, those provided with the diagram were more likely to jump into the implementation stage of problem solving early without fully analyzing and understanding the problem, which can increase the likelihood of mistakes in solutions.
NASA Astrophysics Data System (ADS)
Reiter, D. T.; Rodi, W. L.
2015-12-01
Constructing 3D Earth models through the joint inversion of large geophysical data sets presents numerous theoretical and practical challenges, especially when diverse types of data and model parameters are involved. Among the challenges are the computational complexity associated with large data and model vectors and the need to unify differing model parameterizations, forward modeling methods and regularization schemes within a common inversion framework. The challenges can be addressed in part by decomposing the inverse problem into smaller, simpler inverse problems that can be solved separately, providing one knows how to merge the separate inversion results into an optimal solution of the full problem. We have formulated an approach to the decomposition of large inverse problems based on the augmented Lagrangian technique from optimization theory. As commonly done, we define a solution to the full inverse problem as the Earth model minimizing an objective function motivated, for example, by a Bayesian inference formulation. Our decomposition approach recasts the minimization problem equivalently as the minimization of component objective functions, corresponding to specified data subsets, subject to the constraints that the minimizing models be equal. A standard optimization algorithm solves the resulting constrained minimization problems by alternating between the separate solution of the component problems and the updating of Lagrange multipliers that serve to steer the individual solution models toward a common model solving the full problem. We are applying our inversion method to the reconstruction of the·crust and upper-mantle seismic velocity structure across Eurasia.· Data for the inversion comprise a large set of P and S body-wave travel times·and fundamental and first-higher mode Rayleigh-wave group velocities.
Biomimetics: its practice and theory.
Vincent, Julian F V; Bogatyreva, Olga A; Bogatyrev, Nikolaj R; Bowyer, Adrian; Pahl, Anja-Karina
2006-08-22
Biomimetics, a name coined by Otto Schmitt in the 1950s for the transfer of ideas and analogues from biology to technology, has produced some significant and successful devices and concepts in the past 50 years, but is still empirical. We show that TRIZ, the Russian system of problem solving, can be adapted to illuminate and manipulate this process of transfer. Analysis using TRIZ shows that there is only 12% similarity between biology and technology in the principles which solutions to problems illustrate, and while technology solves problems largely by manipulating usage of energy, biology uses information and structure, two factors largely ignored by technology.
Lexical Problems in Large Distributed Information Systems.
ERIC Educational Resources Information Center
Berkovich, Simon Ya; Shneiderman, Ben
1980-01-01
Suggests a unified concept of a lexical subsystem as part of an information system to deal with lexical problems in local and network environments. The linguistic and control functions of the lexical subsystems in solving problems for large computer systems are described, and references are included. (Author/BK)
Students' Explanations in Complex Learning of Disciplinary Programming
ERIC Educational Resources Information Center
Vieira, Camilo
2016-01-01
Computational Science and Engineering (CSE) has been denominated as the third pillar of science and as a set of important skills to solve the problems of a global society. Along with the theoretical and the experimental approaches, computation offers a third alternative to solve complex problems that require processing large amounts of data, or…
Computer-Based Assessment of Complex Problem Solving: Concept, Implementation, and Application
ERIC Educational Resources Information Center
Greiff, Samuel; Wustenberg, Sascha; Holt, Daniel V.; Goldhammer, Frank; Funke, Joachim
2013-01-01
Complex Problem Solving (CPS) skills are essential to successfully deal with environments that change dynamically and involve a large number of interconnected and partially unknown causal influences. The increasing importance of such skills in the 21st century requires appropriate assessment and intervention methods, which in turn rely on adequate…
Results and Implications of a Problem-Solving Treatment Program for Obesity.
ERIC Educational Resources Information Center
Mahoney, B. K.; And Others
Data are from a large scale experimental study which was designed to evaluate a multimethod problem solving approach to obesity. Obese adult volunteers (N=90) were randomly assigned to three groups: maximal treatment, minimal treatment, and no treatment control. In the two treatment groups, subjects were exposed to bibliographic material and…
The Development of Complex Problem Solving in Adolescence: A Latent Growth Curve Analysis
ERIC Educational Resources Information Center
Frischkorn, Gidon T.; Greiff, Samuel; Wüstenberg, Sascha
2014-01-01
Complex problem solving (CPS) as a cross-curricular competence has recently attracted more attention in educational psychology as indicated by its implementation in international educational large-scale assessments such as the Programme for International Student Assessment. However, research on the development of CPS is scarce, and the few…
ERIC Educational Resources Information Center
Mashood, K. K.; Singh, Vijay A.
2013-01-01
Research suggests that problem-solving skills are transferable across domains. This claim, however, needs further empirical substantiation. We suggest correlation studies as a methodology for making preliminary inferences about transfer. The correlation of the physics performance of students with their performance in chemistry and mathematics in…
Show, Don't Tell: Using Photographic "Snapsignments" to Advance and Assess Creative Problem Solving
ERIC Educational Resources Information Center
Machin, Jane E.
2016-01-01
Traditional assignments that aim to develop and evaluate creative problem solving skills are frequently foregone in large marketing classes due to the daunting grading prospect they present. Here, a new assessment method is introduced: the "snapsignment." Through photography, individual projects can be assigned that promote higher order…
The Effects of Feedback during Exploratory Mathematics Problem Solving: Prior Knowledge Matters
ERIC Educational Resources Information Center
Fyfe, Emily R.; Rittle-Johnson, Bethany; DeCaro, Marci S.
2012-01-01
Providing exploratory activities prior to explicit instruction can facilitate learning. However, the level of guidance provided during the exploration has largely gone unstudied. In this study, we examined the effects of 1 form of guidance, feedback, during exploratory mathematics problem solving for children with varying levels of prior domain…
Summer Proceedings 2016: The Center for Computing Research at Sandia National Laboratories
DOE Office of Scientific and Technical Information (OSTI.GOV)
Carleton, James Brian; Parks, Michael L.
Solving sparse linear systems from the discretization of elliptic partial differential equations (PDEs) is an important building block in many engineering applications. Sparse direct solvers can solve general linear systems, but are usually slower and use much more memory than effective iterative solvers. To overcome these two disadvantages, a hierarchical solver (LoRaSp) based on H2-matrices was introduced in [22]. Here, we have developed a parallel version of the algorithm in LoRaSp to solve large sparse matrices on distributed memory machines. On a single processor, the factorization time of our parallel solver scales almost linearly with the problem size for three-dimensionalmore » problems, as opposed to the quadratic scalability of many existing sparse direct solvers. Moreover, our solver leads to almost constant numbers of iterations, when used as a preconditioner for Poisson problems. On more than one processor, our algorithm has significant speedups compared to sequential runs. With this parallel algorithm, we are able to solve large problems much faster than many existing packages as demonstrated by the numerical experiments.« less
Six-Degree-of-Freedom Trajectory Optimization Utilizing a Two-Timescale Collocation Architecture
NASA Technical Reports Server (NTRS)
Desai, Prasun N.; Conway, Bruce A.
2005-01-01
Six-degree-of-freedom (6DOF) trajectory optimization of a reentry vehicle is solved using a two-timescale collocation methodology. This class of 6DOF trajectory problems are characterized by two distinct timescales in their governing equations, where a subset of the states have high-frequency dynamics (the rotational equations of motion) while the remaining states (the translational equations of motion) vary comparatively slowly. With conventional collocation methods, the 6DOF problem size becomes extraordinarily large and difficult to solve. Utilizing the two-timescale collocation architecture, the problem size is reduced significantly. The converged solution shows a realistic landing profile and captures the appropriate high-frequency rotational dynamics. A large reduction in the overall problem size (by 55%) is attained with the two-timescale architecture as compared to the conventional single-timescale collocation method. Consequently, optimum 6DOF trajectory problems can now be solved efficiently using collocation, which was not previously possible for a system with two distinct timescales in the governing states.
The mathematical statement for the solving of the problem of N-version software system design
NASA Astrophysics Data System (ADS)
Kovalev, I. V.; Kovalev, D. I.; Zelenkov, P. V.; Voroshilova, A. A.
2015-10-01
The N-version programming, as a methodology of the fault-tolerant software systems design, allows successful solving of the mentioned tasks. The use of N-version programming approach turns out to be effective, since the system is constructed out of several parallel executed versions of some software module. Those versions are written to meet the same specification but by different programmers. The problem of developing an optimal structure of N-version software system presents a kind of very complex optimization problem. This causes the use of deterministic optimization methods inappropriate for solving the stated problem. In this view, exploiting heuristic strategies looks more rational. In the field of pseudo-Boolean optimization theory, the so called method of varied probabilities (MVP) has been developed to solve problems with a large dimensionality.
He, Qiang; Hu, Xiangtao; Ren, Hong; Zhang, Hongqi
2015-11-01
A novel artificial fish swarm algorithm (NAFSA) is proposed for solving large-scale reliability-redundancy allocation problem (RAP). In NAFSA, the social behaviors of fish swarm are classified in three ways: foraging behavior, reproductive behavior, and random behavior. The foraging behavior designs two position-updating strategies. And, the selection and crossover operators are applied to define the reproductive ability of an artificial fish. For the random behavior, which is essentially a mutation strategy, the basic cloud generator is used as the mutation operator. Finally, numerical results of four benchmark problems and a large-scale RAP are reported and compared. NAFSA shows good performance in terms of computational accuracy and computational efficiency for large scale RAP. Copyright © 2015 ISA. Published by Elsevier Ltd. All rights reserved.
Estimation and Compression over Large Alphabets
ERIC Educational Resources Information Center
Acharya, Jayadev
2014-01-01
Compression, estimation, and prediction are basic problems in Information theory, statistics and machine learning. These problems have been extensively studied in all these fields, though the primary focus in a large portion of the work has been on understanding and solving the problems in the asymptotic regime, "i.e." the alphabet size…
AI tools in computer based problem solving
NASA Technical Reports Server (NTRS)
Beane, Arthur J.
1988-01-01
The use of computers to solve value oriented, deterministic, algorithmic problems, has evolved a structured life cycle model of the software process. The symbolic processing techniques used, primarily in research, for solving nondeterministic problems, and those for which an algorithmic solution is unknown, have evolved a different model, much less structured. Traditionally, the two approaches have been used completely independently. With the advent of low cost, high performance 32 bit workstations executing identical software with large minicomputers and mainframes, it became possible to begin to merge both models into a single extended model of computer problem solving. The implementation of such an extended model on a VAX family of micro/mini/mainframe systems is described. Examples in both development and deployment of applications involving a blending of AI and traditional techniques are given.
NASA Astrophysics Data System (ADS)
Prather, Edward E.; Wallace, Colin Scott
2018-06-01
We present an instructional framework that allowed a first time physics instructor to improve students quantitative problem solving abilities by more than a letter grade over what was achieved by students in an experienced instructor’s course. This instructional framework uses a Think-Pair-Share approach to foster collaborative quantitative problem solving during the lecture portion of a large enrollment introductory calculus-based mechanics course. Through the development of carefully crafted and sequenced TPS questions, we engage students in rich discussions on key problem solving issues that we typically only hear about when a student comes for help during office hours. Current work in the sophomore E&M course illustrates that this framework is generalizable to classes beyond the introductory level and for topics beyond mechanics.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Moryakov, A. V., E-mail: sailor@yauza.ru; Pylyov, S. S.
This paper presents the formulation of the problem and the methodical approach for solving large systems of linear differential equations describing nonstationary processes with the use of CUDA technology; this approach is implemented in the ANGEL program. Results for a test problem on transport of radioactive products over loops of a nuclear power plant are given. The possibilities for the use of the ANGEL program for solving various problems that simulate arbitrary nonstationary processes are discussed.
ERIC Educational Resources Information Center
Lau, Matthew Y.; Sieler, Jay D.; Muyskens, Paul; Canter, Andrea; VanKeuren, Barbara; Marston, Doug
2006-01-01
The Minneapolis Public School System has been implementing an intervention-based approach to special education placement. This Problem-Solving Model (PSM) was designed to de-emphasize the role of norm-referenced tests and to provide early instructional interventions. The basic outline of the PSM is to define the problem, determine the best…
Improving insight and non-insight problem solving with brief interventions.
Wen, Ming-Ching; Butler, Laurie T; Koutstaal, Wilma
2013-02-01
Developing brief training interventions that benefit different forms of problem solving is challenging. In earlier research, Chrysikou (2006) showed that engaging in a task requiring generation of alternative uses of common objects improved subsequent insight problem solving. These benefits were attributed to a form of implicit transfer of processing involving enhanced construction of impromptu, on-the-spot or 'ad hoc' goal-directed categorizations of the problem elements. Following this, it is predicted that the alternative uses exercise should benefit abilities that govern goal-directed behaviour, such as fluid intelligence and executive functions. Similarly, an indirect intervention - self-affirmation (SA) - that has been shown to enhance cognitive and executive performance after self-regulation challenge and when under stereotype threat, may also increase adaptive goal-directed thinking and likewise should bolster problem-solving performance. In Experiment 1, brief single-session interventions, involving either alternative uses generation or SA, significantly enhanced both subsequent insight and visual-spatial fluid reasoning problem solving. In Experiment 2, we replicated the finding of benefits of both alternative uses generation and SA on subsequent insight problem-solving performance, and demonstrated that the underlying mechanism likely involves improved executive functioning. Even brief cognitive- and social-psychological interventions may substantially bolster different types of problem solving and may exert largely similar facilitatory effects on goal-directed behaviours. © 2012 The British Psychological Society.
NASA Astrophysics Data System (ADS)
Chen, Zhongzhou; Demirci, Neset; Choi, Youn-Jeng; Pritchard, David E.
2017-06-01
Previous research on problem diagrams suggested that including a supportive diagram, one that does not provide necessary problem solving information, may bring little, or even negative, benefit to students' problem solving success. We tested the usefulness of problem diagrams on 12 different physics problems (6A/B experiments) in our massive open online course. By analyzing over 8000 student responses in total, we found that including a problem diagram that contains no significant additional information only slightly improves the first attempt correct rate for the few most spatially complex problems, and has little impact on either the final correct percentage or the time spent on solving the problem. On the other hand, in half of the cases, removing the diagram significantly increased the fraction of students' drawing their own diagrams during problem solving. The increase in drawing behavior is largely independent of students' physics abilities. In summary, our results suggest that for many physics problems, the benefit of a diagram is exceedingly small and may not justify the effort of creating one.
The fastclime Package for Linear Programming and Large-Scale Precision Matrix Estimation in R.
Pang, Haotian; Liu, Han; Vanderbei, Robert
2014-02-01
We develop an R package fastclime for solving a family of regularized linear programming (LP) problems. Our package efficiently implements the parametric simplex algorithm, which provides a scalable and sophisticated tool for solving large-scale linear programs. As an illustrative example, one use of our LP solver is to implement an important sparse precision matrix estimation method called CLIME (Constrained L 1 Minimization Estimator). Compared with existing packages for this problem such as clime and flare, our package has three advantages: (1) it efficiently calculates the full piecewise-linear regularization path; (2) it provides an accurate dual certificate as stopping criterion; (3) it is completely coded in C and is highly portable. This package is designed to be useful to statisticians and machine learning researchers for solving a wide range of problems.
A family of conjugate gradient methods for large-scale nonlinear equations.
Feng, Dexiang; Sun, Min; Wang, Xueyong
2017-01-01
In this paper, we present a family of conjugate gradient projection methods for solving large-scale nonlinear equations. At each iteration, it needs low storage and the subproblem can be easily solved. Compared with the existing solution methods for solving the problem, its global convergence is established without the restriction of the Lipschitz continuity on the underlying mapping. Preliminary numerical results are reported to show the efficiency of the proposed method.
Designs for Operationalizing Collaborative Problem Solving for Automated Assessment
ERIC Educational Resources Information Center
Scoular, Claire; Care, Esther; Hesse, Friedrich W.
2017-01-01
Collaborative problem solving is a complex skill set that draws on social and cognitive factors. The construct remains in its infancy due to lack of empirical evidence that can be drawn upon for validation. The differences and similarities between two large-scale initiatives that reflect this state of the art, in terms of underlying assumptions…
General Problem-Solving Styles and Problem-Solving Approaches in Video Games
ERIC Educational Resources Information Center
Hamlen, Karla R.
2018-01-01
Video game play is a pervasive recreational activity, particularly among college students. While there is a large research base focused on educational video game play and uses of games in the classroom, there is much less research focused on cognitive strategies and entertainment video game play. The purpose of this study was to investigate…
Elementary Teachers' Perspectives of Mathematics Problem Solving Strategies
ERIC Educational Resources Information Center
Bruun, Faye
2013-01-01
Participants in this study were asked to report what strategies were most often used in their attempts to foster their students' problem solving abilities. Participants included 70 second through fifth-grade elementary teachers from 42 schools in a large state of the south central region in the U.S. Data analyses of the interviews revealed that…
VET Workers' Problem-Solving Skills in Technology-Rich Environments: European Approach
ERIC Educational Resources Information Center
Hämäläinen, Raija; Cincinnato, Sebastiano; Malin, Antero; De Wever, Bram
2014-01-01
The European workplace is challenging VET adults' problem-solving skills in technology-rich environments (TREs). So far, no international large-scale assessment data has been available for VET. The PIAAC data comprise the most comprehensive source of information on adults' skills to date. The present study (N = 50 369) focuses on gaining insight…
ERIC Educational Resources Information Center
Greiff, Samuel; Wustenberg, Sascha; Molnar, Gyongyver; Fischer, Andreas; Funke, Joachim; Csapo, Beno
2013-01-01
Innovative assessments of cross-curricular competencies such as complex problem solving (CPS) have currently received considerable attention in large-scale educational studies. This study investigated the nature of CPS by applying a state-of-the-art approach to assess CPS in high school. We analyzed whether two processes derived from cognitive…
ERIC Educational Resources Information Center
Choi, Jung-Min
2010-01-01
The primary concern in current interaction design is focused on how to help users solve problems and achieve goals more easily and efficiently. While users' sufficient knowledge acquisition of operating a product or system is considered important, their acquisition of problem-solving knowledge in the task domain has largely been disregarded. As a…
ERIC Educational Resources Information Center
Heylen, Christel; Smet, Marc; Buelens, Hermans; Sloten, Jos Vander
2007-01-01
A present-day engineer has a large scientific knowledge; he is a team-player, eloquent communicator and life-long learner. At the Katholieke Universiteit Leuven, the course "Problem Solving and Engineering Design" introduces engineering students from the first semester onwards into real engineering practice and teamwork. Working in small…
Assessment of Complex Problem Solving: What We Know and What We Don't Know
ERIC Educational Resources Information Center
Herde, Christoph Nils; Wüstenberg, Sascha; Greiff, Samuel
2016-01-01
Complex Problem Solving (CPS) is seen as a cross-curricular 21st century skill that has attracted interest in large-scale-assessments. In the Programme for International Student Assessment (PISA) 2012, CPS was assessed all over the world to gain information on students' skills to acquire and apply knowledge while dealing with nontransparent…
A numerical projection technique for large-scale eigenvalue problems
NASA Astrophysics Data System (ADS)
Gamillscheg, Ralf; Haase, Gundolf; von der Linden, Wolfgang
2011-10-01
We present a new numerical technique to solve large-scale eigenvalue problems. It is based on the projection technique, used in strongly correlated quantum many-body systems, where first an effective approximate model of smaller complexity is constructed by projecting out high energy degrees of freedom and in turn solving the resulting model by some standard eigenvalue solver. Here we introduce a generalization of this idea, where both steps are performed numerically and which in contrast to the standard projection technique converges in principle to the exact eigenvalues. This approach is not just applicable to eigenvalue problems encountered in many-body systems but also in other areas of research that result in large-scale eigenvalue problems for matrices which have, roughly speaking, mostly a pronounced dominant diagonal part. We will present detailed studies of the approach guided by two many-body models.
Solving large scale traveling salesman problems by chaotic neurodynamics.
Hasegawa, Mikio; Ikeguch, Tohru; Aihara, Kazuyuki
2002-03-01
We propose a novel approach for solving large scale traveling salesman problems (TSPs) by chaotic dynamics. First, we realize the tabu search on a neural network, by utilizing the refractory effects as the tabu effects. Then, we extend it to a chaotic neural network version. We propose two types of chaotic searching methods, which are based on two different tabu searches. While the first one requires neurons of the order of n2 for an n-city TSP, the second one requires only n neurons. Moreover, an automatic parameter tuning method of our chaotic neural network is presented for easy application to various problems. Last, we show that our method with n neurons is applicable to large TSPs such as an 85,900-city problem and exhibits better performance than the conventional stochastic searches and the tabu searches.
Total-variation based velocity inversion with Bregmanized operator splitting algorithm
NASA Astrophysics Data System (ADS)
Zand, Toktam; Gholami, Ali
2018-04-01
Many problems in applied geophysics can be formulated as a linear inverse problem. The associated problems, however, are large-scale and ill-conditioned. Therefore, regularization techniques are needed to be employed for solving them and generating a stable and acceptable solution. We consider numerical methods for solving such problems in this paper. In order to tackle the ill-conditioning of the problem we use blockiness as a prior information of the subsurface parameters and formulate the problem as a constrained total variation (TV) regularization. The Bregmanized operator splitting (BOS) algorithm as a combination of the Bregman iteration and the proximal forward backward operator splitting method is developed to solve the arranged problem. Two main advantages of this new algorithm are that no matrix inversion is required and that a discrepancy stopping criterion is used to stop the iterations, which allow efficient solution of large-scale problems. The high performance of the proposed TV regularization method is demonstrated using two different experiments: 1) velocity inversion from (synthetic) seismic data which is based on Born approximation, 2) computing interval velocities from RMS velocities via Dix formula. Numerical examples are presented to verify the feasibility of the proposed method for high-resolution velocity inversion.
A scalable parallel algorithm for multiple objective linear programs
NASA Technical Reports Server (NTRS)
Wiecek, Malgorzata M.; Zhang, Hong
1994-01-01
This paper presents an ADBASE-based parallel algorithm for solving multiple objective linear programs (MOLP's). Job balance, speedup and scalability are of primary interest in evaluating efficiency of the new algorithm. Implementation results on Intel iPSC/2 and Paragon multiprocessors show that the algorithm significantly speeds up the process of solving MOLP's, which is understood as generating all or some efficient extreme points and unbounded efficient edges. The algorithm gives specially good results for large and very large problems. Motivation and justification for solving such large MOLP's are also included.
Exact solution of large asymmetric traveling salesman problems.
Miller, D L; Pekny, J F
1991-02-15
The traveling salesman problem is one of a class of difficult problems in combinatorial optimization that is representative of a large number of important scientific and engineering problems. A survey is given of recent applications and methods for solving large problems. In addition, an algorithm for the exact solution of the asymmetric traveling salesman problem is presented along with computational results for several classes of problems. The results show that the algorithm performs remarkably well for some classes of problems, determining an optimal solution even for problems with large numbers of cities, yet for other classes, even small problems thwart determination of a provably optimal solution.
Inverse problems in the design, modeling and testing of engineering systems
NASA Technical Reports Server (NTRS)
Alifanov, Oleg M.
1991-01-01
Formulations, classification, areas of application, and approaches to solving different inverse problems are considered for the design of structures, modeling, and experimental data processing. Problems in the practical implementation of theoretical-experimental methods based on solving inverse problems are analyzed in order to identify mathematical models of physical processes, aid in input data preparation for design parameter optimization, help in design parameter optimization itself, and to model experiments, large-scale tests, and real tests of engineering systems.
Fuchs, Lynn S; Seethaler, Pamela M; Powell, Sarah R; Fuchs, Douglas; Hamlett, Carol L; Fletcher, Jack M
2008-01-01
This study assessed the effects of preventative tutoring on the math problem solving of third-grade students with math and reading difficulties. Students (n = 35) were assigned randomly to continue in their general education math program or to receive secondary preventative tutoring 3 times per week, 30 min per session, for 12 weeks. Schema-broadening tutoring taught students to (a) focus on the mathematical structure of 3 problem types; (b) recognize problems as belonging to those 3 problem-type schemas; (c) solve the 3 word-problem types; and (d) transfer solution methods to problems that include irrelevant information, 2-digit operands, missing information in the first or second positions in the algebraic equation, or relevant information in charts, graphs, and pictures. Also, students were taught to perform the calculation and algebraic skills foundational for problem solving. Analyses of variance revealed statistically significant effects on a wide range of word problems, with large effect sizes. Findings support the efficacy of the tutoring protocol for preventing word-problem deficits among third-grade students with math and reading deficits.
Fuchs, Lynn S.; Seethaler, Pamela M.; Powell, Sarah R.; Fuchs, Douglas; Hamlett, Carol L.; Fletcher, Jack M.
2009-01-01
This study assessed the effects of preventative tutoring on the math problem solving of third-grade students with math and reading difficulties. Students (n = 35) were assigned randomly to continue in their general education math program or to receive secondary preventative tutoring 3 times per week, 30 min per session, for 12 weeks. Schema-broadening tutoring taught students to (a) focus on the mathematical structure of 3 problem types; (b) recognize problems as belonging to those 3 problem-type schemas; (c) solve the 3 word-problem types; and (d) transfer solution methods to problems that include irrelevant information, 2-digit operands, missing information in the first or second positions in the algebraic equation, or relevant information in charts, graphs, and pictures. Also, students were taught to perform the calculation and algebraic skills foundational for problem solving. Analyses of variance revealed statistically significant effects on a wide range of word problems, with large effect sizes. Findings support the efficacy of the tutoring protocol for preventing word-problem deficits among third-grade students with math and reading deficits. PMID:20209074
Anderson, William L; Mitchell, Steven M; Osgood, Marcy P
2008-01-01
For the past 3 yr, faculty at the University of New Mexico, Department of Biochemistry and Molecular Biology have been using interactive online Problem-Based Learning (PBL) case discussions in our large-enrollment classes. We have developed an illustrative tracking method to monitor student use of problem-solving strategies to provide targeted help to groups and to individual students. This method of assessing performance has a high interrater reliability, and senior students, with training, can serve as reliable graders. We have been able to measure improvements in many students' problem-solving strategies, but, not unexpectedly, there is a population of students who consistently apply the same failing strategy when there is no faculty intervention. This new methodology provides an effective tool to direct faculty to constructively intercede in this area of student development.
Biomimetics: its practice and theory
Vincent, Julian F.V; Bogatyreva, Olga A; Bogatyrev, Nikolaj R; Bowyer, Adrian; Pahl, Anja-Karina
2006-01-01
Biomimetics, a name coined by Otto Schmitt in the 1950s for the transfer of ideas and analogues from biology to technology, has produced some significant and successful devices and concepts in the past 50 years, but is still empirical. We show that TRIZ, the Russian system of problem solving, can be adapted to illuminate and manipulate this process of transfer. Analysis using TRIZ shows that there is only 12% similarity between biology and technology in the principles which solutions to problems illustrate, and while technology solves problems largely by manipulating usage of energy, biology uses information and structure, two factors largely ignored by technology. PMID:16849244
Software environment for implementing engineering applications on MIMD computers
NASA Technical Reports Server (NTRS)
Lopez, L. A.; Valimohamed, K. A.; Schiff, S.
1990-01-01
In this paper the concept for a software environment for developing engineering application systems for multiprocessor hardware (MIMD) is presented. The philosophy employed is to solve the largest problems possible in a reasonable amount of time, rather than solve existing problems faster. In the proposed environment most of the problems concerning parallel computation and handling of large distributed data spaces are hidden from the application program developer, thereby facilitating the development of large-scale software applications. Applications developed under the environment can be executed on a variety of MIMD hardware; it protects the application software from the effects of a rapidly changing MIMD hardware technology.
Can goal-free problems facilitating students' flexible thinking?
NASA Astrophysics Data System (ADS)
Maulidya, Sity Rahmy; Hasanah, Rusi Ulfa; Retnowati, Endah
2017-08-01
Problem solving is the key of doing and also learning mathematics. It takes also the fundamental role of developing mathematical knowledge. Responding to the current reform movement in mathematics, students are expected to learn to be a flexible thinker. The ability to think flexible is challenged by the globalisation, hence influence mathematics education. A flexible thinking includes ability to apply knowledge in different contexts rather than simply use it in similar context when it is studied. Arguably problem solving activities can contribute to the development of the ability to apply skills to unfamiliar situations. Accordingly, an appropriate classroom instructional strategy must be developed. A cognitive load theory suggests that by reducing extraneous cognitive load during learning could enhance transfer learning. A goal-free problem strategy that is developed based in cognitive load theory have been showed to be effective for transfer learning. This strategy enables students to learn a large numbers of problem solving moves from a mathematics problem. The instruction in a goal-free problem directs students to `calculate as many solution as you can' rather than to calculate a single given goal. Many experiment research evident goal-free problem enhance learning. This literature review will discuss evidence goal-free problem facilitate students to solve problems flexibly and thus enhance their problem solving skills, including how its implication in the classroom.
The pseudo-Boolean optimization approach to form the N-version software structure
NASA Astrophysics Data System (ADS)
Kovalev, I. V.; Kovalev, D. I.; Zelenkov, P. V.; Voroshilova, A. A.
2015-10-01
The problem of developing an optimal structure of N-version software system presents a kind of very complex optimization problem. This causes the use of deterministic optimization methods inappropriate for solving the stated problem. In this view, exploiting heuristic strategies looks more rational. In the field of pseudo-Boolean optimization theory, the so called method of varied probabilities (MVP) has been developed to solve problems with a large dimensionality. Some additional modifications of MVP have been made to solve the problem of N-version systems design. Those algorithms take into account the discovered specific features of the objective function. The practical experiments have shown the advantage of using these algorithm modifications because of reducing a search space.
NASA Technical Reports Server (NTRS)
Ferencz, Donald C.; Viterna, Larry A.
1991-01-01
ALPS is a computer program which can be used to solve general linear program (optimization) problems. ALPS was designed for those who have minimal linear programming (LP) knowledge and features a menu-driven scheme to guide the user through the process of creating and solving LP formulations. Once created, the problems can be edited and stored in standard DOS ASCII files to provide portability to various word processors or even other linear programming packages. Unlike many math-oriented LP solvers, ALPS contains an LP parser that reads through the LP formulation and reports several types of errors to the user. ALPS provides a large amount of solution data which is often useful in problem solving. In addition to pure linear programs, ALPS can solve for integer, mixed integer, and binary type problems. Pure linear programs are solved with the revised simplex method. Integer or mixed integer programs are solved initially with the revised simplex, and the completed using the branch-and-bound technique. Binary programs are solved with the method of implicit enumeration. This manual describes how to use ALPS to create, edit, and solve linear programming problems. Instructions for installing ALPS on a PC compatible computer are included in the appendices along with a general introduction to linear programming. A programmers guide is also included for assistance in modifying and maintaining the program.
NASA Astrophysics Data System (ADS)
Rakkapao, S.; Pengpan, T.; Srikeaw, S.; Prasitpong, S.
2014-01-01
This study aims to investigate the use of the predict-observe-explain (POE) approach integrated into large lecture classes on forces and motion. It is compared to the instructor-led problem-solving method using model analysis. The samples are science (SC, N = 420) and engineering (EN, N = 434) freshmen, from Prince of Songkla University, Thailand. Research findings from the force and motion conceptual evaluation indicate that the multimedia-supported POE method promotes students’ learning better than the problem-solving method, in particular for the velocity and acceleration concepts. There is a small shift of the students’ model states after the problem-solving instruction. Moreover, by using model analysis instructors are able to investigate students’ misconceptions and evaluate teaching methods. It benefits instructors in organizing subsequent instructional materials.
An investigation of the effects of interventions on problem-solving strategies and abilities
NASA Astrophysics Data System (ADS)
Cox, Charles Terrence, Jr.
Problem-solving has been described as being the "heart" of the chemistry classroom, and students' development of problem-solving skills is essential for their success in chemistry. Despite the importance of problem-solving, there has been little research within the chemistry domain, largely because of the lack of tools to collect data for large populations. Problem-solving was assessed using a software package known as IMMEX (for Interactive Multimedia Exercises) which has an HTML tracking feature that allows for collection of problem-solving data in the background as students work the problems. The primary goal of this research was to develop methods (known as interventions) that could promote improvements in students' problem-solving and most notably aid in their transition from the novice to competent level. Three intervention techniques that were incorporated within the chemistry curricula: collaborative grouping (face-to-face and distance), concept mapping, and peer-led team learning. The face-to-face collaborative grouping intervention was designed to probe the factors affecting the quality of the group interaction. Students' logical reasoning abilities were measured using the Group Assessment of Logical Thinking (GALT) test which classifies students as formal, transitional, or concrete. These classifications essentially provide a basis for identifying scientific aptitude. These designations were used as the basis for forming collaborative groups of two students. The six possibilities (formal-formal, formal-transitional, etc.) were formed to determine how the group composition influences the gains in student abilities observed from collaborative grouping interventions. Students were given three assignments (an individual pre-collaborative, an individual post collaborative, and a collaborative assignment) each requiring them to work an IMMEX problem set. Similar gains in performance of 10% gains were observed for each group with two exceptions. The transitional students who were paired with concrete students had a 15% gain, and the concrete students paired with other concrete students had only a marginal gain. In fact, there was no statistical difference in the pre-collaborative and post-collaborative student abilities for concrete-concrete groups. The distance collaborative intervention was completed using a new interface for the IMMEX software designed to mimic face-to-face collaboration. A stereochemistry problem set which had a solved rate of 28% prior to collaboration was chosen for incorporation into this distance collaboration study. (Abstract shortened by UMI.)
ERIC Educational Resources Information Center
Van Dooren, Wim; De Bock, Dirk; Verschaffel, Lieven
2010-01-01
This study builds on two lines of research that have so far developed largely separately: the use of additive methods to solve proportional word problems and the use of proportional methods to solve additive word problems. We investigated the development with age of both kinds of erroneous solution methods. We gave a test containing missing-value…
Custers, Eugène J F M
2013-08-01
Recently, human reasoning, problem solving, and decision making have been viewed as products of two separate systems: "System 1," the unconscious, intuitive, or nonanalytic system, and "System 2," the conscious, analytic, or reflective system. This view has penetrated the medical education literature, yet the idea of two independent dichotomous cognitive systems is not entirely without problems.This article outlines the difficulties of this "two-system view" and presents an alternative, developed by K.R. Hammond and colleagues, called cognitive continuum theory (CCT). CCT is featured by three key assumptions. First, human reasoning, problem solving, and decision making can be arranged on a cognitive continuum, with pure intuition at one end, pure analysis at the other, and a large middle ground called "quasirationality." Second, the nature and requirements of the cognitive task, as perceived by the person performing the task, determine to a large extent whether a task will be approached more intuitively or more analytically. Third, for optimal task performance, this approach needs to match the cognitive properties and requirements of the task. Finally, the author makes a case that CCT is better able than a two-system view to describe medical problem solving and clinical reasoning and that it provides clear clues for how to organize training in clinical reasoning.
Empirical results on scheduling and dynamic backtracking
NASA Technical Reports Server (NTRS)
Boddy, Mark S.; Goldman, Robert P.
1994-01-01
At the Honeywell Technology Center (HTC), we have been working on a scheduling problem related to commercial avionics. This application is large, complex, and hard to solve. To be a little more concrete: 'large' means almost 20,000 activities, 'complex' means several activity types, periodic behavior, and assorted types of temporal constraints, and 'hard to solve' means that we have been unable to eliminate backtracking through the use of search heuristics. At this point, we can generate solutions, where solutions exist, or report failure and sometimes why the system failed. To the best of our knowledge, this is among the largest and most complex scheduling problems to have been solved as a constraint satisfaction problem, at least that has appeared in the published literature. This abstract is a preliminary report on what we have done and how. In the next section, we present our approach to treating scheduling as a constraint satisfaction problem. The following sections present the application in more detail and describe how we solve scheduling problems in the application domain. The implemented system makes use of Ginsberg's Dynamic Backtracking algorithm, with some minor extensions to improve its utility for scheduling. We describe those extensions and the performance of the resulting system. The paper concludes with some general remarks, open questions and plans for future work.
A globally convergent LCL method for nonlinear optimization.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Friedlander, M. P.; Saunders, M. A.; Mathematics and Computer Science
2005-01-01
For optimization problems with nonlinear constraints, linearly constrained Lagrangian (LCL) methods solve a sequence of subproblems of the form 'minimize an augmented Lagrangian function subject to linearized constraints.' Such methods converge rapidly near a solution but may not be reliable from arbitrary starting points. Nevertheless, the well-known software package MINOS has proved effective on many large problems. Its success motivates us to derive a related LCL algorithm that possesses three important properties: it is globally convergent, the subproblem constraints are always feasible, and the subproblems may be solved inexactly. The new algorithm has been implemented in Matlab, with an optionmore » to use either MINOS or SNOPT (Fortran codes) to solve the linearly constrained subproblems. Only first derivatives are required. We present numerical results on a subset of the COPS, HS, and CUTE test problems, which include many large examples. The results demonstrate the robustness and efficiency of the stabilized LCL procedure.« less
Construction, testing and development of large wind energy facilities
NASA Technical Reports Server (NTRS)
Windheim, R. (Editor); Cuntze, R. (Editor)
1982-01-01
Building large rotor blades and control of oscillations in large facilities are discussed. It is concluded that the technical problems in the design of large rotor blades and control of oscillations can be solved.
Chen, Zhe; Honomichl, Ryan; Kennedy, Diane; Tan, Enda
2016-06-01
The present study examines 5- to 8-year-old children's relation reasoning in solving matrix completion tasks. This study incorporates a componential analysis, an eye-tracking method, and a microgenetic approach, which together allow an investigation of the cognitive processing strategies involved in the development and learning of children's relational thinking. Developmental differences in problem-solving performance were largely due to deficiencies in engaging the processing strategies that are hypothesized to facilitate problem-solving performance. Feedback designed to highlight the relations between objects within the matrix improved 5- and 6-year-olds' problem-solving performance, as well as their use of appropriate processing strategies. Furthermore, children who engaged the processing strategies early on in the task were more likely to solve subsequent problems in later phases. These findings suggest that encoding relations, integrating rules, completing the model, and generalizing strategies across tasks are critical processing components that underlie relational thinking. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
Aras, N; Altinel, I K; Oommen, J
2003-01-01
In addition to the classical heuristic algorithms of operations research, there have also been several approaches based on artificial neural networks for solving the traveling salesman problem. Their efficiency, however, decreases as the problem size (number of cities) increases. A technique to reduce the complexity of a large-scale traveling salesman problem (TSP) instance is to decompose or partition it into smaller subproblems. We introduce an all-neural decomposition heuristic that is based on a recent self-organizing map called KNIES, which has been successfully implemented for solving both the Euclidean traveling salesman problem and the Euclidean Hamiltonian path problem. Our solution for the Euclidean TSP proceeds by solving the Euclidean HPP for the subproblems, and then patching these solutions together. No such all-neural solution has ever been reported.
NASA Astrophysics Data System (ADS)
Plestenjak, Bor; Gheorghiu, Călin I.; Hochstenbach, Michiel E.
2015-10-01
In numerous science and engineering applications a partial differential equation has to be solved on some fairly regular domain that allows the use of the method of separation of variables. In several orthogonal coordinate systems separation of variables applied to the Helmholtz, Laplace, or Schrödinger equation leads to a multiparameter eigenvalue problem (MEP); important cases include Mathieu's system, Lamé's system, and a system of spheroidal wave functions. Although multiparameter approaches are exploited occasionally to solve such equations numerically, MEPs remain less well known, and the variety of available numerical methods is not wide. The classical approach of discretizing the equations using standard finite differences leads to algebraic MEPs with large matrices, which are difficult to solve efficiently. The aim of this paper is to change this perspective. We show that by combining spectral collocation methods and new efficient numerical methods for algebraic MEPs it is possible to solve such problems both very efficiently and accurately. We improve on several previous results available in the literature, and also present a MATLAB toolbox for solving a wide range of problems.
Unequal-area, fixed-shape facility layout problems using the firefly algorithm
NASA Astrophysics Data System (ADS)
Ingole, Supriya; Singh, Dinesh
2017-07-01
In manufacturing industries, the facility layout design is a very important task, as it is concerned with the overall manufacturing cost and profit of the industry. The facility layout problem (FLP) is solved by arranging the departments or facilities of known dimensions on the available floor space. The objective of this article is to implement the firefly algorithm (FA) for solving unequal-area, fixed-shape FLPs and optimizing the costs of total material handling and transportation between the facilities. The FA is a nature-inspired algorithm and can be used for combinatorial optimization problems. Benchmark problems from the previous literature are solved using the FA. To check its effectiveness, it is implemented to solve large-sized FLPs. Computational results obtained using the FA show that the algorithm is less time consuming and the total layout costs for FLPs are better than the best results achieved so far.
A Game Based e-Learning System to Teach Artificial Intelligence in the Computer Sciences Degree
ERIC Educational Resources Information Center
de Castro-Santos, Amable; Fajardo, Waldo; Molina-Solana, Miguel
2017-01-01
Our students taking the Artificial Intelligence and Knowledge Engineering courses often encounter a large number of problems to solve which are not directly related to the subject to be learned. To solve this problem, we have developed a game based e-learning system. The elected game, that has been implemented as an e-learning system, allows to…
Vieira, J; Cunha, M C
2011-01-01
This article describes a solution method of solving large nonlinear problems in two steps. The two steps solution approach takes advantage of handling smaller and simpler models and having better starting points to improve solution efficiency. The set of nonlinear constraints (named as complicating constraints) which makes the solution of the model rather complex and time consuming is eliminated from step one. The complicating constraints are added only in the second step so that a solution of the complete model is then found. The solution method is applied to a large-scale problem of conjunctive use of surface water and groundwater resources. The results obtained are compared with solutions determined with the direct solve of the complete model in one single step. In all examples the two steps solution approach allowed a significant reduction of the computation time. This potential gain of efficiency of the two steps solution approach can be extremely important for work in progress and it can be particularly useful for cases where the computation time would be a critical factor for having an optimized solution in due time.
1984-04-01
numerical solution, of sstem ot stiff Wh-f Cr ODs. Fro- qontl. a substantial portia of the total computationskwok and cooap required! to solve stiff...exep, possl- bly, foreciadalms of problem. That is% a syste of linewat o nonlinear algebrac equa- tion mumt be solved at auk step of the numerical ...onjugate gradient method [431 is a mall-know ezuze, have prove to be particularly -2- efecti for solving the linear stwem that &ise in the numerical
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dul, F.A.; Arczewski, K.
1994-03-01
Although it has been stated that [open quotes]an attempt to solve (very large problems) by subspace iterations seems futile[close quotes], we will show that the statement is not true, especially for extremely large eigenproblems. In this paper a new two-phase subspace iteration/Rayleigh quotient/conjugate gradient method for generalized, large, symmetric eigenproblems Ax = [lambda]Bx is presented. It has the ability of solving extremely large eigenproblems, N = 216,000, for example, and finding a large number of leftmost or rightmost eigenpairs, up to 1000 or more. Multiple eigenpairs, even those with multiplicity 100, can be easily found. The use of the proposedmore » method for solving the big full eigenproblems (N [approximately] 10[sup 3]), as well as for large weakly non-symmetric eigenproblems, have been considered also. The proposed method is fully iterative; thus the factorization of matrices ins avoided. The key idea consists in joining two methods: subspace and Rayleigh quotient iterations. The systems of indefinite and almost singular linear equations (a - [sigma]B)x = By are solved by various iterative conjugate gradient method can be used without danger of breaking down due to its property that may be called [open quotes]self-correction towards the eigenvector,[close quotes] discovered recently by us. The use of various preconditioners (SSOR and IC) has also been considered. The main features of the proposed method have been analyzed in detail. Comparisons with other methods, such as, accelerated subspace iteration, Lanczos, Davidson, TLIME, TRACMN, and SRQMCG, are presented. The results of numerical tests for various physical problems (acoustic, vibrations of structures, quantum chemistry) are presented as well. 40 refs., 12 figs., 2 tabs.« less
Klegeris, Andis; Hurren, Heather
2011-12-01
Problem-based learning (PBL) can be described as a learning environment where the problem drives the learning. This technique usually involves learning in small groups, which are supervised by tutors. It is becoming evident that PBL in a small-group setting has a robust positive effect on student learning and skills, including better problem-solving skills and an increase in overall motivation. However, very little research has been done on the educational benefits of PBL in a large classroom setting. Here, we describe a PBL approach (using tutorless groups) that was introduced as a supplement to standard didactic lectures in University of British Columbia Okanagan undergraduate biochemistry classes consisting of 45-85 students. PBL was chosen as an effective method to assist students in learning biochemical and physiological processes. By monitoring student attendance and using informal and formal surveys, we demonstrated that PBL has a significant positive impact on student motivation to attend and participate in the course work. Student responses indicated that PBL is superior to traditional lecture format with regard to the understanding of course content and retention of information. We also demonstrated that student problem-solving skills are significantly improved, but additional controlled studies are needed to determine how much PBL exercises contribute to this improvement. These preliminary data indicated several positive outcomes of using PBL in a large classroom setting, although further studies aimed at assessing student learning are needed to further justify implementation of this technique in courses delivered to large undergraduate classes.
Active subspace: toward scalable low-rank learning.
Liu, Guangcan; Yan, Shuicheng
2012-12-01
We address the scalability issues in low-rank matrix learning problems. Usually these problems resort to solving nuclear norm regularized optimization problems (NNROPs), which often suffer from high computational complexities if based on existing solvers, especially in large-scale settings. Based on the fact that the optimal solution matrix to an NNROP is often low rank, we revisit the classic mechanism of low-rank matrix factorization, based on which we present an active subspace algorithm for efficiently solving NNROPs by transforming large-scale NNROPs into small-scale problems. The transformation is achieved by factorizing the large solution matrix into the product of a small orthonormal matrix (active subspace) and another small matrix. Although such a transformation generally leads to nonconvex problems, we show that a suboptimal solution can be found by the augmented Lagrange alternating direction method. For the robust PCA (RPCA) (Candès, Li, Ma, & Wright, 2009 ) problem, a typical example of NNROPs, theoretical results verify the suboptimality of the solution produced by our algorithm. For the general NNROPs, we empirically show that our algorithm significantly reduces the computational complexity without loss of optimality.
NASA Technical Reports Server (NTRS)
Sohn, Andrew; Biswas, Rupak
1996-01-01
Solving the hard Satisfiability Problem is time consuming even for modest-sized problem instances. Solving the Random L-SAT Problem is especially difficult due to the ratio of clauses to variables. This report presents a parallel synchronous simulated annealing method for solving the Random L-SAT Problem on a large-scale distributed-memory multiprocessor. In particular, we use a parallel synchronous simulated annealing procedure, called Generalized Speculative Computation, which guarantees the same decision sequence as sequential simulated annealing. To demonstrate the performance of the parallel method, we have selected problem instances varying in size from 100-variables/425-clauses to 5000-variables/21,250-clauses. Experimental results on the AP1000 multiprocessor indicate that our approach can satisfy 99.9 percent of the clauses while giving almost a 70-fold speedup on 500 processors.
NASA Astrophysics Data System (ADS)
Lin, Y.; O'Malley, D.; Vesselinov, V. V.
2015-12-01
Inverse modeling seeks model parameters given a set of observed state variables. However, for many practical problems due to the facts that the observed data sets are often large and model parameters are often numerous, conventional methods for solving the inverse modeling can be computationally expensive. We have developed a new, computationally-efficient Levenberg-Marquardt method for solving large-scale inverse modeling. Levenberg-Marquardt methods require the solution of a dense linear system of equations which can be prohibitively expensive to compute for large-scale inverse problems. Our novel method projects the original large-scale linear problem down to a Krylov subspace, such that the dimensionality of the measurements can be significantly reduced. Furthermore, instead of solving the linear system for every Levenberg-Marquardt damping parameter, we store the Krylov subspace computed when solving the first damping parameter and recycle it for all the following damping parameters. The efficiency of our new inverse modeling algorithm is significantly improved by using these computational techniques. We apply this new inverse modeling method to invert for a random transitivity field. Our algorithm is fast enough to solve for the distributed model parameters (transitivity) at each computational node in the model domain. The inversion is also aided by the use regularization techniques. The algorithm is coded in Julia and implemented in the MADS computational framework (http://mads.lanl.gov). Julia is an advanced high-level scientific programing language that allows for efficient memory management and utilization of high-performance computational resources. By comparing with a Levenberg-Marquardt method using standard linear inversion techniques, our Levenberg-Marquardt method yields speed-up ratio of 15 in a multi-core computational environment and a speed-up ratio of 45 in a single-core computational environment. Therefore, our new inverse modeling method is a powerful tool for large-scale applications.
A multilevel correction adaptive finite element method for Kohn-Sham equation
NASA Astrophysics Data System (ADS)
Hu, Guanghui; Xie, Hehu; Xu, Fei
2018-02-01
In this paper, an adaptive finite element method is proposed for solving Kohn-Sham equation with the multilevel correction technique. In the method, the Kohn-Sham equation is solved on a fixed and appropriately coarse mesh with the finite element method in which the finite element space is kept improving by solving the derived boundary value problems on a series of adaptively and successively refined meshes. A main feature of the method is that solving large scale Kohn-Sham system is avoided effectively, and solving the derived boundary value problems can be handled efficiently by classical methods such as the multigrid method. Hence, the significant acceleration can be obtained on solving Kohn-Sham equation with the proposed multilevel correction technique. The performance of the method is examined by a variety of numerical experiments.
An interior-point method-based solver for simulation of aircraft parts riveting
NASA Astrophysics Data System (ADS)
Stefanova, Maria; Yakunin, Sergey; Petukhova, Margarita; Lupuleac, Sergey; Kokkolaras, Michael
2018-05-01
The particularities of the aircraft parts riveting process simulation necessitate the solution of a large amount of contact problems. A primal-dual interior-point method-based solver is proposed for solving such problems efficiently. The proposed method features a worst case polynomial complexity bound ? on the number of iterations, where n is the dimension of the problem and ε is a threshold related to desired accuracy. In practice, the convergence is often faster than this worst case bound, which makes the method applicable to large-scale problems. The computational challenge is solving the system of linear equations because the associated matrix is ill conditioned. To that end, the authors introduce a preconditioner and a strategy for determining effective initial guesses based on the physics of the problem. Numerical results are compared with ones obtained using the Goldfarb-Idnani algorithm. The results demonstrate the efficiency of the proposed method.
Recursive partitioned inversion of large (1500 x 1500) symmetric matrices
NASA Technical Reports Server (NTRS)
Putney, B. H.; Brownd, J. E.; Gomez, R. A.
1976-01-01
A recursive algorithm was designed to invert large, dense, symmetric, positive definite matrices using small amounts of computer core, i.e., a small fraction of the core needed to store the complete matrix. The described algorithm is a generalized Gaussian elimination technique. Other algorithms are also discussed for the Cholesky decomposition and step inversion techniques. The purpose of the inversion algorithm is to solve large linear systems of normal equations generated by working geodetic problems. The algorithm was incorporated into a computer program called SOLVE. In the past the SOLVE program has been used in obtaining solutions published as the Goddard earth models.
Parallel Optimization of Polynomials for Large-scale Problems in Stability and Control
NASA Astrophysics Data System (ADS)
Kamyar, Reza
In this thesis, we focus on some of the NP-hard problems in control theory. Thanks to the converse Lyapunov theory, these problems can often be modeled as optimization over polynomials. To avoid the problem of intractability, we establish a trade off between accuracy and complexity. In particular, we develop a sequence of tractable optimization problems --- in the form of Linear Programs (LPs) and/or Semi-Definite Programs (SDPs) --- whose solutions converge to the exact solution of the NP-hard problem. However, the computational and memory complexity of these LPs and SDPs grow exponentially with the progress of the sequence - meaning that improving the accuracy of the solutions requires solving SDPs with tens of thousands of decision variables and constraints. Setting up and solving such problems is a significant challenge. The existing optimization algorithms and software are only designed to use desktop computers or small cluster computers --- machines which do not have sufficient memory for solving such large SDPs. Moreover, the speed-up of these algorithms does not scale beyond dozens of processors. This in fact is the reason we seek parallel algorithms for setting-up and solving large SDPs on large cluster- and/or super-computers. We propose parallel algorithms for stability analysis of two classes of systems: 1) Linear systems with a large number of uncertain parameters; 2) Nonlinear systems defined by polynomial vector fields. First, we develop a distributed parallel algorithm which applies Polya's and/or Handelman's theorems to some variants of parameter-dependent Lyapunov inequalities with parameters defined over the standard simplex. The result is a sequence of SDPs which possess a block-diagonal structure. We then develop a parallel SDP solver which exploits this structure in order to map the computation, memory and communication to a distributed parallel environment. Numerical tests on a supercomputer demonstrate the ability of the algorithm to efficiently utilize hundreds and potentially thousands of processors, and analyze systems with 100+ dimensional state-space. Furthermore, we extend our algorithms to analyze robust stability over more complicated geometries such as hypercubes and arbitrary convex polytopes. Our algorithms can be readily extended to address a wide variety of problems in control such as Hinfinity synthesis for systems with parametric uncertainty and computing control Lyapunov functions.
Neural Activity When People Solve Verbal Problems with Insight
Bowden, Edward M; Haberman, Jason; Frymiare, Jennifer L; Arambel-Liu, Stella; Greenblatt, Richard; Reber, Paul J
2004-01-01
People sometimes solve problems with a unique process called insight, accompanied by an “Aha!” experience. It has long been unclear whether different cognitive and neural processes lead to insight versus noninsight solutions, or if solutions differ only in subsequent subjective feeling. Recent behavioral studies indicate distinct patterns of performance and suggest differential hemispheric involvement for insight and noninsight solutions. Subjects solved verbal problems, and after each correct solution indicated whether they solved with or without insight. We observed two objective neural correlates of insight. Functional magnetic resonance imaging (Experiment 1) revealed increased activity in the right hemisphere anterior superior temporal gyrus for insight relative to noninsight solutions. The same region was active during initial solving efforts. Scalp electroencephalogram recordings (Experiment 2) revealed a sudden burst of high-frequency (gamma-band) neural activity in the same area beginning 0.3 s prior to insight solutions. This right anterior temporal area is associated with making connections across distantly related information during comprehension. Although all problem solving relies on a largely shared cortical network, the sudden flash of insight occurs when solvers engage distinct neural and cognitive processes that allow them to see connections that previously eluded them. PMID:15094802
ERIC Educational Resources Information Center
Klegeris, Andis; Bahniwal, Manpreet; Hurren, Heather
2013-01-01
Problem-based learning (PBL) was originally introduced in medical education programs as a form of small-group learning, but its use has now spread to large undergraduate classrooms in various other disciplines. Introduction of new teaching techniques, including PBL-based methods, needs to be justified by demonstrating the benefits of such…
NASA Astrophysics Data System (ADS)
Konno, Yohko; Suzuki, Keiji
This paper describes an approach to development of a solution algorithm of a general-purpose for large scale problems using “Local Clustering Organization (LCO)” as a new solution for Job-shop scheduling problem (JSP). Using a performance effective large scale scheduling in the study of usual LCO, a solving JSP keep stability induced better solution is examined. In this study for an improvement of a performance of a solution for JSP, processes to a optimization by LCO is examined, and a scheduling solution-structure is extended to a new solution-structure based on machine-division. A solving method introduced into effective local clustering for the solution-structure is proposed as an extended LCO. An extended LCO has an algorithm which improves scheduling evaluation efficiently by clustering of parallel search which extends over plural machines. A result verified by an application of extended LCO on various scale of problems proved to conduce to minimizing make-span and improving on the stable performance.
Sparse deconvolution for the large-scale ill-posed inverse problem of impact force reconstruction
NASA Astrophysics Data System (ADS)
Qiao, Baijie; Zhang, Xingwu; Gao, Jiawei; Liu, Ruonan; Chen, Xuefeng
2017-01-01
Most previous regularization methods for solving the inverse problem of force reconstruction are to minimize the l2-norm of the desired force. However, these traditional regularization methods such as Tikhonov regularization and truncated singular value decomposition, commonly fail to solve the large-scale ill-posed inverse problem in moderate computational cost. In this paper, taking into account the sparse characteristic of impact force, the idea of sparse deconvolution is first introduced to the field of impact force reconstruction and a general sparse deconvolution model of impact force is constructed. Second, a novel impact force reconstruction method based on the primal-dual interior point method (PDIPM) is proposed to solve such a large-scale sparse deconvolution model, where minimizing the l2-norm is replaced by minimizing the l1-norm. Meanwhile, the preconditioned conjugate gradient algorithm is used to compute the search direction of PDIPM with high computational efficiency. Finally, two experiments including the small-scale or medium-scale single impact force reconstruction and the relatively large-scale consecutive impact force reconstruction are conducted on a composite wind turbine blade and a shell structure to illustrate the advantage of PDIPM. Compared with Tikhonov regularization, PDIPM is more efficient, accurate and robust whether in the single impact force reconstruction or in the consecutive impact force reconstruction.
An efficient strongly coupled immersed boundary method for deforming bodies
NASA Astrophysics Data System (ADS)
Goza, Andres; Colonius, Tim
2016-11-01
Immersed boundary methods treat the fluid and immersed solid with separate domains. As a result, a nonlinear interface constraint must be satisfied when these methods are applied to flow-structure interaction problems. This typically results in a large nonlinear system of equations that is difficult to solve efficiently. Often, this system is solved with a block Gauss-Seidel procedure, which is easy to implement but can require many iterations to converge for small solid-to-fluid mass ratios. Alternatively, a Newton-Raphson procedure can be used to solve the nonlinear system. This typically leads to convergence in a small number of iterations for arbitrary mass ratios, but involves the use of large Jacobian matrices. We present an immersed boundary formulation that, like the Newton-Raphson approach, uses a linearization of the system to perform iterations. It therefore inherits the same favorable convergence behavior. However, we avoid large Jacobian matrices by using a block LU factorization of the linearized system. We derive our method for general deforming surfaces and perform verification on 2D test problems of flow past beams. These test problems involve large amplitude flapping and a wide range of mass ratios. This work was partially supported by the Jet Propulsion Laboratory and Air Force Office of Scientific Research.
An evolutionary algorithm for large traveling salesman problems.
Tsai, Huai-Kuang; Yang, Jinn-Moon; Tsai, Yuan-Fang; Kao, Cheng-Yan
2004-08-01
This work proposes an evolutionary algorithm, called the heterogeneous selection evolutionary algorithm (HeSEA), for solving large traveling salesman problems (TSP). The strengths and limitations of numerous well-known genetic operators are first analyzed, along with local search methods for TSPs from their solution qualities and mechanisms for preserving and adding edges. Based on this analysis, a new approach, HeSEA is proposed which integrates edge assembly crossover (EAX) and Lin-Kernighan (LK) local search, through family competition and heterogeneous pairing selection. This study demonstrates experimentally that EAX and LK can compensate for each other's disadvantages. Family competition and heterogeneous pairing selections are used to maintain the diversity of the population, which is especially useful for evolutionary algorithms in solving large TSPs. The proposed method was evaluated on 16 well-known TSPs in which the numbers of cities range from 318 to 13509. Experimental results indicate that HeSEA performs well and is very competitive with other approaches. The proposed method can determine the optimum path when the number of cities is under 10,000 and the mean solution quality is within 0.0074% above the optimum for each test problem. These findings imply that the proposed method can find tours robustly with a fixed small population and a limited family competition length in reasonable time, when used to solve large TSPs.
Solving optimization problems on computational grids.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wright, S. J.; Mathematics and Computer Science
2001-05-01
Multiprocessor computing platforms, which have become more and more widely available since the mid-1980s, are now heavily used by organizations that need to solve very demanding computational problems. Parallel computing is now central to the culture of many research communities. Novel parallel approaches were developed for global optimization, network optimization, and direct-search methods for nonlinear optimization. Activity was particularly widespread in parallel branch-and-bound approaches for various problems in combinatorial and network optimization. As the cost of personal computers and low-end workstations has continued to fall, while the speed and capacity of processors and networks have increased dramatically, 'cluster' platforms havemore » become popular in many settings. A somewhat different type of parallel computing platform know as a computational grid (alternatively, metacomputer) has arisen in comparatively recent times. Broadly speaking, this term refers not to a multiprocessor with identical processing nodes but rather to a heterogeneous collection of devices that are widely distributed, possibly around the globe. The advantage of such platforms is obvious: they have the potential to deliver enormous computing power. Just as obviously, however, the complexity of grids makes them very difficult to use. The Condor team, headed by Miron Livny at the University of Wisconsin, were among the pioneers in providing infrastructure for grid computations. More recently, the Globus project has developed technologies to support computations on geographically distributed platforms consisting of high-end computers, storage and visualization devices, and other scientific instruments. In 1997, we started the metaneos project as a collaborative effort between optimization specialists and the Condor and Globus groups. Our aim was to address complex, difficult optimization problems in several areas, designing and implementing the algorithms and the software infrastructure need to solve these problems on computational grids. This article describes some of the results we have obtained during the first three years of the metaneos project. Our efforts have led to development of the runtime support library MW for implementing algorithms with master-worker control structure on Condor platforms. This work is discussed here, along with work on algorithms and codes for integer linear programming, the quadratic assignment problem, and stochastic linear programmming. Our experiences in the metaneos project have shown that cheap, powerful computational grids can be used to tackle large optimization problems of various types. In an industrial or commercial setting, the results demonstrate that one may not have to buy powerful computational servers to solve many of the large problems arising in areas such as scheduling, portfolio optimization, or logistics; the idle time on employee workstations (or, at worst, an investment in a modest cluster of PCs) may do the job. For the optimization research community, our results motivate further work on parallel, grid-enabled algorithms for solving very large problems of other types. The fact that very large problems can be solved cheaply allows researchers to better understand issues of 'practical' complexity and of the role of heuristics.« less
Ant colony optimization for solving university facility layout problem
NASA Astrophysics Data System (ADS)
Mohd Jani, Nurul Hafiza; Mohd Radzi, Nor Haizan; Ngadiman, Mohd Salihin
2013-04-01
Quadratic Assignment Problems (QAP) is classified as the NP hard problem. It has been used to model a lot of problem in several areas such as operational research, combinatorial data analysis and also parallel and distributed computing, optimization problem such as graph portioning and Travel Salesman Problem (TSP). In the literature, researcher use exact algorithm, heuristics algorithm and metaheuristic approaches to solve QAP problem. QAP is largely applied in facility layout problem (FLP). In this paper we used QAP to model university facility layout problem. There are 8 facilities that need to be assigned to 8 locations. Hence we have modeled a QAP problem with n ≤ 10 and developed an Ant Colony Optimization (ACO) algorithm to solve the university facility layout problem. The objective is to assign n facilities to n locations such that the minimum product of flows and distances is obtained. Flow is the movement from one to another facility, whereas distance is the distance between one locations of a facility to other facilities locations. The objective of the QAP is to obtain minimum total walking (flow) of lecturers from one destination to another (distance).
Solving geosteering inverse problems by stochastic Hybrid Monte Carlo method
Shen, Qiuyang; Wu, Xuqing; Chen, Jiefu; ...
2017-11-20
The inverse problems arise in almost all fields of science where the real-world parameters are extracted from a set of measured data. The geosteering inversion plays an essential role in the accurate prediction of oncoming strata as well as a reliable guidance to adjust the borehole position on the fly to reach one or more geological targets. This mathematical treatment is not easy to solve, which requires finding an optimum solution among a large solution space, especially when the problem is non-linear and non-convex. Nowadays, a new generation of logging-while-drilling (LWD) tools has emerged on the market. The so-called azimuthalmore » resistivity LWD tools have azimuthal sensitivity and a large depth of investigation. Hence, the associated inverse problems become much more difficult since the earth model to be inverted will have more detailed structures. The conventional deterministic methods are incapable to solve such a complicated inverse problem, where they suffer from the local minimum trap. Alternatively, stochastic optimizations are in general better at finding global optimal solutions and handling uncertainty quantification. In this article, we investigate the Hybrid Monte Carlo (HMC) based statistical inversion approach and suggest that HMC based inference is more efficient in dealing with the increased complexity and uncertainty faced by the geosteering problems.« less
Can microbes economically remove sulfur
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fox, J.L.
Researchers have reported that refiners who now rely on costly physic-chemical procedures to desulfurize petroleum will soon have an alternative microbial-enzyme-based approach to this process. This new approach is still under development and considerable number chemical engineering problems need to be solved before this process is ready for large-scale use. This paper reviews the several research projects dedicated solving the problems that keep a biotechnology-based alternative from competing with chemical desulfurization.
How doctors learn: the role of clinical problems across the medical school-to-practice continuum.
Slotnick, H B
1996-01-01
The author proposes a theory of how physicians learn that uses clinical problem solving as its central feature. His theory, which integrates insights from Maslow, Schön, Norman, and others, claims that physicians-in-training and practicing physicians learn largely by deriving insights from clinical experience. These insights allow the learner to solve future problems and thereby address the learner's basic human needs for security, affiliation, and self-esteem. Ensuring that students gain such insights means that the proper roles of the teacher are (1) to select problems for students to solve and offer guidance on how to solve them, and (2) to serve as a role model of how to reflect on the problem, its solution, and the solution's effectiveness. Three principles guide instruction within its framework for learning: (1) learners, whether physicians-in-training or practicing physicians, seek to solve problems they recognize they have; (2) learners want to be involved in their own learning; and (3) instruction must both be time-efficient and also demonstrate the range of ways in which students can apply what they learn. The author concludes by applying the theory to an aspect of undergraduate education and to the general process of continuing medical education.
Finite element modeling of electromagnetic fields and waves using NASTRAN
NASA Technical Reports Server (NTRS)
Moyer, E. Thomas, Jr.; Schroeder, Erwin
1989-01-01
The various formulations of Maxwell's equations are reviewed with emphasis on those formulations which most readily form analogies with Navier's equations. Analogies involving scalar and vector potentials and electric and magnetic field components are presented. Formulations allowing for media with dielectric and conducting properties are emphasized. It is demonstrated that many problems in electromagnetism can be solved using the NASTRAN finite element code. Several fundamental problems involving time harmonic solutions of Maxwell's equations with known analytic solutions are solved using NASTRAN to demonstrate convergence and mesh requirements. Mesh requirements are studied as a function of frequency, conductivity, and dielectric properties. Applications in both low frequency and high frequency are highlighted. The low frequency problems demonstrate the ability to solve problems involving media inhomogeneity and unbounded domains. The high frequency applications demonstrate the ability to handle problems with large boundary to wavelength ratios.
Mathematical programming formulations for satellite synthesis
NASA Technical Reports Server (NTRS)
Bhasin, Puneet; Reilly, Charles H.
1987-01-01
The problem of satellite synthesis can be described as optimally allotting locations and sometimes frequencies and polarizations, to communication satellites so that interference from unwanted satellite signals does not exceed a specified threshold. In this report, mathematical programming models and optimization methods are used to solve satellite synthesis problems. A nonlinear programming formulation which is solved using Zoutendijk's method and a gradient search method is described. Nine mixed integer programming models are considered. Results of computer runs with these nine models and five geographically compatible scenarios are presented and evaluated. A heuristic solution procedure is also used to solve two of the models studied. Heuristic solutions to three large synthesis problems are presented. The results of our analysis show that the heuristic performs very well, both in terms of solution quality and solution time, on the two models to which it was applied. It is concluded that the heuristic procedure is the best of the methods considered for solving satellite synthesis problems.
Simultaneous optimization of loading pattern and burnable poison placement for PWRs
DOE Office of Scientific and Technical Information (OSTI.GOV)
Alim, F.; Ivanov, K.; Yilmaz, S.
2006-07-01
To solve in-core fuel management optimization problem, GARCO-PSU (Genetic Algorithm Reactor Core Optimization - Pennsylvania State Univ.) is developed. This code is applicable for all types and geometry of PWR core structures with unlimited number of fuel assembly (FA) types in the inventory. For this reason an innovative genetic algorithm is developed with modifying the classical representation of the genotype. In-core fuel management heuristic rules are introduced into GARCO. The core re-load design optimization has two parts, loading pattern (LP) optimization and burnable poison (BP) placement optimization. These parts depend on each other, but it is difficult to solve themore » combined problem due to its large size. Separating the problem into two parts provides a practical way to solve the problem. However, the result of this method does not reflect the real optimal solution. GARCO-PSU achieves to solve LP optimization and BP placement optimization simultaneously in an efficient manner. (authors)« less
Surveying college introductory physics students’ attitudes and approaches to problem solving
NASA Astrophysics Data System (ADS)
Mason, Andrew J.; Singh, Chandralekha
2016-09-01
Students’ attitudes and approaches to problem solving in physics can greatly impact their actual problem solving practices and also influence their motivation to learn and ultimately the development of expertise. We developed and validated an attitudes and approaches to problem solving (AAPS) survey and administered it to students in the introductory physics courses in a typical large research university in the US. Here, we discuss the development and validation of the survey and analysis of the student responses to the survey questions in introductory physics courses. The introductory physics students’ responses to the survey questions were also compared with those of physics faculty members and physics PhD students. We find that introductory students are in general less expert-like than the physics faculty members and PhD students. Moreover, on some AAPS survey questions, the responses of students and faculty have unexpected trends. Those trends were interpreted via individual interviews, which helped clarify reasons for those survey responses.
Optimal Full Information Synthesis for Flexible Structures Implemented on Cray Supercomputers
NASA Technical Reports Server (NTRS)
Lind, Rick; Balas, Gary J.
1995-01-01
This paper considers an algorithm for synthesis of optimal controllers for full information feedback. The synthesis procedure reduces to a single linear matrix inequality which may be solved via established convex optimization algorithms. The computational cost of the optimization is investigated. It is demonstrated the problem dimension and corresponding matrices can become large for practical engineering problems. This algorithm represents a process that is impractical for standard workstations for large order systems. A flexible structure is presented as a design example. Control synthesis requires several days on a workstation but may be solved in a reasonable amount of time using a Cray supercomputer.
Development and Applications of a Modular Parallel Process for Large Scale Fluid/Structures Problems
NASA Technical Reports Server (NTRS)
Guruswamy, Guru P.; Kwak, Dochan (Technical Monitor)
2002-01-01
A modular process that can efficiently solve large scale multidisciplinary problems using massively parallel supercomputers is presented. The process integrates disciplines with diverse physical characteristics by retaining the efficiency of individual disciplines. Computational domain independence of individual disciplines is maintained using a meta programming approach. The process integrates disciplines without affecting the combined performance. Results are demonstrated for large scale aerospace problems on several supercomputers. The super scalability and portability of the approach is demonstrated on several parallel computers.
Development and Applications of a Modular Parallel Process for Large Scale Fluid/Structures Problems
NASA Technical Reports Server (NTRS)
Guruswamy, Guru P.; Byun, Chansup; Kwak, Dochan (Technical Monitor)
2001-01-01
A modular process that can efficiently solve large scale multidisciplinary problems using massively parallel super computers is presented. The process integrates disciplines with diverse physical characteristics by retaining the efficiency of individual disciplines. Computational domain independence of individual disciplines is maintained using a meta programming approach. The process integrates disciplines without affecting the combined performance. Results are demonstrated for large scale aerospace problems on several supercomputers. The super scalability and portability of the approach is demonstrated on several parallel computers.
NASA Astrophysics Data System (ADS)
Ushijima, Timothy T.; Yeh, William W.-G.
2013-10-01
An optimal experimental design algorithm is developed to select locations for a network of observation wells that provide maximum information about unknown groundwater pumping in a confined, anisotropic aquifer. The design uses a maximal information criterion that chooses, among competing designs, the design that maximizes the sum of squared sensitivities while conforming to specified design constraints. The formulated optimization problem is non-convex and contains integer variables necessitating a combinatorial search. Given a realistic large-scale model, the size of the combinatorial search required can make the problem difficult, if not impossible, to solve using traditional mathematical programming techniques. Genetic algorithms (GAs) can be used to perform the global search; however, because a GA requires a large number of calls to a groundwater model, the formulated optimization problem still may be infeasible to solve. As a result, proper orthogonal decomposition (POD) is applied to the groundwater model to reduce its dimensionality. Then, the information matrix in the full model space can be searched without solving the full model. Results from a small-scale test case show identical optimal solutions among the GA, integer programming, and exhaustive search methods. This demonstrates the GA's ability to determine the optimal solution. In addition, the results show that a GA with POD model reduction is several orders of magnitude faster in finding the optimal solution than a GA using the full model. The proposed experimental design algorithm is applied to a realistic, two-dimensional, large-scale groundwater problem. The GA converged to a solution for this large-scale problem.
Games that Enlist Collective Intelligence to Solve Complex Scientific Problems.
Burnett, Stephen; Furlong, Michelle; Melvin, Paul Guy; Singiser, Richard
2016-03-01
There is great value in employing the collective problem-solving power of large groups of people. Technological advances have allowed computer games to be utilized by a diverse population to solve problems. Science games are becoming more popular and cover various areas such as sequence alignments, DNA base-pairing, and protein and RNA folding. While these tools have been developed for the general population, they can also be used effectively in the classroom to teach students about various topics. Many games also employ a social component that entices students to continue playing and thereby to continue learning. The basic functions of game play and the potential of game play as a tool in the classroom are discussed in this article.
Games that Enlist Collective Intelligence to Solve Complex Scientific Problems
Burnett, Stephen; Furlong, Michelle; Melvin, Paul Guy; Singiser, Richard
2016-01-01
There is great value in employing the collective problem-solving power of large groups of people. Technological advances have allowed computer games to be utilized by a diverse population to solve problems. Science games are becoming more popular and cover various areas such as sequence alignments, DNA base-pairing, and protein and RNA folding. While these tools have been developed for the general population, they can also be used effectively in the classroom to teach students about various topics. Many games also employ a social component that entices students to continue playing and thereby to continue learning. The basic functions of game play and the potential of game play as a tool in the classroom are discussed in this article. PMID:27047610
NASA Astrophysics Data System (ADS)
Chaves-González, José M.; Vega-Rodríguez, Miguel A.; Gómez-Pulido, Juan A.; Sánchez-Pérez, Juan M.
2011-08-01
This article analyses the use of a novel parallel evolutionary strategy to solve complex optimization problems. The work developed here has been focused on a relevant real-world problem from the telecommunication domain to verify the effectiveness of the approach. The problem, known as frequency assignment problem (FAP), basically consists of assigning a very small number of frequencies to a very large set of transceivers used in a cellular phone network. Real data FAP instances are very difficult to solve due to the NP-hard nature of the problem, therefore using an efficient parallel approach which makes the most of different evolutionary strategies can be considered as a good way to obtain high-quality solutions in short periods of time. Specifically, a parallel hyper-heuristic based on several meta-heuristics has been developed. After a complete experimental evaluation, results prove that the proposed approach obtains very high-quality solutions for the FAP and beats any other result published.
A General-Purpose Optimization Engine for Multi-Disciplinary Design Applications
NASA Technical Reports Server (NTRS)
Patnaik, Surya N.; Hopkins, Dale A.; Berke, Laszlo
1996-01-01
A general purpose optimization tool for multidisciplinary applications, which in the literature is known as COMETBOARDS, is being developed at NASA Lewis Research Center. The modular organization of COMETBOARDS includes several analyzers and state-of-the-art optimization algorithms along with their cascading strategy. The code structure allows quick integration of new analyzers and optimizers. The COMETBOARDS code reads input information from a number of data files, formulates a design as a set of multidisciplinary nonlinear programming problems, and then solves the resulting problems. COMETBOARDS can be used to solve a large problem which can be defined through multiple disciplines, each of which can be further broken down into several subproblems. Alternatively, a small portion of a large problem can be optimized in an effort to improve an existing system. Some of the other unique features of COMETBOARDS include design variable formulation, constraint formulation, subproblem coupling strategy, global scaling technique, analysis approximation, use of either sequential or parallel computational modes, and so forth. The special features and unique strengths of COMETBOARDS assist convergence and reduce the amount of CPU time used to solve the difficult optimization problems of aerospace industries. COMETBOARDS has been successfully used to solve a number of problems, including structural design of space station components, design of nozzle components of an air-breathing engine, configuration design of subsonic and supersonic aircraft, mixed flow turbofan engines, wave rotor topped engines, and so forth. This paper introduces the COMETBOARDS design tool and its versatility, which is illustrated by citing examples from structures, aircraft design, and air-breathing propulsion engine design.
Robust penalty method for structural synthesis
NASA Technical Reports Server (NTRS)
Kamat, M. P.
1983-01-01
The Sequential Unconstrained Minimization Technique (SUMT) offers an easy way of solving nonlinearly constrained problems. However, this algorithm frequently suffers from the need to minimize an ill-conditioned penalty function. An ill-conditioned minimization problem can be solved very effectively by posing the problem as one of integrating a system of stiff differential equations utilizing concepts from singular perturbation theory. This paper evaluates the robustness and the reliability of such a singular perturbation based SUMT algorithm on two different problems of structural optimization of widely separated scales. The report concludes that whereas conventional SUMT can be bogged down by frequent ill-conditioning, especially in large scale problems, the singular perturbation SUMT has no such difficulty in converging to very accurate solutions.
Implementing and Assessing Computational Modeling in Introductory Mechanics
ERIC Educational Resources Information Center
Caballero, Marcos D.; Kohlmyer, Matthew A.; Schatz, Michael F.
2012-01-01
Students taking introductory physics are rarely exposed to computational modeling. In a one-semester large lecture introductory calculus-based mechanics course at Georgia Tech, students learned to solve physics problems using the VPython programming environment. During the term, 1357 students in this course solved a suite of 14 computational…
A subgradient approach for constrained binary optimization via quantum adiabatic evolution
NASA Astrophysics Data System (ADS)
Karimi, Sahar; Ronagh, Pooya
2017-08-01
Outer approximation method has been proposed for solving the Lagrangian dual of a constrained binary quadratic programming problem via quantum adiabatic evolution in the literature. This should be an efficient prescription for solving the Lagrangian dual problem in the presence of an ideally noise-free quantum adiabatic system. However, current implementations of quantum annealing systems demand methods that are efficient at handling possible sources of noise. In this paper, we consider a subgradient method for finding an optimal primal-dual pair for the Lagrangian dual of a constrained binary polynomial programming problem. We then study the quadratic stable set (QSS) problem as a case study. We see that this method applied to the QSS problem can be viewed as an instance-dependent penalty-term approach that avoids large penalty coefficients. Finally, we report our experimental results of using the D-Wave 2X quantum annealer and conclude that our approach helps this quantum processor to succeed more often in solving these problems compared to the usual penalty-term approaches.
Cooley, R.L.; Hill, M.C.
1992-01-01
Three methods of solving nonlinear least-squares problems were compared for robustness and efficiency using a series of hypothetical and field problems. A modified Gauss-Newton/full Newton hybrid method (MGN/FN) and an analogous method for which part of the Hessian matrix was replaced by a quasi-Newton approximation (MGN/QN) solved some of the problems with appreciably fewer iterations than required using only a modified Gauss-Newton (MGN) method. In these problems, model nonlinearity and a large variance for the observed data apparently caused MGN to converge more slowly than MGN/FN or MGN/QN after the sum of squared errors had almost stabilized. Other problems were solved as efficiently with MGN as with MGN/FN or MGN/QN. Because MGN/FN can require significantly more computer time per iteration and more computer storage for transient problems, it is less attractive for a general purpose algorithm than MGN/QN.
NASA Technical Reports Server (NTRS)
Han, Chia Yung; Wan, Liqun; Wee, William G.
1990-01-01
A knowledge-based interactive problem solving environment called KIPSE1 is presented. The KIPSE1 is a system built on a commercial expert system shell, the KEE system. This environment gives user capability to carry out exploratory data analysis and pattern classification tasks. A good solution often consists of a sequence of steps with a set of methods used at each step. In KIPSE1, solution is represented in the form of a decision tree and each node of the solution tree represents a partial solution to the problem. Many methodologies are provided at each node to the user such that the user can interactively select the method and data sets to test and subsequently examine the results. Otherwise, users are allowed to make decisions at various stages of problem solving to subdivide the problem into smaller subproblems such that a large problem can be handled and a better solution can be found.
NASA Astrophysics Data System (ADS)
Fazayeli, Saeed; Eydi, Alireza; Kamalabadi, Isa Nakhai
2017-07-01
Nowadays, organizations have to compete with different competitors in regional, national and international levels, so they have to improve their competition capabilities to survive against competitors. Undertaking activities on a global scale requires a proper distribution system which could take advantages of different transportation modes. Accordingly, the present paper addresses a location-routing problem on multimodal transportation network. The introduced problem follows four objectives simultaneously which form main contribution of the paper; determining multimodal routes between supplier and distribution centers, locating mode changing facilities, locating distribution centers, and determining product delivery tours from the distribution centers to retailers. An integer linear programming is presented for the problem, and a genetic algorithm with a new chromosome structure proposed to solve the problem. Proposed chromosome structure consists of two different parts for multimodal transportation and location-routing parts of the model. Based on published data in the literature, two numerical cases with different sizes generated and solved. Also, different cost scenarios designed to better analyze model and algorithm performance. Results show that algorithm can effectively solve large-size problems within a reasonable time which GAMS software failed to reach an optimal solution even within much longer times.
NASA Astrophysics Data System (ADS)
Fazayeli, Saeed; Eydi, Alireza; Kamalabadi, Isa Nakhai
2018-07-01
Nowadays, organizations have to compete with different competitors in regional, national and international levels, so they have to improve their competition capabilities to survive against competitors. Undertaking activities on a global scale requires a proper distribution system which could take advantages of different transportation modes. Accordingly, the present paper addresses a location-routing problem on multimodal transportation network. The introduced problem follows four objectives simultaneously which form main contribution of the paper; determining multimodal routes between supplier and distribution centers, locating mode changing facilities, locating distribution centers, and determining product delivery tours from the distribution centers to retailers. An integer linear programming is presented for the problem, and a genetic algorithm with a new chromosome structure proposed to solve the problem. Proposed chromosome structure consists of two different parts for multimodal transportation and location-routing parts of the model. Based on published data in the literature, two numerical cases with different sizes generated and solved. Also, different cost scenarios designed to better analyze model and algorithm performance. Results show that algorithm can effectively solve large-size problems within a reasonable time which GAMS software failed to reach an optimal solution even within much longer times.
Solving Large Problems Quickly: Progress in 2001-2003
NASA Technical Reports Server (NTRS)
Mowry, Todd C.; Colohan, Christopher B.; Brown, Angela Demke; Steffan, J. Gregory; Zhai, Antonia
2004-01-01
This document describes the progress we have made and the lessons we have learned in 2001 through 2003 under the NASA grant entitled "Solving Important Problems Faster". The long-term goal of this research is to accelerate large, irregular scientific applications which have enormous data sets and which are difficult to parallelize. To accomplish this goal, we are exploring two complementary techniques: (i) using compiler-inserted prefetching to automatically hide the I/O latency of accessing these large data sets from disk; and (ii) using thread-level data speculation to enable the optimistic parallelization of applications despite uncertainty as to whether data dependences exist between the resulting threads which would normally make them unsafe to execute in parallel. Overall, we made significant progress in 2001 through 2003, and the project has gone well.
A numerical scheme to solve unstable boundary value problems
NASA Technical Reports Server (NTRS)
Kalnay Derivas, E.
1975-01-01
A new iterative scheme for solving boundary value problems is presented. It consists of the introduction of an artificial time dependence into a modified version of the system of equations. Then explicit forward integrations in time are followed by explicit integrations backwards in time. The method converges under much more general conditions than schemes based in forward time integrations (false transient schemes). In particular it can attain a steady state solution of an elliptical system of equations even if the solution is unstable, in which case other iterative schemes fail to converge. The simplicity of its use makes it attractive for solving large systems of nonlinear equations.
NASA Astrophysics Data System (ADS)
Sanan, P.; Schnepp, S. M.; May, D.; Schenk, O.
2014-12-01
Geophysical applications require efficient forward models for non-linear Stokes flow on high resolution spatio-temporal domains. The bottleneck in applying the forward model is solving the linearized, discretized Stokes problem which takes the form of a large, indefinite (saddle point) linear system. Due to the heterogeniety of the effective viscosity in the elliptic operator, devising effective preconditioners for saddle point problems has proven challenging and highly problem-dependent. Nevertheless, at least three approaches show promise for preconditioning these difficult systems in an algorithmically scalable way using multigrid and/or domain decomposition techniques. The first is to work with a hierarchy of coarser or smaller saddle point problems. The second is to use the Schur complement method to decouple and sequentially solve for the pressure and velocity. The third is to use the Schur decomposition to devise preconditioners for the full operator. These involve sub-solves resembling inexact versions of the sequential solve. The choice of approach and sub-methods depends crucially on the motivating physics, the discretization, and available computational resources. Here we examine the performance trade-offs for preconditioning strategies applied to idealized models of mantle convection and lithospheric dynamics, characterized by large viscosity gradients. Due to the arbitrary topological structure of the viscosity field in geodynamical simulations, we utilize low order, inf-sup stable mixed finite element spatial discretizations which are suitable when sharp viscosity variations occur in element interiors. Particular attention is paid to possibilities within the decoupled and approximate Schur complement factorization-based monolithic approaches to leverage recently-developed flexible, communication-avoiding, and communication-hiding Krylov subspace methods in combination with `heavy' smoothers, which require solutions of large per-node sub-problems, well-suited to solution on hybrid computational clusters. To manage the combinatorial explosion of solver options (which include hybridizations of all the approaches mentioned above), we leverage the modularity of the PETSc library.
Li, Yongming; Tong, Shaocheng
The problem of active fault-tolerant control (FTC) is investigated for the large-scale nonlinear systems in nonstrict-feedback form. The nonstrict-feedback nonlinear systems considered in this paper consist of unstructured uncertainties, unmeasured states, unknown interconnected terms, and actuator faults (e.g., bias fault and gain fault). A state observer is designed to solve the unmeasurable state problem. Neural networks (NNs) are used to identify the unknown lumped nonlinear functions so that the problems of unstructured uncertainties and unknown interconnected terms can be solved. By combining the adaptive backstepping design principle with the combination Nussbaum gain function property, a novel NN adaptive output-feedback FTC approach is developed. The proposed FTC controller can guarantee that all signals in all subsystems are bounded, and the tracking errors for each subsystem converge to a small neighborhood of zero. Finally, numerical results of practical examples are presented to further demonstrate the effectiveness of the proposed control strategy.The problem of active fault-tolerant control (FTC) is investigated for the large-scale nonlinear systems in nonstrict-feedback form. The nonstrict-feedback nonlinear systems considered in this paper consist of unstructured uncertainties, unmeasured states, unknown interconnected terms, and actuator faults (e.g., bias fault and gain fault). A state observer is designed to solve the unmeasurable state problem. Neural networks (NNs) are used to identify the unknown lumped nonlinear functions so that the problems of unstructured uncertainties and unknown interconnected terms can be solved. By combining the adaptive backstepping design principle with the combination Nussbaum gain function property, a novel NN adaptive output-feedback FTC approach is developed. The proposed FTC controller can guarantee that all signals in all subsystems are bounded, and the tracking errors for each subsystem converge to a small neighborhood of zero. Finally, numerical results of practical examples are presented to further demonstrate the effectiveness of the proposed control strategy.
Townsend, E; Hawton, K; Altman, D G; Arensman, E; Gunnell, D; Hazell, P; House, A; Van Heeringen, K
2001-08-01
Brief problem-solving therapy is regarded as a pragmatic treatment for deliberate self-harm (DSH) patients. A recent meta-analysis of randomized controlled trials (RCTs) evaluating this approach indicated a trend towards reduced repetition of DSH but the pooled odds ratio was not statistically significant. We have now examined other important outcomes using this procedure, namely depression, hopelessness and improvement in problems. Six trials in which problem-solving therapy was compared with control treatment were identified from an extensive literature review of RCTs of treatments for DSH patients. Data concerning depression, hopelessness and improvement in problems were extracted. Where relevant statistical data (e.g. standard deviations) were missing these were imputed using various statistical methods. Results were pooled using meta-analytical procedures. At follow-up, patients who were offered problem-solving therapy had significantly greater improvement in scores for depression (standardized mean difference = -0.36; 95% CI -0.61 to -0.11) and hopelessness (weighted mean difference =-3.2; 95% CI -4.0 to -2.41), and significantly more reported improvement in their problems (odds ratio = 2.31; 95% CI 1.29 to 4.13), than patients who were in the control treatment groups. Problem-solving therapy for DSH patients appears to produce better results than control treatment with regard to improvement in depression, hopelessness and problems. It is desirable that this finding is confirmed in a large trial, which will also allow adequate testing of the impact of this treatment on repetition of DSH.
NASA Astrophysics Data System (ADS)
Çağlayan Mercan, Fatih
2012-06-01
This study examines the epistemic beliefs about justification employed by physics undergraduate and graduate students and faculty in the context of solving a standard classical physics problem and a frontier physics problem. Data were collected by a think-aloud problem solving session followed by a semi-structured interview conducted with 50 participants, 10 participants at freshmen, seniors, masters, PhD, and faculty levels. Seven modes of justification were identified and used for exploring the relationships between each justification mode and problem context, and expertise level. The data showed that justification modes were not mutually exclusive and many respondents combined different modes in their responses in both problem contexts. Success on solving the standard classical physics problem was not related to any of the justification modes and was independent of expertise level. The strength of the association across the problem contexts for the authoritative, rational, and empirical justification modes fell in the medium range and for the modeling justification mode fell in the large range of practical significance. Expertise level was not related with the empirical and religious justification modes. The strength of the association between the expertise level and the authoritative, rational, experiential, and relativistic justification modes fell in the medium range, and the modeling justification mode fell in the large range of practical significance. The results provide support for the importance of context for the epistemic beliefs about justification and are discussed in terms of the implications for teaching and learning science.
A feasible DY conjugate gradient method for linear equality constraints
NASA Astrophysics Data System (ADS)
LI, Can
2017-09-01
In this paper, we propose a feasible conjugate gradient method for solving linear equality constrained optimization problem. The method is an extension of the Dai-Yuan conjugate gradient method proposed by Dai and Yuan to linear equality constrained optimization problem. It can be applied to solve large linear equality constrained problem due to lower storage requirement. An attractive property of the method is that the generated direction is always feasible and descent direction. Under mild conditions, the global convergence of the proposed method with exact line search is established. Numerical experiments are also given which show the efficiency of the method.
Final Report---Optimization Under Nonconvexity and Uncertainty: Algorithms and Software
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jeff Linderoth
2011-11-06
the goal of this work was to develop new algorithmic techniques for solving large-scale numerical optimization problems, focusing on problems classes that have proven to be among the most challenging for practitioners: those involving uncertainty and those involving nonconvexity. This research advanced the state-of-the-art in solving mixed integer linear programs containing symmetry, mixed integer nonlinear programs, and stochastic optimization problems. The focus of the work done in the continuation was on Mixed Integer Nonlinear Programs (MINLP)s and Mixed Integer Linear Programs (MILP)s, especially those containing a great deal of symmetry.
Solving satisfiability problems using a novel microarray-based DNA computer.
Lin, Che-Hsin; Cheng, Hsiao-Ping; Yang, Chang-Biau; Yang, Chia-Ning
2007-01-01
An algorithm based on a modified sticker model accompanied with an advanced MEMS-based microarray technology is demonstrated to solve SAT problem, which has long served as a benchmark in DNA computing. Unlike conventional DNA computing algorithms needing an initial data pool to cover correct and incorrect answers and further executing a series of separation procedures to destroy the unwanted ones, we built solutions in parts to satisfy one clause in one step, and eventually solve the entire Boolean formula through steps. No time-consuming sample preparation procedures and delicate sample applying equipment were required for the computing process. Moreover, experimental results show the bound DNA sequences can sustain the chemical solutions during computing processes such that the proposed method shall be useful in dealing with large-scale problems.
Insightful problem solving in an Asian elephant.
Foerder, Preston; Galloway, Marie; Barthel, Tony; Moore, Donald E; Reiss, Diana
2011-01-01
The "aha" moment or the sudden arrival of the solution to a problem is a common human experience. Spontaneous problem solving without evident trial and error behavior in humans and other animals has been referred to as insight. Surprisingly, elephants, thought to be highly intelligent, have failed to exhibit insightful problem solving in previous cognitive studies. We tested whether three Asian elephants (Elephas maximus) would use sticks or other objects to obtain food items placed out-of-reach and overhead. Without prior trial and error behavior, a 7-year-old male Asian elephant showed spontaneous problem solving by moving a large plastic cube, on which he then stood, to acquire the food. In further testing he showed behavioral flexibility, using this technique to reach other items and retrieving the cube from various locations to use as a tool to acquire food. In the cube's absence, he generalized this tool utilization technique to other objects and, when given smaller objects, stacked them in an attempt to reach the food. The elephant's overall behavior was consistent with the definition of insightful problem solving. Previous failures to demonstrate this ability in elephants may have resulted not from a lack of cognitive ability but from the presentation of tasks requiring trunk-held sticks as potential tools, thereby interfering with the trunk's use as a sensory organ to locate the targeted food.
Direct Solve of Electrically Large Integral Equations for Problem Sizes to 1M Unknowns
NASA Technical Reports Server (NTRS)
Shaeffer, John
2008-01-01
Matrix methods for solving integral equations via direct solve LU factorization are presently limited to weeks to months of very expensive supercomputer time for problems sizes of several hundred thousand unknowns. This report presents matrix LU factor solutions for electromagnetic scattering problems for problem sizes to one million unknowns with thousands of right hand sides that run in mere days on PC level hardware. This EM solution is accomplished by utilizing the numerical low rank nature of spatially blocked unknowns using the Adaptive Cross Approximation for compressing the rank deficient blocks of the system Z matrix, the L and U factors, the right hand side forcing function and the final current solution. This compressed matrix solution is applied to a frequency domain EM solution of Maxwell's equations using standard Method of Moments approach. Compressed matrix storage and operations count leads to orders of magnitude reduction in memory and run time.
Quantum Heterogeneous Computing for Satellite Positioning Optimization
NASA Astrophysics Data System (ADS)
Bass, G.; Kumar, V.; Dulny, J., III
2016-12-01
Hard optimization problems occur in many fields of academic study and practical situations. We present results in which quantum heterogeneous computing is used to solve a real-world optimization problem: satellite positioning. Optimization problems like this can scale very rapidly with problem size, and become unsolvable with traditional brute-force methods. Typically, such problems have been approximately solved with heuristic approaches; however, these methods can take a long time to calculate and are not guaranteed to find optimal solutions. Quantum computing offers the possibility of producing significant speed-up and improved solution quality. There are now commercially available quantum annealing (QA) devices that are designed to solve difficult optimization problems. These devices have 1000+ quantum bits, but they have significant hardware size and connectivity limitations. We present a novel heterogeneous computing stack that combines QA and classical machine learning and allows the use of QA on problems larger than the quantum hardware could solve in isolation. We begin by analyzing the satellite positioning problem with a heuristic solver, the genetic algorithm. The classical computer's comparatively large available memory can explore the full problem space and converge to a solution relatively close to the true optimum. The QA device can then evolve directly to the optimal solution within this more limited space. Preliminary experiments, using the Quantum Monte Carlo (QMC) algorithm to simulate QA hardware, have produced promising results. Working with problem instances with known global minima, we find a solution within 8% in a matter of seconds, and within 5% in a few minutes. Future studies include replacing QMC with commercially available quantum hardware and exploring more problem sets and model parameters. Our results have important implications for how heterogeneous quantum computing can be used to solve difficult optimization problems in any field.
2012-09-03
prac- tice to solve these initial value problems. Additionally, the predictor / corrector methods are combined with adaptive stepsize and adaptive ...for implementing a numerical path tracking algorithm is to decide which predictor / corrector method to employ, how large to take the step ∆t, and what...the endgame algorithm . Output: A steady state solution Set ǫ = 1 while ǫ >= ǫend do set the stepsize ∆ǫ by using adaptive stepsize control algorithm
NASA Astrophysics Data System (ADS)
Khairnasov, K. Z.
2018-04-01
The paper presents a mathematical model for solving the problem of behavior of shell configurations under the action of static and dynamic impacts. The problem is solved in geometrically nonlinear statement with regard to the finite element method. The composite structures with different material layers are considered. The obtained equations are used to study the behavior of shell configurations under the action of dynamic loads. The results agree well with the experimental data.
Large-scale inverse model analyses employing fast randomized data reduction
NASA Astrophysics Data System (ADS)
Lin, Youzuo; Le, Ellen B.; O'Malley, Daniel; Vesselinov, Velimir V.; Bui-Thanh, Tan
2017-08-01
When the number of observations is large, it is computationally challenging to apply classical inverse modeling techniques. We have developed a new computationally efficient technique for solving inverse problems with a large number of observations (e.g., on the order of 107 or greater). Our method, which we call the randomized geostatistical approach (RGA), is built upon the principal component geostatistical approach (PCGA). We employ a data reduction technique combined with the PCGA to improve the computational efficiency and reduce the memory usage. Specifically, we employ a randomized numerical linear algebra technique based on a so-called "sketching" matrix to effectively reduce the dimension of the observations without losing the information content needed for the inverse analysis. In this way, the computational and memory costs for RGA scale with the information content rather than the size of the calibration data. Our algorithm is coded in Julia and implemented in the MADS open-source high-performance computational framework (http://mads.lanl.gov). We apply our new inverse modeling method to invert for a synthetic transmissivity field. Compared to a standard geostatistical approach (GA), our method is more efficient when the number of observations is large. Most importantly, our method is capable of solving larger inverse problems than the standard GA and PCGA approaches. Therefore, our new model inversion method is a powerful tool for solving large-scale inverse problems. The method can be applied in any field and is not limited to hydrogeological applications such as the characterization of aquifer heterogeneity.
a Novel Discrete Optimal Transport Method for Bayesian Inverse Problems
NASA Astrophysics Data System (ADS)
Bui-Thanh, T.; Myers, A.; Wang, K.; Thiery, A.
2017-12-01
We present the Augmented Ensemble Transform (AET) method for generating approximate samples from a high-dimensional posterior distribution as a solution to Bayesian inverse problems. Solving large-scale inverse problems is critical for some of the most relevant and impactful scientific endeavors of our time. Therefore, constructing novel methods for solving the Bayesian inverse problem in more computationally efficient ways can have a profound impact on the science community. This research derives the novel AET method for exploring a posterior by solving a sequence of linear programming problems, resulting in a series of transport maps which map prior samples to posterior samples, allowing for the computation of moments of the posterior. We show both theoretical and numerical results, indicating this method can offer superior computational efficiency when compared to other SMC methods. Most of this efficiency is derived from matrix scaling methods to solve the linear programming problem and derivative-free optimization for particle movement. We use this method to determine inter-well connectivity in a reservoir and the associated uncertainty related to certain parameters. The attached file shows the difference between the true parameter and the AET parameter in an example 3D reservoir problem. The error is within the Morozov discrepancy allowance with lower computational cost than other particle methods.
Efficient Implementation of an Optimal Interpolator for Large Spatial Data Sets
NASA Technical Reports Server (NTRS)
Memarsadeghi, Nargess; Mount, David M.
2007-01-01
Scattered data interpolation is a problem of interest in numerous areas such as electronic imaging, smooth surface modeling, and computational geometry. Our motivation arises from applications in geology and mining, which often involve large scattered data sets and a demand for high accuracy. The method of choice is ordinary kriging. This is because it is a best unbiased estimator. Unfortunately, this interpolant is computationally very expensive to compute exactly. For n scattered data points, computing the value of a single interpolant involves solving a dense linear system of size roughly n x n. This is infeasible for large n. In practice, kriging is solved approximately by local approaches that are based on considering only a relatively small'number of points that lie close to the query point. There are many problems with this local approach, however. The first is that determining the proper neighborhood size is tricky, and is usually solved by ad hoc methods such as selecting a fixed number of nearest neighbors or all the points lying within a fixed radius. Such fixed neighborhood sizes may not work well for all query points, depending on local density of the point distribution. Local methods also suffer from the problem that the resulting interpolant is not continuous. Meyer showed that while kriging produces smooth continues surfaces, it has zero order continuity along its borders. Thus, at interface boundaries where the neighborhood changes, the interpolant behaves discontinuously. Therefore, it is important to consider and solve the global system for each interpolant. However, solving such large dense systems for each query point is impractical. Recently a more principled approach to approximating kriging has been proposed based on a technique called covariance tapering. The problems arise from the fact that the covariance functions that are used in kriging have global support. Our implementations combine, utilize, and enhance a number of different approaches that have been introduced in literature for solving large linear systems for interpolation of scattered data points. For very large systems, exact methods such as Gaussian elimination are impractical since they require 0(n(exp 3)) time and 0(n(exp 2)) storage. As Billings et al. suggested, we use an iterative approach. In particular, we use the SYMMLQ method, for solving the large but sparse ordinary kriging systems that result from tapering. The main technical issue that need to be overcome in our algorithmic solution is that the points' covariance matrix for kriging should be symmetric positive definite. The goal of tapering is to obtain a sparse approximate representation of the covariance matrix while maintaining its positive definiteness. Furrer et al. used tapering to obtain a sparse linear system of the form Ax = b, where A is the tapered symmetric positive definite covariance matrix. Thus, Cholesky factorization could be used to solve their linear systems. They implemented an efficient sparse Cholesky decomposition method. They also showed if these tapers are used for a limited class of covariance models, the solution of the system converges to the solution of the original system. Matrix A in the ordinary kriging system, while symmetric, is not positive definite. Thus, their approach is not applicable to the ordinary kriging system. Therefore, we use tapering only to obtain a sparse linear system. Then, we use SYMMLQ to solve the ordinary kriging system. We show that solving large kriging systems becomes practical via tapering and iterative methods, and results in lower estimation errors compared to traditional local approaches, and significant memory savings compared to the original global system. We also developed a more efficient variant of the sparse SYMMLQ method for large ordinary kriging systems. This approach adaptively finds the correct local neighborhood for each query point in the interpolation process.
Jump phenomena. [large amplitude responses of nonlinear systems
NASA Technical Reports Server (NTRS)
Reiss, E. L.
1980-01-01
The paper considers jump phenomena composed of large amplitude responses of nonlinear systems caused by small amplitude disturbances. Physical problems where large jumps in the solution amplitude are important features of the response are described, including snap buckling of elastic shells, chemical reactions leading to combustion and explosion, and long-term climatic changes of the earth's atmosphere. A new method of rational functions was then developed which consists of representing the solutions of the jump problems as rational functions of the small disturbance parameter; this method can solve jump problems explicitly.
Detwiler, R.L.; Mehl, S.; Rajaram, H.; Cheung, W.W.
2002-01-01
Numerical solution of large-scale ground water flow and transport problems is often constrained by the convergence behavior of the iterative solvers used to solve the resulting systems of equations. We demonstrate the ability of an algebraic multigrid algorithm (AMG) to efficiently solve the large, sparse systems of equations that result from computational models of ground water flow and transport in large and complex domains. Unlike geometric multigrid methods, this algorithm is applicable to problems in complex flow geometries, such as those encountered in pore-scale modeling of two-phase flow and transport. We integrated AMG into MODFLOW 2000 to compare two- and three-dimensional flow simulations using AMG to simulations using PCG2, a preconditioned conjugate gradient solver that uses the modified incomplete Cholesky preconditioner and is included with MODFLOW 2000. CPU times required for convergence with AMG were up to 140 times faster than those for PCG2. The cost of this increased speed was up to a nine-fold increase in required random access memory (RAM) for the three-dimensional problems and up to a four-fold increase in required RAM for the two-dimensional problems. We also compared two-dimensional numerical simulations of steady-state transport using AMG and the generalized minimum residual method with an incomplete LU-decomposition preconditioner. For these transport simulations, AMG yielded increased speeds of up to 17 times with only a 20% increase in required RAM. The ability of AMG to solve flow and transport problems in large, complex flow systems and its ready availability make it an ideal solver for use in both field-scale and pore-scale modeling.
Solving large sparse eigenvalue problems on supercomputers
NASA Technical Reports Server (NTRS)
Philippe, Bernard; Saad, Youcef
1988-01-01
An important problem in scientific computing consists in finding a few eigenvalues and corresponding eigenvectors of a very large and sparse matrix. The most popular methods to solve these problems are based on projection techniques on appropriate subspaces. The main attraction of these methods is that they only require the use of the matrix in the form of matrix by vector multiplications. The implementations on supercomputers of two such methods for symmetric matrices, namely Lanczos' method and Davidson's method are compared. Since one of the most important operations in these two methods is the multiplication of vectors by the sparse matrix, methods of performing this operation efficiently are discussed. The advantages and the disadvantages of each method are compared and implementation aspects are discussed. Numerical experiments on a one processor CRAY 2 and CRAY X-MP are reported. Possible parallel implementations are also discussed.
Discovering Motifs in Biological Sequences Using the Micron Automata Processor.
Roy, Indranil; Aluru, Srinivas
2016-01-01
Finding approximately conserved sequences, called motifs, across multiple DNA or protein sequences is an important problem in computational biology. In this paper, we consider the (l, d) motif search problem of identifying one or more motifs of length l present in at least q of the n given sequences, with each occurrence differing from the motif in at most d substitutions. The problem is known to be NP-complete, and the largest solved instance reported to date is (26,11). We propose a novel algorithm for the (l,d) motif search problem using streaming execution over a large set of non-deterministic finite automata (NFA). This solution is designed to take advantage of the micron automata processor, a new technology close to deployment that can simultaneously execute multiple NFA in parallel. We demonstrate the capability for solving much larger instances of the (l, d) motif search problem using the resources available within a single automata processor board, by estimating run-times for problem instances (39,18) and (40,17). The paper serves as a useful guide to solving problems using this new accelerator technology.
Linear solver performance in elastoplastic problem solution on GPU cluster
NASA Astrophysics Data System (ADS)
Khalevitsky, Yu. V.; Konovalov, A. V.; Burmasheva, N. V.; Partin, A. S.
2017-12-01
Applying the finite element method to severe plastic deformation problems involves solving linear equation systems. While the solution procedure is relatively hard to parallelize and computationally intensive by itself, a long series of large scale systems need to be solved for each problem. When dealing with fine computational meshes, such as in the simulations of three-dimensional metal matrix composite microvolume deformation, tens and hundreds of hours may be needed to complete the whole solution procedure, even using modern supercomputers. In general, one of the preconditioned Krylov subspace methods is used in a linear solver for such problems. The method convergence highly depends on the operator spectrum of a problem stiffness matrix. In order to choose the appropriate method, a series of computational experiments is used. Different methods may be preferable for different computational systems for the same problem. In this paper we present experimental data obtained by solving linear equation systems from an elastoplastic problem on a GPU cluster. The data can be used to substantiate the choice of the appropriate method for a linear solver to use in severe plastic deformation simulations.
Berteletti, Ilaria; Prado, Jérôme; Booth, James R
2014-08-01
Greater skill in solving single-digit multiplication problems requires a progressive shift from a reliance on numerical to verbal mechanisms over development. Children with mathematical learning disability (MD), however, are thought to suffer from a specific impairment in numerical mechanisms. Here we tested the hypothesis that this impairment might prevent MD children from transitioning toward verbal mechanisms when solving single-digit multiplication problems. Brain activations during multiplication problems were compared in MD and typically developing (TD) children (3rd to 7th graders) in numerical and verbal regions which were individuated by independent localizer tasks. We used small (e.g., 2 × 3) and large (e.g., 7 × 9) problems as these problems likely differ in their reliance on verbal versus numerical mechanisms. Results indicate that MD children have reduced activations in both the verbal (i.e., left inferior frontal gyrus and left middle temporal to superior temporal gyri) and the numerical (i.e., right superior parietal lobule including intra-parietal sulcus) regions suggesting that both mechanisms are impaired. Moreover, the only reliable activation observed for MD children was in the numerical region when solving small problems. This suggests that MD children could effectively engage numerical mechanisms only for the easier problems. Conversely, TD children showed a modulation of activation with problem size in the verbal regions. This suggests that TD children were effectively engaging verbal mechanisms for the easier problems. Moreover, TD children with better language skills were more effective at engaging verbal mechanisms. In conclusion, results suggest that the numerical- and language-related processes involved in solving multiplication problems are impaired in MD children. Published by Elsevier Ltd.
Evaluation of the eigenvalue method in the solution of transient heat conduction problems
NASA Astrophysics Data System (ADS)
Landry, D. W.
1985-01-01
The eigenvalue method is evaluated to determine the advantages and disadvantages of the method as compared to fully explicit, fully implicit, and Crank-Nicolson methods. Time comparisons and accuracy comparisons are made in an effort to rank the eigenvalue method in relation to the comparison schemes. The eigenvalue method is used to solve the parabolic heat equation in multidimensions with transient temperatures. Extensions into three dimensions are made to determine the method's feasibility in handling large geometry problems requiring great numbers of internal mesh points. The eigenvalue method proves to be slightly better in accuracy than the comparison routines because of an exact treatment, as opposed to a numerical approximation, of the time derivative in the heat equation. It has the potential of being a very powerful routine in solving long transient type problems. The method is not well suited to finely meshed grid arrays or large regions because of the time and memory requirements necessary for calculating large sets of eigenvalues and eigenvectors.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Baker, Kyri; Toomey, Bridget
Evolving power systems with increasing levels of stochasticity call for a need to solve optimal power flow problems with large quantities of random variables. Weather forecasts, electricity prices, and shifting load patterns introduce higher levels of uncertainty and can yield optimization problems that are difficult to solve in an efficient manner. Solution methods for single chance constraints in optimal power flow problems have been considered in the literature, ensuring single constraints are satisfied with a prescribed probability; however, joint chance constraints, ensuring multiple constraints are simultaneously satisfied, have predominantly been solved via scenario-based approaches or by utilizing Boole's inequality asmore » an upper bound. In this paper, joint chance constraints are used to solve an AC optimal power flow problem while preventing overvoltages in distribution grids under high penetrations of photovoltaic systems. A tighter version of Boole's inequality is derived and used to provide a new upper bound on the joint chance constraint, and simulation results are shown demonstrating the benefit of the proposed upper bound. The new framework allows for a less conservative and more computationally efficient solution to considering joint chance constraints, specifically regarding preventing overvoltages.« less
Studies of Verbal Problem Solving. 1. Two Performance-Aiding Programs
1977-09-01
5. Effective Date of Contract 6. Contract Expiration Date 7. Amount of Contract 8. Contract Number 9. Principal Investigator 10 . Scientific...7 10 3. A Mystery Solved by Propositional Calculus 16 1v I. INTRODUCTION Certain intellectual tasks, such as estimating and combining prob...aids or without any special training, they get the right answer within 15 minutes or so (about 80% of one large psychology class solved it). A few
NASA Astrophysics Data System (ADS)
Pei, Jun; Liu, Xinbao; Pardalos, Panos M.; Fan, Wenjuan; Wang, Ling; Yang, Shanlin
2016-03-01
Motivated by applications in manufacturing industry, we consider a supply chain scheduling problem, where each job is characterised by non-identical sizes, different release times and unequal processing times. The objective is to minimise the makespan by making batching and sequencing decisions. The problem is formalised as a mixed integer programming model and proved to be strongly NP-hard. Some structural properties are presented for both the general case and a special case. Based on these properties, a lower bound is derived, and a novel two-phase heuristic (TP-H) is developed to solve the problem, which guarantees to obtain a worst case performance ratio of ?. Computational experiments with a set of different sizes of random instances are conducted to evaluate the proposed approach TP-H, which is superior to another two heuristics proposed in the literature. Furthermore, the experimental results indicate that TP-H can effectively and efficiently solve large-size problems in a reasonable time.
An optical solution for the traveling salesman problem.
Haist, Tobias; Osten, Wolfgang
2007-08-06
We introduce an optical method based on white light interferometry in order to solve the well-known NP-complete traveling salesman problem. To our knowledge it is the first time that a method for the reduction of non-polynomial time to quadratic time has been proposed. We will show that this achievement is limited by the number of available photons for solving the problem. It will turn out that this number of photons is proportional to N(N) for a traveling salesman problem with N cities and that for large numbers of cities the method in practice therefore is limited by the signal-to-noise ratio. The proposed method is meant purely as a gedankenexperiment.
Knowledge-based control for robot self-localization
NASA Technical Reports Server (NTRS)
Bennett, Bonnie Kathleen Holte
1993-01-01
Autonomous robot systems are being proposed for a variety of missions including the Mars rover/sample return mission. Prior to any other mission objectives being met, an autonomous robot must be able to determine its own location. This will be especially challenging because location sensors like GPS, which are available on Earth, will not be useful, nor will INS sensors because their drift is too large. Another approach to self-localization is required. In this paper, we describe a novel approach to localization by applying a problem solving methodology. The term 'problem solving' implies a computational technique based on logical representational and control steps. In this research, these steps are derived from observing experts solving localization problems. The objective is not specifically to simulate human expertise but rather to apply its techniques where appropriate for computational systems. In doing this, we describe a model for solving the problem and a system built on that model, called localization control and logic expert (LOCALE), which is a demonstration of concept for the approach and the model. The results of this work represent the first successful solution to high-level control aspects of the localization problem.
Linear complementarity formulation for 3D frictional sliding problems
Kaven, Joern; Hickman, Stephen H.; Davatzes, Nicholas C.; Mutlu, Ovunc
2012-01-01
Frictional sliding on quasi-statically deforming faults and fractures can be modeled efficiently using a linear complementarity formulation. We review the formulation in two dimensions and expand the formulation to three-dimensional problems including problems of orthotropic friction. This formulation accurately reproduces analytical solutions to static Coulomb friction sliding problems. The formulation accounts for opening displacements that can occur near regions of non-planarity even under large confining pressures. Such problems are difficult to solve owing to the coupling of relative displacements and tractions; thus, many geomechanical problems tend to neglect these effects. Simple test cases highlight the importance of including friction and allowing for opening when solving quasi-static fault mechanics models. These results also underscore the importance of considering the effects of non-planarity in modeling processes associated with crustal faulting.
Students Advise Fortune 500 Company: Designing a Problem-Based Learning Community
ERIC Educational Resources Information Center
Brzovic, Kathy; Matz, S. Irene
2009-01-01
This article describes the process of planning and implementing a problem-based learning community. Business and communication students from a large university in the Western United States competed in teams to solve an authentic business problem posed by a Fortune 500 company. The company's willingness to adopt some of their recommendations…
Wade, Shari L.; Stancin, Terry; Kirkwood, Michael; Brown, Tanya Maines; Rochester, Mayo Clinic; McMullen, Kendra M.; Taylor, H. Gerry
2013-01-01
Objective To test the efficacy of Counselor-Assisted Problem Solving (CAPS) versus an internet resources comparison (IRC) condition in reducing behavior problems in adolescents following traumatic brain injury (TBI). Design Randomized clinical trial with interviewers naïve to treatment condition. Setting Three large tertiary children's hospitals and two general hospitals with pediatric commitment. Participants 132 children ages 12-17 years hospitalized during the previous 6 months for moderate to severe TBI. Interventions Participants in CAPS (n = 65) completed 8-12 online modules providing training in problem solving, communication skills, and self-regulation and subsequent synchronous videoconferences with a therapist. Participants in the IRC group (n = 67) received links to internet resources about pediatric TBI. Main Outcome Measures Child Behavior Checklist (CBCL) administered before and after completion of treatment (i.e., approximately six months after treatment initiation). Results Post hoc analysis of covariance (ANCOVA), controlling for pre-treatment scores, was used to examine group differences in behavior problems in the entire sample and among older (n=59) and younger adolescents (n=53). Among older but not younger adolescents, CAPS resulted in greater improvements on multiple dimensions of externalizing behavior problems than did IRC. Conclusion Online problem-solving therapy may be effective in reducing behavior problems in older adolescent survivors of moderate-severe TBI. PMID:23640543
Large-N -approximated field theory for multipartite entanglement
NASA Astrophysics Data System (ADS)
Facchi, P.; Florio, G.; Parisi, G.; Pascazio, S.; Scardicchio, A.
2015-12-01
We try to characterize the statistics of multipartite entanglement of the random states of an n -qubit system. Unable to solve the problem exactly we generalize it, replacing complex numbers with real vectors with Nc components (the original problem is recovered for Nc=2 ). Studying the leading diagrams in the large-Nc approximation, we unearth the presence of a phase transition and, in an explicit example, show that the so-called entanglement frustration disappears in the large-Nc limit.
Computer problem-solving coaches for introductory physics: Design and usability studies
NASA Astrophysics Data System (ADS)
Ryan, Qing X.; Frodermann, Evan; Heller, Kenneth; Hsu, Leonardo; Mason, Andrew
2016-06-01
The combination of modern computing power, the interactivity of web applications, and the flexibility of object-oriented programming may finally be sufficient to create computer coaches that can help students develop metacognitive problem-solving skills, an important competence in our rapidly changing technological society. However, no matter how effective such coaches might be, they will only be useful if they are attractive to students. We describe the design and testing of a set of web-based computer programs that act as personal coaches to students while they practice solving problems from introductory physics. The coaches are designed to supplement regular human instruction, giving students access to effective forms of practice outside class. We present results from large-scale usability tests of the computer coaches and discuss their implications for future versions of the coaches.
Bicriteria Network Optimization Problem using Priority-based Genetic Algorithm
NASA Astrophysics Data System (ADS)
Gen, Mitsuo; Lin, Lin; Cheng, Runwei
Network optimization is being an increasingly important and fundamental issue in the fields such as engineering, computer science, operations research, transportation, telecommunication, decision support systems, manufacturing, and airline scheduling. In many applications, however, there are several criteria associated with traversing each edge of a network. For example, cost and flow measures are both important in the networks. As a result, there has been recent interest in solving Bicriteria Network Optimization Problem. The Bicriteria Network Optimization Problem is known a NP-hard. The efficient set of paths may be very large, possibly exponential in size. Thus the computational effort required to solve it can increase exponentially with the problem size in the worst case. In this paper, we propose a genetic algorithm (GA) approach used a priority-based chromosome for solving the bicriteria network optimization problem including maximum flow (MXF) model and minimum cost flow (MCF) model. The objective is to find the set of Pareto optimal solutions that give possible maximum flow with minimum cost. This paper also combines Adaptive Weight Approach (AWA) that utilizes some useful information from the current population to readjust weights for obtaining a search pressure toward a positive ideal point. Computer simulations show the several numerical experiments by using some difficult-to-solve network design problems, and show the effectiveness of the proposed method.
Enhanced algorithms for stochastic programming
DOE Office of Scientific and Technical Information (OSTI.GOV)
Krishna, Alamuru S.
1993-09-01
In this dissertation, we present some of the recent advances made in solving two-stage stochastic linear programming problems of large size and complexity. Decomposition and sampling are two fundamental components of techniques to solve stochastic optimization problems. We describe improvements to the current techniques in both these areas. We studied different ways of using importance sampling techniques in the context of Stochastic programming, by varying the choice of approximation functions used in this method. We have concluded that approximating the recourse function by a computationally inexpensive piecewise-linear function is highly efficient. This reduced the problem from finding the mean ofmore » a computationally expensive functions to finding that of a computationally inexpensive function. Then we implemented various variance reduction techniques to estimate the mean of a piecewise-linear function. This method achieved similar variance reductions in orders of magnitude less time than, when we directly applied variance-reduction techniques directly on the given problem. In solving a stochastic linear program, the expected value problem is usually solved before a stochastic solution and also to speed-up the algorithm by making use of the information obtained from the solution of the expected value problem. We have devised a new decomposition scheme to improve the convergence of this algorithm.« less
Influence of the large-small split effect on strategy choice in complex subtraction.
Xiang, Yan Hui; Wu, Hao; Shang, Rui Hong; Chao, Xiaomei; Ren, Ting Ting; Zheng, Li Ling; Mo, Lei
2018-04-01
Two main theories have been used to explain the arithmetic split effect: decision-making process theory and strategy choice theory. Using the inequality paradigm, previous studies have confirmed that individuals tend to adopt a plausibility-checking strategy and a whole-calculation strategy to solve large and small split problems in complex addition arithmetic, respectively. This supports strategy choice theory, but it is unknown whether this theory also explains performance in solving different split problems in complex subtraction arithmetic. This study used small, intermediate and large split sizes, with each split condition being further divided into problems requiring and not requiring borrowing. The reaction times (RTs) for large and intermediate splits were significantly shorter than those for small splits, while accuracy was significantly higher for large and middle splits than for small splits, reflecting no speed-accuracy trade-off. Further, RTs and accuracy differed significantly between the borrow and no-borrow conditions only for small splits. This study indicates that strategy choice theory is suitable to explain the split effect in complex subtraction arithmetic. That is, individuals tend to choose the plausibility-checking strategy or the whole-calculation strategy according to the split size. © 2016 International Union of Psychological Science.
Structure preserving parallel algorithms for solving the Bethe–Salpeter eigenvalue problem
Shao, Meiyue; da Jornada, Felipe H.; Yang, Chao; ...
2015-10-02
The Bethe–Salpeter eigenvalue problem is a dense structured eigenvalue problem arising from discretized Bethe–Salpeter equation in the context of computing exciton energies and states. A computational challenge is that at least half of the eigenvalues and the associated eigenvectors are desired in practice. In this paper, we establish the equivalence between Bethe–Salpeter eigenvalue problems and real Hamiltonian eigenvalue problems. Based on theoretical analysis, structure preserving algorithms for a class of Bethe–Salpeter eigenvalue problems are proposed. We also show that for this class of problems all eigenvalues obtained from the Tamm–Dancoff approximation are overestimated. In order to solve large scale problemsmore » of practical interest, we discuss parallel implementations of our algorithms targeting distributed memory systems. Finally, several numerical examples are presented to demonstrate the efficiency and accuracy of our algorithms.« less
Shaw, William S; Feuerstein, Michael; Miller, Virginia I; Wood, Patricia M
2003-08-01
Improving health and work outcomes for individuals with work related upper extremity disorders (WRUEDs) may require a broad assessment of potential return to work barriers by engaging workers in collaborative problem solving. In this study, half of all nurse case managers from a large workers' compensation system were randomly selected and invited to participate in a randomized, controlled trial of an integrated case management (ICM) approach for WRUEDs. The focus of ICM was problem solving skills training and workplace accommodation. Volunteer nurses attended a 2 day ICM training workshop including instruction in a 6 step process to engage clients in problem solving to overcome barriers to recovery. A chart review of WRUED case management reports (n = 70) during the following 2 years was conducted to extract case managers' reports of barriers to recovery and return to work. Case managers documented from 0 to 21 barriers per case (M = 6.24, SD = 4.02) within 5 domains: signs and symptoms (36%), work environment (27%), medical care (13%), functional limitations (12%), and coping (12%). Compared with case managers who did not receive the training (n = 67), workshop participants identified more barriers related to signs and symptoms, work environment, functional limitations, and coping (p < .05), but not to medical care. Problem solving skills training may help focus case management services on the most salient recovery factors affecting return to work.
P1 Nonconforming Finite Element Method for the Solution of Radiation Transport Problems
NASA Technical Reports Server (NTRS)
Kang, Kab S.
2002-01-01
The simulation of radiation transport in the optically thick flux-limited diffusion regime has been identified as one of the most time-consuming tasks within large simulation codes. Due to multimaterial complex geometry, the radiation transport system must often be solved on unstructured grids. In this paper, we investigate the behavior and the benefits of the unstructured P(sub 1) nonconforming finite element method, which has proven to be flexible and effective on related transport problems, in solving unsteady implicit nonlinear radiation diffusion problems using Newton and Picard linearization methods. Key words. nonconforrning finite elements, radiation transport, inexact Newton linearization, multigrid preconditioning
The use of Lanczos's method to solve the large generalized symmetric definite eigenvalue problem
NASA Technical Reports Server (NTRS)
Jones, Mark T.; Patrick, Merrell L.
1989-01-01
The generalized eigenvalue problem, Kx = Lambda Mx, is of significant practical importance, especially in structural enginering where it arises as the vibration and buckling problem. A new algorithm, LANZ, based on Lanczos's method is developed. LANZ uses a technique called dynamic shifting to improve the efficiency and reliability of the Lanczos algorithm. A new algorithm for solving the tridiagonal matrices that arise when using Lanczos's method is described. A modification of Parlett and Scott's selective orthogonalization algorithm is proposed. Results from an implementation of LANZ on a Convex C-220 show it to be superior to a subspace iteration code.
Aerospace applications of integer and combinatorial optimization
NASA Technical Reports Server (NTRS)
Padula, S. L.; Kincaid, R. K.
1995-01-01
Research supported by NASA Langley Research Center includes many applications of aerospace design optimization and is conducted by teams of applied mathematicians and aerospace engineers. This paper investigates the benefits from this combined expertise in solving combinatorial optimization problems. Applications range from the design of large space antennas to interior noise control. A typical problem, for example, seeks the optimal locations for vibration-damping devices on a large space structure and is expressed as a mixed/integer linear programming problem with more than 1500 design variables.
[Research on non-rigid registration of multi-modal medical image based on Demons algorithm].
Hao, Peibo; Chen, Zhen; Jiang, Shaofeng; Wang, Yang
2014-02-01
Non-rigid medical image registration is a popular subject in the research areas of the medical image and has an important clinical value. In this paper we put forward an improved algorithm of Demons, together with the conservation of gray model and local structure tensor conservation model, to construct a new energy function processing multi-modal registration problem. We then applied the L-BFGS algorithm to optimize the energy function and solve complex three-dimensional data optimization problem. And finally we used the multi-scale hierarchical refinement ideas to solve large deformation registration. The experimental results showed that the proposed algorithm for large de formation and multi-modal three-dimensional medical image registration had good effects.
Analysis of Algorithms: Coping with Hard Problems
ERIC Educational Resources Information Center
Kolata, Gina Bari
1974-01-01
Although today's computers can perform as many as one million operations per second, there are many problems that are still too large to be solved in a straightforward manner. Recent work indicates that many approximate solutions are useful and more efficient than exact solutions. (Author/RH)
Li, Yongming; Ma, Zhiyao; Tong, Shaocheng
2017-09-01
The problem of adaptive fuzzy output-constrained tracking fault-tolerant control (FTC) is investigated for the large-scale stochastic nonlinear systems of pure-feedback form. The nonlinear systems considered in this paper possess the unstructured uncertainties, unknown interconnected terms and unknown nonaffine nonlinear faults. The fuzzy logic systems are employed to identify the unknown lumped nonlinear functions so that the problems of structured uncertainties can be solved. An adaptive fuzzy state observer is designed to solve the nonmeasurable state problem. By combining the barrier Lyapunov function theory, adaptive decentralized and stochastic control principles, a novel fuzzy adaptive output-constrained FTC approach is constructed. All the signals in the closed-loop system are proved to be bounded in probability and the system outputs are constrained in a given compact set. Finally, the applicability of the proposed controller is well carried out by a simulation example.
Matching by linear programming and successive convexification.
Jiang, Hao; Drew, Mark S; Li, Ze-Nian
2007-06-01
We present a novel convex programming scheme to solve matching problems, focusing on the challenging problem of matching in a large search range and with cluttered background. Matching is formulated as metric labeling with L1 regularization terms, for which we propose a novel linear programming relaxation method and an efficient successive convexification implementation. The unique feature of the proposed relaxation scheme is that a much smaller set of basis labels is used to represent the original label space. This greatly reduces the size of the searching space. A successive convexification scheme solves the labeling problem in a coarse to fine manner. Importantly, the original cost function is reconvexified at each stage, in the new focus region only, and the focus region is updated so as to refine the searching result. This makes the method well-suited for large label set matching. Experiments demonstrate successful applications of the proposed matching scheme in object detection, motion estimation, and tracking.
NASA Astrophysics Data System (ADS)
Kuncoro, K. S.; Junaedi, I.; Dwijanto
2018-03-01
This study aimed to reveal the effectiveness of Project Based Learning with Resource Based Learning approach computer-aided program and analyzed problem-solving abilities in terms of problem-solving steps based on Polya stages. The research method used was mixed method with sequential explanatory design. The subject of this research was the students of math semester 4. The results showed that the S-TPS (Strong Top Problem Solving) and W-TPS (Weak Top Problem Solving) had good problem-solving abilities in each problem-solving indicator. The problem-solving ability of S-MPS (Strong Middle Problem Solving) and (Weak Middle Problem Solving) in each indicator was good. The subject of S-BPS (Strong Bottom Problem Solving) had a difficulty in solving the problem with computer program, less precise in writing the final conclusion and could not reflect the problem-solving process using Polya’s step. While the Subject of W-BPS (Weak Bottom Problem Solving) had not been able to meet almost all the indicators of problem-solving. The subject of W-BPS could not precisely made the initial table of completion so that the completion phase with Polya’s step was constrained.
Bakallbashi, Eni; Vyas, Anjali; Vaswani, Nikita; Rosales, David; Russell, David; Dowding, Dawn; Bernstein, Michael; Abdelaal, Hany; Hawkey, Regina
2015-01-01
An internal employee challenge competition is a way to promote staff engagement and generate innovative business solutions. This Spotlight on Leadership focuses on the approach that a large not-for-profit healthcare organization, the Visiting Nurse Service of New York, took in designing and executing an innovation challenge. The challenge leveraged internal staff expertise and promoted wide participation. This model is 1 that can be replicated by organizations as leaders work to engage employees at the point of service in organization-wide problem solving.
Engineering neural systems for high-level problem solving.
Sylvester, Jared; Reggia, James
2016-07-01
There is a long-standing, sometimes contentious debate in AI concerning the relative merits of a symbolic, top-down approach vs. a neural, bottom-up approach to engineering intelligent machine behaviors. While neurocomputational methods excel at lower-level cognitive tasks (incremental learning for pattern classification, low-level sensorimotor control, fault tolerance and processing of noisy data, etc.), they are largely non-competitive with top-down symbolic methods for tasks involving high-level cognitive problem solving (goal-directed reasoning, metacognition, planning, etc.). Here we take a step towards addressing this limitation by developing a purely neural framework named galis. Our goal in this work is to integrate top-down (non-symbolic) control of a neural network system with more traditional bottom-up neural computations. galis is based on attractor networks that can be "programmed" with temporal sequences of hand-crafted instructions that control problem solving by gating the activity retention of, communication between, and learning done by other neural networks. We demonstrate the effectiveness of this approach by showing that it can be applied successfully to solve sequential card matching problems, using both human performance and a top-down symbolic algorithm as experimental controls. Solving this kind of problem makes use of top-down attention control and the binding together of visual features in ways that are easy for symbolic AI systems but not for neural networks to achieve. Our model can not only be instructed on how to solve card matching problems successfully, but its performance also qualitatively (and sometimes quantitatively) matches the performance of both human subjects that we had perform the same task and the top-down symbolic algorithm that we used as an experimental control. We conclude that the core principles underlying the galis framework provide a promising approach to engineering purely neurocomputational systems for problem-solving tasks that in people require higher-level cognitive functions. Copyright © 2016 Elsevier Ltd. All rights reserved.
Hybrid Microgrid Configuration Optimization with Evolutionary Algorithms
NASA Astrophysics Data System (ADS)
Lopez, Nicolas
This dissertation explores the Renewable Energy Integration Problem, and proposes a Genetic Algorithm embedded with a Monte Carlo simulation to solve large instances of the problem that are impractical to solve via full enumeration. The Renewable Energy Integration Problem is defined as finding the optimum set of components to supply the electric demand to a hybrid microgrid. The components considered are solar panels, wind turbines, diesel generators, electric batteries, connections to the power grid and converters, which can be inverters and/or rectifiers. The methodology developed is explained as well as the combinatorial formulation. In addition, 2 case studies of a single objective optimization version of the problem are presented, in order to minimize cost and to minimize global warming potential (GWP) followed by a multi-objective implementation of the offered methodology, by utilizing a non-sorting Genetic Algorithm embedded with a monte Carlo Simulation. The method is validated by solving a small instance of the problem with known solution via a full enumeration algorithm developed by NREL in their software HOMER. The dissertation concludes that the evolutionary algorithms embedded with Monte Carlo simulation namely modified Genetic Algorithms are an efficient form of solving the problem, by finding approximate solutions in the case of single objective optimization, and by approximating the true Pareto front in the case of multiple objective optimization of the Renewable Energy Integration Problem.
NASA Astrophysics Data System (ADS)
Iswari, T.; Asih, A. M. S.
2018-04-01
In the logistics system, transportation plays an important role to connect every element in the supply chain, but it can produces the greatest cost. Therefore, it is important to make the transportation costs as minimum as possible. Reducing the transportation cost can be done in several ways. One of the ways to minimizing the transportation cost is by optimizing the routing of its vehicles. It refers to Vehicle Routing Problem (VRP). The most common type of VRP is Capacitated Vehicle Routing Problem (CVRP). In CVRP, the vehicles have their own capacity and the total demands from the customer should not exceed the capacity of the vehicle. CVRP belongs to the class of NP-hard problems. These NP-hard problems make it more complex to solve such that exact algorithms become highly time-consuming with the increases in problem sizes. Thus, for large-scale problem instances, as typically found in industrial applications, finding an optimal solution is not practicable. Therefore, this paper uses two kinds of metaheuristics approach to solving CVRP. Those are Genetic Algorithm and Particle Swarm Optimization. This paper compares the results of both algorithms and see the performance of each algorithm. The results show that both algorithms perform well in solving CVRP but still needs to be improved. From algorithm testing and numerical example, Genetic Algorithm yields a better solution than Particle Swarm Optimization in total distance travelled.
NASA Astrophysics Data System (ADS)
Karimi, Hamed; Rosenberg, Gili; Katzgraber, Helmut G.
2017-10-01
We present and apply a general-purpose, multistart algorithm for improving the performance of low-energy samplers used for solving optimization problems. The algorithm iteratively fixes the value of a large portion of the variables to values that have a high probability of being optimal. The resulting problems are smaller and less connected, and samplers tend to give better low-energy samples for these problems. The algorithm is trivially parallelizable since each start in the multistart algorithm is independent, and could be applied to any heuristic solver that can be run multiple times to give a sample. We present results for several classes of hard problems solved using simulated annealing, path-integral quantum Monte Carlo, parallel tempering with isoenergetic cluster moves, and a quantum annealer, and show that the success metrics and the scaling are improved substantially. When combined with this algorithm, the quantum annealer's scaling was substantially improved for native Chimera graph problems. In addition, with this algorithm the scaling of the time to solution of the quantum annealer is comparable to the Hamze-de Freitas-Selby algorithm on the weak-strong cluster problems introduced by Boixo et al. Parallel tempering with isoenergetic cluster moves was able to consistently solve three-dimensional spin glass problems with 8000 variables when combined with our method, whereas without our method it could not solve any.
ERIC Educational Resources Information Center
DOLBY, J.L.; AND OTHERS
THE STUDY IS CONCERNED WITH THE LINGUISTIC PROBLEM INVOLVED IN TEXT COMPRESSION--EXTRACTING, INDEXING, AND THE AUTOMATIC CREATION OF SPECIAL-PURPOSE CITATION DICTIONARIES. IN SPITE OF EARLY SUCCESS IN USING LARGE-SCALE COMPUTERS TO AUTOMATE CERTAIN HUMAN TASKS, THESE PROBLEMS REMAIN AMONG THE MOST DIFFICULT TO SOLVE. ESSENTIALLY, THE PROBLEM IS TO…
Mexican Hat Wavelet Kernel ELM for Multiclass Classification.
Wang, Jie; Song, Yi-Fan; Ma, Tian-Lei
2017-01-01
Kernel extreme learning machine (KELM) is a novel feedforward neural network, which is widely used in classification problems. To some extent, it solves the existing problems of the invalid nodes and the large computational complexity in ELM. However, the traditional KELM classifier usually has a low test accuracy when it faces multiclass classification problems. In order to solve the above problem, a new classifier, Mexican Hat wavelet KELM classifier, is proposed in this paper. The proposed classifier successfully improves the training accuracy and reduces the training time in the multiclass classification problems. Moreover, the validity of the Mexican Hat wavelet as a kernel function of ELM is rigorously proved. Experimental results on different data sets show that the performance of the proposed classifier is significantly superior to the compared classifiers.
Hanten, Gerri; Cook, Lori; Orsten, Kimberley; Chapman, Sandra B; Li, Xiaoqi; Wilde, Elisabeth A; Schnelle, Kathleen P; Levin, Harvey S
2011-02-01
Social problem solving was assessed in 28 youth ages 12-19 years (15 with moderate to severe traumatic brain injury (TBI), 13 uninjured) using a naturalistic, computerized virtual reality (VR) version of the Interpersonal Negotiations Strategy interview (Yeates, Schultz, & Selman, 1991). In each scenario, processing load condition was varied in terms of number of characters and amount of information. Adolescents viewed animated scenarios depicting social conflict in a virtual microworld environment from an avatar's viewpoint, and were questioned on four problem solving steps: defining the problem, generating solutions, selecting solutions, and evaluating the likely outcome. Scoring was based on a developmental scale in which responses were judged as impulsive, unilateral, reciprocal, or collaborative, in order of increasing score. Adolescents with TBI were significantly impaired on the summary VR-Social Problem Solving (VR-SPS) score in Condition A (2 speakers, no irrelevant information), p=0.005; in Condition B (2 speakers+irrelevant information), p=0.035; and Condition C (4 speakers+irrelevant information), p=0.008. Effect sizes (Cohen's D) were large (A=1.40, B=0.96, C=1.23). Significant group differences were strongest and most consistent for defining the problems and evaluating outcomes. The relation of task performance to cortical thickness of specific brain regions was also explored, with significant relations found with orbitofrontal regions, the frontal pole, the cuneus, and the temporal pole. Results are discussed in the context of specific cognitive and neural mechanisms underlying social problem solving deficits after childhood TBI. Copyright © 2010 Elsevier Ltd. All rights reserved.
Hanten, Gerri; Cook, Lori; Orsten, Kimberley; Chapman, Sandra B.; Li, Xiaoqi; Wilde, Elisabeth A.; Schnelle, Kathleen P.; Levin, Harvey S.
2011-01-01
Social problem solving was assessed in 28 youth ages 12–19 years (15 with moderate to severe traumatic brain injury (TBI), 13 uninjured) using a naturalistic, computerized virtual reality (VR) version of the Interpersonal Negotiations Strategy interview (Yeates, Schultz, & Selman, 1991). In each scenario, processing load condition was varied in terms of number of characters and amount of information. Adolescents viewed animated scenarios depicting social conflict in a virtual microworld environment from an avatar’s viewpoint, and were questioned on four problem solving steps: defining the problem, generating solutions, selecting solutions, and evaluating the likely outcome. Scoring was based on a developmental scale in which responses were judged as impulsive, unilateral, reciprocal, or collaborative, in order of increasing score. Adolescents with TBI were significantly impaired on the summary VR-Social Problem Solving (VR-SPS) score in Condition A (2 speakers, no irrelevant information), p = 0.005; in Condition B (2 speakers + irrelevant information), p = 0.035; and Condition C (4 speakers + irrelevant information), p = 0.008. Effect sizes (Cohen’s d) were large (A = 1.40, B = 0.96, C = 1.23). Significant group differences were strongest and most consistent for defining the problems and evaluating outcomes. The relation of task performance to cortical thickness of specific brain regions was also explored, with significant relations found with orbitofrontal regions, the frontal pole, the cuneus, and the temporal pole. Results are discussed in the context of specific cognitive and neural mechanisms underlying social problem solving deficits after childhood TBI. PMID:21147137
Coordinated interaction of two hydraulic cylinders when moving large-sized objects
NASA Astrophysics Data System (ADS)
Kreinin, G. V.; Misyurin, S. Yu; Lunev, A. V.
2017-12-01
The problem of the choice of parameters and the control scheme of the dynamics system for the coordinated displacement of a large mass object by two hydraulic piston type engines is considered. As a first stage, the problem is solved with respect to a system in which a heavy load of relatively large geometric dimensions is lifted or lowered in the progressive motion by two unidirectional hydraulic cylinders while maintaining the plane of the lifted object in a strictly horizontal position.
Loosli, Gaelle; Canu, Stephane; Ong, Cheng Soon
2016-06-01
This paper presents a theoretical foundation for an SVM solver in Kreĭn spaces. Up to now, all methods are based either on the matrix correction, or on non-convex minimization, or on feature-space embedding. Here we justify and evaluate a solution that uses the original (indefinite) similarity measure, in the original Kreĭn space. This solution is the result of a stabilization procedure. We establish the correspondence between the stabilization problem (which has to be solved) and a classical SVM based on minimization (which is easy to solve). We provide simple equations to go from one to the other (in both directions). This link between stabilization and minimization problems is the key to obtain a solution in the original Kreĭn space. Using KSVM, one can solve SVM with usually troublesome kernels (large negative eigenvalues or large numbers of negative eigenvalues). We show experiments showing that our algorithm KSVM outperforms all previously proposed approaches to deal with indefinite matrices in SVM-like kernel methods.
Multigrid method for stability problems
NASA Technical Reports Server (NTRS)
Ta'asan, Shlomo
1988-01-01
The problem of calculating the stability of steady state solutions of differential equations is addressed. Leading eigenvalues of large matrices that arise from discretization are calculated, and an efficient multigrid method for solving these problems is presented. The resulting grid functions are used as initial approximations for appropriate eigenvalue problems. The method employs local relaxation on all levels together with a global change on the coarsest level only, which is designed to separate the different eigenfunctions as well as to update their corresponding eigenvalues. Coarsening is done using the FAS formulation in a nonstandard way in which the right-hand side of the coarse grid equations involves unknown parameters to be solved on the coarse grid. This leads to a new multigrid method for calculating the eigenvalues of symmetric problems. Numerical experiments with a model problem are presented which demonstrate the effectiveness of the method.
Lattice gas methods for computational aeroacoustics
NASA Technical Reports Server (NTRS)
Sparrow, Victor W.
1995-01-01
This paper presents the lattice gas solution to the category 1 problems of the ICASE/LaRC Workshop on Benchmark Problems in Computational Aeroacoustics. The first and second problems were solved for Delta t = Delta x = 1, and additionally the second problem was solved for Delta t = 1/4 and Delta x = 1/2. The results are striking: even for these large time and space grids the lattice gas numerical solutions are almost indistinguishable from the analytical solutions. A simple bug in the Mathematica code was found in the solutions submitted for comparison, and the comparison plots shown at the end of this volume show the bug. An Appendix to the present paper shows an example lattice gas solution with and without the bug.
NASA Astrophysics Data System (ADS)
Shiangjen, Kanokwatt; Chaijaruwanich, Jeerayut; Srisujjalertwaja, Wijak; Unachak, Prakarn; Somhom, Samerkae
2018-02-01
This article presents an efficient heuristic placement algorithm, namely, a bidirectional heuristic placement, for solving the two-dimensional rectangular knapsack packing problem. The heuristic demonstrates ways to maximize space utilization by fitting the appropriate rectangle from both sides of the wall of the current residual space layer by layer. The iterative local search along with a shift strategy is developed and applied to the heuristic to balance the exploitation and exploration tasks in the solution space without the tuning of any parameters. The experimental results on many scales of packing problems show that this approach can produce high-quality solutions for most of the benchmark datasets, especially for large-scale problems, within a reasonable duration of computational time.
Estimation of the Thermal Process in the Honeycomb Panel by a Monte Carlo Method
NASA Astrophysics Data System (ADS)
Gusev, S. A.; Nikolaev, V. N.
2018-01-01
A new Monte Carlo method for estimating the thermal state of the heat insulation containing honeycomb panels is proposed in the paper. The heat transfer in the honeycomb panel is described by a boundary value problem for a parabolic equation with discontinuous diffusion coefficient and boundary conditions of the third kind. To obtain an approximate solution, it is proposed to use the smoothing of the diffusion coefficient. After that, the obtained problem is solved on the basis of the probability representation. The probability representation is the expectation of the functional of the diffusion process corresponding to the boundary value problem. The process of solving the problem is reduced to numerical statistical modelling of a large number of trajectories of the diffusion process corresponding to the parabolic problem. It was used earlier the Euler method for this object, but that requires a large computational effort. In this paper the method is modified by using combination of the Euler and the random walk on moving spheres methods. The new approach allows us to significantly reduce the computation costs.
Toward Solving the Problem of Problem Solving: An Analysis Framework
ERIC Educational Resources Information Center
Roesler, Rebecca A.
2016-01-01
Teaching is replete with problem solving. Problem solving as a skill, however, is seldom addressed directly within music teacher education curricula, and research in music education has not examined problem solving systematically. A framework detailing problem-solving component skills would provide a needed foundation. I observed problem solving…
Sweat, Noah W; Bates, Larry W; Hendricks, Peter S
2016-01-01
Developing methods for improving creativity is of broad interest. Classic psychedelics may enhance creativity; however, the underlying mechanisms of action are unknown. This study was designed to assess whether a relationship exists between naturalistic classic psychedelic use and heightened creative problem-solving ability and if so, whether this is mediated by lifetime mystical experience. Participants (N = 68) completed a survey battery assessing lifetime mystical experience and circumstances surrounding the most memorable experience. They were then administered a functional fixedness task in which faster completion times indicate greater creative problem-solving ability. Participants reporting classic psychedelic use concurrent with mystical experience (n = 11) exhibited significantly faster times on the functional fixedness task (Cohen's d = -.87; large effect) and significantly greater lifetime mystical experience (Cohen's d = .93; large effect) than participants not reporting classic psychedelic use concurrent with mystical experience. However, lifetime mystical experience was unrelated to completion times on the functional fixedness task (standardized β = -.06), and was therefore not a significant mediator. Classic psychedelic use may increase creativity independent of its effects on mystical experience. Maximizing the likelihood of mystical experience may need not be a goal of psychedelic interventions designed to boost creativity.
Scalable software architectures for decision support.
Musen, M A
1999-12-01
Interest in decision-support programs for clinical medicine soared in the 1970s. Since that time, workers in medical informatics have been particularly attracted to rule-based systems as a means of providing clinical decision support. Although developers have built many successful applications using production rules, they also have discovered that creation and maintenance of large rule bases is quite problematic. In the 1980s, several groups of investigators began to explore alternative programming abstractions that can be used to build decision-support systems. As a result, the notions of "generic tasks" and of reusable problem-solving methods became extremely influential. By the 1990s, academic centers were experimenting with architectures for intelligent systems based on two classes of reusable components: (1) problem-solving methods--domain-independent algorithms for automating stereotypical tasks--and (2) domain ontologies that captured the essential concepts (and relationships among those concepts) in particular application areas. This paper highlights how developers can construct large, maintainable decision-support systems using these kinds of building blocks. The creation of domain ontologies and problem-solving methods is the fundamental end product of basic research in medical informatics. Consequently, these concepts need more attention by our scientific community.
Dawson, Katie; Joscelyne, Amy; Meijer, Catherine; Steel, Zachary; Silove, Derrick; Bryant, Richard A
2018-03-01
To evaluate the relative efficacies of trauma-focused cognitive behavior therapy and problem-solving therapy in treating post-traumatic stress disorder in children affected by civil conflict in Aceh, Indonesia. A controlled trial of children with post-traumatic stress disorder ( N = 64) randomized children to either five individual weekly sessions of trauma-focused cognitive behavior therapy or problem-solving therapy provided by lay-counselors who were provided with brief training. Children were assessed by blind independent assessors at pretreatment, posttreatment and 3-month follow-up on post-traumatic stress disorder, depression and anger, as well as caregiver ratings of the child's post-traumatic stress disorder levels. Intent-to-treat analyses indicated no significant linear time × treatment condition interaction effects for post-traumatic stress disorder at follow-up ( t(129.05) = -0.55, p = 0.58), indicating the two conditions did not differ. Across both conditions, there were significant reductions in post-traumatic stress disorder on self-reported ( t(131.26) = -9.26, p < 0.001) and caregiver-reported ( t(170.65) = 3.53, p = 0.001) measures and anger ( t(127.66) = -7.14, p < 0.001). Across both conditions, there was a large effect size for self-reported post-traumatic stress disorder (cognitive behavior therapy: 3.73, 95% confidence interval = [2.75, 3.97]; problem-solving: 2.68, 95% confidence interval = [2.07, 3.29]). These findings suggest that trauma-focused cognitive behavior therapy and problem-solving approaches are comparably successful in reducing post-traumatic stress disorder and anger in treating mental health in children in a post-conflict setting. This pattern may reflect the benefits of non-specific therapy effects or gains associated with trauma-focused or problem-solving approaches.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Heydari, M.H., E-mail: heydari@stu.yazd.ac.ir; The Laboratory of Quantum Information Processing, Yazd University, Yazd; Hooshmandasl, M.R., E-mail: hooshmandasl@yazd.ac.ir
Because of the nonlinearity, closed-form solutions of many important stochastic functional equations are virtually impossible to obtain. Thus, numerical solutions are a viable alternative. In this paper, a new computational method based on the generalized hat basis functions together with their stochastic operational matrix of Itô-integration is proposed for solving nonlinear stochastic Itô integral equations in large intervals. In the proposed method, a new technique for computing nonlinear terms in such problems is presented. The main advantage of the proposed method is that it transforms problems under consideration into nonlinear systems of algebraic equations which can be simply solved. Errormore » analysis of the proposed method is investigated and also the efficiency of this method is shown on some concrete examples. The obtained results reveal that the proposed method is very accurate and efficient. As two useful applications, the proposed method is applied to obtain approximate solutions of the stochastic population growth models and stochastic pendulum problem.« less
Serang, Oliver
2012-01-01
Linear programming (LP) problems are commonly used in analysis and resource allocation, frequently surfacing as approximations to more difficult problems. Existing approaches to LP have been dominated by a small group of methods, and randomized algorithms have not enjoyed popularity in practice. This paper introduces a novel randomized method of solving LP problems by moving along the facets and within the interior of the polytope along rays randomly sampled from the polyhedral cones defined by the bounding constraints. This conic sampling method is then applied to randomly sampled LPs, and its runtime performance is shown to compare favorably to the simplex and primal affine-scaling algorithms, especially on polytopes with certain characteristics. The conic sampling method is then adapted and applied to solve a certain quadratic program, which compute a projection onto a polytope; the proposed method is shown to outperform the proprietary software Mathematica on large, sparse QP problems constructed from mass spectometry-based proteomics. PMID:22952741
NASA Technical Reports Server (NTRS)
Korivi, V. M.; Taylor, A. C., III; Newman, P. A.; Hou, G. J.-W.; Jones, H. E.
1992-01-01
An incremental strategy is presented for iteratively solving very large systems of linear equations, which are associated with aerodynamic sensitivity derivatives for advanced CFD codes. It is shown that the left-hand side matrix operator and the well-known factorization algorithm used to solve the nonlinear flow equations can also be used to efficiently solve the linear sensitivity equations. Two airfoil problems are considered as an example: subsonic low Reynolds number laminar flow and transonic high Reynolds number turbulent flow.
Hoppmann, Christiane A; Coats, Abby Heckman; Blanchard-Fields, Fredda
2008-07-01
Qualitative interviews on family and financial problems from 332 adolescents, young, middle-aged, and older adults, demonstrated that developmentally relevant goals predicted problem-solving strategy use over and above problem domain. Four focal goals concerned autonomy, generativity, maintaining good relationships with others, and changing another person. We examined both self- and other-focused problem-solving strategies. Autonomy goals were associated with self-focused instrumental problem solving and generative goals were related to other-focused instrumental problem solving in family and financial problems. Goals of changing another person were related to other-focused instrumental problem solving in the family domain only. The match between goals and strategies, an indicator of problem-solving adaptiveness, showed that young individuals displayed the greatest match between autonomy goals and self-focused problem solving, whereas older adults showed a greater match between generative goals and other-focused problem solving. Findings speak to the importance of considering goals in investigations of age-related differences in everyday problem solving.
Engineering Antifragile Systems: A Change In Design Philosophy
NASA Technical Reports Server (NTRS)
Jones, Kennie H.
2014-01-01
While technology has made astounding advances in the last century, problems are confronting the engineering community that must be solved. Cost and schedule of producing large systems are increasing at an unsustainable rate and these systems often do not perform as intended. New systems are required that may not be achieved by current methods. To solve these problems, NASA is working to infuse concepts from Complexity Science into the engineering process. Some of these problems may be solved by a change in design philosophy. Instead of designing systems to meet known requirements that will always lead to fragile systems at some degree, systems should be designed wherever possible to be antifragile: designing cognitive cyberphysical systems that can learn from their experience, adapt to unforeseen events they face in their environment, and grow stronger in the face of adversity. Several examples are presented of on ongoing research efforts to employ this philosophy.
Solving a real-world problem using an evolving heuristically driven schedule builder.
Hart, E; Ross, P; Nelson, J
1998-01-01
This work addresses the real-life scheduling problem of a Scottish company that must produce daily schedules for the catching and transportation of large numbers of live chickens. The problem is complex and highly constrained. We show that it can be successfully solved by division into two subproblems and solving each using a separate genetic algorithm (GA). We address the problem of whether this produces locally optimal solutions and how to overcome this. We extend the traditional approach of evolving a "permutation + schedule builder" by concentrating on evolving the schedule builder itself. This results in a unique schedule builder being built for each daily scheduling problem, each individually tailored to deal with the particular features of that problem. This results in a robust, fast, and flexible system that can cope with most of the circumstances imaginable at the factory. We also compare the performance of a GA approach to several other evolutionary methods and show that population-based methods are superior to both hill-climbing and simulated annealing in the quality of solutions produced. Population-based methods also have the distinct advantage of producing multiple, equally fit solutions, which is of particular importance when considering the practical aspects of the problem.
Parallel computation with molecular-motor-propelled agents in nanofabricated networks.
Nicolau, Dan V; Lard, Mercy; Korten, Till; van Delft, Falco C M J M; Persson, Malin; Bengtsson, Elina; Månsson, Alf; Diez, Stefan; Linke, Heiner; Nicolau, Dan V
2016-03-08
The combinatorial nature of many important mathematical problems, including nondeterministic-polynomial-time (NP)-complete problems, places a severe limitation on the problem size that can be solved with conventional, sequentially operating electronic computers. There have been significant efforts in conceiving parallel-computation approaches in the past, for example: DNA computation, quantum computation, and microfluidics-based computation. However, these approaches have not proven, so far, to be scalable and practical from a fabrication and operational perspective. Here, we report the foundations of an alternative parallel-computation system in which a given combinatorial problem is encoded into a graphical, modular network that is embedded in a nanofabricated planar device. Exploring the network in a parallel fashion using a large number of independent, molecular-motor-propelled agents then solves the mathematical problem. This approach uses orders of magnitude less energy than conventional computers, thus addressing issues related to power consumption and heat dissipation. We provide a proof-of-concept demonstration of such a device by solving, in a parallel fashion, the small instance {2, 5, 9} of the subset sum problem, which is a benchmark NP-complete problem. Finally, we discuss the technical advances necessary to make our system scalable with presently available technology.
Grand challenges for biological engineering
Yoon, Jeong-Yeol; Riley, Mark R
2009-01-01
Biological engineering will play a significant role in solving many of the world's problems in medicine, agriculture, and the environment. Recently the U.S. National Academy of Engineering (NAE) released a document "Grand Challenges in Engineering," covering broad realms of human concern from sustainability, health, vulnerability and the joy of living. Biological engineers, having tools and techniques at the interface between living and non-living entities, will play a prominent role in forging a better future. The 2010 Institute of Biological Engineering (IBE) conference in Cambridge, MA, USA will address, in part, the roles of biological engineering in solving the challenges presented by the NAE. This letter presents a brief outline of how biological engineers are working to solve these large scale and integrated problems of our society. PMID:19772647
Automation of multi-agent control for complex dynamic systems in heterogeneous computational network
NASA Astrophysics Data System (ADS)
Oparin, Gennady; Feoktistov, Alexander; Bogdanova, Vera; Sidorov, Ivan
2017-01-01
The rapid progress of high-performance computing entails new challenges related to solving large scientific problems for various subject domains in a heterogeneous distributed computing environment (e.g., a network, Grid system, or Cloud infrastructure). The specialists in the field of parallel and distributed computing give the special attention to a scalability of applications for problem solving. An effective management of the scalable application in the heterogeneous distributed computing environment is still a non-trivial issue. Control systems that operate in networks, especially relate to this issue. We propose a new approach to the multi-agent management for the scalable applications in the heterogeneous computational network. The fundamentals of our approach are the integrated use of conceptual programming, simulation modeling, network monitoring, multi-agent management, and service-oriented programming. We developed a special framework for an automation of the problem solving. Advantages of the proposed approach are demonstrated on the parametric synthesis example of the static linear regulator for complex dynamic systems. Benefits of the scalable application for solving this problem include automation of the multi-agent control for the systems in a parallel mode with various degrees of its detailed elaboration.
Insight and search in Katona's five-square problem.
Ollinger, Michael; Jones, Gary; Knoblich, Günther
2014-01-01
Insights are often productive outcomes of human thinking. We provide a cognitive model that explains insight problem solving by the interplay of problem space search and representational change, whereby the problem space is constrained or relaxed based on the problem representation. By introducing different experimental conditions that either constrained the initial search space or helped solvers to initiate a representational change, we investigated the interplay of problem space search and representational change in Katona's five-square problem. Testing 168 participants, we demonstrated that independent hints relating to the initial search space and to representational change had little effect on solution rates. However, providing both hints caused a significant increase in solution rates. Our results show the interplay between problem space search and representational change in insight problem solving: The initial problem space can be so large that people fail to encounter impasse, but even when representational change is achieved the resulting problem space can still provide a major obstacle to finding the solution.
Modern architectures for intelligent systems: reusable ontologies and problem-solving methods.
Musen, M. A.
1998-01-01
When interest in intelligent systems for clinical medicine soared in the 1970s, workers in medical informatics became particularly attracted to rule-based systems. Although many successful rule-based applications were constructed, development and maintenance of large rule bases remained quite problematic. In the 1980s, an entire industry dedicated to the marketing of tools for creating rule-based systems rose and fell, as workers in medical informatics began to appreciate deeply why knowledge acquisition and maintenance for such systems are difficult problems. During this time period, investigators began to explore alternative programming abstractions that could be used to develop intelligent systems. The notions of "generic tasks" and of reusable problem-solving methods became extremely influential. By the 1990s, academic centers were experimenting with architectures for intelligent systems based on two classes of reusable components: (1) domain-independent problem-solving methods-standard algorithms for automating stereotypical tasks--and (2) domain ontologies that captured the essential concepts (and relationships among those concepts) in particular application areas. This paper will highlight how intelligent systems for diverse tasks can be efficiently automated using these kinds of building blocks. The creation of domain ontologies and problem-solving methods is the fundamental end product of basic research in medical informatics. Consequently, these concepts need more attention by our scientific community. PMID:9929181
Modern architectures for intelligent systems: reusable ontologies and problem-solving methods.
Musen, M A
1998-01-01
When interest in intelligent systems for clinical medicine soared in the 1970s, workers in medical informatics became particularly attracted to rule-based systems. Although many successful rule-based applications were constructed, development and maintenance of large rule bases remained quite problematic. In the 1980s, an entire industry dedicated to the marketing of tools for creating rule-based systems rose and fell, as workers in medical informatics began to appreciate deeply why knowledge acquisition and maintenance for such systems are difficult problems. During this time period, investigators began to explore alternative programming abstractions that could be used to develop intelligent systems. The notions of "generic tasks" and of reusable problem-solving methods became extremely influential. By the 1990s, academic centers were experimenting with architectures for intelligent systems based on two classes of reusable components: (1) domain-independent problem-solving methods-standard algorithms for automating stereotypical tasks--and (2) domain ontologies that captured the essential concepts (and relationships among those concepts) in particular application areas. This paper will highlight how intelligent systems for diverse tasks can be efficiently automated using these kinds of building blocks. The creation of domain ontologies and problem-solving methods is the fundamental end product of basic research in medical informatics. Consequently, these concepts need more attention by our scientific community.
Optimum oil production planning using infeasibility driven evolutionary algorithm.
Singh, Hemant Kumar; Ray, Tapabrata; Sarker, Ruhul
2013-01-01
In this paper, we discuss a practical oil production planning optimization problem. For oil wells with insufficient reservoir pressure, gas is usually injected to artificially lift oil, a practice commonly referred to as enhanced oil recovery (EOR). The total gas that can be used for oil extraction is constrained by daily availability limits. The oil extracted from each well is known to be a nonlinear function of the gas injected into the well and varies between wells. The problem is to identify the optimal amount of gas that needs to be injected into each well to maximize the amount of oil extracted subject to the constraint on the total daily gas availability. The problem has long been of practical interest to all major oil exploration companies as it has the potential to derive large financial benefit. In this paper, an infeasibility driven evolutionary algorithm is used to solve a 56 well reservoir problem which demonstrates its efficiency in solving constrained optimization problems. Furthermore, a multi-objective formulation of the problem is posed and solved using a number of algorithms, which eliminates the need for solving the (single objective) problem on a regular basis. Lastly, a modified single objective formulation of the problem is also proposed, which aims to maximize the profit instead of the quantity of oil. It is shown that even with a lesser amount of oil extracted, more economic benefits can be achieved through the modified formulation.
Teaching basic science to optimize transfer.
Norman, Geoff
2009-09-01
Basic science teachers share the concern that much of what they teach is soon forgotten. Although some evidence suggests that relatively little basic science is forgotten, it may not appear so, as students commonly have difficulty using these concepts to solve or explain clinical problems: This phenomenon, using a concept learned in one context to solve a problem in a different context, is known to cognitive psychologists as transfer. The psychology literature shows that transfer is difficult; typically, even though students may know a concept, fewer than 30% will be able to use it to solve new problems. However a number of strategies to improve transfer can be adopted at the time of initial teaching of the concept, in the use of exemplars to illustrate the concept, and in practice with additional problems. In this article, we review the literature in psychology to identify practical strategies to improve transfer. Critical review of psychology literature to identify factors that enhance or impede transfer. There are a number of strategies available to teachers to facilitate transfer. These include active problem-solving at the time of initial learning, imbedding the concept in a problem context, using everyday analogies, and critically, practice with multiple dissimilar problems. Further, mixed practice, where problems illustrating different concepts are mixed together, and distributed practice, spread out over time, can result in significant and large gains. Transfer is difficult, but specific teaching strategies can enhance this skill by factors of two or three.
Resources in Technology: Problem-Solving.
ERIC Educational Resources Information Center
Technology Teacher, 1986
1986-01-01
This instructional module examines a key function of science and technology: problem solving. It studies the meaning of problem solving, looks at techniques for problem solving, examines case studies that exemplify the problem-solving approach, presents problems for the reader to solve, and provides a student self-quiz. (Author/CT)
More than Just "Plug-and-Chug": Exploring How Physics Students Make Sense with Equations
ERIC Educational Resources Information Center
Kuo, Eric
2013-01-01
Although a large part the Physics Education Research (PER) literature investigates students' conceptual understanding in physics, these investigations focus on qualitative, conceptual reasoning. Even in modeling expert problem solving, attention to conceptual understanding means a focus on initial qualitative analysis of the problem; the equations…
Teaching Database Design with Constraint-Based Tutors
ERIC Educational Resources Information Center
Mitrovic, Antonija; Suraweera, Pramuditha
2016-01-01
Design tasks are difficult to teach, due to large, unstructured solution spaces, underspecified problems, non-existent problem solving algorithms and stopping criteria. In this paper, we comment on our approach to develop KERMIT, a constraint-based tutor that taught database design. In later work, we re-implemented KERMIT as EER-Tutor, and…
Algorithms for Mathematical Programming with Emphasis on Bi-level Models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goldfarb, Donald; Iyengar, Garud
2014-05-22
The research supported by this grant was focused primarily on first-order methods for solving large scale and structured convex optimization problems and convex relaxations of nonconvex problems. These include optimal gradient methods, operator and variable splitting methods, alternating direction augmented Lagrangian methods, and block coordinate descent methods.
Private Education as a Policy Tool in Turkey
ERIC Educational Resources Information Center
Cinoglu, Mustafa
2006-01-01
This paper discusses privatization as policy tool to solve educational problems in Turkey. Turkey, as a developing country, is faced with many problems in education. Large class size, low enrollment rate, girl's education, high illiteracy rate, religious education, textbooks, curriculum and multicultural education are some of the important…
Two New PRP Conjugate Gradient Algorithms for Minimization Optimization Models.
Yuan, Gonglin; Duan, Xiabin; Liu, Wenjie; Wang, Xiaoliang; Cui, Zengru; Sheng, Zhou
2015-01-01
Two new PRP conjugate Algorithms are proposed in this paper based on two modified PRP conjugate gradient methods: the first algorithm is proposed for solving unconstrained optimization problems, and the second algorithm is proposed for solving nonlinear equations. The first method contains two aspects of information: function value and gradient value. The two methods both possess some good properties, as follows: 1) βk ≥ 0 2) the search direction has the trust region property without the use of any line search method 3) the search direction has sufficient descent property without the use of any line search method. Under some suitable conditions, we establish the global convergence of the two algorithms. We conduct numerical experiments to evaluate our algorithms. The numerical results indicate that the first algorithm is effective and competitive for solving unconstrained optimization problems and that the second algorithm is effective for solving large-scale nonlinear equations.
Two New PRP Conjugate Gradient Algorithms for Minimization Optimization Models
Yuan, Gonglin; Duan, Xiabin; Liu, Wenjie; Wang, Xiaoliang; Cui, Zengru; Sheng, Zhou
2015-01-01
Two new PRP conjugate Algorithms are proposed in this paper based on two modified PRP conjugate gradient methods: the first algorithm is proposed for solving unconstrained optimization problems, and the second algorithm is proposed for solving nonlinear equations. The first method contains two aspects of information: function value and gradient value. The two methods both possess some good properties, as follows: 1)β k ≥ 0 2) the search direction has the trust region property without the use of any line search method 3) the search direction has sufficient descent property without the use of any line search method. Under some suitable conditions, we establish the global convergence of the two algorithms. We conduct numerical experiments to evaluate our algorithms. The numerical results indicate that the first algorithm is effective and competitive for solving unconstrained optimization problems and that the second algorithm is effective for solving large-scale nonlinear equations. PMID:26502409
Multi-period natural gas market modeling Applications, stochastic extensions and solution approaches
NASA Astrophysics Data System (ADS)
Egging, Rudolf Gerardus
This dissertation develops deterministic and stochastic multi-period mixed complementarity problems (MCP) for the global natural gas market, as well as solution approaches for large-scale stochastic MCP. The deterministic model is unique in the combination of the level of detail of the actors in the natural gas markets and the transport options, the detailed regional and global coverage, the multi-period approach with endogenous capacity expansions for transportation and storage infrastructure, the seasonal variation in demand and the representation of market power according to Nash-Cournot theory. The model is applied to several scenarios for the natural gas market that cover the formation of a cartel by the members of the Gas Exporting Countries Forum, a low availability of unconventional gas in the United States, and cost reductions in long-distance gas transportation. 1 The results provide insights in how different regions are affected by various developments, in terms of production, consumption, traded volumes, prices and profits of market participants. The stochastic MCP is developed and applied to a global natural gas market problem with four scenarios for a time horizon until 2050 with nineteen regions and containing 78,768 variables. The scenarios vary in the possibility of a gas market cartel formation and varying depletion rates of gas reserves in the major gas importing regions. Outcomes for hedging decisions of market participants show some significant shifts in the timing and location of infrastructure investments, thereby affecting local market situations. A first application of Benders decomposition (BD) is presented to solve a large-scale stochastic MCP for the global gas market with many hundreds of first-stage capacity expansion variables and market players exerting various levels of market power. The largest problem solved successfully using BD contained 47,373 variables of which 763 first-stage variables, however using BD did not result in shorter solution times relative to solving the extensive-forms. Larger problems, up to 117,481 variables, were solved in extensive-form, but not when applying BD due to numerical issues. It is discussed how BD could significantly reduce the solution time of large-scale stochastic models, but various challenges remain and more research is needed to assess the potential of Benders decomposition for solving large-scale stochastic MCP. 1 www.gecforum.org
NASA Astrophysics Data System (ADS)
Kobak, B. V.; Zhukovskiy, A. G.; Kuzin, A. P.
2018-05-01
This paper considers one of the classical NP complete problems - an inhomogeneous minimax problem. When solving such large-scale problem, there appear difficulties in obtaining an exact solution. Therefore, let us propose getting an optimum solution in an acceptable time. Among a wide range of genetic algorithm models, let us choose the modified Goldberg model, which earlier was successfully used by authors in solving NP complete problems. The classical Goldberg model uses a single-point crossover and a singlepoint mutation, which somewhat decreases the accuracy of the obtained results. In the article, let us propose using a full two-point crossover with various mutations previously researched. In addition, the work studied the necessary probability to apply it to the crossover in order to obtain results that are more accurate. Results of the computation experiment showed that the higher the probability of a crossover, the higher the quality of both the average results and the best solutions. In addition, it was found out that the higher the values of the number of individuals and the number of repetitions, the closer both the average results and the best solutions to the optimum. The paper shows how the use of a full two-point crossover increases the accuracy of solving an inhomogeneous minimax problem, while the time for getting the solution increases, but remains polynomial.
A Cognitive Analysis of Students’ Mathematical Problem Solving Ability on Geometry
NASA Astrophysics Data System (ADS)
Rusyda, N. A.; Kusnandi, K.; Suhendra, S.
2017-09-01
The purpose of this research is to analyze of mathematical problem solving ability of students in one of secondary school on geometry. This research was conducted by using quantitative approach with descriptive method. Population in this research was all students of that school and the sample was twenty five students that was chosen by purposive sampling technique. Data of mathematical problem solving were collected through essay test. The results showed the percentage of achievement of mathematical problem solving indicators of students were: 1) solve closed mathematical problems with context in math was 50%; 2) solve the closed mathematical problems with the context beyond mathematics was 24%; 3) solving open mathematical problems with contexts in mathematics was 35%; And 4) solving open mathematical problems with contexts outside mathematics was 44%. Based on the percentage, it can be concluded that the level of achievement of mathematical problem solving ability in geometry still low. This is because students are not used to solving problems that measure mathematical problem solving ability, weaknesses remember previous knowledge, and lack of problem solving framework. So the students’ ability of mathematical problems solving need to be improved with implement appropriate learning strategy.
Computation of Pressurized Gas Bearings Using CE/SE Method
NASA Technical Reports Server (NTRS)
Cioc, Sorin; Dimofte, Florin; Keith, Theo G., Jr.; Fleming, David P.
2003-01-01
The space-time conservation element and solution element (CE/SE) method is extended to compute compressible viscous flows in pressurized thin fluid films. This numerical scheme has previously been used successfully to solve a wide variety of compressible flow problems, including flows with large and small discontinuities. In this paper, the method is applied to calculate the pressure distribution in a hybrid gas journal bearing. The formulation of the problem is presented, including the modeling of the feeding system. the numerical results obtained are compared with experimental data. Good agreement between the computed results and the test data were obtained, and thus validate the CE/SE method to solve such problems.
A Roadmap of Innovative Nuclear Energy System
NASA Astrophysics Data System (ADS)
Sekimoto, Hiroshi
2017-01-01
Nuclear is a dense energy without CO2 emission. It can be used for more than 100,000 years using fast breeder reactors with uranium from the sea. However, it raises difficult problems associated with severe accidents, spent fuel waste and nuclear threats, which should be solved with acceptable costs. Some innovative reactors have attracted interest, and many designs have been proposed for small reactors. These reactors are considered much safer than conventional large reactors and have fewer technical obstructions. Breed-and-burn reactors have high potential to solve all inherent problems for peaceful use of nuclear energy. However, they have some technical problems with materials. A roadmap for innovative reactors is presented herein.
Combining constraint satisfaction and local improvement algorithms to construct anaesthetists' rotas
NASA Technical Reports Server (NTRS)
Smith, Barbara M.; Bennett, Sean
1992-01-01
A system is described which was built to compile weekly rotas for the anaesthetists in a large hospital. The rota compilation problem is an optimization problem (the number of tasks which cannot be assigned to an anaesthetist must be minimized) and was formulated as a constraint satisfaction problem (CSP). The forward checking algorithm is used to find a feasible rota, but because of the size of the problem, it cannot find an optimal (or even a good enough) solution in an acceptable time. Instead, an algorithm was devised which makes local improvements to a feasible solution. The algorithm makes use of the constraints as expressed in the CSP to ensure that feasibility is maintained, and produces very good rotas which are being used by the hospital involved in the project. It is argued that formulation as a constraint satisfaction problem may be a good approach to solving discrete optimization problems, even if the resulting CSP is too large to be solved exactly in an acceptable time. A CSP algorithm may be able to produce a feasible solution which can then be improved, giving a good, if not provably optimal, solution.
An unbalanced spectra classification method based on entropy
NASA Astrophysics Data System (ADS)
Liu, Zhong-bao; Zhao, Wen-juan
2017-05-01
How to solve the problem of distinguishing the minority spectra from the majority of the spectra is quite important in astronomy. In view of this, an unbalanced spectra classification method based on entropy (USCM) is proposed in this paper to deal with the unbalanced spectra classification problem. USCM greatly improves the performances of the traditional classifiers on distinguishing the minority spectra as it takes the data distribution into consideration in the process of classification. However, its time complexity is exponential with the training size, and therefore, it can only deal with the problem of small- and medium-scale classification. How to solve the large-scale classification problem is quite important to USCM. It can be easily obtained by mathematical computation that the dual form of USCM is equivalent to the minimum enclosing ball (MEB), and core vector machine (CVM) is introduced, USCM based on CVM is proposed to deal with the large-scale classification problem. Several comparative experiments on the 4 subclasses of K-type spectra, 3 subclasses of F-type spectra and 3 subclasses of G-type spectra from Sloan Digital Sky Survey (SDSS) verify USCM and USCM based on CVM perform better than kNN (k nearest neighbor) and SVM (support vector machine) in dealing with the problem of rare spectra mining respectively on the small- and medium-scale datasets and the large-scale datasets.
Review on solving the forward problem in EEG source analysis
Hallez, Hans; Vanrumste, Bart; Grech, Roberta; Muscat, Joseph; De Clercq, Wim; Vergult, Anneleen; D'Asseler, Yves; Camilleri, Kenneth P; Fabri, Simon G; Van Huffel, Sabine; Lemahieu, Ignace
2007-01-01
Background The aim of electroencephalogram (EEG) source localization is to find the brain areas responsible for EEG waves of interest. It consists of solving forward and inverse problems. The forward problem is solved by starting from a given electrical source and calculating the potentials at the electrodes. These evaluations are necessary to solve the inverse problem which is defined as finding brain sources which are responsible for the measured potentials at the EEG electrodes. Methods While other reviews give an extensive summary of the both forward and inverse problem, this review article focuses on different aspects of solving the forward problem and it is intended for newcomers in this research field. Results It starts with focusing on the generators of the EEG: the post-synaptic potentials in the apical dendrites of pyramidal neurons. These cells generate an extracellular current which can be modeled by Poisson's differential equation, and Neumann and Dirichlet boundary conditions. The compartments in which these currents flow can be anisotropic (e.g. skull and white matter). In a three-shell spherical head model an analytical expression exists to solve the forward problem. During the last two decades researchers have tried to solve Poisson's equation in a realistically shaped head model obtained from 3D medical images, which requires numerical methods. The following methods are compared with each other: the boundary element method (BEM), the finite element method (FEM) and the finite difference method (FDM). In the last two methods anisotropic conducting compartments can conveniently be introduced. Then the focus will be set on the use of reciprocity in EEG source localization. It is introduced to speed up the forward calculations which are here performed for each electrode position rather than for each dipole position. Solving Poisson's equation utilizing FEM and FDM corresponds to solving a large sparse linear system. Iterative methods are required to solve these sparse linear systems. The following iterative methods are discussed: successive over-relaxation, conjugate gradients method and algebraic multigrid method. Conclusion Solving the forward problem has been well documented in the past decades. In the past simplified spherical head models are used, whereas nowadays a combination of imaging modalities are used to accurately describe the geometry of the head model. Efforts have been done on realistically describing the shape of the head model, as well as the heterogenity of the tissue types and realistically determining the conductivity. However, the determination and validation of the in vivo conductivity values is still an important topic in this field. In addition, more studies have to be done on the influence of all the parameters of the head model and of the numerical techniques on the solution of the forward problem. PMID:18053144
An adaptive grid algorithm for one-dimensional nonlinear equations
NASA Technical Reports Server (NTRS)
Gutierrez, William E.; Hills, Richard G.
1990-01-01
Richards' equation, which models the flow of liquid through unsaturated porous media, is highly nonlinear and difficult to solve. Step gradients in the field variables require the use of fine grids and small time step sizes. The numerical instabilities caused by the nonlinearities often require the use of iterative methods such as Picard or Newton interation. These difficulties result in large CPU requirements in solving Richards equation. With this in mind, adaptive and multigrid methods are investigated for use with nonlinear equations such as Richards' equation. Attention is focused on one-dimensional transient problems. To investigate the use of multigrid and adaptive grid methods, a series of problems are studied. First, a multigrid program is developed and used to solve an ordinary differential equation, demonstrating the efficiency with which low and high frequency errors are smoothed out. The multigrid algorithm and an adaptive grid algorithm is used to solve one-dimensional transient partial differential equations, such as the diffusive and convective-diffusion equations. The performance of these programs are compared to that of the Gauss-Seidel and tridiagonal methods. The adaptive and multigrid schemes outperformed the Gauss-Seidel algorithm, but were not as fast as the tridiagonal method. The adaptive grid scheme solved the problems slightly faster than the multigrid method. To solve nonlinear problems, Picard iterations are introduced into the adaptive grid and tridiagonal methods. Burgers' equation is used as a test problem for the two algorithms. Both methods obtain solutions of comparable accuracy for similar time increments. For the Burgers' equation, the adaptive grid method finds the solution approximately three times faster than the tridiagonal method. Finally, both schemes are used to solve the water content formulation of the Richards' equation. For this problem, the adaptive grid method obtains a more accurate solution in fewer work units and less computation time than required by the tridiagonal method. The performance of the adaptive grid method tends to degrade as the solution process proceeds in time, but still remains faster than the tridiagonal scheme.
Optimal placement of tuning masses on truss structures by genetic algorithms
NASA Technical Reports Server (NTRS)
Ponslet, Eric; Haftka, Raphael T.; Cudney, Harley H.
1993-01-01
Optimal placement of tuning masses, actuators and other peripherals on large space structures is a combinatorial optimization problem. This paper surveys several techniques for solving this problem. The genetic algorithm approach to the solution of the placement problem is described in detail. An example of minimizing the difference between the two lowest frequencies of a laboratory truss by adding tuning masses is used for demonstrating some of the advantages of genetic algorithms. The relative efficiencies of different codings are compared using the results of a large number of optimization runs.
NASA Astrophysics Data System (ADS)
Mukherjee, Anamitra; Patel, Niravkumar D.; Bishop, Chris; Dagotto, Elbio
2015-06-01
Lattice spin-fermion models are important to study correlated systems where quantum dynamics allows for a separation between slow and fast degrees of freedom. The fast degrees of freedom are treated quantum mechanically while the slow variables, generically referred to as the "spins," are treated classically. At present, exact diagonalization coupled with classical Monte Carlo (ED + MC) is extensively used to solve numerically a general class of lattice spin-fermion problems. In this common setup, the classical variables (spins) are treated via the standard MC method while the fermion problem is solved by exact diagonalization. The "traveling cluster approximation" (TCA) is a real space variant of the ED + MC method that allows to solve spin-fermion problems on lattice sizes with up to 103 sites. In this publication, we present a novel reorganization of the TCA algorithm in a manner that can be efficiently parallelized. This allows us to solve generic spin-fermion models easily on 104 lattice sites and with some effort on 105 lattice sites, representing the record lattice sizes studied for this family of models.
The min-conflicts heuristic: Experimental and theoretical results
NASA Technical Reports Server (NTRS)
Minton, Steven; Philips, Andrew B.; Johnston, Mark D.; Laird, Philip
1991-01-01
This paper describes a simple heuristic method for solving large-scale constraint satisfaction and scheduling problems. Given an initial assignment for the variables in a problem, the method operates by searching through the space of possible repairs. The search is guided by an ordering heuristic, the min-conflicts heuristic, that attempts to minimize the number of constraint violations after each step. We demonstrate empirically that the method performs orders of magnitude better than traditional backtracking techniques on certain standard problems. For example, the one million queens problem can be solved rapidly using our approach. We also describe practical scheduling applications where the method has been successfully applied. A theoretical analysis is presented to explain why the method works so well on certain types of problems and to predict when it is likely to be most effective.
Cloud-based large-scale air traffic flow optimization
NASA Astrophysics Data System (ADS)
Cao, Yi
The ever-increasing traffic demand makes the efficient use of airspace an imperative mission, and this paper presents an effort in response to this call. Firstly, a new aggregate model, called Link Transmission Model (LTM), is proposed, which models the nationwide traffic as a network of flight routes identified by origin-destination pairs. The traversal time of a flight route is assumed to be the mode of distribution of historical flight records, and the mode is estimated by using Kernel Density Estimation. As this simplification abstracts away physical trajectory details, the complexity of modeling is drastically decreased, resulting in efficient traffic forecasting. The predicative capability of LTM is validated against recorded traffic data. Secondly, a nationwide traffic flow optimization problem with airport and en route capacity constraints is formulated based on LTM. The optimization problem aims at alleviating traffic congestions with minimal global delays. This problem is intractable due to millions of variables. A dual decomposition method is applied to decompose the large-scale problem such that the subproblems are solvable. However, the whole problem is still computational expensive to solve since each subproblem is an smaller integer programming problem that pursues integer solutions. Solving an integer programing problem is known to be far more time-consuming than solving its linear relaxation. In addition, sequential execution on a standalone computer leads to linear runtime increase when the problem size increases. To address the computational efficiency problem, a parallel computing framework is designed which accommodates concurrent executions via multithreading programming. The multithreaded version is compared with its monolithic version to show decreased runtime. Finally, an open-source cloud computing framework, Hadoop MapReduce, is employed for better scalability and reliability. This framework is an "off-the-shelf" parallel computing model that can be used for both offline historical traffic data analysis and online traffic flow optimization. It provides an efficient and robust platform for easy deployment and implementation. A small cloud consisting of five workstations was configured and used to demonstrate the advantages of cloud computing in dealing with large-scale parallelizable traffic problems.
Acceleration of aircraft-level Traffic Flow Management
NASA Astrophysics Data System (ADS)
Rios, Joseph Lucio
This dissertation describes novel approaches to solving large-scale, high fidelity, aircraft-level Traffic Flow Management scheduling problems. Depending on the methods employed, solving these problems to optimality can take longer than the length of the planning horizon in question. Research in this domain typically focuses on the quality of the modeling used to describe the problem and the benefits achieved from the optimized solution, often treating computational aspects as secondary or tertiary. The work presented here takes the complementary view and considers the computational aspect as the primary concern. To this end, a previously published model for solving this Traffic Flow Management scheduling problem is used as starting point for this study. The model proposed by Bertsimas and Stock-Patterson is a binary integer program taking into account all major resource capacities and the trajectories of each flight to decide which flights should be held in which resource for what amount of time in order to satisfy all capacity requirements. For large instances, the solve time using state-of-the-art solvers is prohibitive for use within a potential decision support tool. With this dissertation, however, it will be shown that solving can be achieved in reasonable time for instances of real-world size. Five other techniques developed and tested for this dissertation will be described in detail. These are heuristic methods that provide good results. Performance is measured in terms of runtime and "optimality gap." We then describe the most successful method presented in this dissertation: Dantzig-Wolfe Decomposition. Results indicate that a parallel implementation of Dantzig-Wolfe Decomposition optimally solves the original problem in much reduced time and with better integrality and smaller optimality gap than any of the heuristic methods or state-of-the-art, commercial solvers. The solution quality improves in every measureable way as the number of subproblems solved in parallel increases. A maximal decomposition provides the best results of any method tested. The convergence qualities of Dantzig-Wolfe Decomposition have been criticized in the past, so we examine what makes the Bertsimas-Stock Patterson model so amenable to use of this method. These mathematical qualities of the model are generalized to provide guidance on other problems that may benefit from massively parallel Dantzig-Wolfe Decomposition. This result, together with the development of the software, and the experimental results indicating the feasibility of real-time, nationwide Traffic Flow Management scheduling represent the major contributions of this dissertation.
Cross-borehole flowmeter tests for transient heads in heterogeneous aquifers.
Le Borgne, Tanguy; Paillet, Frederick; Bour, Olivier; Caudal, Jean-Pierre
2006-01-01
Cross-borehole flowmeter tests have been proposed as an efficient method to investigate preferential flowpaths in heterogeneous aquifers, which is a major task in the characterization of fractured aquifers. Cross-borehole flowmeter tests are based on the idea that changing the pumping conditions in a given aquifer will modify the hydraulic head distribution in large-scale flowpaths, producing measurable changes in the vertical flow profiles in observation boreholes. However, inversion of flow measurements to derive flowpath geometry and connectivity and to characterize their hydraulic properties is still a subject of research. In this study, we propose a framework for cross-borehole flowmeter test interpretation that is based on a two-scale conceptual model: discrete fractures at the borehole scale and zones of interconnected fractures at the aquifer scale. We propose that the two problems may be solved independently. The first inverse problem consists of estimating the hydraulic head variations that drive the transient borehole flow observed in the cross-borehole flowmeter experiments. The second inverse problem is related to estimating the geometry and hydraulic properties of large-scale flowpaths in the region between pumping and observation wells that are compatible with the head variations deduced from the first problem. To solve the borehole-scale problem, we treat the transient flow data as a series of quasi-steady flow conditions and solve for the hydraulic head changes in individual fractures required to produce these data. The consistency of the method is verified using field experiments performed in a fractured-rock aquifer.
Multi-GPU implementation of a VMAT treatment plan optimization algorithm.
Tian, Zhen; Peng, Fei; Folkerts, Michael; Tan, Jun; Jia, Xun; Jiang, Steve B
2015-06-01
Volumetric modulated arc therapy (VMAT) optimization is a computationally challenging problem due to its large data size, high degrees of freedom, and many hardware constraints. High-performance graphics processing units (GPUs) have been used to speed up the computations. However, GPU's relatively small memory size cannot handle cases with a large dose-deposition coefficient (DDC) matrix in cases of, e.g., those with a large target size, multiple targets, multiple arcs, and/or small beamlet size. The main purpose of this paper is to report an implementation of a column-generation-based VMAT algorithm, previously developed in the authors' group, on a multi-GPU platform to solve the memory limitation problem. While the column-generation-based VMAT algorithm has been previously developed, the GPU implementation details have not been reported. Hence, another purpose is to present detailed techniques employed for GPU implementation. The authors also would like to utilize this particular problem as an example problem to study the feasibility of using a multi-GPU platform to solve large-scale problems in medical physics. The column-generation approach generates VMAT apertures sequentially by solving a pricing problem (PP) and a master problem (MP) iteratively. In the authors' method, the sparse DDC matrix is first stored on a CPU in coordinate list format (COO). On the GPU side, this matrix is split into four submatrices according to beam angles, which are stored on four GPUs in compressed sparse row format. Computation of beamlet price, the first step in PP, is accomplished using multi-GPUs. A fast inter-GPU data transfer scheme is accomplished using peer-to-peer access. The remaining steps of PP and MP problems are implemented on CPU or a single GPU due to their modest problem scale and computational loads. Barzilai and Borwein algorithm with a subspace step scheme is adopted here to solve the MP problem. A head and neck (H&N) cancer case is then used to validate the authors' method. The authors also compare their multi-GPU implementation with three different single GPU implementation strategies, i.e., truncating DDC matrix (S1), repeatedly transferring DDC matrix between CPU and GPU (S2), and porting computations involving DDC matrix to CPU (S3), in terms of both plan quality and computational efficiency. Two more H&N patient cases and three prostate cases are used to demonstrate the advantages of the authors' method. The authors' multi-GPU implementation can finish the optimization process within ∼ 1 min for the H&N patient case. S1 leads to an inferior plan quality although its total time was 10 s shorter than the multi-GPU implementation due to the reduced matrix size. S2 and S3 yield the same plan quality as the multi-GPU implementation but take ∼4 and ∼6 min, respectively. High computational efficiency was consistently achieved for the other five patient cases tested, with VMAT plans of clinically acceptable quality obtained within 23-46 s. Conversely, to obtain clinically comparable or acceptable plans for all six of these VMAT cases that the authors have tested in this paper, the optimization time needed in a commercial TPS system on CPU was found to be in an order of several minutes. The results demonstrate that the multi-GPU implementation of the authors' column-generation-based VMAT optimization can handle the large-scale VMAT optimization problem efficiently without sacrificing plan quality. The authors' study may serve as an example to shed some light on other large-scale medical physics problems that require multi-GPU techniques.
Solving lot-sizing problem with quantity discount and transportation cost
NASA Astrophysics Data System (ADS)
Lee, Amy H. I.; Kang, He-Yau; Lai, Chun-Mei
2013-04-01
Owing to today's increasingly competitive market and ever-changing manufacturing environment, the inventory problem is becoming more complicated to solve. The incorporation of heuristics methods has become a new trend to tackle the complex problem in the past decade. This article considers a lot-sizing problem, and the objective is to minimise total costs, where the costs include ordering, holding, purchase and transportation costs, under the requirement that no inventory shortage is allowed in the system. We first formulate the lot-sizing problem as a mixed integer programming (MIP) model. Next, an efficient genetic algorithm (GA) model is constructed for solving large-scale lot-sizing problems. An illustrative example with two cases in a touch panel manufacturer is used to illustrate the practicality of these models, and a sensitivity analysis is applied to understand the impact of the changes in parameters to the outcomes. The results demonstrate that both the MIP model and the GA model are effective and relatively accurate tools for determining the replenishment for touch panel manufacturing for multi-periods with quantity discount and batch transportation. The contributions of this article are to construct an MIP model to obtain an optimal solution when the problem is not too complicated itself and to present a GA model to find a near-optimal solution efficiently when the problem is complicated.
Theory and computation of optimal low- and medium-thrust transfers
NASA Technical Reports Server (NTRS)
Chuang, C.-H.
1994-01-01
This report describes the current state of development of methods for calculating optimal orbital transfers with large numbers of burns. Reported on first is the homotopy-motivated and so-called direction correction method. So far this method has been partially tested with one solver; the final step has yet to be implemented. Second is the patched transfer method. This method is rooted in some simplifying approximations made on the original optimal control problem. The transfer is broken up into single-burn segments, each single-burn solved as a predictor step and the whole problem then solved with a corrector step.
NASA Astrophysics Data System (ADS)
Yang, Bo; Wang, Mi; Xu, Wen; Li, Deren; Gong, Jianya; Pi, Yingdong
2017-12-01
The potential of large-scale block adjustment (BA) without ground control points (GCPs) has long been a concern among photogrammetric researchers, which is of effective guiding significance for global mapping. However, significant problems with the accuracy and efficiency of this method remain to be solved. In this study, we analyzed the effects of geometric errors on BA, and then developed a step-wise BA method to conduct integrated processing of large-scale ZY-3 satellite images without GCPs. We first pre-processed the BA data, by adopting a geometric calibration (GC) method based on the viewing-angle model to compensate for systematic errors, such that the BA input images were of good initial geometric quality. The second step was integrated BA without GCPs, in which a series of technical methods were used to solve bottleneck problems and ensure accuracy and efficiency. The BA model, based on virtual control points (VCPs), was constructed to address the rank deficiency problem caused by lack of absolute constraints. We then developed a parallel matching strategy to improve the efficiency of tie points (TPs) matching, and adopted a three-array data structure based on sparsity to relieve the storage and calculation burden of the high-order modified equation. Finally, we used the conjugate gradient method to improve the speed of solving the high-order equations. To evaluate the feasibility of the presented large-scale BA method, we conducted three experiments on real data collected by the ZY-3 satellite. The experimental results indicate that the presented method can effectively improve the geometric accuracies of ZY-3 satellite images. This study demonstrates the feasibility of large-scale mapping without GCPs.
Efficient Implementation of an Optimal Interpolator for Large Spatial Data Sets
NASA Technical Reports Server (NTRS)
Memarsadeghi, Nargess; Mount, David M.
2007-01-01
Interpolating scattered data points is a problem of wide ranging interest. A number of approaches for interpolation have been proposed both from theoretical domains such as computational geometry and in applications' fields such as geostatistics. Our motivation arises from geological and mining applications. In many instances data can be costly to compute and are available only at nonuniformly scattered positions. Because of the high cost of collecting measurements, high accuracy is required in the interpolants. One of the most popular interpolation methods in this field is called ordinary kriging. It is popular because it is a best linear unbiased estimator. The price for its statistical optimality is that the estimator is computationally very expensive. This is because the value of each interpolant is given by the solution of a large dense linear system. In practice, kriging problems have been solved approximately by restricting the domain to a small local neighborhood of points that lie near the query point. Determining the proper size for this neighborhood is a solved by ad hoc methods, and it has been shown that this approach leads to undesirable discontinuities in the interpolant. Recently a more principled approach to approximating kriging has been proposed based on a technique called covariance tapering. This process achieves its efficiency by replacing the large dense kriging system with a much sparser linear system. This technique has been applied to a restriction of our problem, called simple kriging, which is not unbiased for general data sets. In this paper we generalize these results by showing how to apply covariance tapering to the more general problem of ordinary kriging. Through experimentation we demonstrate the space and time efficiency and accuracy of approximating ordinary kriging through the use of covariance tapering combined with iterative methods for solving large sparse systems. We demonstrate our approach on large data sizes arising both from synthetic sources and from real applications.
Application of symbolic/numeric matrix solution techniques to the NASTRAN program
NASA Technical Reports Server (NTRS)
Buturla, E. M.; Burroughs, S. H.
1977-01-01
The matrix solving algorithm of any finite element algorithm is extremely important since solution of the matrix equations requires a large amount of elapse time due to null calculations and excessive input/output operations. An alternate method of solving the matrix equations is presented. A symbolic processing step followed by numeric solution yields the solution very rapidly and is especially useful for nonlinear problems.
Hoppmann, Christiane A; Blanchard-Fields, Fredda
2011-09-01
Problem-solving does not take place in isolation and often involves social others such as spouses. Using repeated daily life assessments from 98 older spouses (M age = 72 years; M marriage length = 42 years), the present study examined theoretical notions from social-contextual models of coping regarding (a) the origins of problem-solving variability and (b) associations between problem-solving and specific problem-, person-, and couple- characteristics. Multilevel models indicate that the lion's share of variability in everyday problem-solving is located at the level of the problem situation. Importantly, participants reported more proactive emotion regulation and collaborative problem-solving for social than nonsocial problems. We also found person-specific consistencies in problem-solving. That is, older spouses high in Neuroticism reported more problems across the study period as well as less instrumental problem-solving and more passive emotion regulation than older spouses low in Neuroticism. Contrary to expectations, relationship satisfaction was unrelated to problem-solving in the present sample. Results are in line with the stress and coping literature in demonstrating that everyday problem-solving is a dynamic process that has to be viewed in the broader context in which it occurs. Our findings also complement previous laboratory-based work on everyday problem-solving by underscoring the benefits of examining everyday problem-solving as it unfolds in spouses' own environment.
Resource Letter RPS-1: Research in problem solving
NASA Astrophysics Data System (ADS)
Hsu, Leonardo; Brewe, Eric; Foster, Thomas M.; Harper, Kathleen A.
2004-09-01
This Resource Letter provides a guide to the literature on research in problem solving, especially in physics. The references were compiled with two audiences in mind: physicists who are (or might become) engaged in research on problem solving, and physics instructors who are interested in using research results to improve their students' learning of problem solving. In addition to general references, journal articles and books are cited for the following topics: cognitive aspects of problem solving, expert-novice problem-solver characteristics, problem solving in mathematics, alternative problem types, curricular interventions, and the use of computers in problem solving.
Topological defects in alternative theories to cosmic inflation and string cosmology
NASA Astrophysics Data System (ADS)
Alexander, Stephon H. S.
The physics of the Early Universe is described in terms of the inflationary paradigm, which is based on a marriage between Einstein's general theory of relativity minimally coupled to quantum field theory. Inflation was posed to solve some of the outstanding problems of the Standard Big Bang Cosmology (SBB) such as the horizon, formation of structure and monopole problems. Despite its observational and theoretical successes, inflation is plagued with fine tuning and initial singularity problems. On the other hand, superstring/M theory, a theory of quantum gravity, possesses symmetries which naturally avoid space-time singularities. This thesis investigates alternative theories to cosmic inflation for solving the initial singularity, horizon and monopole problems, making use of topological defects. It was proposed by Dvali, Liu and Vaschaspati that the monopole problem can be solved without inflation if domain walls "sweep" up the monopoles in the early universe, thus reducing their number density significantly. Necessary for this mechanism to work is the presence of an attractive force between the monopole and the domain wall as well as a channel for the monopole's unwinding. We show numerically and analytically in two field theory models that for global defects the attraction is a universal result but the unwinding is model specific. The second part of this thesis investigates a string/M theory inspired model for solving the horizon problem. It was proposed by Moffat, Albrecht and Magueijo that the horizon problem is solved with a "phase transition" associated with a varying speed of light before the surface of last scattering. We provide a string/M theory mechanism based on assuming that our space-time is a D-3 brane probing a bulk supergravity black hole bulk background. This mechanism provides the necessary time variation of the velocity of light to solve the horizon problem. We suggest a mechanism which stablilizes the speed of light on the D-3 brane. We finally address the cosmological initial singularity problem using the target space duality inherent in string/M theory. It was suggested by Brandenberger and Vafa that superstring theory can solve the singularity problem and in addition explain why only three spatial dimensions can become large. We show that under specific conditions this mechanism still persists when including the effects of D-branes.
Paul B. Hamel
2008-01-01
Cerulean Warbler is a bird with problems; this migratorybird lives in environments on which large numbersof people depend for an adequate productivelivelihood, energy, high quality wood products, coffee,and cacao. Solving the biological problems of thisspecies in its complex...
NASA Technical Reports Server (NTRS)
Iida, H. T.
1966-01-01
Computational procedure reduces the numerical effort whenever the method of finite differences is used to solve ablation problems for which the surface recession is large relative to the initial slab thickness. The number of numerical operations required for a given maximum space mesh size is reduced.
Students’ difficulties in probabilistic problem-solving
NASA Astrophysics Data System (ADS)
Arum, D. P.; Kusmayadi, T. A.; Pramudya, I.
2018-03-01
There are many errors can be identified when students solving mathematics problems, particularly in solving the probabilistic problem. This present study aims to investigate students’ difficulties in solving the probabilistic problem. It focuses on analyzing and describing students errors during solving the problem. This research used the qualitative method with case study strategy. The subjects in this research involve ten students of 9th grade that were selected by purposive sampling. Data in this research involve students’ probabilistic problem-solving result and recorded interview regarding students’ difficulties in solving the problem. Those data were analyzed descriptively using Miles and Huberman steps. The results show that students have difficulties in solving the probabilistic problem and can be divided into three categories. First difficulties relate to students’ difficulties in understanding the probabilistic problem. Second, students’ difficulties in choosing and using appropriate strategies for solving the problem. Third, students’ difficulties with the computational process in solving the problem. Based on the result seems that students still have difficulties in solving the probabilistic problem. It means that students have not able to use their knowledge and ability for responding probabilistic problem yet. Therefore, it is important for mathematics teachers to plan probabilistic learning which could optimize students probabilistic thinking ability.
Neural Networks For Demodulation Of Phase-Modulated Signals
NASA Technical Reports Server (NTRS)
Altes, Richard A.
1995-01-01
Hopfield neural networks proposed for demodulating quadrature phase-shift-keyed (QPSK) signals carrying digital information. Networks solve nonlinear integral equations prior demodulation circuits cannot solve. Consists of set of N operational amplifiers connected in parallel, with weighted feedback from output terminal of each amplifier to input terminals of other amplifiers. Used to solve signal processing problems. Implemented as analog very-large-scale integrated circuit that achieves rapid convergence. Alternatively, implemented as digital simulation of such circuit. Also used to improve phase estimation performance over that of phase-locked loop.
Mega-prizes in medicine: big cash awards may stimulate useful and rapid therapeutic innovation.
Charlton, Bruce G
2007-01-01
Following Horrobin's suggestion of 1986, I argue that offering very large prizes (tens of millions of US dollars, or more) for solving specific therapeutic problems, would be an excellent strategy for promoting the rapid development of effective new treatments. The two mainstream ways of paying for medical research are funding the process with grants or funding the outcome via patent protection. When grants are used to fund the process of research the result tends to be 'pure' science, guided by intrinsic scientific objectives. Practical results, such as useful therapeutic advances, are a by-product. Patent-seeking research, by contrast, is more focused on technology than science. It seeks practical results; and aims to pay for itself (and make a profit) in the long term by generating a patentable product or procedure. Prize-seeking research is subject to different incentives and applicable to different situations than either process-funded or patent-seeking research. Prize seeking researchers have a strong incentive to solve the specified problem as rapidly as possible, but the problem may be solved using old ideas that are scientifically mundane or unpatentable technologies and methods. Prizes therefore seem to generate solutions which are incremental extensions, new applications or novel combinations of already-existing technologies. The main use of mega-prizes in medicine would be to accelerate therapeutic progress in stagnant fields of research and to address urgent problems. For example, medical charities focused on specific diseases should consider accumulating their resources until they can offer a mega-prize for solving a clinical problem of special concern to their patients. Prize money should be big enough to pay for the research and development, the evaluation of the new treatment in a clinical trial, and with a large profit left-over to compensate for the intrinsic risk of competing. Sufficiently large amounts of money, and the prestige and publicity derived from winning a mega-prize, could rapidly mobilize research efforts to discover a whole range of scientifically un-glamorous but clinically-useful therapeutic breakthroughs.
A Multilevel, Hierarchical Sampling Technique for Spatially Correlated Random Fields
Osborn, Sarah; Vassilevski, Panayot S.; Villa, Umberto
2017-10-26
In this paper, we propose an alternative method to generate samples of a spatially correlated random field with applications to large-scale problems for forward propagation of uncertainty. A classical approach for generating these samples is the Karhunen--Loève (KL) decomposition. However, the KL expansion requires solving a dense eigenvalue problem and is therefore computationally infeasible for large-scale problems. Sampling methods based on stochastic partial differential equations provide a highly scalable way to sample Gaussian fields, but the resulting parametrization is mesh dependent. We propose a multilevel decomposition of the stochastic field to allow for scalable, hierarchical sampling based on solving amore » mixed finite element formulation of a stochastic reaction-diffusion equation with a random, white noise source function. Lastly, numerical experiments are presented to demonstrate the scalability of the sampling method as well as numerical results of multilevel Monte Carlo simulations for a subsurface porous media flow application using the proposed sampling method.« less
A Multilevel, Hierarchical Sampling Technique for Spatially Correlated Random Fields
DOE Office of Scientific and Technical Information (OSTI.GOV)
Osborn, Sarah; Vassilevski, Panayot S.; Villa, Umberto
In this paper, we propose an alternative method to generate samples of a spatially correlated random field with applications to large-scale problems for forward propagation of uncertainty. A classical approach for generating these samples is the Karhunen--Loève (KL) decomposition. However, the KL expansion requires solving a dense eigenvalue problem and is therefore computationally infeasible for large-scale problems. Sampling methods based on stochastic partial differential equations provide a highly scalable way to sample Gaussian fields, but the resulting parametrization is mesh dependent. We propose a multilevel decomposition of the stochastic field to allow for scalable, hierarchical sampling based on solving amore » mixed finite element formulation of a stochastic reaction-diffusion equation with a random, white noise source function. Lastly, numerical experiments are presented to demonstrate the scalability of the sampling method as well as numerical results of multilevel Monte Carlo simulations for a subsurface porous media flow application using the proposed sampling method.« less
Parameter optimization of differential evolution algorithm for automatic playlist generation problem
NASA Astrophysics Data System (ADS)
Alamag, Kaye Melina Natividad B.; Addawe, Joel M.
2017-11-01
With the digitalization of music, the number of collection of music increased largely and there is a need to create lists of music that filter the collection according to user preferences, thus giving rise to the Automatic Playlist Generation Problem (APGP). Previous attempts to solve this problem include the use of search and optimization algorithms. If a music database is very large, the algorithm to be used must be able to search the lists thoroughly taking into account the quality of the playlist given a set of user constraints. In this paper we perform an evolutionary meta-heuristic optimization algorithm, Differential Evolution (DE) using different combination of parameter values and select the best performing set when used to solve four standard test functions. Performance of the proposed algorithm is then compared with normal Genetic Algorithm (GA) and a hybrid GA with Tabu Search. Numerical simulations are carried out to show better results from Differential Evolution approach with the optimized parameter values.
NASA Astrophysics Data System (ADS)
Adams, Wendy Kristine
The purpose of my research was to produce a problem solving evaluation tool for physics. To do this it was necessary to gain a thorough understanding of how students solve problems. Although physics educators highly value problem solving and have put extensive effort into understanding successful problem solving, there is currently no efficient way to evaluate problem solving skill. Attempts have been made in the past; however, knowledge of the principles required to solve the subject problem are so absolutely critical that they completely overshadow any other skills students may use when solving a problem. The work presented here is unique because the evaluation tool removes the requirement that the student already have a grasp of physics concepts. It is also unique because I picked a wide range of people and picked a wide range of tasks for evaluation. This is an important design feature that helps make things emerge more clearly. This dissertation includes an extensive literature review of problem solving in physics, math, education and cognitive science as well as descriptions of studies involving student use of interactive computer simulations, the design and validation of a beliefs about physics survey and finally the design of the problem solving evaluation tool. I have successfully developed and validated a problem solving evaluation tool that identifies 44 separate assets (skills) necessary for solving problems. Rigorous validation studies, including work with an independent interviewer, show these assets identified by this content-free evaluation tool are the same assets that students use to solve problems in mechanics and quantum mechanics. Understanding this set of component assets will help teachers and researchers address problem solving within the classroom.
Blanchard-Fields, Fredda; Mienaltowski, Andrew; Seay, Renee Baldi
2007-01-01
Using the Everyday Problem Solving Inventory of Cornelius and Caspi, we examined differences in problem-solving strategy endorsement and effectiveness in two domains of everyday functioning (instrumental or interpersonal, and a mixture of the two domains) and for four strategies (avoidance-denial, passive dependence, planful problem solving, and cognitive analysis). Consistent with past research, our research showed that older adults were more problem focused than young adults in their approach to solving instrumental problems, whereas older adults selected more avoidant-denial strategies than young adults when solving interpersonal problems. Overall, older adults were also more effective than young adults when solving everyday problems, in particular for interpersonal problems.
Spontaneous gestures influence strategy choices in problem solving.
Alibali, Martha W; Spencer, Robert C; Knox, Lucy; Kita, Sotaro
2011-09-01
Do gestures merely reflect problem-solving processes, or do they play a functional role in problem solving? We hypothesized that gestures highlight and structure perceptual-motor information, and thereby make such information more likely to be used in problem solving. Participants in two experiments solved problems requiring the prediction of gear movement, either with gesture allowed or with gesture prohibited. Such problems can be correctly solved using either a perceptual-motor strategy (simulation of gear movements) or an abstract strategy (the parity strategy). Participants in the gesture-allowed condition were more likely to use perceptual-motor strategies than were participants in the gesture-prohibited condition. Gesture promoted use of perceptual-motor strategies both for participants who talked aloud while solving the problems (Experiment 1) and for participants who solved the problems silently (Experiment 2). Thus, spontaneous gestures influence strategy choices in problem solving.
Dixon-Gordon, Katherine L; Chapman, Alexander L; Lovasz, Nathalie; Walters, Kris
2011-10-01
Borderline personality disorder (BPD) is associated with poor social problem solving and problems with emotion regulation. In this study, the social problem-solving performance of undergraduates with high (n = 26), mid (n = 32), or low (n = 29) levels of BPD features was assessed with the Social Problem-Solving Inventory-Revised and using the means-ends problem-solving procedure before and after a social rejection stressor. The high-BP group, but not the low-BP group, showed a significant reduction in relevant solutions to social problems and more inappropriate solutions following the negative emotion induction. Increases in self-reported negative emotions during the emotion induction mediated the relationship between BP features and reductions in social problem-solving performance. In addition, the high-BP group demonstrated trait deficits in social problem solving on the Social Problem-Solving Inventory-Revised. These findings suggest that future research must examine social problem solving under differing emotional conditions, and that clinical interventions to improve social problem solving among persons with BP features should focus on responses to emotional contexts.
Dynamically programmable cache
NASA Astrophysics Data System (ADS)
Nakkar, Mouna; Harding, John A.; Schwartz, David A.; Franzon, Paul D.; Conte, Thomas
1998-10-01
Reconfigurable machines have recently been used as co- processors to accelerate the execution of certain algorithms or program subroutines. The problems with the above approach include high reconfiguration time and limited partial reconfiguration. By far the most critical problems are: (1) the small on-chip memory which results in slower execution time, and (2) small FPGA areas that cannot implement large subroutines. Dynamically Programmable Cache (DPC) is a novel architecture for embedded processors which offers solutions to the above problems. To solve memory access problems, DPC processors merge reconfigurable arrays with the data cache at various cache levels to create a multi-level reconfigurable machines. As a result DPC machines have both higher data accessibility and FPGA memory bandwidth. To solve the limited FPGA resource problem, DPC processors implemented multi-context switching (Virtualization) concept. Virtualization allows implementation of large subroutines with fewer FPGA cells. Additionally, DPC processors can parallelize the execution of several operations resulting in faster execution time. In this paper, the speedup improvement for DPC machines are shown to be 5X faster than an Altera FLEX10K FPGA chip and 2X faster than a Sun Ultral SPARC station for two different algorithms (convolution and motion estimation).
High-order Path Integral Monte Carlo methods for solving strongly correlated fermion problems
NASA Astrophysics Data System (ADS)
Chin, Siu A.
2015-03-01
In solving for the ground state of a strongly correlated many-fermion system, the conventional second-order Path Integral Monte Carlo method is plagued with the sign problem. This is due to the large number of anti-symmetric free fermion propagators that are needed to extract the square of the ground state wave function at large imaginary time. In this work, I show that optimized fourth-order Path Integral Monte Carlo methods, which uses no more than 5 free-fermion propagators, in conjunction with the use of the Hamiltonian energy estimator, can yield accurate ground state energies for quantum dots with up to 20 polarized electrons. The correlations are directly built-in and no explicit wave functions are needed. This work is supported by the Qatar National Research Fund NPRP GRANT #5-674-1-114.
An Investigation of Secondary Teachers’ Understanding and Belief on Mathematical Problem Solving
NASA Astrophysics Data System (ADS)
Yuli Eko Siswono, Tatag; Wachidul Kohar, Ahmad; Kurniasari, Ika; Puji Astuti, Yuliani
2016-02-01
Weaknesses on problem solving of Indonesian students as reported by recent international surveys give rise to questions on how Indonesian teachers bring out idea of problem solving in mathematics lesson. An explorative study was undertaken to investigate how secondary teachers who teach mathematics at junior high school level understand and show belief toward mathematical problem solving. Participants were teachers from four cities in East Java province comprising 45 state teachers and 25 private teachers. Data was obtained through questionnaires and written test. The results of this study point out that the teachers understand pedagogical problem solving knowledge well as indicated by high score of observed teachers‘ responses showing understanding on problem solving as instruction as well as implementation of problem solving in teaching practice. However, they less understand on problem solving content knowledge such as problem solving strategies and meaning of problem itself. Regarding teacher's difficulties, teachers admitted to most frequently fail in (1) determining a precise mathematical model or strategies when carrying out problem solving steps which is supported by data of test result that revealed transformation error as the most frequently observed errors in teachers’ work and (2) choosing suitable real situation when designing context-based problem solving task. Meanwhile, analysis of teacher's beliefs on problem solving shows that teachers tend to view both mathematics and how students should learn mathematics as body static perspective, while they tend to believe to apply idea of problem solving as dynamic approach when teaching mathematics.
ERIC Educational Resources Information Center
Hayel Al-Srour, Nadia; Al-Ali, Safa M.; Al-Oweidi, Alia
2016-01-01
The present study aims to detect the impact of teacher training on creative writing and problem-solving using both Futuristic scenarios program to solve problems creatively, and creative problem solving. To achieve the objectives of the study, the sample was divided into two groups, the first consist of 20 teachers, and 23 teachers to second…
Inflationary dynamics for matrix eigenvalue problems
Heller, Eric J.; Kaplan, Lev; Pollmann, Frank
2008-01-01
Many fields of science and engineering require finding eigenvalues and eigenvectors of large matrices. The solutions can represent oscillatory modes of a bridge, a violin, the disposition of electrons around an atom or molecule, the acoustic modes of a concert hall, or hundreds of other physical quantities. Often only the few eigenpairs with the lowest or highest frequency (extremal solutions) are needed. Methods that have been developed over the past 60 years to solve such problems include the Lanczos algorithm, Jacobi–Davidson techniques, and the conjugate gradient method. Here, we present a way to solve the extremal eigenvalue/eigenvector problem, turning it into a nonlinear classical mechanical system with a modified Lagrangian constraint. The constraint induces exponential inflationary growth of the desired extremal solutions. PMID:18511564
NASA Astrophysics Data System (ADS)
Palacio-Cayetano, Joycelin
"Problem-solving through reflective thinking should be both the method and valuable outcome of science instruction in America's schools" proclaimed John Dewey (Gabel, 1995). If the development of problem-solving is a primary goal of science education, more problem-solving opportunities must be an integral part of K-16 education. To examine the effective use of technology in developing and assessing problem-solving skills, a problem-solving authoring, learning, and assessment software, the UCLA IMMEX Program-Interactive Multimedia Exercises-was investigated. This study was a twenty-week quasi-experimental study that was implemented as a control-group time series design among 120 tenth grade students. Both the experimental group (n = 60) and the control group (n = 60) participated in a problem-based learning curriculum; however, the experimental group received regular intensive experiences with IMMEX problem-solving and the control group did not. Problem-solving pretest and posttest were administered to all students. The instruments used were a 35-item Processes of Biological Inquiry Test and an IMMEX problem-solving assessment test, True Roots. Students who participated in the IMMEX Program achieved significant (p <.05) gains in problem-solving skills on both problem-solving assessment instruments. This study provided evidence that IMMEX software is highly efficient in evaluating salient elements of problem-solving. Outputs of students' problem-solving strategies revealed that unsuccessful problem solvers primarily used the following four strategies: (1) no data search strategy, students simply guessed; (2) limited data search strategy leading to insufficient data and premature closing; (3) irrelevant data search strategy, students focus in areas bearing no substantive data; and (4) extensive data search strategy with inadequate integration and analysis. On the contrary, successful problem solvers used the following strategies; (1) focused search strategy coupled with the ability to fill in knowledge gaps by accessing the appropriate resources; (2) targeted search strategy coupled with high level of analytical and integration skills; and (3) focused search strategy coupled with superior discrimination, analytical, and integration skills. The strategies of students who were successful and unsuccessful solving IMMEX problems were consistent with those of expert and novice problem solvers identified in the literature on problem-solving.
Resing, Wilma C M; Bakker, Merel; Pronk, Christine M E; Elliott, Julian G
2017-01-01
The current study investigated developmental trajectories of analogical reasoning performance of 104 7- and 8-year-old children. We employed a microgenetic research method and multilevel analysis to examine the influence of several background variables and experimental treatment on the children's developmental trajectories. Our participants were divided into two treatment groups: repeated practice alone and repeated practice with training. Each child received an initial working memory assessment and was subsequently asked to solve figural analogies on each of several sessions. We examined children's analogical problem-solving behavior and their subsequent verbal accounts of their employed solving processes. We also investigated the influence of verbal and visual-spatial working memory capacity and initial variability in strategy use on analogical reasoning development. Results indicated that children in both treatment groups improved but that gains were greater for those who had received training. Training also reduced the influence of children's initial variability in the use of analogical strategies with the degree of improvement in reasoning largely unrelated to working memory capacity. Findings from this study demonstrate the value of a microgenetic research method and the use of multilevel analysis to examine inter- and intra-individual change in problem-solving processes. Copyright © 2016 Elsevier Inc. All rights reserved.
Quantum algorithm for energy matching in hard optimization problems
NASA Astrophysics Data System (ADS)
Baldwin, C. L.; Laumann, C. R.
2018-06-01
We consider the ability of local quantum dynamics to solve the "energy-matching" problem: given an instance of a classical optimization problem and a low-energy state, find another macroscopically distinct low-energy state. Energy matching is difficult in rugged optimization landscapes, as the given state provides little information about the distant topography. Here, we show that the introduction of quantum dynamics can provide a speedup over classical algorithms in a large class of hard optimization problems. Tunneling allows the system to explore the optimization landscape while approximately conserving the classical energy, even in the presence of large barriers. Specifically, we study energy matching in the random p -spin model of spin-glass theory. Using perturbation theory and exact diagonalization, we show that introducing a transverse field leads to three sharp dynamical phases, only one of which solves the matching problem: (1) a small-field "trapped" phase, in which tunneling is too weak for the system to escape the vicinity of the initial state; (2) a large-field "excited" phase, in which the field excites the system into high-energy states, effectively forgetting the initial energy; and (3) the intermediate "tunneling" phase, in which the system succeeds at energy matching. The rate at which distant states are found in the tunneling phase, although exponentially slow in system size, is exponentially faster than classical search algorithms.
In search of the 'Aha!' experience: Elucidating the emotionality of insight problem-solving.
Shen, Wangbing; Yuan, Yuan; Liu, Chang; Luo, Jing
2016-05-01
Although the experience of insight has long been noted, the essence of the 'Aha!' experience, reflecting a sudden change in the brain that accompanies an insight solution, remains largely unknown. This work aimed to uncover the mystery of the 'Aha!' experience through three studies. In Study 1, participants were required to solve a set of verbal insight problems and then subjectively report their affective experience when solving the problem. The participants were found to have experienced many types of emotions, with happiness the most frequently reported one. Multidimensional scaling was employed in Study 2 to simplify the dimensions of these reported emotions. The results showed that these different types of emotions could be clearly placed in two-dimensional space and that components constituting the 'Aha!' experience mainly reflected positive emotion and approached cognition. To validate previous findings, in Study 3, participants were asked to select the most appropriate emotional item describing their feelings at the time the problem was solved. The results of this study replicated the multidimensional construct consisting of approached cognition and positive affect. These three studies provide the first direct evidence of the essence of the 'Aha!' The potential significance of the findings was discussed. © 2015 The British Psychological Society.
Search and Coherence-Building in Intuition and Insight Problem Solving.
Öllinger, Michael; von Müller, Albrecht
2017-01-01
Coherence-building is a key concept for a better understanding of the underlying mechanisms of intuition and insight problem solving. There are several accounts that address certain aspects of coherence-building. However, there is still no proper framework defining the general principles of coherence-building. We propose a four-stage model of coherence-building. The first stage starts with spreading activation restricted by constraints. This dynamic is a well-defined rule based process. The second stage is characterized by detecting a coherent state. We adopted a fluency account assuming that the ease of information processing indicates the realization of a coherent state. The third stage is designated to evaluate the result of the coherence-building process and assess whether the given problem is solved or not. If the coherent state does not fit the requirements of the task, the process re-enters at stage 1. These three stages characterize intuition. For insight problem solving a fourth stage is necessary, which restructures the given representation after repeated failure, so that a new search space results. The new search space enables new coherent states. We provide a review of the most important findings, outline our model, present a large number of examples, deduce potential new paradigms and measures that might help to decipher the underlying cognitive processes.
Young doctors' problem solving strategies on call may be improved.
Michelsen, Jens; Malchow-Møller, Axel; Charles, Peder; Eika, Berit
2013-03-01
The first year following graduation from medical school is challenging as learning from books changes to workplace-based learning. Analysis and reflection on experience may ease this transition. We used Significant Event Analysis (SEA) as a tool to explore what pre-registration house officers (PRHOs) consider successful and problematic events, and to identify what problem-solving strategies they employ. A senior house officer systematically led the PRHO through the SEA of one successful and one problematic event following a night call. The PRHO wrote answers to questions about diagnosis, what happened, how he or she contributed and what knowledge-gaining activities the PRHO would prioritise before the next call. By using an inductive, thematic data analysis, we identified five problem-solving strategies: non-analytical reasoning, analytical reasoning, communication with patients, communication with colleagues and professional behaviour. On average, 1.5 strategies were used in the successful events and 1.2 strategies in the problematic events. Most PRHOs were unable to suggest activities other than reading textbooks. SEA was valuable for the identification of PRHOs' problem-solving strategies in a natural setting. PRHOs should be assisted in increasing their repertoire of strategies, and they should also be helped to "learn to learn" as they were largely unable to point to new learning strategies. not relevant. not relevant.
Dornburg, Courtney C; Stevens, Susan M; Hendrickson, Stacey M L; Davidson, George S
2009-08-01
An experiment was conducted to compare the effectiveness of individual versus group electronic brainstorming to address difficult, real-world challenges. Although industrial reliance on electronic communications has become ubiquitous, empirical and theoretical understanding of the bounds of its effectiveness have been limited. Previous research using short-term laboratory experiments have engaged small groups of students in answering questions irrelevant to an industrial setting. The present experiment extends current findings beyond the laboratory to larger groups of real-world employees addressing organization-relevant challenges during the course of 4 days. Employees and contractors at a national laboratory participated, either in a group setting or individually, in an electronic brainstorm to pose solutions to a real-world problem. The data demonstrate that (for this design) individuals perform at least as well as groups in producing quantity of electronic ideas, regardless of brainstorming duration. However, when judged with respect to quality along three dimensions (originality, feasibility, and effectiveness), the individuals significantly (p < .05) outperformed the group. When quality is used to benchmark success, these data indicate that work-relevant challenges are better solved by aggregating electronic individual responses rather than by electronically convening a group. This research suggests that industrial reliance on electronic problem-solving groups should be tempered, and large nominal groups may be more appropriate corporate problem-solving vehicles.
Metcalfe, Arron W. S.; Ashkenazi, Sarit; Rosenberg-Lee, Miriam; Menon, Vinod
2013-01-01
Baddeley and Hitch’s multi-component working memory (WM) model has played an enduring and influential role in our understanding of cognitive abilities. Very little is known, however, about the neural basis of this multi-component WM model and the differential role each component plays in mediating arithmetic problem solving abilities in children. Here, we investigate the neural basis of the central executive (CE), phonological (PL) and visuo-spatial (VS) components of WM during a demanding mental arithmetic task in 7–9 year old children (N=74). The VS component was the strongest predictor of math ability in children and was associated with increased arithmetic complexity-related responses in left dorsolateral and right ventrolateral prefrontal cortices as well as bilateral intra-parietal sulcus and supramarginal gyrus in posterior parietal cortex. Critically, VS, CE and PL abilities were associated with largely distinct patterns of brain response. Overlap between VS and CE components was observed in left supramarginal gyrus and no overlap was observed between VS and PL components. Our findings point to a central role of visuo-spatial WM during arithmetic problem-solving in young grade-school children and highlight the usefulness of the multi-component Baddeley and Hitch WM model in fractionating the neural correlates of arithmetic problem solving during development. PMID:24212504
Search and Coherence-Building in Intuition and Insight Problem Solving
Öllinger, Michael; von Müller, Albrecht
2017-01-01
Coherence-building is a key concept for a better understanding of the underlying mechanisms of intuition and insight problem solving. There are several accounts that address certain aspects of coherence-building. However, there is still no proper framework defining the general principles of coherence-building. We propose a four-stage model of coherence-building. The first stage starts with spreading activation restricted by constraints. This dynamic is a well-defined rule based process. The second stage is characterized by detecting a coherent state. We adopted a fluency account assuming that the ease of information processing indicates the realization of a coherent state. The third stage is designated to evaluate the result of the coherence-building process and assess whether the given problem is solved or not. If the coherent state does not fit the requirements of the task, the process re-enters at stage 1. These three stages characterize intuition. For insight problem solving a fourth stage is necessary, which restructures the given representation after repeated failure, so that a new search space results. The new search space enables new coherent states. We provide a review of the most important findings, outline our model, present a large number of examples, deduce potential new paradigms and measures that might help to decipher the underlying cognitive processes. PMID:28611702
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chen, Yi; Jakeman, John; Gittelson, Claude
2015-01-08
In this paper we present a localized polynomial chaos expansion for partial differential equations (PDE) with random inputs. In particular, we focus on time independent linear stochastic problems with high dimensional random inputs, where the traditional polynomial chaos methods, and most of the existing methods, incur prohibitively high simulation cost. Furthermore, the local polynomial chaos method employs a domain decomposition technique to approximate the stochastic solution locally. In each subdomain, a subdomain problem is solved independently and, more importantly, in a much lower dimensional random space. In a postprocesing stage, accurate samples of the original stochastic problems are obtained frommore » the samples of the local solutions by enforcing the correct stochastic structure of the random inputs and the coupling conditions at the interfaces of the subdomains. Overall, the method is able to solve stochastic PDEs in very large dimensions by solving a collection of low dimensional local problems and can be highly efficient. In our paper we present the general mathematical framework of the methodology and use numerical examples to demonstrate the properties of the method.« less
The application of dynamic programming in production planning
NASA Astrophysics Data System (ADS)
Wu, Run
2017-05-01
Nowadays, with the popularity of the computers, various industries and fields are widely applying computer information technology, which brings about huge demand for a variety of application software. In order to develop software meeting various needs with most economical cost and best quality, programmers must design efficient algorithms. A superior algorithm can not only soul up one thing, but also maximize the benefits and generate the smallest overhead. As one of the common algorithms, dynamic programming algorithms are used to solving problems with some sort of optimal properties. When solving problems with a large amount of sub-problems that needs repetitive calculations, the ordinary sub-recursive method requires to consume exponential time, and dynamic programming algorithm can reduce the time complexity of the algorithm to the polynomial level, according to which we can conclude that dynamic programming algorithm is a very efficient compared to other algorithms reducing the computational complexity and enriching the computational results. In this paper, we expound the concept, basic elements, properties, core, solving steps and difficulties of the dynamic programming algorithm besides, establish the dynamic programming model of the production planning problem.
Revisiting software specification and design for large astronomy projects
NASA Astrophysics Data System (ADS)
Wiant, Scott; Berukoff, Steven
2016-07-01
The separation of science and engineering in the delivery of software systems overlooks the true nature of the problem being solved and the organization that will solve it. Use of a systems engineering approach to managing the requirements flow between these two groups as between a customer and contractor has been used with varying degrees of success by well-known entities such as the U.S. Department of Defense. However, treating science as the customer and engineering as the contractor fosters unfavorable consequences that can be avoided and opportunities that are missed. For example, the "problem" being solved is only partially specified through the requirements generation process since it focuses on detailed specification guiding the parties to a technical solution. Equally important is the portion of the problem that will be solved through the definition of processes and staff interacting through them. This interchange between people and processes is often underrepresented and under appreciated. By concentrating on the full problem and collaborating on a strategy for its solution a science-implementing organization can realize the benefits of driving towards common goals (not just requirements) and a cohesive solution to the entire problem. The initial phase of any project when well executed is often the most difficult yet most critical and thus it is essential to employ a methodology that reinforces collaboration and leverages the full suite of capabilities within the team. This paper describes an integrated approach to specifying the needs induced by a problem and the design of its solution.
ERIC Educational Resources Information Center
Aljaberi, Nahil M.; Gheith, Eman
2016-01-01
This study aims to investigate the ability of pre-service class teacher at University of Petrain solving mathematical problems using Polya's Techniques, their level of problem solving skills in daily-life issues. The study also investigates the correlation between their ability to solve mathematical problems and their level of problem solving…
Solving Navier-Stokes equations on a massively parallel processor; The 1 GFLOP performance
DOE Office of Scientific and Technical Information (OSTI.GOV)
Saati, A.; Biringen, S.; Farhat, C.
This paper reports on experience in solving large-scale fluid dynamics problems on the Connection Machine model CM-2. The authors have implemented a parallel version of the MacCormack scheme for the solution of the Navier-Stokes equations. By using triad floating point operations and reducing the number of interprocessor communications, they have achieved a sustained performance rate of 1.42 GFLOPS.
The Association between Motivation, Affect, and Self-regulated Learning When Solving Problems.
Baars, Martine; Wijnia, Lisette; Paas, Fred
2017-01-01
Self-regulated learning (SRL) skills are essential for learning during school years, particularly in complex problem-solving domains, such as biology and math. Although a lot of studies have focused on the cognitive resources that are needed for learning to solve problems in a self-regulated way, affective and motivational resources have received much less research attention. The current study investigated the relation between affect (i.e., Positive Affect and Negative Affect Scale), motivation (i.e., autonomous and controlled motivation), mental effort, SRL skills, and problem-solving performance when learning to solve biology problems in a self-regulated online learning environment. In the learning phase, secondary education students studied video-modeling examples of how to solve hereditary problems, solved hereditary problems which they chose themselves from a set of problems with different complexity levels (i.e., five levels). In the posttest, students solved hereditary problems, self-assessed their performance, and chose a next problem from the set of problems but did not solve these problems. The results from this study showed that negative affect, inaccurate self-assessments during the posttest, and higher perceptions of mental effort during the posttest were negatively associated with problem-solving performance after learning in a self-regulated way.
An approach to solving large reliability models
NASA Technical Reports Server (NTRS)
Boyd, Mark A.; Veeraraghavan, Malathi; Dugan, Joanne Bechta; Trivedi, Kishor S.
1988-01-01
This paper describes a unified approach to the problem of solving large realistic reliability models. The methodology integrates behavioral decomposition, state trunction, and efficient sparse matrix-based numerical methods. The use of fault trees, together with ancillary information regarding dependencies to automatically generate the underlying Markov model state space is proposed. The effectiveness of this approach is illustrated by modeling a state-of-the-art flight control system and a multiprocessor system. Nonexponential distributions for times to failure of components are assumed in the latter example. The modeling tool used for most of this analysis is HARP (the Hybrid Automated Reliability Predictor).
Algorithms for solving large sparse systems of simultaneous linear equations on vector processors
NASA Technical Reports Server (NTRS)
David, R. E.
1984-01-01
Very efficient algorithms for solving large sparse systems of simultaneous linear equations have been developed for serial processing computers. These involve a reordering of matrix rows and columns in order to obtain a near triangular pattern of nonzero elements. Then an LU factorization is developed to represent the matrix inverse in terms of a sequence of elementary Gaussian eliminations, or pivots. In this paper it is shown how these algorithms are adapted for efficient implementation on vector processors. Results obtained on the CYBER 200 Model 205 are presented for a series of large test problems which show the comparative advantages of the triangularization and vector processing algorithms.
A multilevel probabilistic beam search algorithm for the shortest common supersequence problem.
Gallardo, José E
2012-01-01
The shortest common supersequence problem is a classical problem with many applications in different fields such as planning, Artificial Intelligence and especially in Bioinformatics. Due to its NP-hardness, we can not expect to efficiently solve this problem using conventional exact techniques. This paper presents a heuristic to tackle this problem based on the use at different levels of a probabilistic variant of a classical heuristic known as Beam Search. The proposed algorithm is empirically analysed and compared to current approaches in the literature. Experiments show that it provides better quality solutions in a reasonable time for medium and large instances of the problem. For very large instances, our heuristic also provides better solutions, but required execution times may increase considerably.
Experimental Matching of Instances to Heuristics for Constraint Satisfaction Problems.
Moreno-Scott, Jorge Humberto; Ortiz-Bayliss, José Carlos; Terashima-Marín, Hugo; Conant-Pablos, Santiago Enrique
2016-01-01
Constraint satisfaction problems are of special interest for the artificial intelligence and operations research community due to their many applications. Although heuristics involved in solving these problems have largely been studied in the past, little is known about the relation between instances and the respective performance of the heuristics used to solve them. This paper focuses on both the exploration of the instance space to identify relations between instances and good performing heuristics and how to use such relations to improve the search. Firstly, the document describes a methodology to explore the instance space of constraint satisfaction problems and evaluate the corresponding performance of six variable ordering heuristics for such instances in order to find regions on the instance space where some heuristics outperform the others. Analyzing such regions favors the understanding of how these heuristics work and contribute to their improvement. Secondly, we use the information gathered from the first stage to predict the most suitable heuristic to use according to the features of the instance currently being solved. This approach proved to be competitive when compared against the heuristics applied in isolation on both randomly generated and structured instances of constraint satisfaction problems.
Experimental Matching of Instances to Heuristics for Constraint Satisfaction Problems
Moreno-Scott, Jorge Humberto; Ortiz-Bayliss, José Carlos; Terashima-Marín, Hugo; Conant-Pablos, Santiago Enrique
2016-01-01
Constraint satisfaction problems are of special interest for the artificial intelligence and operations research community due to their many applications. Although heuristics involved in solving these problems have largely been studied in the past, little is known about the relation between instances and the respective performance of the heuristics used to solve them. This paper focuses on both the exploration of the instance space to identify relations between instances and good performing heuristics and how to use such relations to improve the search. Firstly, the document describes a methodology to explore the instance space of constraint satisfaction problems and evaluate the corresponding performance of six variable ordering heuristics for such instances in order to find regions on the instance space where some heuristics outperform the others. Analyzing such regions favors the understanding of how these heuristics work and contribute to their improvement. Secondly, we use the information gathered from the first stage to predict the most suitable heuristic to use according to the features of the instance currently being solved. This approach proved to be competitive when compared against the heuristics applied in isolation on both randomly generated and structured instances of constraint satisfaction problems. PMID:26949383
A Simple Label Switching Algorithm for Semisupervised Structural SVMs.
Balamurugan, P; Shevade, Shirish; Sundararajan, S
2015-10-01
In structured output learning, obtaining labeled data for real-world applications is usually costly, while unlabeled examples are available in abundance. Semisupervised structured classification deals with a small number of labeled examples and a large number of unlabeled structured data. In this work, we consider semisupervised structural support vector machines with domain constraints. The optimization problem, which in general is not convex, contains the loss terms associated with the labeled and unlabeled examples, along with the domain constraints. We propose a simple optimization approach that alternates between solving a supervised learning problem and a constraint matching problem. Solving the constraint matching problem is difficult for structured prediction, and we propose an efficient and effective label switching method to solve it. The alternating optimization is carried out within a deterministic annealing framework, which helps in effective constraint matching and avoiding poor local minima, which are not very useful. The algorithm is simple and easy to implement. Further, it is suitable for any structured output learning problem where exact inference is available. Experiments on benchmark sequence labeling data sets and a natural language parsing data set show that the proposed approach, though simple, achieves comparable generalization performance.
Dynamic simulation solves process control problem in Oman
DOE Office of Scientific and Technical Information (OSTI.GOV)
NONE
1998-11-16
A dynamic simulation study solved the process control problems for a Saih Rawl, Oman, gas compressor station operated by Petroleum Development of Oman (PDO). PDO encountered persistent compressor failure that caused frequent facility shutdowns, oil production deferment, and gas flaring. It commissioned MSE (Consultants) Ltd., U.K., to find a solution for the problem. Saih Rawl, about 40 km from Qarn Alam, produces oil and associated gas from a large number of low and high-pressure wells. Oil and gas are separated in three separators. The oil is pumped to Qarn Alam for treatment and export. Associated gas is compressed in twomore » parallel trains. Train K-1115 is a 350,000 standard cu m/day, four-stage reciprocating compressor driven by a fixed-speed electric motor. Train K-1120 is a 1 million standard cu m/day, four-stage reciprocating compressor driven by a fixed-speed electric motor. Train K-1120 is a 1 million standard cu m/day, four-stage centrifugal compressor driven by a variable-speed motor. The paper describes tripping and surging problems with the gas compressor and the control simplifications that solved the problem.« less
NASA Astrophysics Data System (ADS)
Lateh, Masitah Abdul; Kamilah Muda, Azah; Yusof, Zeratul Izzah Mohd; Azilah Muda, Noor; Sanusi Azmi, Mohd
2017-09-01
The emerging era of big data for past few years has led to large and complex data which needed faster and better decision making. However, the small dataset problems still arise in a certain area which causes analysis and decision are hard to make. In order to build a prediction model, a large sample is required as a training sample of the model. Small dataset is insufficient to produce an accurate prediction model. This paper will review an artificial data generation approach as one of the solution to solve the small dataset problem.
Functional reasoning in diagnostic problem solving
NASA Technical Reports Server (NTRS)
Sticklen, Jon; Bond, W. E.; Stclair, D. C.
1988-01-01
This work is one facet of an integrated approach to diagnostic problem solving for aircraft and space systems currently under development. The authors are applying a method of modeling and reasoning about deep knowledge based on a functional viewpoint. The approach recognizes a level of device understanding which is intermediate between a compiled level of typical Expert Systems, and a deep level at which large-scale device behavior is derived from known properties of device structure and component behavior. At this intermediate functional level, a device is modeled in three steps. First, a component decomposition of the device is defined. Second, the functionality of each device/subdevice is abstractly identified. Third, the state sequences which implement each function are specified. Given a functional representation and a set of initial conditions, the functional reasoner acts as a consequence finder. The output of the consequence finder can be utilized in diagnostic problem solving. The paper also discussed ways in which this functional approach may find application in the aerospace field.
Fast scaffolding with small independent mixed integer programs
Salmela, Leena; Mäkinen, Veli; Välimäki, Niko; Ylinen, Johannes; Ukkonen, Esko
2011-01-01
Motivation: Assembling genomes from short read data has become increasingly popular, but the problem remains computationally challenging especially for larger genomes. We study the scaffolding phase of sequence assembly where preassembled contigs are ordered based on mate pair data. Results: We present MIP Scaffolder that divides the scaffolding problem into smaller subproblems and solves these with mixed integer programming. The scaffolding problem can be represented as a graph and the biconnected components of this graph can be solved independently. We present a technique for restricting the size of these subproblems so that they can be solved accurately with mixed integer programming. We compare MIP Scaffolder to two state of the art methods, SOPRA and SSPACE. MIP Scaffolder is fast and produces better or as good scaffolds as its competitors on large genomes. Availability: The source code of MIP Scaffolder is freely available at http://www.cs.helsinki.fi/u/lmsalmel/mip-scaffolder/. Contact: leena.salmela@cs.helsinki.fi PMID:21998153
NASA Astrophysics Data System (ADS)
Ushijima, T.; Yeh, W.
2013-12-01
An optimal experimental design algorithm is developed to select locations for a network of observation wells that provides the maximum information about unknown hydraulic conductivity in a confined, anisotropic aquifer. The design employs a maximal information criterion that chooses, among competing designs, the design that maximizes the sum of squared sensitivities while conforming to specified design constraints. Because that the formulated problem is non-convex and contains integer variables (necessitating a combinatorial search), for a realistically-scaled model, the problem may be difficult, if not impossible, to solve through traditional mathematical programming techniques. Genetic Algorithms (GAs) are designed to search out the global optimum; however because a GA requires a large number of calls to a groundwater model, the formulated optimization problem may still be infeasible to solve. To overcome this, Proper Orthogonal Decomposition (POD) is applied to the groundwater model to reduce its dimension. The information matrix in the full model space can then be searched without solving the full model.
Multilevel acceleration of scattering-source iterations with application to electron transport
Drumm, Clif; Fan, Wesley
2017-08-18
Acceleration/preconditioning strategies available in the SCEPTRE radiation transport code are described. A flexible transport synthetic acceleration (TSA) algorithm that uses a low-order discrete-ordinates (S N) or spherical-harmonics (P N) solve to accelerate convergence of a high-order S N source-iteration (SI) solve is described. Convergence of the low-order solves can be further accelerated by applying off-the-shelf incomplete-factorization or algebraic-multigrid methods. Also available is an algorithm that uses a generalized minimum residual (GMRES) iterative method rather than SI for convergence, using a parallel sweep-based solver to build up a Krylov subspace. TSA has been applied as a preconditioner to accelerate the convergencemore » of the GMRES iterations. The methods are applied to several problems involving electron transport and problems with artificial cross sections with large scattering ratios. These methods were compared and evaluated by considering material discontinuities and scattering anisotropy. Observed accelerations obtained are highly problem dependent, but speedup factors around 10 have been observed in typical applications.« less
NASA Astrophysics Data System (ADS)
Gross, Lutz; Altinay, Cihan; Fenwick, Joel; Smith, Troy
2014-05-01
The program package escript has been designed for solving mathematical modeling problems using python, see Gross et al. (2013). Its development and maintenance has been funded by the Australian Commonwealth to provide open source software infrastructure for the Australian Earth Science community (recent funding by the Australian Geophysical Observing System EIF (AGOS) and the AuScope Collaborative Research Infrastructure Scheme (CRIS)). The key concepts of escript are based on the terminology of spatial functions and partial differential equations (PDEs) - an approach providing abstraction from the underlying spatial discretization method (i.e. the finite element method (FEM)). This feature presents a programming environment to the user which is easy to use even for complex models. Due to the fact that implementations are independent from data structures simulations are easily portable across desktop computers and scalable compute clusters without modifications to the program code. escript has been successfully applied in a variety of applications including modeling mantel convection, melting processes, volcanic flow, earthquakes, faulting, multi-phase flow, block caving and mineralization (see Poulet et al. 2013). The recent escript release (see Gross et al. (2013)) provides an open framework for solving joint inversion problems for geophysical data sets (potential field, seismic and electro-magnetic). The strategy bases on the idea to formulate the inversion problem as an optimization problem with PDE constraints where the cost function is defined by the data defect and the regularization term for the rock properties, see Gross & Kemp (2013). This approach of first-optimize-then-discretize avoids the assemblage of the - in general- dense sensitivity matrix as used in conventional approaches where discrete programming techniques are applied to the discretized problem (first-discretize-then-optimize). In this paper we will discuss the mathematical framework for inversion and appropriate solution schemes in escript. We will also give a brief introduction into escript's open framework for defining and solving geophysical inversion problems. Finally we will show some benchmark results to demonstrate the computational scalability of the inversion method across a large number of cores and compute nodes in a parallel computing environment. References: - L. Gross et al. (2013): Escript Solving Partial Differential Equations in Python Version 3.4, The University of Queensland, https://launchpad.net/escript-finley - L. Gross and C. Kemp (2013) Large Scale Joint Inversion of Geophysical Data using the Finite Element Method in escript. ASEG Extended Abstracts 2013, http://dx.doi.org/10.1071/ASEG2013ab306 - T. Poulet, L. Gross, D. Georgiev, J. Cleverley (2012): escript-RT: Reactive transport simulation in Python using escript, Computers & Geosciences, Volume 45, 168-176. http://dx.doi.org/10.1016/j.cageo.2011.11.005.
Research on the precise positioning of customers in large data environment
NASA Astrophysics Data System (ADS)
Zhou, Xu; He, Lili
2018-04-01
Customer positioning has always been a problem that enterprises focus on. In this paper, FCM clustering algorithm is used to cluster customer groups. However, due to the traditional FCM clustering algorithm, which is susceptible to the influence of the initial clustering center and easy to fall into the local optimal problem, the short board of FCM is solved by the gray optimization algorithm (GWO) to achieve efficient and accurate handling of a large number of retailer data.
NASA Technical Reports Server (NTRS)
Hofmann, R.
1980-01-01
The STEALTH code system, which solves large strain, nonlinear continuum mechanics problems, was rigorously structured in both overall design and programming standards. The design is based on the theoretical elements of analysis while the programming standards attempt to establish a parallelism between physical theory, programming structure, and documentation. These features have made it easy to maintain, modify, and transport the codes. It has also guaranteed users a high level of quality control and quality assurance.
Intelligent Design: Student Perceptions of Teaching and Learning in Large Social Work Classes
ERIC Educational Resources Information Center
Moulding, Nicole Therese
2010-01-01
Research into the effects of large classes demonstrates that students are disadvantaged in terms of higher order learning because interactions between teachers and students occur at lower cognitive levels. This has significance for social work education, with its emphasis on the development of critical thinking and problem solving, both higher…
Extraction of a group-pair relation: problem-solving relation from web-board documents.
Pechsiri, Chaveevan; Piriyakul, Rapepun
2016-01-01
This paper aims to extract a group-pair relation as a Problem-Solving relation, for example a DiseaseSymptom-Treatment relation and a CarProblem-Repair relation, between two event-explanation groups, a problem-concept group as a symptom/CarProblem-concept group and a solving-concept group as a treatment-concept/repair concept group from hospital-web-board and car-repair-guru-web-board documents. The Problem-Solving relation (particularly Symptom-Treatment relation) including the graphical representation benefits non-professional persons by supporting knowledge of primarily solving problems. The research contains three problems: how to identify an EDU (an Elementary Discourse Unit, which is a simple sentence) with the event concept of either a problem or a solution; how to determine a problem-concept EDU boundary and a solving-concept EDU boundary as two event-explanation groups, and how to determine the Problem-Solving relation between these two event-explanation groups. Therefore, we apply word co-occurrence to identify a problem-concept EDU and a solving-concept EDU, and machine-learning techniques to solve a problem-concept EDU boundary and a solving-concept EDU boundary. We propose using k-mean and Naïve Bayes to determine the Problem-Solving relation between the two event-explanation groups involved with clustering features. In contrast to previous works, the proposed approach enables group-pair relation extraction with high accuracy.
NASA Astrophysics Data System (ADS)
Nasution, M. L.; Yerizon, Y.; Gusmiyanti, R.
2018-04-01
One of the purpose mathematic learning is to develop problem solving abilities. Problem solving is obtained through experience in questioning non-routine. Improving students’ mathematical problem-solving abilities required an appropriate strategy in learning activities one of them is models problem based learning (PBL). Thus, the purpose of this research is to determine whether the problem solving abilities of mathematical students’ who learn to use PBL better than on the ability of students’ mathematical problem solving by applying conventional learning. This research included quasi experiment with static group design and population is students class XI MIA SMAN 1 Lubuk Alung. Class experiment in the class XI MIA 5 and class control in the class XI MIA 6. The instrument of final test students’ mathematical problem solving used essay form. The result of data final test in analyzed with t-test. The result is students’ mathematical problem solving abilities with PBL better then on the ability of students’ mathematical problem solving by applying conventional learning. It’s seen from the high percentage achieved by the group of students who learn to use PBL for each indicator of students’ mathematical problem solving.
Introduction to bioinformatics.
Can, Tolga
2014-01-01
Bioinformatics is an interdisciplinary field mainly involving molecular biology and genetics, computer science, mathematics, and statistics. Data intensive, large-scale biological problems are addressed from a computational point of view. The most common problems are modeling biological processes at the molecular level and making inferences from collected data. A bioinformatics solution usually involves the following steps: Collect statistics from biological data. Build a computational model. Solve a computational modeling problem. Test and evaluate a computational algorithm. This chapter gives a brief introduction to bioinformatics by first providing an introduction to biological terminology and then discussing some classical bioinformatics problems organized by the types of data sources. Sequence analysis is the analysis of DNA and protein sequences for clues regarding function and includes subproblems such as identification of homologs, multiple sequence alignment, searching sequence patterns, and evolutionary analyses. Protein structures are three-dimensional data and the associated problems are structure prediction (secondary and tertiary), analysis of protein structures for clues regarding function, and structural alignment. Gene expression data is usually represented as matrices and analysis of microarray data mostly involves statistics analysis, classification, and clustering approaches. Biological networks such as gene regulatory networks, metabolic pathways, and protein-protein interaction networks are usually modeled as graphs and graph theoretic approaches are used to solve associated problems such as construction and analysis of large-scale networks.
Self-Scheduling Parallel Methods for Multiple Serial Codes with Application to WOPWOP
NASA Technical Reports Server (NTRS)
Long, Lyle N.; Brentner, Kenneth S.
2000-01-01
This paper presents a scheme for efficiently running a large number of serial jobs on parallel computers. Two examples are given of computer programs that run relatively quickly, but often they must be run numerous times to obtain all the results needed. It is very common in science and engineering to have codes that are not massive computing challenges in themselves, but due to the number of instances that must be run, they do become large-scale computing problems. The two examples given here represent common problems in aerospace engineering: aerodynamic panel methods and aeroacoustic integral methods. The first example simply solves many systems of linear equations. This is representative of an aerodynamic panel code where someone would like to solve for numerous angles of attack. The complete code for this first example is included in the appendix so that it can be readily used by others as a template. The second example is an aeroacoustics code (WOPWOP) that solves the Ffowcs Williams Hawkings equation to predict the far-field sound due to rotating blades. In this example, one quite often needs to compute the sound at numerous observer locations, hence parallelization is utilized to automate the noise computation for a large number of observers.
ERIC Educational Resources Information Center
American School and University, 1984
1984-01-01
A large urban school board has solved the security problem for electronically stored student records by refusing to file them on a dialup system and by requiring a notarized affidavit of custody before release to a parent or guardian. (TE)
Using a general problem-solving strategy to promote transfer.
Youssef-Shalala, Amina; Ayres, Paul; Schubert, Carina; Sweller, John
2014-09-01
Cognitive load theory was used to hypothesize that a general problem-solving strategy based on a make-as-many-moves-as-possible heuristic could facilitate problem solutions for transfer problems. In four experiments, school students were required to learn about a topic through practice with a general problem-solving strategy, through a conventional problem solving strategy or by studying worked examples. In Experiments 1 and 2 using junior high school students learning geometry, low knowledge students in the general problem-solving group scored significantly higher on near or far transfer tests than the conventional problem-solving group. In Experiment 3, an advantage for a general problem-solving group over a group presented worked examples was obtained on far transfer tests using the same curriculum materials, again presented to junior high school students. No differences between conditions were found in Experiments 1, 2, or 3 using test problems similar to the acquisition problems. Experiment 4 used senior high school students studying economics and found the general problem-solving group scored significantly higher than the conventional problem-solving group on both similar and transfer tests. It was concluded that the general problem-solving strategy was helpful for novices, but not for students that had access to domain-specific knowledge. PsycINFO Database Record (c) 2014 APA, all rights reserved.
NASA Astrophysics Data System (ADS)
Hafner, Robert; Stewart, Jim
Past problem-solving research has provided a basis for helping students structure their knowledge and apply appropriate problem-solving strategies to solve problems for which their knowledge (or mental models) of scientific phenomena is adequate (model-using problem solving). This research examines how problem solving in the domain of Mendelian genetics proceeds in situations where solvers' mental models are insufficient to solve problems at hand (model-revising problem solving). Such situations require solvers to use existing models to recognize anomalous data and to revise those models to accommodate the data. The study was conducted in the context of 9-week high school genetics course and addressed: the heuristics charactenstic of successful model-revising problem solving: the nature of the model revisions, made by students as well as the nature of model development across problem types; and the basis upon which solvers decide that a revised model is sufficient (that t has both predictive and explanatory power).
Azad, Gazi F.; Kim, Mina; Marcus, Steven C.; Mandell, David S.; Sheridan, Susan M.
2016-01-01
Effective parent-teacher communication involves problem-solving concerns about students. Few studies have examined problem solving interactions between parents and teachers of children with autism spectrum disorder (ASD), with a particular focus on identifying communication barriers and strategies for improving them. This study examined the problem-solving behaviors of parents and teachers of children with ASD. Participants included 18 teachers and 39 parents of children with ASD. Parent-teacher dyads were prompted to discuss and provide a solution for a problem that a student experienced at home and at school. Parents and teachers also reported on their problem-solving behaviors. Results showed that parents and teachers displayed limited use of the core elements of problem-solving. Teachers displayed more problem-solving behaviors than parents. Both groups reported engaging in more problem-solving behaviors than they were observed to display during their discussions. Our findings suggest that teacher and parent training programs should include collaborative approaches to problem-solving. PMID:28392604
Azad, Gazi F; Kim, Mina; Marcus, Steven C; Mandell, David S; Sheridan, Susan M
2016-12-01
Effective parent-teacher communication involves problem-solving concerns about students. Few studies have examined problem solving interactions between parents and teachers of children with autism spectrum disorder (ASD), with a particular focus on identifying communication barriers and strategies for improving them. This study examined the problem-solving behaviors of parents and teachers of children with ASD. Participants included 18 teachers and 39 parents of children with ASD. Parent-teacher dyads were prompted to discuss and provide a solution for a problem that a student experienced at home and at school. Parents and teachers also reported on their problem-solving behaviors. Results showed that parents and teachers displayed limited use of the core elements of problem-solving. Teachers displayed more problem-solving behaviors than parents. Both groups reported engaging in more problem-solving behaviors than they were observed to display during their discussions. Our findings suggest that teacher and parent training programs should include collaborative approaches to problem-solving.
NASA Astrophysics Data System (ADS)
Rr Chusnul, C.; Mardiyana, S., Dewi Retno
2017-12-01
Problem solving is the basis of mathematics learning. Problem solving teaches us to clarify an issue coherently in order to avoid misunderstanding information. Sometimes there may be mistakes in problem solving due to misunderstanding the issue, choosing a wrong concept or misapplied concept. The problem-solving test was carried out after students were given treatment on learning by using cooperative learning of TTW type. The purpose of this study was to elucidate student problem regarding to problem solving errors after learning by using cooperative learning of TTW type. Newman stages were used to identify problem solving errors in this study. The new research used a descriptive method to find out problem solving errors in students. The subject in this study were students of Vocational Senior High School (SMK) in 10th grade. Test and interview was conducted for data collection. Thus, the results of this study suggested problem solving errors in students after learning by using cooperative learning of TTW type for Newman stages.
Rejection Sensitivity and Depression: Indirect Effects Through Problem Solving.
Kraines, Morganne A; Wells, Tony T
2017-01-01
Rejection sensitivity (RS) and deficits in social problem solving are risk factors for depression. Despite their relationship to depression and the potential connection between them, no studies have examined RS and social problem solving together in the context of depression. As such, we examined RS, five facets of social problem solving, and symptoms of depression in a young adult sample. A total of 180 participants completed measures of RS, social problem solving, and depressive symptoms. We used bootstrapping to examine the indirect effect of RS on depressive symptoms through problem solving. RS was positively associated with depressive symptoms. A negative problem orientation, impulsive/careless style, and avoidance style of social problem solving were positively associated with depressive symptoms, and a positive problem orientation was negatively associated with depressive symptoms. RS demonstrated an indirect effect on depressive symptoms through two social problem-solving facets: the tendency to view problems as threats to one's well-being and an avoidance problem-solving style characterized by procrastination, passivity, or overdependence on others. These results are consistent with prior research that found a positive association between RS and depression symptoms, but this is the first study to implicate specific problem-solving deficits in the relationship between RS and depression. Our results suggest that depressive symptoms in high RS individuals may result from viewing problems as threats and taking an avoidant, rather than proactive, approach to dealing with problems. These findings may have implications for problem-solving interventions for rejection sensitive individuals.
Parallel-vector solution of large-scale structural analysis problems on supercomputers
NASA Technical Reports Server (NTRS)
Storaasli, Olaf O.; Nguyen, Duc T.; Agarwal, Tarun K.
1989-01-01
A direct linear equation solution method based on the Choleski factorization procedure is presented which exploits both parallel and vector features of supercomputers. The new equation solver is described, and its performance is evaluated by solving structural analysis problems on three high-performance computers. The method has been implemented using Force, a generic parallel FORTRAN language.
Multilevel decomposition of complete vehicle configuration in a parallel computing environment
NASA Technical Reports Server (NTRS)
Bhatt, Vinay; Ragsdell, K. M.
1989-01-01
This research summarizes various approaches to multilevel decomposition to solve large structural problems. A linear decomposition scheme based on the Sobieski algorithm is selected as a vehicle for automated synthesis of a complete vehicle configuration in a parallel processing environment. The research is in a developmental state. Preliminary numerical results are presented for several example problems.
ERIC Educational Resources Information Center
Didion, Catherine Jay; Guenther, Rita S.; Gunderson, Victoria
2012-01-01
Scientists, engineers, and medical professionals play a vital role in building the 21st- century science and technology enterprises that will create solutions and jobs critical to solving the large, complex, and interdisciplinary problems faced by society: problems in energy, sustainability, the environment, water, food, disease, and healthcare.…
Mathematical Problem Solving: A Review of the Literature.
ERIC Educational Resources Information Center
Funkhouser, Charles
The major perspectives on problem solving of the twentieth century are reviewed--associationism, Gestalt psychology, and cognitive science. The results of the review on teaching problem solving and the uses of computers to teach problem solving are included. Four major issues related to the teaching of problem solving are discussed: (1)…
Teaching Problem Solving Skills to Elementary Age Students with Autism
ERIC Educational Resources Information Center
Cote, Debra L.; Jones, Vita L.; Barnett, Crystal; Pavelek, Karin; Nguyen, Hoang; Sparks, Shannon L.
2014-01-01
Students with disabilities need problem-solving skills to promote their success in solving the problems of daily life. The research into problem-solving instruction has been limited for students with autism. Using a problem-solving intervention and the Self Determined Learning Model of Instruction, three elementary age students with autism were…
Learning problem-solving skills in a distance education physics course
NASA Astrophysics Data System (ADS)
Rampho, G. J.; Ramorola, M. Z.
2017-10-01
In this paper we present the results of a study on the effectiveness of combinations of delivery modes of distance education in learning problem-solving skills in a distance education introductory physics course. A problem-solving instruction with the explicit teaching of a problem-solving strategy and worked-out examples were implemented in the course. The study used the ex post facto research design with stratified sampling to investigate the effect of the learning of a problem-solving strategy on the problem-solving performance. The number of problems attempted and the mean frequency of using a strategy in solving problems in the three course presentation modes were compared. The finding of the study indicated that combining the different course presentation modes had no statistically significant effect in the learning of problem-solving skills in the distance education course.
NASA Astrophysics Data System (ADS)
Tian, Zhang; Yanfeng, Gong
2017-05-01
In order to solve the contradiction between demand and distribution range of primary energy resource, Ultra High Voltage (UHV) power grids should be developed rapidly to meet development of energy bases and accessing of large-scale renewable energy. This paper reviewed the latest research processes of AC/DC transmission technologies, summarized the characteristics of AC/DC power grids, concluded that China’s power grids certainly enter a new period of large -scale hybrid UHV AC/DC power grids and characteristics of “strong DC and weak AC” becomes increasingly pro minent; possible problems in operation of AC/DC power grids was discussed, and interaction or effect between AC/DC power grids was made an intensive study of; according to above problems in operation of power grids, preliminary scheme is summarized as fo llows: strengthening backbone structures, enhancing AC/DC transmission technologies, promoting protection measures of clean energ y accessing grids, and taking actions to solve stability problems of voltage and frequency etc. It’s valuable for making hybrid UHV AC/DC power grids adapt to operating mode of large power grids, thus guaranteeing security and stability of power system.
Functional fixedness in a technologically sparse culture.
German, Tim P; Barrett, H Clark
2005-01-01
Problem solving can be inefficient when the solution requires subjects to generate an atypical function for an object and the object's typical function has been primed. Subjects become "fixed" on the design function of the object, and problem solving suffers relative to control conditions in which the object's function is not demonstrated. In the current study, such functional fixedness was demonstrated in a sample of adolescents (mean age of 16 years) among the Shuar of Ecuadorian Amazonia, whose technologically sparse culture provides limited access to large numbers of artifacts with highly specialized functions. This result suggests that design function may universally be the core property of artifact concepts in human semantic memory.
Coherent mode decomposition using mixed Wigner functions of Hermite-Gaussian beams.
Tanaka, Takashi
2017-04-15
A new method of coherent mode decomposition (CMD) is proposed that is based on a Wigner-function representation of Hermite-Gaussian beams. In contrast to the well-known method using the cross spectral density (CSD), it directly determines the mode functions and their weights without solving the eigenvalue problem. This facilitates the CMD of partially coherent light whose Wigner functions (and thus CSDs) are not separable, in which case the conventional CMD requires solving an eigenvalue problem with a large matrix and thus is numerically formidable. An example is shown regarding the CMD of synchrotron radiation, one of the most important applications of the proposed method.
An edge-based solution-adaptive method applied to the AIRPLANE code
NASA Technical Reports Server (NTRS)
Biswas, Rupak; Thomas, Scott D.; Cliff, Susan E.
1995-01-01
Computational methods to solve large-scale realistic problems in fluid flow can be made more efficient and cost effective by using them in conjunction with dynamic mesh adaption procedures that perform simultaneous coarsening and refinement to capture flow features of interest. This work couples the tetrahedral mesh adaption scheme, 3D_TAG, with the AIRPLANE code to solve complete aircraft configuration problems in transonic and supersonic flow regimes. Results indicate that the near-field sonic boom pressure signature of a cone-cylinder is improved, the oblique and normal shocks are better resolved on a transonic wing, and the bow shock ahead of an unstarted inlet is better defined.
NASA Astrophysics Data System (ADS)
Sun, Y. S.; Zhang, L.; Xu, B.; Zhang, Y.
2018-04-01
The accurate positioning of optical satellite image without control is the precondition for remote sensing application and small/medium scale mapping in large abroad areas or with large-scale images. In this paper, aiming at the geometric features of optical satellite image, based on a widely used optimization method of constraint problem which is called Alternating Direction Method of Multipliers (ADMM) and RFM least-squares block adjustment, we propose a GCP independent block adjustment method for the large-scale domestic high resolution optical satellite image - GISIBA (GCP-Independent Satellite Imagery Block Adjustment), which is easy to parallelize and highly efficient. In this method, the virtual "average" control points are built to solve the rank defect problem and qualitative and quantitative analysis in block adjustment without control. The test results prove that the horizontal and vertical accuracy of multi-covered and multi-temporal satellite images are better than 10 m and 6 m. Meanwhile the mosaic problem of the adjacent areas in large area DOM production can be solved if the public geographic information data is introduced as horizontal and vertical constraints in the block adjustment process. Finally, through the experiments by using GF-1 and ZY-3 satellite images over several typical test areas, the reliability, accuracy and performance of our developed procedure will be presented and studied in this paper.
Improved Quasi-Newton method via PSB update for solving systems of nonlinear equations
NASA Astrophysics Data System (ADS)
Mamat, Mustafa; Dauda, M. K.; Waziri, M. Y.; Ahmad, Fadhilah; Mohamad, Fatma Susilawati
2016-10-01
The Newton method has some shortcomings which includes computation of the Jacobian matrix which may be difficult or even impossible to compute and solving the Newton system in every iteration. Also, the common setback with some quasi-Newton methods is that they need to compute and store an n × n matrix at each iteration, this is computationally costly for large scale problems. To overcome such drawbacks, an improved Method for solving systems of nonlinear equations via PSB (Powell-Symmetric-Broyden) update is proposed. In the proposed method, the approximate Jacobian inverse Hk of PSB is updated and its efficiency has improved thereby require low memory storage, hence the main aim of this paper. The preliminary numerical results show that the proposed method is practically efficient when applied on some benchmark problems.
Lusby, Richard Martin; Schwierz, Martin; Range, Troels Martin; Larsen, Jesper
2016-11-01
The aim of this paper is to provide an improved method for solving the so-called dynamic patient admission scheduling (DPAS) problem. This is a complex scheduling problem that involves assigning a set of patients to hospital beds over a given time horizon in such a way that several quality measures reflecting patient comfort and treatment efficiency are maximized. Consideration must be given to uncertainty in the length of stays of patients as well as the possibility of emergency patients. We develop an adaptive large neighborhood search (ALNS) procedure to solve the problem. This procedure utilizes a Simulated Annealing framework. We thoroughly test the performance of the proposed ALNS approach on a set of 450 publicly available problem instances. A comparison with the current state-of-the-art indicates that the proposed methodology provides solutions that are of comparable quality for small and medium sized instances (up to 1000 patients); the two approaches provide solutions that differ in quality by approximately 1% on average. The ALNS procedure does, however, provide solutions in a much shorter time frame. On larger instances (between 1000-4000 patients) the improvement in solution quality by the ALNS procedure is substantial, approximately 3-14% on average, and as much as 22% on a single instance. The time taken to find such results is, however, in the worst case, a factor 12 longer on average than the time limit which is granted to the current state-of-the-art. The proposed ALNS procedure is an efficient and flexible method for solving the DPAS problem. Copyright © 2016 Elsevier B.V. All rights reserved.
Mukherjee, Anamitra; Patel, Niravkumar D.; Bishop, Chris; ...
2015-06-08
Lattice spin-fermion models are quite important to study correlated systems where quantum dynamics allows for a separation between slow and fast degrees of freedom. The fast degrees of freedom are treated quantum mechanically while the slow variables, generically referred to as the “spins,” are treated classically. At present, exact diagonalization coupled with classical Monte Carlo (ED + MC) is extensively used to solve numerically a general class of lattice spin-fermion problems. In this common setup, the classical variables (spins) are treated via the standard MC method while the fermion problem is solved by exact diagonalization. The “traveling cluster approximation” (TCA)more » is a real space variant of the ED + MC method that allows to solve spin-fermion problems on lattice sizes with up to 10 3 sites. In this paper, we present a novel reorganization of the TCA algorithm in a manner that can be efficiently parallelized. Finally, this allows us to solve generic spin-fermion models easily on 10 4 lattice sites and with some effort on 10 5 lattice sites, representing the record lattice sizes studied for this family of models.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mukherjee, Anamitra; Patel, Niravkumar D.; Bishop, Chris
Lattice spin-fermion models are quite important to study correlated systems where quantum dynamics allows for a separation between slow and fast degrees of freedom. The fast degrees of freedom are treated quantum mechanically while the slow variables, generically referred to as the “spins,” are treated classically. At present, exact diagonalization coupled with classical Monte Carlo (ED + MC) is extensively used to solve numerically a general class of lattice spin-fermion problems. In this common setup, the classical variables (spins) are treated via the standard MC method while the fermion problem is solved by exact diagonalization. The “traveling cluster approximation” (TCA)more » is a real space variant of the ED + MC method that allows to solve spin-fermion problems on lattice sizes with up to 10 3 sites. In this paper, we present a novel reorganization of the TCA algorithm in a manner that can be efficiently parallelized. Finally, this allows us to solve generic spin-fermion models easily on 10 4 lattice sites and with some effort on 10 5 lattice sites, representing the record lattice sizes studied for this family of models.« less
The Association between Motivation, Affect, and Self-regulated Learning When Solving Problems
Baars, Martine; Wijnia, Lisette; Paas, Fred
2017-01-01
Self-regulated learning (SRL) skills are essential for learning during school years, particularly in complex problem-solving domains, such as biology and math. Although a lot of studies have focused on the cognitive resources that are needed for learning to solve problems in a self-regulated way, affective and motivational resources have received much less research attention. The current study investigated the relation between affect (i.e., Positive Affect and Negative Affect Scale), motivation (i.e., autonomous and controlled motivation), mental effort, SRL skills, and problem-solving performance when learning to solve biology problems in a self-regulated online learning environment. In the learning phase, secondary education students studied video-modeling examples of how to solve hereditary problems, solved hereditary problems which they chose themselves from a set of problems with different complexity levels (i.e., five levels). In the posttest, students solved hereditary problems, self-assessed their performance, and chose a next problem from the set of problems but did not solve these problems. The results from this study showed that negative affect, inaccurate self-assessments during the posttest, and higher perceptions of mental effort during the posttest were negatively associated with problem-solving performance after learning in a self-regulated way. PMID:28848467
Abdulmalik, Jibril; Ani, Cornelius; Ajuwon, Ademola J; Omigbodun, Olayinka
2016-01-01
Aggressive patterns of behavior often start early in childhood, and tend to remain stable into adulthood. The negative consequences include poor academic performance, disciplinary problems and encounters with the juvenile justice system. Early school intervention programs can alter this trajectory for aggressive children. However, there are no studies evaluating the feasibility of such interventions in Africa. This study therefore, assessed the effect of group-based problem-solving interventions on aggressive behaviors among primary school pupils in Ibadan, Nigeria. This was an intervention study with treatment and wait-list control groups. Two public primary schools in Ibadan Nigeria were randomly allocated to an intervention group and a waiting list control group. Teachers rated male Primary five pupils in the two schools on aggressive behaviors and the top 20 highest scorers in each school were selected. Pupils in the intervention school received 6 twice-weekly sessions of group-based intervention, which included problem-solving skills, calming techniques and attribution retraining. Outcome measures were; teacher rated aggressive behaviour (TRAB), self-rated aggression scale (SRAS), strengths and difficulties questionnaire (SDQ), attitude towards aggression questionnaire (ATAQ), and social cognition and attribution scale (SCAS). The participants were aged 12 years (SD = 1.2, range 9-14 years). Both groups had similar socio-demographic backgrounds and baseline measures of aggressive behaviors. Controlling for baseline scores, the intervention group had significantly lower scores on TRAB and SRAS 1-week post intervention with large Cohen's effect sizes of 1.2 and 0.9 respectively. The other outcome measures were not significantly different between the groups post-intervention. Group-based problem solving intervention for aggressive behaviors among primary school students showed significant reductions in both teachers' and students' rated aggressive behaviours with large effect sizes. However, this was a small exploratory trial whose findings may not be generalizable, but it demonstrates that psychological interventions for children with high levels of aggressive behaviour are feasible and potentially effective in Nigeria.
NASA Technical Reports Server (NTRS)
Johnston, William E.; Gannon, Dennis; Nitzberg, Bill; Feiereisen, William (Technical Monitor)
2000-01-01
The term "Grid" refers to distributed, high performance computing and data handling infrastructure that incorporates geographically and organizationally dispersed, heterogeneous resources that are persistent and supported. The vision for NASN's Information Power Grid - a computing and data Grid - is that it will provide significant new capabilities to scientists and engineers by facilitating routine construction of information based problem solving environments / frameworks that will knit together widely distributed computing, data, instrument, and human resources into just-in-time systems that can address complex and large-scale computing and data analysis problems. IPG development and deployment is addressing requirements obtained by analyzing a number of different application areas, in particular from the NASA Aero-Space Technology Enterprise. This analysis has focussed primarily on two types of users: The scientist / design engineer whose primary interest is problem solving (e.g., determining wing aerodynamic characteristics in many different operating environments), and whose primary interface to IPG will be through various sorts of problem solving frameworks. The second type of user if the tool designer: The computational scientists who convert physics and mathematics into code that can simulate the physical world. These are the two primary users of IPG, and they have rather different requirements. This paper describes the current state of IPG (the operational testbed), the set of capabilities being put into place for the operational prototype IPG, as well as some of the longer term R&D tasks.
An experience sampling study of learning, affect, and the demands control support model.
Daniels, Kevin; Boocock, Grahame; Glover, Jane; Holland, Julie; Hartley, Ruth
2009-07-01
The demands control support model (R. A. Karasek & T. Theorell, 1990) indicates that job control and social support enable workers to engage in problem solving. In turn, problem solving is thought to influence learning and well-being (e.g., anxious affect, activated pleasant affect). Two samples (N = 78, N = 106) provided data up to 4 times per day for up to 5 working days. The extent to which job control was used for problem solving was assessed by measuring the extent to which participants changed aspects of their work activities to solve problems. The extent to which social support was used to solve problems was assessed by measuring the extent to which participants discussed problems to solve problems. Learning mediated the relationship between changing aspects of work activities to solve problems and activated pleasant affect. Learning also mediated the relationship between discussing problems to solve problems and activated pleasant affect. The findings indicated that how individuals use control and support to respond to problem-solving demands is associated with organizational and individual phenomena, such as learning and affective well-being.
NASA Technical Reports Server (NTRS)
Chaplygin, S.
1944-01-01
A brief summary of the contents of this paper is presented here. In part I the differential equations of the problem of a gas flow in two dimensions is derived and the particular integrals by which the problem on jets is solved are given. Use is made of the same independent variables as Molenbroek used, but it is found to be more suitable to consider other functions. The stream function and velocity potential corresponding to the problem are given in the form of series. The investigation on the convergence of these series in connection with certain properties of the functions entering them forms the subject of part II. In part III the problem of the outflow of a gas from an infinite vessel with plane walls is solved. In part IV the impact of a gas jet on a plate is considered and the limiting case where the jet expands to infinity changing into a gas flow is taken up in more detail. This also solved the equivalent problem of the resistance of a gaseous medium to the motion of a plate. Finally, in part V, an approximate method is presented that permits a simpler solution of the problem of jet flows in the case where the velocities of the gas (velocities of the particles in the gas) are not very large.
Generating effective project scheduling heuristics by abstraction and reconstitution
NASA Technical Reports Server (NTRS)
Janakiraman, Bhaskar; Prieditis, Armand
1992-01-01
A project scheduling problem consists of a finite set of jobs, each with fixed integer duration, requiring one or more resources such as personnel or equipment, and each subject to a set of precedence relations, which specify allowable job orderings, and a set of mutual exclusion relations, which specify jobs that cannot overlap. No job can be interrupted once started. The objective is to minimize project duration. This objective arises in nearly every large construction project--from software to hardware to buildings. Because such project scheduling problems are NP-hard, they are typically solved by branch-and-bound algorithms. In these algorithms, lower-bound duration estimates (admissible heuristics) are used to improve efficiency. One way to obtain an admissible heuristic is to remove (abstract) all resources and mutual exclusion constraints and then obtain the minimal project duration for the abstracted problem; this minimal duration is the admissible heuristic. Although such abstracted problems can be solved efficiently, they yield inaccurate admissible heuristics precisely because those constraints that are central to solving the original problem are abstracted. This paper describes a method to reconstitute the abstracted constraints back into the solution to the abstracted problem while maintaining efficiency, thereby generating better admissible heuristics. Our results suggest that reconstitution can make good admissible heuristics even better.
Isospin symmetry breaking and large-scale shell-model calculations with the Sakurai-Sugiura method
NASA Astrophysics Data System (ADS)
Mizusaki, Takahiro; Kaneko, Kazunari; Sun, Yang; Tazaki, Shigeru
2015-05-01
Recently isospin symmetry breaking for mass 60-70 region has been investigated based on large-scale shell-model calculations in terms of mirror energy differences (MED), Coulomb energy differences (CED) and triplet energy differences (TED). Behind these investigations, we have encountered a subtle problem in numerical calculations for odd-odd N = Z nuclei with large-scale shell-model calculations. Here we focus on how to solve this subtle problem by the Sakurai-Sugiura (SS) method, which has been recently proposed as a new diagonalization method and has been successfully applied to nuclear shell-model calculations.
What Does (and Doesn't) Make Analogical Problem Solving Easy? A Complexity-Theoretic Perspective
ERIC Educational Resources Information Center
Wareham, Todd; Evans, Patricia; van Rooij, Iris
2011-01-01
Solving new problems can be made easier if one can build on experiences with other problems one has already successfully solved. The ability to exploit earlier problem-solving experiences in solving new problems seems to require several cognitive sub-abilities. Minimally, one needs to be able to retrieve relevant knowledge of earlier solved…
ERIC Educational Resources Information Center
Kamis, Arnold; Khan, Beverly K.
2009-01-01
How do we model and improve technical problem solving, such as network subnetting? This paper reports an experimental study that tested several hypotheses derived from Kolb's experiential learning cycle and Huber's problem solving model. As subjects solved a network subnetting problem, they mapped their mental processes according to Huber's…
ERIC Educational Resources Information Center
Paraschiv, Irina; Olley, J. Gregory
This paper describes the "Problem Solving for Life" training program which trains adolescents and adults with mental retardation in skills for solving social problems. The program requires group participants to solve social problems by practicing two prerequisite skills (relaxation and positive self-statements) and four problem solving steps: (1)…
Young Children's Analogical Problem Solving: Gaining Insights from Video Displays
ERIC Educational Resources Information Center
Chen, Zhe; Siegler, Robert S.
2013-01-01
This study examined how toddlers gain insights from source video displays and use the insights to solve analogous problems. Two- to 2.5-year-olds viewed a source video illustrating a problem-solving strategy and then attempted to solve analogous problems. Older but not younger toddlers extracted the problem-solving strategy depicted in the video…
Investigating Problem-Solving Perseverance Using Lesson Study
ERIC Educational Resources Information Center
Bieda, Kristen N.; Huhn, Craig
2017-01-01
Problem solving has long been a focus of research and curriculum reform (Kilpatrick 1985; Lester 1994; NCTM 1989, 2000; CCSSI 2010). The importance of problem solving is not new, but the Common Core introduced the idea of making sense of problems and persevering in solving them (CCSSI 2010, p. 6) as an aspect of problem solving. Perseverance is…
Problem-solving deficits in Iranian people with borderline personality disorder.
Akbari Dehaghi, Ashraf; Kaviani, Hossein; Tamanaeefar, Shima
2014-01-01
Interventions for people suffering from borderline personality disorder (BPD), such as dialectical behavior therapy, often include a problem-solving component. However, there is an absence of published studies examining the problem-solving abilities of this client group in Iran. The study compared inpatients and outpatients with BPD and a control group on problem-solving capabilities in an Iranian sample. It was hypothesized that patients with BPD would have more deficiencies in this area. Fifteen patients with BPD were compared to 15 healthy participants. Means-ends problem-solving task (MEPS) was used to measure problem-solving skills in both groups. BPD group reported less effective strategies in solving problems as opposed to the healthy group. Compared to the control group, participants with BPD provided empirical support for the use of problem-solving interventions with people suffering from BPD. The findings supported the idea that a problem-solving intervention can be efficiently applied either as a stand-alone therapy or in conjunction with other available psychotherapies to treat people with BPD.
Impulsivity as a mediator in the relationship between problem solving and suicidal ideation.
Gonzalez, Vivian M; Neander, Lucía L
2018-03-15
This study examined whether three facets of impulsivity previously shown to be associated with suicidal ideation and attempts (negative urgency, lack of premeditation, and lack of perseverance) help to account for the established association between problem solving deficits and suicidal ideation. Emerging adult college student drinkers with a history of at least passive suicidal ideation (N = 387) completed measures of problem solving, impulsivity, and suicidal ideation. A path analysis was conducted to examine the mediating role of impulsivity variables in the association between problem solving (rational problem solving, positive and negative problem orientation, and avoidance style) and suicidal ideation. Direct and indirect associations through impulsivity, particularly negative urgency, were found between problem solving and severity of suicidal ideation. Interventions aimed at teaching problem solving skills, as well as self-efficacy and optimism for solving life problems, may help to reduce impulsivity and suicidal ideation. © 2018 Wiley Periodicals, Inc.
Improving mathematical problem solving skills through visual media
NASA Astrophysics Data System (ADS)
Widodo, S. A.; Darhim; Ikhwanudin, T.
2018-01-01
The purpose of this article was to find out the enhancement of students’ mathematical problem solving by using visual learning media. The ability to solve mathematical problems is the ability possessed by students to solve problems encountered, one of the problem-solving model of Polya. This preliminary study was not to make a model, but it only took a conceptual approach by comparing the various literature of problem-solving skills by linking visual learning media. The results of the study indicated that the use of learning media had not been appropriated so that the ability to solve mathematical problems was not optimal. The inappropriateness of media use was due to the instructional media that was not adapted to the characteristics of the learners. Suggestions that can be given is the need to develop visual media to increase the ability to solve problems.
Explicit solutions for exit-only radioactive decay chains
NASA Astrophysics Data System (ADS)
Yuan, Ding; Kernan, Warnick
2007-05-01
In this study, we extended Bateman's [Proc. Cambridge Philos. Soc. 15, 423 (1910)] original work for solving radioactive decay chains and explicitly derived analytic solutions for generic exit-only radioactive decay problems under given initial conditions. Instead of using the conventional Laplace transform for solving Bateman's equations, we used a much simpler algebraic approach. Finally, we discuss methods of breaking down certain classes of large decay chains into collections of simpler chains for easy handling.
Implementing and Bounding a Cascade Heuristic for Large-Scale Optimization
2017-06-01
solving the monolith, we develop a method for producing lower bounds to the optimal objective function value. To do this, we solve a new integer...as developing and analyzing methods for producing lower bounds to the optimal objective function value of the seminal problem monolith, which this...length of the window decreases, the end effects of the model typically increase (Zerr, 2016). There are four primary methods for correcting end
ERIC Educational Resources Information Center
Limin, Chen; Van Dooren, Wim; Verschaffel, Lieven
2013-01-01
The goal of the present study is to investigate the relationship between pupils' problem posing and problem solving abilities, their beliefs about problem posing and problem solving, and their general mathematics abilities, in a Chinese context. Five instruments, i.e., a problem posing test, a problem solving test, a problem posing questionnaire,…
ERIC Educational Resources Information Center
Higgins, Karen M.
This study investigated the effects of Oregon's Lane County "Problem Solving in Mathematics" (PSM) materials on middle-school students' attitudes, beliefs, and abilities in problem solving and mathematics. The instructional approach advocated in PSM includes: the direct teaching of five problem-solving skills, weekly challenge problems,…
A Large-scale Distributed Indexed Learning Framework for Data that Cannot Fit into Memory
2015-03-27
learn a classifier. Integrating three learning techniques (online, semi-supervised and active learning ) together with a selective sampling with minimum communication between the server and the clients solved this problem.
A progress report on estuary modeling by the finite-element method
Gray, William G.
1978-01-01
Various schemes are investigated for finite-element modeling of two-dimensional surface-water flows. The first schemes investigated combine finite-element spatial discretization with split-step time stepping schemes that have been found useful in finite-difference computations. Because of the large number of numerical integrations performed in space and the large sparse matrices solved, these finite-element schemes were found to be economically uncompetitive with finite-difference schemes. A very promising leapfrog scheme is proposed which, when combined with a novel very fast spatial integration procedure, eliminates the need to solve any matrices at all. Additional problems attacked included proper propagation of waves and proper specification of the normal flow-boundary condition. This report indicates work in progress and does not come to a definitive conclusion as to the best approach for finite-element modeling of surface-water problems. The results presented represent findings obtained between September 1973 and July 1976. (Woodard-USGS)
Crowdsourced 'R&D' and medical research.
Callaghan, Christian William
2015-09-01
Crowdsourced R&D, a research methodology increasingly applied to medical research, has properties well suited to large-scale medical data collection and analysis, as well as enabling rapid research responses to crises such as disease outbreaks. Multidisciplinary literature offers diverse perspectives of crowdsourced R&D as a useful large-scale medical data collection and research problem-solving methodology. Crowdsourced R&D has demonstrated 'proof of concept' in a host of different biomedical research applications. A wide range of quality and ethical issues relate to crowdsourced R&D. The rapid growth in applications of crowdsourced R&D in medical research is predicted by an increasing body of multidisciplinary theory. Further research in areas such as artificial intelligence may allow better coordination and management of the high volumes of medical data and problem-solving inputs generated by the crowdsourced R&D process. © The Author 2015. Published by Oxford University Press. All rights reserved. For permissions, please e-mail: journals.permissions@oup.com.
Experimentally Building a Qualitative Understanding of Newton's Second Law
NASA Astrophysics Data System (ADS)
Gates, Joshua
2014-12-01
Newton's second law is one of the cornerstones of the introductory physics curriculum, but it can still trouble a large number of students well after its introduction, hobbling their ability to apply the concept to problem solving1 and to related concepts, such as momentum, circular motion, and orbits. While there are several possibilities for lab activities addressing the functional relationship among net force, mass, and acceleration, the qualitative understanding of the connection between forces and acceleration can still be lacking,2 leading to poor performance in problem solving and in assessments such as the Force Concept Inventory3 and Force and Motion Conceptual Evaluation.4 There is a need for strong conceptual understanding of the relationships between net force and acceleration and between acceleration and velocity in order to effectively address common force-motion misconceptions;5 there is a large literature concerning student understanding of force and motion.6
NASA Technical Reports Server (NTRS)
Nguyen, D. T.; Watson, Willie R. (Technical Monitor)
2005-01-01
The overall objectives of this research work are to formulate and validate efficient parallel algorithms, and to efficiently design/implement computer software for solving large-scale acoustic problems, arised from the unified frameworks of the finite element procedures. The adopted parallel Finite Element (FE) Domain Decomposition (DD) procedures should fully take advantages of multiple processing capabilities offered by most modern high performance computing platforms for efficient parallel computation. To achieve this objective. the formulation needs to integrate efficient sparse (and dense) assembly techniques, hybrid (or mixed) direct and iterative equation solvers, proper pre-conditioned strategies, unrolling strategies, and effective processors' communicating schemes. Finally, the numerical performance of the developed parallel finite element procedures will be evaluated by solving series of structural, and acoustic (symmetrical and un-symmetrical) problems (in different computing platforms). Comparisons with existing "commercialized" and/or "public domain" software are also included, whenever possible.
Towards a global human embryonic stem cell bank.
Lott, Jason P; Savulescu, Julian
2007-08-01
An increasingly unbridgeable gap exists between the supply and demand of transplantable organs. Human embryonic stem cell technology could solve the organ shortage problem by restoring diseased or damaged tissue across a range of common conditions. However, such technology faces several largely ignored immunological challenges in delivering cell lines to large populations. We address some of these challenges and argue in favor of encouraging contribution or intentional creation of embryos from which widely immunocompatible stem cell lines could be derived. Further, we argue that current immunological constraints in tissue transplantation demand the creation of a global stem cell bank, which may hold particular promise for minority populations and other sub-groups currently marginalized from organ procurement and allocation systems. Finally, we conclude by offering a number of practical and ethically oriented recommendations for constructing a human embryonic stem cell bank that we hope will help solve the ongoing organ shortage problem.
NASA Astrophysics Data System (ADS)
Zhang, Jia-shi; Yang, Xi-xiang
2017-11-01
The stratospheric airship has the characteristics of large inertia, long time delay and large disturbance of wind field , so the trajectory control is very difficult .Build the lateral three degrees of freedom dynamic model which consider the wind interference , the dynamics equation is linearized by the small perturbation theory, propose a trajectory control method Combine with the sliding mode control and prediction, design the trajectory controller , takes the HAA airship as the reference to carry out simulation analysis. Results show that the improved sliding mode control with front-feedback method not only can solve well control problems of airship trajectory in wind field, but also can effectively improve the control accuracy of the traditional sliding mode control method, solved problems that using the traditional sliding mode control to control. It provides a useful reference for dynamic modeling and trajectory control of stratospheric airship.
PetIGA: A framework for high-performance isogeometric analysis
Dalcin, Lisandro; Collier, Nathaniel; Vignal, Philippe; ...
2016-05-25
We present PetIGA, a code framework to approximate the solution of partial differential equations using isogeometric analysis. PetIGA can be used to assemble matrices and vectors which come from a Galerkin weak form, discretized with Non-Uniform Rational B-spline basis functions. We base our framework on PETSc, a high-performance library for the scalable solution of partial differential equations, which simplifies the development of large-scale scientific codes, provides a rich environment for prototyping, and separates parallelism from algorithm choice. We describe the implementation of PetIGA, and exemplify its use by solving a model nonlinear problem. To illustrate the robustness and flexibility ofmore » PetIGA, we solve some challenging nonlinear partial differential equations that include problems in both solid and fluid mechanics. Lastly, we show strong scaling results on up to 4096 cores, which confirm the suitability of PetIGA for large scale simulations.« less
Student’s scheme in solving mathematics problems
NASA Astrophysics Data System (ADS)
Setyaningsih, Nining; Juniati, Dwi; Suwarsono
2018-03-01
The purpose of this study was to investigate students’ scheme in solving mathematics problems. Scheme are data structures for representing the concepts stored in memory. In this study, we used it in solving mathematics problems, especially ratio and proportion topics. Scheme is related to problem solving that assumes that a system is developed in the human mind by acquiring a structure in which problem solving procedures are integrated with some concepts. The data were collected by interview and students’ written works. The results of this study revealed are students’ scheme in solving the problem of ratio and proportion as follows: (1) the content scheme, where students can describe the selected components of the problem according to their prior knowledge, (2) the formal scheme, where students can explain in construct a mental model based on components that have been selected from the problem and can use existing schemes to build planning steps, create something that will be used to solve problems and (3) the language scheme, where students can identify terms, or symbols of the components of the problem.Therefore, by using the different strategies to solve the problems, the students’ scheme in solving the ratio and proportion problems will also differ.
Evaluation of Genetic Algorithm Concepts Using Model Problems. Part 2; Multi-Objective Optimization
NASA Technical Reports Server (NTRS)
Holst, Terry L.; Pulliam, Thomas H.
2003-01-01
A genetic algorithm approach suitable for solving multi-objective optimization problems is described and evaluated using a series of simple model problems. Several new features including a binning selection algorithm and a gene-space transformation procedure are included. The genetic algorithm is suitable for finding pareto optimal solutions in search spaces that are defined by any number of genes and that contain any number of local extrema. Results indicate that the genetic algorithm optimization approach is flexible in application and extremely reliable, providing optimal results for all optimization problems attempted. The binning algorithm generally provides pareto front quality enhancements and moderate convergence efficiency improvements for most of the model problems. The gene-space transformation procedure provides a large convergence efficiency enhancement for problems with non-convoluted pareto fronts and a degradation in efficiency for problems with convoluted pareto fronts. The most difficult problems --multi-mode search spaces with a large number of genes and convoluted pareto fronts-- require a large number of function evaluations for GA convergence, but always converge.
ERIC Educational Resources Information Center
Scherer, Ronny; Tiemann, Rudiger
2012-01-01
The ability to solve complex scientific problems is regarded as one of the key competencies in science education. Until now, research on problem solving focused on the relationship between analytical and complex problem solving, but rarely took into account the structure of problem-solving processes and metacognitive aspects. This paper,…
Same Old Problem, New Name? Alerting Students to the Nature of the Problem-Solving Process
ERIC Educational Resources Information Center
Yerushalmi, Edit; Magen, Esther
2006-01-01
Students frequently misconceive the process of problem-solving, expecting the linear process required for solving an exercise, rather than the convoluted search process required to solve a genuine problem. In this paper we present an activity designed to foster in students realization and appreciation of the nature of the problem-solving process,…
ERIC Educational Resources Information Center
Gustafsson, Peter; Jonsson, Gunnar; Enghag, Margareta
2015-01-01
The problem-solving process is investigated for five groups of students when solving context-rich problems in an introductory physics course included in an engineering programme. Through transcripts of their conversation, the paths in the problem-solving process have been traced and related to a general problem-solving model. All groups exhibit…
Klein, Daniel N.; Leon, Andrew C.; Li, Chunshan; D’Zurilla, Thomas J.; Black, Sarah R.; Vivian, Dina; Dowling, Frank; Arnow, Bruce A.; Manber, Rachel; Markowitz, John C.; Kocsis, James H.
2011-01-01
Objective Depression is associated with poor social problem-solving, and psychotherapies that focus on problem-solving skills are efficacious in treating depression. We examined the associations between treatment, social problem solving, and depression in a randomized clinical trial testing the efficacy of psychotherapy augmentation for chronically depressed patients who failed to fully respond to an initial trial of pharmacotherapy (Kocsis et al., 2009). Method Participants with chronic depression (n = 491) received Cognitive Behavioral Analysis System of Psychotherapy (CBASP), which emphasizes interpersonal problem-solving, plus medication; Brief Supportive Psychotherapy (BSP) plus medication; or medication alone for 12 weeks. Results CBASP plus pharmacotherapy was associated with significantly greater improvement in social problem solving than BSP plus pharmacotherapy, and a trend for greater improvement in problem solving than pharmacotherapy alone. In addition, change in social problem solving predicted subsequent change in depressive symptoms over time. However, the magnitude of the associations between changes in social problem solving and subsequent depressive symptoms did not differ across treatment conditions. Conclusions It does not appear that improved social problem solving is a mechanism that uniquely distinguishes CBASP from other treatment approaches. PMID:21500885
Implementing thinking aloud pair and Pólya problem solving strategies in fractions
NASA Astrophysics Data System (ADS)
Simpol, N. S. H.; Shahrill, M.; Li, H.-C.; Prahmana, R. C. I.
2017-12-01
This study implemented two pedagogical strategies, the Thinking Aloud Pair Problem Solving and Pólya’s Problem Solving, to support students’ learning of fractions. The participants were 51 students (ages 11-13) from two Year 7 classes in a government secondary school in Brunei Darussalam. A mixed method design was employed in the present study, with data collected from the pre- and post-tests, problem solving behaviour questionnaire and interviews. The study aimed to explore if there were differences in the students’ problem solving behaviour before and after the implementation of the problem solving strategies. Results from the Wilcoxon Signed Rank Test revealed a significant difference in the test results regarding student problem solving behaviour, z = -3.68, p = .000, with a higher mean score for the post-test (M = 95.5, SD = 13.8) than for the pre-test (M = 88.9, SD = 15.2). This implied that there was improvement in the students’ problem solving performance from the pre-test to the post-test. Results from the questionnaire showed that more than half of the students increased scores in all four stages of the Pólya’s problem solving strategy, which provided further evidence of the students’ improvement in problem solving.
The nonequilibrium quantum many-body problem as a paradigm for extreme data science
NASA Astrophysics Data System (ADS)
Freericks, J. K.; Nikolić, B. K.; Frieder, O.
2014-12-01
Generating big data pervades much of physics. But some problems, which we call extreme data problems, are too large to be treated within big data science. The nonequilibrium quantum many-body problem on a lattice is just such a problem, where the Hilbert space grows exponentially with system size and rapidly becomes too large to fit on any computer (and can be effectively thought of as an infinite-sized data set). Nevertheless, much progress has been made with computational methods on this problem, which serve as a paradigm for how one can approach and attack extreme data problems. In addition, viewing these physics problems from a computer-science perspective leads to new approaches that can be tried to solve more accurately and for longer times. We review a number of these different ideas here.
NASA Astrophysics Data System (ADS)
Liu, Lu; Tong, Yibin; Zhao, Zhigang; Zhang, Xuefen
2018-03-01
Large-scale access of distributed residential photovoltaic (PV) in rural areas has solved the voltage problem to a certain extent. However, due to the intermittency of PV and the particularity of rural residents’ power load, the problem of low voltage in the evening peak remains to be resolved. This paper proposes to solve the problem by accessing residential energy storage. Firstly, the influence of access location and capacity of energy storage on voltage distribution in rural distribution network is analyzed. Secondly, the relation between the storage capacity and load capacity is deduced for four typical load and energy storage cases when the voltage deviation meets the demand. Finally, the optimal storage position and capacity are obtained by using PSO and power flow simulation.
Indirect addressing and load balancing for faster solution to Mandelbrot Set on SIMD architectures
NASA Technical Reports Server (NTRS)
Tomboulian, Sherryl
1989-01-01
SIMD computers with local indirect addressing allow programs to have queues and buffers, making certain kinds of problems much more efficient. Examined here are a class of problems characterized by computations on data points where the computation is identical, but the convergence rate is data dependent. Normally, in this situation, the algorithm time is governed by the maximum number of iterations required by each point. Using indirect addressing allows a processor to proceed to the next data point when it is done, reducing the overall number of iterations required to approach the mean convergence rate when a sufficiently large problem set is solved. Load balancing techniques can be applied for additional performance improvement. Simulations of this technique applied to solving Mandelbrot Sets indicate significant performance gains.
NASA Astrophysics Data System (ADS)
Negrello, Camille; Gosselet, Pierre; Rey, Christian
2018-05-01
An efficient method for solving large nonlinear problems combines Newton solvers and Domain Decomposition Methods (DDM). In the DDM framework, the boundary conditions can be chosen to be primal, dual or mixed. The mixed approach presents the advantage to be eligible for the research of an optimal interface parameter (often called impedance) which can increase the convergence rate. The optimal value for this parameter is often too expensive to be computed exactly in practice: an approximate version has to be sought for, along with a compromise between efficiency and computational cost. In the context of parallel algorithms for solving nonlinear structural mechanical problems, we propose a new heuristic for the impedance which combines short and long range effects at a low computational cost.
Freiberger, Manuel; Egger, Herbert; Liebmann, Manfred; Scharfetter, Hermann
2011-11-01
Image reconstruction in fluorescence optical tomography is a three-dimensional nonlinear ill-posed problem governed by a system of partial differential equations. In this paper we demonstrate that a combination of state of the art numerical algorithms and a careful hardware optimized implementation allows to solve this large-scale inverse problem in a few seconds on standard desktop PCs with modern graphics hardware. In particular, we present methods to solve not only the forward but also the non-linear inverse problem by massively parallel programming on graphics processors. A comparison of optimized CPU and GPU implementations shows that the reconstruction can be accelerated by factors of about 15 through the use of the graphics hardware without compromising the accuracy in the reconstructed images.
High-frequency CAD-based scattering model: SERMAT
NASA Astrophysics Data System (ADS)
Goupil, D.; Boutillier, M.
1991-09-01
Specifications for an industrial radar cross section (RCS) calculation code are given: it must be able to exchange data with many computer aided design (CAD) systems, it must be fast, and it must have powerful graphic tools. Classical physical optics (PO) and equivalent currents (EC) techniques have proven their efficiency on simple objects for a long time. Difficult geometric problems occur when objects with very complex shapes have to be computed. Only a specific geometric code can solve these problems. We have established that, once these problems have been solved: (1) PO and EC give good results on complex objects of large size compared to wavelength; and (2) the implementation of these objects in a software package (SERMAT) allows fast and sufficiently precise domain RCS calculations to meet industry requirements in the domain of stealth.
Mesoscale modeling: solving complex flows in biology and biotechnology.
Mills, Zachary Grant; Mao, Wenbin; Alexeev, Alexander
2013-07-01
Fluids are involved in practically all physiological activities of living organisms. However, biological and biorelated flows are hard to analyze due to the inherent combination of interdependent effects and processes that occur on a multitude of spatial and temporal scales. Recent advances in mesoscale simulations enable researchers to tackle problems that are central for the understanding of such flows. Furthermore, computational modeling effectively facilitates the development of novel therapeutic approaches. Among other methods, dissipative particle dynamics and the lattice Boltzmann method have become increasingly popular during recent years due to their ability to solve a large variety of problems. In this review, we discuss recent applications of these mesoscale methods to several fluid-related problems in medicine, bioengineering, and biotechnology. Copyright © 2013 Elsevier Ltd. All rights reserved.
The NASA/Baltimore Applications Project: An experiment in technology transfer
NASA Technical Reports Server (NTRS)
Golden, T. S.
1981-01-01
Conclusions drawn from the experiment thus far are presented. The problems of a large city most often do not require highly sophisticated solutions; the simpler the solution, the better. A problem focused approach is a greater help to the city than a product focused approach. Most problem situations involve several individuals or organized groups within the city. Mutual trust and good interpersonal relationships between the technologist and the administrator is as important for solving problems as technological know-how.
Modeling and stress analysis of large format InSb focal plane arrays detector under thermal shock
NASA Astrophysics Data System (ADS)
Zhang, Li-Wen; Meng, Qing-Duan; Zhang, Xiao-Ling; Yu, Qian; Lv, Yan-Qiu; Si, Jun-Jie
2013-09-01
Higher fracture probability, appearing in large format InSb infrared focal plane arrays detector under thermal shock loadings, limits its applicability and suitability for large format equipment, and has been an urgent problem to be solved. In order to understand the fracture mechanism and improve the reliability, three dimensional modeling and stress analysis of large format InSb detector is necessary. However, there are few reports on three dimensional modeling and simulation of large format InSb detector, due to huge meshing numbers and time-consuming operation to solve. To solve the problems, basing on the thermal mismatch displacement formula, an equivalent modeling method is proposed in this paper. With the proposed equivalent modeling method, employing the ANSYS software, three dimensional large format InSb detector is modeled, and the maximum Von Mises stress appearing in InSb chip dependent on array format is researched. According to the maximum Von Mises stress location shift and stress increasing tendency, the adaptability range of the proposed equivalent method is also derived, that is, for 16 × 16, 32 × 32 and 64 × 64 format, its adaptability ranges are not larger than 64 × 64, 256 × 256 and 1024 × 1024 format, respectively. Taking 1024 × 1024 InSb detector as an example, the Von Mises stress distribution appearing in InSb chip, Si readout integrated circuits and indium bump arrays are described, and the causes are discussed in detail. All these will provide a feasible research plan to identify the fracture origins of InSb chip and reduce fracture probability for large format InSb detector.
Jiang, Weili; Shang, Siyuan; Su, Yanjie
2015-01-01
People may experience an “aha” moment, when suddenly realizing a solution of a puzzling problem. This experience is called insight problem solving. Several findings suggest that catecholamine-related genes may contribute to insight problem solving, among which the catechol-O-methyltransferase (COMT) gene is the most promising candidate. The current study examined 753 healthy individuals to determine the associations between 7 candidate single nucleotide polymorphisms on the COMT gene and insight problem-solving performance, while considering gender differences. The results showed that individuals carrying A allele of rs4680 or T allele of rs4633 scored significantly higher on insight problem-solving tasks, and the COMT gene rs5993883 combined with gender interacted with correct solutions of insight problems, specifically showing that this gene only influenced insight problem-solving performance in males. This study presents the first investigation of the genetic impact on insight problem solving and provides evidence that highlights the role that the COMT gene plays in insight problem solving. PMID:26528222
Jiang, Weili; Shang, Siyuan; Su, Yanjie
2015-01-01
People may experience an "aha" moment, when suddenly realizing a solution of a puzzling problem. This experience is called insight problem solving. Several findings suggest that catecholamine-related genes may contribute to insight problem solving, among which the catechol-O-methyltransferase (COMT) gene is the most promising candidate. The current study examined 753 healthy individuals to determine the associations between 7 candidate single nucleotide polymorphisms on the COMT gene and insight problem-solving performance, while considering gender differences. The results showed that individuals carrying A allele of rs4680 or T allele of rs4633 scored significantly higher on insight problem-solving tasks, and the COMT gene rs5993883 combined with gender interacted with correct solutions of insight problems, specifically showing that this gene only influenced insight problem-solving performance in males. This study presents the first investigation of the genetic impact on insight problem solving and provides evidence that highlights the role that the COMT gene plays in insight problem solving.
Understanding Undergraduates’ Problem-Solving Processes †
Nehm, Ross H.
2010-01-01
Fostering effective problem-solving skills is one of the most longstanding and widely agreed upon goals of biology education. Nevertheless, undergraduate biology educators have yet to leverage many major findings about problem-solving processes from the educational and cognitive science research literatures. This article highlights key facets of problem-solving processes and introduces methodologies that may be used to reveal how undergraduate students perceive and represent biological problems. Overall, successful problem-solving entails a keen sensitivity to problem contexts, disciplined internal representation or modeling of the problem, and the principled management and deployment of cognitive resources. Context recognition tasks, problem representation practice, and cognitive resource management receive remarkably little emphasis in the biology curriculum, despite their central roles in problem-solving success. PMID:23653710
Thinking Process of Naive Problem Solvers to Solve Mathematical Problems
ERIC Educational Resources Information Center
Mairing, Jackson Pasini
2017-01-01
Solving problems is not only a goal of mathematical learning. Students acquire ways of thinking, habits of persistence and curiosity, and confidence in unfamiliar situations by learning to solve problems. In fact, there were students who had difficulty in solving problems. The students were naive problem solvers. This research aimed to describe…
Teaching Problem Solving without Modeling through "Thinking Aloud Pair Problem Solving."
ERIC Educational Resources Information Center
Pestel, Beverly C.
1993-01-01
Reviews research relevant to the problem of unsatisfactory student problem-solving abilities and suggests a teaching strategy that addresses the issue. Author explains how she uses teaching aloud problem solving (TAPS) in college chemistry and presents evaluation data. Among the findings are that the TAPS class got fewer problems completely right,…
Social Problem Solving, Conduct Problems, and Callous-Unemotional Traits in Children
ERIC Educational Resources Information Center
Waschbusch, Daniel A.; Walsh, Trudi M.; Andrade, Brendan F.; King, Sara; Carrey, Normand J.
2007-01-01
This study examined the association between social problem solving, conduct problems (CP), and callous-unemotional (CU) traits in elementary age children. Participants were 53 children (40 boys and 13 girls) aged 7-12 years. Social problem solving was evaluated using the Social Problem Solving Test-Revised, which requires children to produce…
NASA Technical Reports Server (NTRS)
Motil, Susan
2003-01-01
People within large organizations have probably already dealt with problems similar to the problems that you face; you can save time and money by taking advantage of that experience and knowledge. Knowledge sharing by mentors can empower less experienced managers who would otherwise not challenge the status quo. Reviews should encourage joint problem solving rather than just reporting. To accomplish this, ensure that the review process is viewed as feedback from independent and supportive experts.
Personality, problem solving, and adolescent substance use.
Jaffee, William B; D'Zurilla, Thomas J
2009-03-01
The major aim of this study was to examine the role of social problem solving in the relationship between personality and substance use in adolescents. Although a number of studies have identified a relationship between personality and substance use, the precise mechanism by which this occurs is not clear. We hypothesized that problem-solving skills could be one such mechanism. More specifically, we sought to determine whether problem solving mediates, moderates, or both mediates and moderates the relationship between different personality traits and substance use. Three hundred and seven adolescents were administered the Substance Use Profile Scale, the Social Problem-Solving Inventory-Revised, and the Personality Experiences Inventory to assess personality, social problem-solving ability, and substance use, respectively. Results showed that the dimension of rational problem solving (i.e., effective problem-solving skills) significantly mediated the relationship between hopelessness and lifetime alcohol and marijuana use. The theoretical and clinical implications of these results were discussed.
Enhancing chemistry problem-solving achievement using problem categorization
NASA Astrophysics Data System (ADS)
Bunce, Diane M.; Gabel, Dorothy L.; Samuel, John V.
The enhancement of chemistry students' skill in problem solving through problem categorization is the focus of this study. Twenty-four students in a freshman chemistry course for health professionals are taught how to solve problems using the explicit method of problem solving (EMPS) (Bunce & Heikkinen, 1986). The EMPS is an organized approach to problem analysis which includes encoding the information given in a problem (Given, Asked For), relating this to what is already in long-term memory (Recall), and planning a solution (Overall Plan) before a mathematical solution is attempted. In addition to the EMPS training, treatment students receive three 40-minute sessions following achievement tests in which they are taught how to categorize problems. Control students use this time to review the EMPS solutions of test questions. Although problem categorization is involved in one section of the EMPS (Recall), treatment students who received specific training in problem categorization demonstrate significantly higher achievement on combination problems (those problems requiring the use of more than one chemical topic for their solution) at (p = 0.01) than their counterparts. Significantly higher achievement for treatment students is also measured on an unannounced test (p = 0.02). Analysis of interview transcripts of both treatment and control students illustrates a Rolodex approach to problem solving employed by all students in this study. The Rolodex approach involves organizing equations used to solve problems on mental index cards and flipping through them, matching units given when a new problem is to be solved. A second phenomenon observed during student interviews is the absence of a link in the conceptual understanding of the chemical concepts involved in a problem and the problem-solving skills employed to correctly solve problems. This study shows that explicit training in categorization skills and the EMPS can lead to higher achievement in complex problem-solving situations (combination problems and unannounced test). However, such achievement may be limited by the lack of linkages between students' conceptual understanding and improved problem-solving skill.
Decision-Making and Problem-Solving Approaches in Pharmacy Education
Martin, Lindsay C.; Holdford, David A.
2016-01-01
Domain 3 of the Center for the Advancement of Pharmacy Education (CAPE) 2013 Educational Outcomes recommends that pharmacy school curricula prepare students to be better problem solvers, but are silent on the type of problems they should be prepared to solve. We identified five basic approaches to problem solving in the curriculum at a pharmacy school: clinical, ethical, managerial, economic, and legal. These approaches were compared to determine a generic process that could be applied to all pharmacy decisions. Although there were similarities in the approaches, generic problem solving processes may not work for all problems. Successful problem solving requires identification of the problems faced and application of the right approach to the situation. We also advocate that the CAPE Outcomes make explicit the importance of different approaches to problem solving. Future pharmacists will need multiple approaches to problem solving to adapt to the complexity of health care. PMID:27170823
Decision-Making and Problem-Solving Approaches in Pharmacy Education.
Martin, Lindsay C; Donohoe, Krista L; Holdford, David A
2016-04-25
Domain 3 of the Center for the Advancement of Pharmacy Education (CAPE) 2013 Educational Outcomes recommends that pharmacy school curricula prepare students to be better problem solvers, but are silent on the type of problems they should be prepared to solve. We identified five basic approaches to problem solving in the curriculum at a pharmacy school: clinical, ethical, managerial, economic, and legal. These approaches were compared to determine a generic process that could be applied to all pharmacy decisions. Although there were similarities in the approaches, generic problem solving processes may not work for all problems. Successful problem solving requires identification of the problems faced and application of the right approach to the situation. We also advocate that the CAPE Outcomes make explicit the importance of different approaches to problem solving. Future pharmacists will need multiple approaches to problem solving to adapt to the complexity of health care.
Social problem-solving in Chinese baccalaureate nursing students.
Fang, Jinbo; Luo, Ying; Li, Yanhua; Huang, Wenxia
2016-11-01
To describe social problem solving in Chinese baccalaureate nursing students. A descriptive cross-sectional study was conducted with a cluster sample of 681 Chinese baccalaureate nursing students. The Chinese version of the Social Problem-Solving scale was used. Descriptive analyses, independent t-test and one-way analysis of variance were applied to analyze the data. The final year nursing students presented the highest scores of positive social problem-solving skills. Students with experiences of self-directed and problem-based learning presented significantly higher scores in Positive Problem Orientation subscale. The group with Critical thinking training experience, however, displayed higher negative problem solving scores compared with nonexperience group. Social problem solving abilities varied based upon teaching-learning strategies. Self-directed and problem-based learning may be recommended as effective way to improve social problem-solving ability. © 2016 Chinese Cochrane Center, West China Hospital of Sichuan University and John Wiley & Sons Australia, Ltd.
Problem Solving and Chemical Equilibrium: Successful versus Unsuccessful Performance.
ERIC Educational Resources Information Center
Camacho, Moises; Good, Ron
1989-01-01
Describes the problem-solving behaviors of experts and novices engaged in solving seven chemical equilibrium problems. Lists 27 behavioral tendencies of successful and unsuccessful problem solvers. Discusses several implications for a problem solving theory, think-aloud techniques, adequacy of the chemistry domain, and chemistry instruction.…
Worry and problem-solving skills and beliefs in primary school children.
Parkinson, Monika; Creswell, Cathy
2011-03-01
To examine the association between worry and problem-solving skills and beliefs (confidence and perceived control) in primary school children. Children (8-11 years) were screened using the Penn State Worry Questionnaire for Children. High (N= 27) and low (N= 30) scorers completed measures of anxiety, problem-solving skills (generating alternative solutions to problems, planfulness, and effectiveness of solutions) and problem-solving beliefs (confidence and perceived control). High and low worry groups differed significantly on measures of anxiety and problem-solving beliefs (confidence and control) but not on problem-solving skills. Consistent with findings with adults, worry in children was associated with cognitive distortions, not skills deficits. Interventions for worried children may benefit from a focus on increasing positive problem-solving beliefs. ©2010 The British Psychological Society.
Numerical optimization in Hilbert space using inexact function and gradient evaluations
NASA Technical Reports Server (NTRS)
Carter, Richard G.
1989-01-01
Trust region algorithms provide a robust iterative technique for solving non-convex unstrained optimization problems, but in many instances it is prohibitively expensive to compute high accuracy function and gradient values for the method. Of particular interest are inverse and parameter estimation problems, since function and gradient evaluations involve numerically solving large systems of differential equations. A global convergence theory is presented for trust region algorithms in which neither function nor gradient values are known exactly. The theory is formulated in a Hilbert space setting so that it can be applied to variational problems as well as the finite dimensional problems normally seen in trust region literature. The conditions concerning allowable error are remarkably relaxed: relative errors in the gradient error condition is automatically satisfied if the error is orthogonal to the gradient approximation. A technique for estimating gradient error and improving the approximation is also presented.
A machine learning approach for efficient uncertainty quantification using multiscale methods
NASA Astrophysics Data System (ADS)
Chan, Shing; Elsheikh, Ahmed H.
2018-02-01
Several multiscale methods account for sub-grid scale features using coarse scale basis functions. For example, in the Multiscale Finite Volume method the coarse scale basis functions are obtained by solving a set of local problems over dual-grid cells. We introduce a data-driven approach for the estimation of these coarse scale basis functions. Specifically, we employ a neural network predictor fitted using a set of solution samples from which it learns to generate subsequent basis functions at a lower computational cost than solving the local problems. The computational advantage of this approach is realized for uncertainty quantification tasks where a large number of realizations has to be evaluated. We attribute the ability to learn these basis functions to the modularity of the local problems and the redundancy of the permeability patches between samples. The proposed method is evaluated on elliptic problems yielding very promising results.
NASA Astrophysics Data System (ADS)
Mushlihuddin, R.; Nurafifah; Irvan
2018-01-01
The student’s low ability in mathematics problem solving proved to the less effective of a learning process in the classroom. Effective learning was a learning that affects student’s math skills, one of which is problem-solving abilities. Problem-solving capability consisted of several stages: understanding the problem, planning the settlement, solving the problem as planned, re-examining the procedure and the outcome. The purpose of this research was to know: (1) was there any influence of PBL model in improving ability Problem solving of student math in a subject of vector analysis?; (2) was the PBL model effective in improving students’ mathematical problem-solving skills in vector analysis courses? This research was a quasi-experiment research. The data analysis techniques performed from the test stages of data description, a prerequisite test is the normality test, and hypothesis test using the ANCOVA test and Gain test. The results showed that: (1) there was an influence of PBL model in improving students’ math problem-solving abilities in vector analysis courses; (2) the PBL model was effective in improving students’ problem-solving skills in vector analysis courses with a medium category.
2014-05-22
Situational problem solving at best, delay and indecision at worst, seem more prevalent than reflective thinking, considered debate, and perseverance.2...international problems whose solution will largely determine our future . . . . The war years were critical . . . but I think the present period is, in...Undersecretary of State Dean Acheson travelled to Mississippi on behalf of the President to deliver “a major foreign policy speech” that
ERIC Educational Resources Information Center
Dufner, Hillrey A.; Alexander, Patricia A.
The differential effects of two different types of problem-solving training on the problem-solving abilities of gifted fourth graders were studied. Two successive classes of gifted fourth graders from Weslaco Independent School District (Texas) were pretested with the Coloured Progressive Matrices (CPM) and Thinking Creatively With Pictures…
Social problem-solving among adolescents treated for depression.
Becker-Weidman, Emily G; Jacobs, Rachel H; Reinecke, Mark A; Silva, Susan G; March, John S
2010-01-01
Studies suggest that deficits in social problem-solving may be associated with increased risk of depression and suicidality in children and adolescents. It is unclear, however, which specific dimensions of social problem-solving are related to depression and suicidality among youth. Moreover, rational problem-solving strategies and problem-solving motivation may moderate or predict change in depression and suicidality among children and adolescents receiving treatment. The effect of social problem-solving on acute treatment outcomes were explored in a randomized controlled trial of 439 clinically depressed adolescents enrolled in the Treatment for Adolescents with Depression Study (TADS). Measures included the Children's Depression Rating Scale-Revised (CDRS-R), the Suicidal Ideation Questionnaire--Grades 7-9 (SIQ-Jr), and the Social Problem-Solving Inventory-Revised (SPSI-R). A random coefficients regression model was conducted to examine main and interaction effects of treatment and SPSI-R subscale scores on outcomes during the 12-week acute treatment stage. Negative problem orientation, positive problem orientation, and avoidant problem-solving style were non-specific predictors of depression severity. In terms of suicidality, avoidant problem-solving style and impulsiveness/carelessness style were predictors, whereas negative problem orientation and positive problem orientation were moderators of treatment outcome. Implications of these findings, limitations, and directions for future research are discussed. Copyright 2009 Elsevier Ltd. All rights reserved.
Step by Step: Biology Undergraduates’ Problem-Solving Procedures during Multiple-Choice Assessment
Prevost, Luanna B.; Lemons, Paula P.
2016-01-01
This study uses the theoretical framework of domain-specific problem solving to explore the procedures students use to solve multiple-choice problems about biology concepts. We designed several multiple-choice problems and administered them on four exams. We trained students to produce written descriptions of how they solved the problem, and this allowed us to systematically investigate their problem-solving procedures. We identified a range of procedures and organized them as domain general, domain specific, or hybrid. We also identified domain-general and domain-specific errors made by students during problem solving. We found that students use domain-general and hybrid procedures more frequently when solving lower-order problems than higher-order problems, while they use domain-specific procedures more frequently when solving higher-order problems. Additionally, the more domain-specific procedures students used, the higher the likelihood that they would answer the problem correctly, up to five procedures. However, if students used just one domain-general procedure, they were as likely to answer the problem correctly as if they had used two to five domain-general procedures. Our findings provide a categorization scheme and framework for additional research on biology problem solving and suggest several important implications for researchers and instructors. PMID:27909021
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aguilo Valentin, Miguel Alejandro
2016-07-01
This study presents a new nonlinear programming formulation for the solution of inverse problems. First, a general inverse problem formulation based on the compliance error functional is presented. The proposed error functional enables the computation of the Lagrange multipliers, and thus the first order derivative information, at the expense of just one model evaluation. Therefore, the calculation of the Lagrange multipliers does not require the solution of the computationally intensive adjoint problem. This leads to significant speedups for large-scale, gradient-based inverse problems.
Disciplinary Foundations for Solving Interdisciplinary Scientific Problems
ERIC Educational Resources Information Center
Zhang, Dongmei; Shen, Ji
2015-01-01
Problem-solving has been one of the major strands in science education research. But much of the problem-solving research has been conducted on discipline-based contexts; little research has been done on how students, especially individuals, solve interdisciplinary problems. To understand how individuals reason about interdisciplinary problems, we…
Engineering students' experiences and perceptions of workplace problem solving
NASA Astrophysics Data System (ADS)
Pan, Rui
In this study, I interviewed 22 engineering Co-Op students about their workplace problem solving experiences and reflections and explored: 1) Of Co-Op students who experienced workplace problem solving, what are the different ways in which students experience workplace problem solving? 2) How do students perceive a) the differences between workplace problem solving and classroom problem solving and b) in what areas are they prepared by their college education to solve workplace problems? To answer my first research question, I analyzed data through the lens of phenomenography and I conducted thematic analysis to answer my second research question. The results of this study have implications for engineering education and engineering practice. Specifically, the results reveal the different ways students experience workplace problem solving, which provide engineering educators and practicing engineers a better understanding of the nature of workplace engineering. In addition, the results indicate that there is still a gap between classroom engineering and workplace engineering. For engineering educators who aspire to prepare students to be future engineers, it is imperative to design problem solving experiences that can better prepare students with workplace competency.
Problem-Solving Deficits in Iranian People with Borderline Personality Disorder
Akbari Dehaghi, Ashraf; Kaviani, Hossein; Tamanaeefar, Shima
2014-01-01
Objective: Interventions for people suffering from borderline personality disorder (BPD), such as dialectical behavior therapy, often include a problem-solving component. However, there is an absence of published studies examining the problem-solving abilities of this client group in Iran. The study compared inpatients and outpatients with BPD and a control group on problem-solving capabilities in an Iranian sample. It was hypothesized that patients with BPD would have more deficiencies in this area. Methods: Fifteen patients with BPD were compared to 15 healthy participants. Means-ends problem-solving task (MEPS) was used to measure problem-solving skills in both groups. Results: BPD group reported less effective strategies in solving problems as opposed to the healthy group. Compared to the control group, participants with BPD provided empirical support for the use of problem-solving interventions with people suffering from BPD. Conclusions: The findings supported the idea that a problem-solving intervention can be efficiently applied either as a stand-alone therapy or in conjunction with other available psychotherapies to treat people with BPD. PMID:25798169
Enhancing memory and imagination improves problem solving among individuals with depression.
McFarland, Craig P; Primosch, Mark; Maxson, Chelsey M; Stewart, Brandon T
2017-08-01
Recent work has revealed links between memory, imagination, and problem solving, and suggests that increasing access to detailed memories can lead to improved imagination and problem-solving performance. Depression is often associated with overgeneral memory and imagination, along with problem-solving deficits. In this study, we tested the hypothesis that an interview designed to elicit detailed recollections would enhance imagination and problem solving among both depressed and nondepressed participants. In a within-subjects design, participants completed a control interview or an episodic specificity induction prior to completing memory, imagination, and problem-solving tasks. Results revealed that compared to the control interview, the episodic specificity induction fostered increased detail generation in memory and imagination and more relevant steps on the problem-solving task among depressed and nondepressed participants. This study builds on previous work by demonstrating that a brief interview can enhance problem solving among individuals with depression and supports the notion that episodic memory plays a key role in problem solving. It should be noted, however, that the results of the interview are relatively short-lived.
Micro-bias and macro-performance.
Seaver, S M D; Moreira, A A; Sales-Pardo, M; Malmgren, R D; Diermeier, D; Amaral, L A N
2009-02-01
We use agent-based modeling to investigate the effect of conservatism and partisanship on the efficiency with which large populations solve the density classification task - a paradigmatic problem for information aggregation and consensus building. We find that conservative agents enhance the populations' ability to efficiently solve the density classification task despite large levels of noise in the system. In contrast, we find that the presence of even a small fraction of partisans holding the minority position will result in deadlock or a consensus on an incorrect answer. Our results provide a possible explanation for the emergence of conservatism and suggest that even low levels of partisanship can lead to significant social costs.
Biala, T A; Jator, S N
2015-01-01
In this article, the boundary value method is applied to solve three dimensional elliptic and hyperbolic partial differential equations. The partial derivatives with respect to two of the spatial variables (y, z) are discretized using finite difference approximations to obtain a large system of ordinary differential equations (ODEs) in the third spatial variable (x). Using interpolation and collocation techniques, a continuous scheme is developed and used to obtain discrete methods which are applied via the Block unification approach to obtain approximations to the resulting large system of ODEs. Several test problems are investigated to elucidate the solution process.
Measuring Family Problem Solving: The Family Problem Solving Diary.
ERIC Educational Resources Information Center
Kieren, Dianne K.
The development and use of the family problem-solving diary are described. The diary is one of several indicators and measures of family problem-solving behavior. It provides a record of each person's perception of day-to-day family problems (what the problem concerns, what happened, who got involved, what those involved did, how the problem…
Parallel Computation of Flow in Heterogeneous Media Modelled by Mixed Finite Elements
NASA Astrophysics Data System (ADS)
Cliffe, K. A.; Graham, I. G.; Scheichl, R.; Stals, L.
2000-11-01
In this paper we describe a fast parallel method for solving highly ill-conditioned saddle-point systems arising from mixed finite element simulations of stochastic partial differential equations (PDEs) modelling flow in heterogeneous media. Each realisation of these stochastic PDEs requires the solution of the linear first-order velocity-pressure system comprising Darcy's law coupled with an incompressibility constraint. The chief difficulty is that the permeability may be highly variable, especially when the statistical model has a large variance and a small correlation length. For reasonable accuracy, the discretisation has to be extremely fine. We solve these problems by first reducing the saddle-point formulation to a symmetric positive definite (SPD) problem using a suitable basis for the space of divergence-free velocities. The reduced problem is solved using parallel conjugate gradients preconditioned with an algebraically determined additive Schwarz domain decomposition preconditioner. The result is a solver which exhibits a good degree of robustness with respect to the mesh size as well as to the variance and to physically relevant values of the correlation length of the underlying permeability field. Numerical experiments exhibit almost optimal levels of parallel efficiency. The domain decomposition solver (DOUG, http://www.maths.bath.ac.uk/~parsoft) used here not only is applicable to this problem but can be used to solve general unstructured finite element systems on a wide range of parallel architectures.
Metcalfe, Arron W S; Ashkenazi, Sarit; Rosenberg-Lee, Miriam; Menon, Vinod
2013-10-01
Baddeley and Hitch's multi-component working memory (WM) model has played an enduring and influential role in our understanding of cognitive abilities. Very little is known, however, about the neural basis of this multi-component WM model and the differential role each component plays in mediating arithmetic problem solving abilities in children. Here, we investigate the neural basis of the central executive (CE), phonological (PL) and visuo-spatial (VS) components of WM during a demanding mental arithmetic task in 7-9 year old children (N=74). The VS component was the strongest predictor of math ability in children and was associated with increased arithmetic complexity-related responses in left dorsolateral and right ventrolateral prefrontal cortices as well as bilateral intra-parietal sulcus and supramarginal gyrus in posterior parietal cortex. Critically, VS, CE and PL abilities were associated with largely distinct patterns of brain response. Overlap between VS and CE components was observed in left supramarginal gyrus and no overlap was observed between VS and PL components. Our findings point to a central role of visuo-spatial WM during arithmetic problem-solving in young grade-school children and highlight the usefulness of the multi-component Baddeley and Hitch WM model in fractionating the neural correlates of arithmetic problem solving during development. Copyright © 2013 Elsevier Ltd. All rights reserved.
Ogawa, Takeshi; Aihara, Takatsugu; Shimokawa, Takeaki; Yamashita, Okito
2018-04-24
Creative insight occurs with an "Aha!" experience when solving a difficult problem. Here, we investigated large-scale networks associated with insight problem solving. We recruited 232 healthy participants aged 21-69 years old. Participants completed a magnetic resonance imaging study (MRI; structural imaging and a 10 min resting-state functional MRI) and an insight test battery (ITB) consisting of written questionnaires (matchstick arithmetic task, remote associates test, and insight problem solving task). To identify the resting-state functional connectivity (RSFC) associated with individual creative insight, we conducted an exploratory voxel-based morphometry (VBM)-constrained RSFC analysis. We identified positive correlations between ITB score and grey matter volume (GMV) in the right insula and middle cingulate cortex/precuneus, and a negative correlation between ITB score and GMV in the left cerebellum crus 1 and right supplementary motor area. We applied seed-based RSFC analysis to whole brain voxels using the seeds obtained from the VBM and identified insight-positive/negative connections, i.e. a positive/negative correlation between the ITB score and individual RSFCs between two brain regions. Insight-specific connections included motor-related regions whereas creative-common connections included a default mode network. Our results indicate that creative insight requires a coupling of multiple networks, such as the default mode, semantic and cerebral-cerebellum networks.
Trumpower, David L; Goldsmith, Timothy E; Guynn, Melissa J
2004-12-01
Solving training problems with nonspecific goals (NG; i.e., solving for all possible unknown values) often results in better transfer than solving training problems with standard goals (SG; i.e., solving for one particular unknown value). In this study, we evaluated an attentional focus explanation of the goal specificity effect. According to the attentional focus view, solving NG problems causes attention to be directed to local relations among successive problem states, whereas solving SG problems causes attention to be directed to relations between the various problem states and the goal state. Attention to the former is thought to enhance structural knowledge about the problem domain and thus promote transfer. Results supported this view because structurally different transfer problems were solved faster following NG training than following SG training. Moreover, structural knowledge representations revealed more links depicting local relations following NG training and more links to the training goal following SG training. As predicted, these effects were obtained only by domain novices.
Problem-Solving After Traumatic Brain Injury in Adolescence: Associations With Functional Outcomes
Wade, Shari L.; Cassedy, Amy E.; Fulks, Lauren E.; Taylor, H. Gerry; Stancin, Terry; Kirkwood, Michael W.; Yeates, Keith O.; Kurowski, Brad G.
2017-01-01
Objective To examine the association of problem-solving with functioning in youth with traumatic brain injury (TBI). Design Cross-sectional evaluation of pretreatment data from a randomized controlled trial. Setting Four children’s hospitals and 1 general hospital, with level 1 trauma units. Participants Youth, ages 11 to 18 years, who sustained moderate or severe TBI in the last 18 months (N=153). Main Outcome Measures Problem-solving skills were assessed using the Social Problem-Solving Inventory (SPSI) and the Dodge Social Information Processing Short Stories. Everyday functioning was assessed based on a structured clinical interview using the Child and Adolescent Functional Assessment Scale (CAFAS) and via adolescent ratings on the Youth Self Report (YSR). Correlations and multiple regression analyses were used to examine associations among measures. Results The TBI group endorsed lower levels of maladaptive problem-solving (negative problem orientation, careless/impulsive responding, and avoidant style) and lower levels of rational problem-solving, resulting in higher total problem-solving scores for the TBI group compared with a normative sample (P<.001). Dodge Social Information Processing Short Stories dimensions were correlated (r=.23–.37) with SPSI subscales in the anticipated direction. Although both maladaptive (P<.001) and adaptive (P=.006) problem-solving composites were associated with overall functioning on the CAFAS, only maladaptive problem-solving (P<.001) was related to the YSR total when outcomes were continuous. For the both CAFAS and YSR logistic models, maladaptive style was significantly associated with greater risk of impairment (P=.001). Conclusions Problem-solving after TBI differs from normative samples and is associated with functional impairments. The relation of problem-solving deficits after TBI with global functioning merits further investigation, with consideration of the potential effects of problem-solving interventions on functional outcomes. PMID:28389109
Problem-Solving After Traumatic Brain Injury in Adolescence: Associations With Functional Outcomes.
Wade, Shari L; Cassedy, Amy E; Fulks, Lauren E; Taylor, H Gerry; Stancin, Terry; Kirkwood, Michael W; Yeates, Keith O; Kurowski, Brad G
2017-08-01
To examine the association of problem-solving with functioning in youth with traumatic brain injury (TBI). Cross-sectional evaluation of pretreatment data from a randomized controlled trial. Four children's hospitals and 1 general hospital, with level 1 trauma units. Youth, ages 11 to 18 years, who sustained moderate or severe TBI in the last 18 months (N=153). Problem-solving skills were assessed using the Social Problem-Solving Inventory (SPSI) and the Dodge Social Information Processing Short Stories. Everyday functioning was assessed based on a structured clinical interview using the Child and Adolescent Functional Assessment Scale (CAFAS) and via adolescent ratings on the Youth Self Report (YSR). Correlations and multiple regression analyses were used to examine associations among measures. The TBI group endorsed lower levels of maladaptive problem-solving (negative problem orientation, careless/impulsive responding, and avoidant style) and lower levels of rational problem-solving, resulting in higher total problem-solving scores for the TBI group compared with a normative sample (P<.001). Dodge Social Information Processing Short Stories dimensions were correlated (r=.23-.37) with SPSI subscales in the anticipated direction. Although both maladaptive (P<.001) and adaptive (P=.006) problem-solving composites were associated with overall functioning on the CAFAS, only maladaptive problem-solving (P<.001) was related to the YSR total when outcomes were continuous. For the both CAFAS and YSR logistic models, maladaptive style was significantly associated with greater risk of impairment (P=.001). Problem-solving after TBI differs from normative samples and is associated with functional impairments. The relation of problem-solving deficits after TBI with global functioning merits further investigation, with consideration of the potential effects of problem-solving interventions on functional outcomes. Copyright © 2017 American Congress of Rehabilitation Medicine. Published by Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Englander, Arnold C.; Englander, Jacob A.
2017-01-01
Interplanetary trajectory optimization problems are highly complex and are characterized by a large number of decision variables and equality and inequality constraints as well as many locally optimal solutions. Stochastic global search techniques, coupled with a large-scale NLP solver, have been shown to solve such problems but are inadequately robust when the problem constraints become very complex. In this work, we present a novel search algorithm that takes advantage of the fact that equality constraints effectively collapse the solution space to lower dimensionality. This new approach walks the filament'' of feasibility to efficiently find the global optimal solution.
ERIC Educational Resources Information Center
Zhang, Yin; Chu, Samuel K. W.
2016-01-01
In recent years, a number of models concerning problem solving systems have been put forward. However, many of them stress on technology and neglect the research of problem solving itself, especially the learning mechanism related to problem solving. In this paper, we analyze the learning mechanism of problem solving, and propose that when…
Perceived problem solving, stress, and health among college students.
Largo-Wight, Erin; Peterson, P Michael; Chen, W William
2005-01-01
To study the relationships among perceived problem solving, stress, and physical health. The Perceived Stress Questionnaire (PSQ), Personal Problem solving Inventory (PSI), and a stress-related physical health symptoms checklist were used to measure perceived stress, problem solving, and health among undergraduate college students (N = 232). Perceived problem-solving ability predicted self-reported physical health symptoms (R2 = .12; P < .001) and perceived stress (R2 = .19; P < .001). Perceived problem solving was a stronger predictor of physical health and perceived stress than were physical activity, alcohol consumption, or social support. Implications for college health promotion are discussed.
NASA Astrophysics Data System (ADS)
Hull, Michael M.; Kuo, Eric; Gupta, Ayush; Elby, Andrew
2013-06-01
Much research in engineering and physics education has focused on improving students’ problem-solving skills. This research has led to the development of step-by-step problem-solving strategies and grading rubrics to assess a student’s expertise in solving problems using these strategies. These rubrics value “communication” between the student’s qualitative description of the physical situation and the student’s formal mathematical descriptions (usually equations) at two points: when initially setting up the equations, and when evaluating the final mathematical answer for meaning and plausibility. We argue that (i) neither the rubrics nor the associated problem-solving strategies explicitly value this kind of communication during mathematical manipulations of the chosen equations, and (ii) such communication is an aspect of problem-solving expertise. To make this argument, we present a case study of two students, Alex and Pat, solving the same kinematics problem in clinical interviews. We argue that Pat’s solution, which connects manipulation of equations to their physical interpretation, is more expertlike than Alex’s solution, which uses equations more algorithmically. We then show that the types of problem-solving rubrics currently available do not discriminate between these two types of solutions. We conclude that problem-solving rubrics should be revised or repurposed to more accurately assess problem-solving expertise.
Solution of matrix equations using sparse techniques
NASA Technical Reports Server (NTRS)
Baddourah, Majdi
1994-01-01
The solution of large systems of matrix equations is key to the solution of a large number of scientific and engineering problems. This talk describes the sparse matrix solver developed at Langley which can routinely solve in excess of 263,000 equations in 40 seconds on one Cray C-90 processor. It appears that for large scale structural analysis applications, sparse matrix methods have a significant performance advantage over other methods.
Examining Tasks that Facilitate the Experience of Incubation While Problem-Solving
ERIC Educational Resources Information Center
Both, Lilly; Needham, Douglas; Wood, Eileen
2004-01-01
The three studies presented here contrasted the problem-solving outcomes of university students when a break was provided or not provided during a problem-solving session. In addition, two studies explored the effect of providing hints (priming) and the placement of hints during the problem-solving session. First, the ability to solve a previously…
Task-driven dictionary learning.
Mairal, Julien; Bach, Francis; Ponce, Jean
2012-04-01
Modeling data with linear combinations of a few elements from a learned dictionary has been the focus of much recent research in machine learning, neuroscience, and signal processing. For signals such as natural images that admit such sparse representations, it is now well established that these models are well suited to restoration tasks. In this context, learning the dictionary amounts to solving a large-scale matrix factorization problem, which can be done efficiently with classical optimization tools. The same approach has also been used for learning features from data for other purposes, e.g., image classification, but tuning the dictionary in a supervised way for these tasks has proven to be more difficult. In this paper, we present a general formulation for supervised dictionary learning adapted to a wide variety of tasks, and present an efficient algorithm for solving the corresponding optimization problem. Experiments on handwritten digit classification, digital art identification, nonlinear inverse image problems, and compressed sensing demonstrate that our approach is effective in large-scale settings, and is well suited to supervised and semi-supervised classification, as well as regression tasks for data that admit sparse representations.
NASA Astrophysics Data System (ADS)
Xu, Ye; Wang, Ling; Wang, Shengyao; Liu, Min
2014-09-01
In this article, an effective hybrid immune algorithm (HIA) is presented to solve the distributed permutation flow-shop scheduling problem (DPFSP). First, a decoding method is proposed to transfer a job permutation sequence to a feasible schedule considering both factory dispatching and job sequencing. Secondly, a local search with four search operators is presented based on the characteristics of the problem. Thirdly, a special crossover operator is designed for the DPFSP, and mutation and vaccination operators are also applied within the framework of the HIA to perform an immune search. The influence of parameter setting on the HIA is investigated based on the Taguchi method of design of experiment. Extensive numerical testing results based on 420 small-sized instances and 720 large-sized instances are provided. The effectiveness of the HIA is demonstrated by comparison with some existing heuristic algorithms and the variable neighbourhood descent methods. New best known solutions are obtained by the HIA for 17 out of 420 small-sized instances and 585 out of 720 large-sized instances.
NASA Astrophysics Data System (ADS)
Jua, S. K.; Sarwanto; Sukarmin
2018-05-01
Problem-solving skills are important skills in physics. However, according to some researchers, the problem-solving skill of Indonesian students’ problem in physics learning is categorized still low. The purpose of this study was to identify the profile of problem-solving skills of students who follow the across the interests program of physics. The subjects of the study were high school students of Social Sciences, grade X. The type of this research was descriptive research. The data which used to analyze the problem-solving skills were obtained through student questionnaires and the test results with impulse materials and collision. From the descriptive analysis results, the percentage of students’ problem-solving skill based on the test was 52.93% and indicators respectively. These results indicated that students’ problem-solving skill is categorized low.
NASA Astrophysics Data System (ADS)
Lin, Youzuo; O'Malley, Daniel; Vesselinov, Velimir V.
2016-09-01
Inverse modeling seeks model parameters given a set of observations. However, for practical problems because the number of measurements is often large and the model parameters are also numerous, conventional methods for inverse modeling can be computationally expensive. We have developed a new, computationally efficient parallel Levenberg-Marquardt method for solving inverse modeling problems with a highly parameterized model space. Levenberg-Marquardt methods require the solution of a linear system of equations which can be prohibitively expensive to compute for moderate to large-scale problems. Our novel method projects the original linear problem down to a Krylov subspace such that the dimensionality of the problem can be significantly reduced. Furthermore, we store the Krylov subspace computed when using the first damping parameter and recycle the subspace for the subsequent damping parameters. The efficiency of our new inverse modeling algorithm is significantly improved using these computational techniques. We apply this new inverse modeling method to invert for random transmissivity fields in 2-D and a random hydraulic conductivity field in 3-D. Our algorithm is fast enough to solve for the distributed model parameters (transmissivity) in the model domain. The algorithm is coded in Julia and implemented in the MADS computational framework (http://mads.lanl.gov). By comparing with Levenberg-Marquardt methods using standard linear inversion techniques such as QR or SVD methods, our Levenberg-Marquardt method yields a speed-up ratio on the order of ˜101 to ˜102 in a multicore computational environment. Therefore, our new inverse modeling method is a powerful tool for characterizing subsurface heterogeneity for moderate to large-scale problems.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chow, Edmond
Solving sparse problems is at the core of many DOE computational science applications. We focus on the challenge of developing sparse algorithms that can fully exploit the parallelism in extreme-scale computing systems, in particular systems with massive numbers of cores per node. Our approach is to express a sparse matrix factorization as a large number of bilinear constraint equations, and then solving these equations via an asynchronous iterative method. The unknowns in these equations are the matrix entries of the factorization that is desired.
Self-interacting inelastic dark matter: a viable solution to the small scale structure problems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Blennow, Mattias; Clementz, Stefan; Herrero-Garcia, Juan, E-mail: emb@kth.se, E-mail: scl@kth.se, E-mail: juan.herrero-garcia@adelaide.edu.au
2017-03-01
Self-interacting dark matter has been proposed as a solution to the small-scale structure problems, such as the observed flat cores in dwarf and low surface brightness galaxies. If scattering takes place through light mediators, the scattering cross section relevant to solve these problems may fall into the non-perturbative regime leading to a non-trivial velocity dependence, which allows compatibility with limits stemming from cluster-size objects. However, these models are strongly constrained by different observations, in particular from the requirements that the decay of the light mediator is sufficiently rapid (before Big Bang Nucleosynthesis) and from direct detection. A natural solution tomore » reconcile both requirements are inelastic endothermic interactions, such that scatterings in direct detection experiments are suppressed or even kinematically forbidden if the mass splitting between the two-states is sufficiently large. Using an exact solution when numerically solving the Schrödinger equation, we study such scenarios and find regions in the parameter space of dark matter and mediator masses, and the mass splitting of the states, where the small scale structure problems can be solved, the dark matter has the correct relic abundance and direct detection limits can be evaded.« less
NASA Astrophysics Data System (ADS)
Rewieński, M.; Lamecki, A.; Mrozowski, M.
2013-09-01
This paper proposes a technique, based on the Inexact Shift-Invert Lanczos (ISIL) method with Inexact Jacobi Orthogonal Component Correction (IJOCC) refinement, and a preconditioned conjugate-gradient (PCG) linear solver with multilevel preconditioner, for finding several eigenvalues for generalized symmetric eigenproblems. Several eigenvalues are found by constructing (with the ISIL process) an extended projection basis. Presented results of numerical experiments confirm the technique can be effectively applied to challenging, large-scale problems characterized by very dense spectra, such as resonant cavities with spatial dimensions which are large with respect to wavelengths of the resonating electromagnetic fields. It is also shown that the proposed scheme based on inexact linear solves delivers superior performance, as compared to methods which rely on exact linear solves, indicating tremendous potential of the 'inexact solve' concept. Finally, the scheme which generates an extended projection basis is found to provide a cost-efficient alternative to classical deflation schemes when several eigenvalues are computed.
ERIC Educational Resources Information Center
Kiliç, Çigdem
2017-01-01
This study examined pre-service primary school teachers' performance in posing problems that require knowledge of problem-solving strategies. Quantitative and qualitative methods were combined. The 120 participants were asked to pose a problem that could be solved by using the find-a-pattern a particular problem-solving strategy. After that,…
ERIC Educational Resources Information Center
Maries, Alexandru; Singh, Chandralekha
2018-01-01
Drawing appropriate diagrams is a useful problem solving heuristic that can transform a problem into a representation that is easier to exploit for solving it. One major focus while helping introductory physics students learn effective problem solving is to help them understand that drawing diagrams can facilitate problem solution. We conducted an…
ERIC Educational Resources Information Center
Sleegers, Peter; Wassink, Hartger; van Veen, Klaas; Imants, Jeroen
2009-01-01
In addition to cognitive research on school leaders' problem solving, this study focuses on the situated and personal nature of problem framing by combining insights from cognitive research on problem solving and sense-making theory. The study reports the results of a case study of two school leaders solving problems in their daily context by…
The Place of Problem Solving in Contemporary Mathematics Curriculum Documents
ERIC Educational Resources Information Center
Stacey, Kaye
2005-01-01
This paper reviews the presentation of problem solving and process aspects of mathematics in curriculum documents from Australia, UK, USA and Singapore. The place of problem solving in the documents is reviewed and contrasted, and illustrative problems from teachers' support materials are used to demonstrate how problem solving is now more often…
Translation among Symbolic Representations in Problem-Solving. Revised.
ERIC Educational Resources Information Center
Shavelson, Richard J.; And Others
This study investigated the relationships among the symbolic representation of problems given to students to solve, the mental representations they use to solve the problems, and the accuracy of their solutions. Twenty eleventh-grade science students were asked to think aloud as they solved problems on the ideal gas laws. The problems were…
Using Students' Representations Constructed during Problem Solving to Infer Conceptual Understanding
ERIC Educational Resources Information Center
Domin, Daniel; Bodner, George
2012-01-01
The differences in the types of representations constructed during successful and unsuccessful problem-solving episodes were investigated within the context of graduate students working on problems that involve concepts from 2D-NMR. Success at problem solving was established by having the participants solve five problems relating to material just…
Errors and Understanding: The Effects of Error-Management Training on Creative Problem-Solving
ERIC Educational Resources Information Center
Robledo, Issac C.; Hester, Kimberly S.; Peterson, David R.; Barrett, Jamie D.; Day, Eric A.; Hougen, Dean P.; Mumford, Michael D.
2012-01-01
People make errors in their creative problem-solving efforts. The intent of this article was to assess whether error-management training would improve performance on creative problem-solving tasks. Undergraduates were asked to solve an educational leadership problem known to call for creative thought where problem solutions were scored for…
Encouraging Sixth-Grade Students' Problem-Solving Performance by Teaching through Problem Solving
ERIC Educational Resources Information Center
Bostic, Jonathan D.; Pape, Stephen J.; Jacobbe, Tim
2016-01-01
This teaching experiment provided students with continuous engagement in a problem-solving based instructional approach during one mathematics unit. Three sections of sixth-grade mathematics were sampled from a school in Florida, U.S.A. and one section was randomly assigned to experience teaching through problem solving. Students' problem-solving…
King Oedipus and the Problem Solving Process.
ERIC Educational Resources Information Center
Borchardt, Donald A.
An analysis of the problem solving process reveals at least three options: (1) finding the cause, (2) solving the problem, and (3) anticipating potential problems. These methods may be illustrated by examining "Oedipus Tyrannus," a play in which a king attempts to deal with a problem that appears to be beyond his ability to solve, and…
Problem Solving with the Elementary Youngster.
ERIC Educational Resources Information Center
Swartz, Vicki
This paper explores research on problem solving and suggests a problem-solving approach to elementary school social studies, using a culture study of the ancient Egyptians and King Tut as a sample unit. The premise is that problem solving is particularly effective in dealing with problems which do not have one simple and correct answer but rather…
ERIC Educational Resources Information Center
Karatas, Ilhan; Baki, Adnan
2013-01-01
Problem solving is recognized as an important life skill involving a range of processes including analyzing, interpreting, reasoning, predicting, evaluating and reflecting. For that reason educating students as efficient problem solvers is an important role of mathematics education. Problem solving skill is the centre of mathematics curriculum.…
The needs analysis of learning Inventive Problem Solving for technical and vocational students
NASA Astrophysics Data System (ADS)
Sai'en, Shanty; Tze Kiong, Tee; Yunos, Jailani Md; Foong, Lee Ming; Heong, Yee Mei; Mohaffyza Mohamad, Mimi
2017-08-01
Malaysian Ministry of Education highlighted in their National Higher Education Strategic plan that higher education’s need to focus adopting 21st century skills in order to increase a graduate’s employability. Current research indicates that most graduate lack of problem solving skills to help them securing the job. Realising the important of this skill hence an alternative way suggested as an option for high institution’s student to solve their problem. This study was undertaken to measure the level of problem solving skills, identify the needs of learning inventive problem solving skills and the needs of developing an Inventive problem solving module. Using a questionnaire, the study sampled 132 students from Faculty of Technical and Vocational Education. Findings indicated that majority of the students fail to define what is an inventive problem and the root cause of a problem. They also unable to state the objectives and goal thus fail to solve the problem. As a result, the students agreed on the developing Inventive Problem Solving Module to assist them.
Holden, Richard J; Rivera-Rodriguez, A Joy; Faye, Héléne; Scanlon, Matthew C; Karsh, Ben-Tzion
2013-08-01
The most common change facing nurses today is new technology, particularly bar coded medication administration technology (BCMA). However, there is a dearth of knowledge on how BCMA alters nursing work. This study investigated how BCMA technology affected nursing work, particularly nurses' operational problem-solving behavior. Cognitive systems engineering observations and interviews were conducted after the implementation of BCMA in three nursing units of a freestanding pediatric hospital. Problem-solving behavior, associated problems, and goals, were specifically defined and extracted from observed episodes of care. Three broad themes regarding BCMA's impact on problem solving were identified. First, BCMA allowed nurses to invent new problem-solving behavior to deal with pre-existing problems. Second, BCMA made it difficult or impossible to apply some problem-solving behaviors that were commonly used pre-BCMA, often requiring nurses to use potentially risky workarounds to achieve their goals. Third, BCMA created new problems that nurses were either able to solve using familiar or novel problem-solving behaviors, or unable to solve effectively. Results from this study shed light on hidden hazards and suggest three critical design needs: (1) ecologically valid design; (2) anticipatory control; and (3) basic usability. Principled studies of the actual nature of clinicians' work, including problem solving, are necessary to uncover hidden hazards and to inform health information technology design and redesign.
Holden, Richard J.; Rivera-Rodriguez, A. Joy; Faye, Héléne; Scanlon, Matthew C.; Karsh, Ben-Tzion
2012-01-01
The most common change facing nurses today is new technology, particularly bar coded medication administration technology (BCMA). However, there is a dearth of knowledge on how BCMA alters nursing work. This study investigated how BCMA technology affected nursing work, particularly nurses’ operational problem-solving behavior. Cognitive systems engineering observations and interviews were conducted after the implementation of BCMA in three nursing units of a freestanding pediatric hospital. Problem-solving behavior, associated problems, and goals, were specifically defined and extracted from observed episodes of care. Three broad themes regarding BCMA’s impact on problem solving were identified. First, BCMA allowed nurses to invent new problem-solving behavior to deal with pre-existing problems. Second, BCMA made it difficult or impossible to apply some problem-solving behaviors that were commonly used pre-BCMA, often requiring nurses to use potentially risky workarounds to achieve their goals. Third, BCMA created new problems that nurses were either able to solve using familiar or novel problem-solving behaviors, or unable to solve effectively. Results from this study shed light on hidden hazards and suggest three critical design needs: (1) ecologically valid design; (2) anticipatory control; and (3) basic usability. Principled studies of the actual nature of clinicians’ work, including problem solving, are necessary to uncover hidden hazards and to inform health information technology design and redesign. PMID:24443642
An Extension of Holographic Moiré to Micromechanics
NASA Astrophysics Data System (ADS)
Sciammarella, C. A.; Sciammarella, F. M.
The electronic Holographic Moiré is an ideal tool for micromechanics studies. It does not require a modification of the surface by the introduction of a reference grating. This is of particular advantage when dealing with materials such as solid propellant grains whose chemical nature and surface finish makes the application of a reference grating very difficult. Traditional electronic Holographic Moiré presents some difficult problems when large magnifications are needed and large rigid body motion takes place. This paper presents developments that solves these problems and extends the application of the technique to micromechanics.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Spencer, Benjamin Whiting; Crane, Nathan K.; Heinstein, Martin W.
2011-03-01
Adagio is a Lagrangian, three-dimensional, implicit code for the analysis of solids and structures. It uses a multi-level iterative solver, which enables it to solve problems with large deformations, nonlinear material behavior, and contact. It also has a versatile library of continuum and structural elements, and an extensive library of material models. Adagio is written for parallel computing environments, and its solvers allow for scalable solutions of very large problems. Adagio uses the SIERRA Framework, which allows for coupling with other SIERRA mechanics codes. This document describes the functionality and input structure for Adagio.
The two-dimensional Stefan problem with slightly varying heat flux
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gammon, J.; Howarth, J.A.
1995-09-01
The authors solve the two-dimensional stefan problem of solidification in a half-space, where the heat flux at the wall is a slightly varying function of positioning along the wall, by means of a large Stefan number approximation (which turns out to be equivalent to a small time solution), and then by means of the Heat Balance Integral Method, which is valid for all time, and which agrees with the large Stefan number solution for small times. A representative solution is given for a particular form of the heat flux perturbation.
Bayindir Çevik, Ayfer; Olgun, Nermin
2015-04-01
This study aimed to determine the relationship between problem-solving and nursing process application skills of nursing. This is a longitudinal and correlational study. The sample included 71 students. An information form, Problem-Solving Inventory, and nursing processes the students presented at the end of clinical courses were used for data collection. Although there was no significant relationship between problem-solving skills and nursing process grades, improving problem-solving skills increased successful grades. Problem-solving skills and nursing process skills can be concomitantly increased. Students were suggested to use critical thinking, practical approaches, and care plans, as well as revising nursing processes in order to improve their problem-solving skills and nursing process application skills. © 2014 NANDA International, Inc.
NASA Astrophysics Data System (ADS)
Udomsungworagul, A.; Charnsethikul, P.
2018-03-01
This article introduces methodology to solve large scale two-phase linear programming with a case of multiple time period animal diet problems under both nutrients in raw materials and finished product demand uncertainties. Assumption of allowing to manufacture multiple product formulas in the same time period and assumption of allowing to hold raw materials and finished products inventory have been added. Dantzig-Wolfe decompositions, Benders decomposition and Column generations technique has been combined and applied to solve the problem. The proposed procedure was programmed using VBA and Solver tool in Microsoft Excel. A case study was used and tested in term of efficiency and effectiveness trade-offs.
A sequential quadratic programming algorithm using an incomplete solution of the subproblem
DOE Office of Scientific and Technical Information (OSTI.GOV)
Murray, W.; Prieto, F.J.
1993-05-01
We analyze sequential quadratic programming (SQP) methods to solve nonlinear constrained optimization problems that are more flexible in their definition than standard SQP methods. The type of flexibility introduced is motivated by the necessity to deviate from the standard approach when solving large problems. Specifically we no longer require a minimizer of the QP subproblem to be determined or particular Lagrange multiplier estimates to be used. Our main focus is on an SQP algorithm that uses a particular augmented Lagrangian merit function. New results are derived for this algorithm under weaker conditions than previously assumed; in particular, it is notmore » assumed that the iterates lie on a compact set.« less
Demonstration of quantum advantage in machine learning
NASA Astrophysics Data System (ADS)
Ristè, Diego; da Silva, Marcus P.; Ryan, Colm A.; Cross, Andrew W.; Córcoles, Antonio D.; Smolin, John A.; Gambetta, Jay M.; Chow, Jerry M.; Johnson, Blake R.
2017-04-01
The main promise of quantum computing is to efficiently solve certain problems that are prohibitively expensive for a classical computer. Most problems with a proven quantum advantage involve the repeated use of a black box, or oracle, whose structure encodes the solution. One measure of the algorithmic performance is the query complexity, i.e., the scaling of the number of oracle calls needed to find the solution with a given probability. Few-qubit demonstrations of quantum algorithms, such as Deutsch-Jozsa and Grover, have been implemented across diverse physical systems such as nuclear magnetic resonance, trapped ions, optical systems, and superconducting circuits. However, at the small scale, these problems can already be solved classically with a few oracle queries, limiting the obtained advantage. Here we solve an oracle-based problem, known as learning parity with noise, on a five-qubit superconducting processor. Executing classical and quantum algorithms using the same oracle, we observe a large gap in query count in favor of quantum processing. We find that this gap grows by orders of magnitude as a function of the error rates and the problem size. This result demonstrates that, while complex fault-tolerant architectures will be required for universal quantum computing, a significant quantum advantage already emerges in existing noisy systems.
NASA Astrophysics Data System (ADS)
O'Malley, D.; Vesselinov, V. V.
2017-12-01
Classical microprocessors have had a dramatic impact on hydrology for decades, due largely to the exponential growth in computing power predicted by Moore's law. However, this growth is not expected to continue indefinitely and has already begun to slow. Quantum computing is an emerging alternative to classical microprocessors. Here, we demonstrated cutting edge inverse model analyses utilizing some of the best available resources in both worlds: high-performance classical computing and a D-Wave quantum annealer. The classical high-performance computing resources are utilized to build an advanced numerical model that assimilates data from O(10^5) observations, including water levels, drawdowns, and contaminant concentrations. The developed model accurately reproduces the hydrologic conditions at a Los Alamos National Laboratory contamination site, and can be leveraged to inform decision-making about site remediation. We demonstrate the use of a D-Wave 2X quantum annealer to solve hydrologic inverse problems. This work can be seen as an early step in quantum-computational hydrology. We compare and contrast our results with an early inverse approach in classical-computational hydrology that is comparable to the approach we use with quantum annealing. Our results show that quantum annealing can be useful for identifying regions of high and low permeability within an aquifer. While the problems we consider are small-scale compared to the problems that can be solved with modern classical computers, they are large compared to the problems that could be solved with early classical CPUs. Further, the binary nature of the high/low permeability problem makes it well-suited to quantum annealing, but challenging for classical computers.
Boosting quantum annealer performance via sample persistence
NASA Astrophysics Data System (ADS)
Karimi, Hamed; Rosenberg, Gili
2017-07-01
We propose a novel method for reducing the number of variables in quadratic unconstrained binary optimization problems, using a quantum annealer (or any sampler) to fix the value of a large portion of the variables to values that have a high probability of being optimal. The resulting problems are usually much easier for the quantum annealer to solve, due to their being smaller and consisting of disconnected components. This approach significantly increases the success rate and number of observations of the best known energy value in samples obtained from the quantum annealer, when compared with calling the quantum annealer without using it, even when using fewer annealing cycles. Use of the method results in a considerable improvement in success metrics even for problems with high-precision couplers and biases, which are more challenging for the quantum annealer to solve. The results are further enhanced by applying the method iteratively and combining it with classical pre-processing. We present results for both Chimera graph-structured problems and embedded problems from a real-world application.
Collis-Romberg Mathematical Problem Solving Profiles.
ERIC Educational Resources Information Center
Collis, K. F.; Romberg, T. A.
Problem solving has become a focus of mathematics programs in Australia in recent years, necessitating the assessment of students' problem-solving abilities. This manual provides a problem-solving assessment and teaching resource package containing four elements: (1) profiles assessment items; (2) profiles diagnostic forms for recording individual…
A cooperative strategy for parameter estimation in large scale systems biology models.
Villaverde, Alejandro F; Egea, Jose A; Banga, Julio R
2012-06-22
Mathematical models play a key role in systems biology: they summarize the currently available knowledge in a way that allows to make experimentally verifiable predictions. Model calibration consists of finding the parameters that give the best fit to a set of experimental data, which entails minimizing a cost function that measures the goodness of this fit. Most mathematical models in systems biology present three characteristics which make this problem very difficult to solve: they are highly non-linear, they have a large number of parameters to be estimated, and the information content of the available experimental data is frequently scarce. Hence, there is a need for global optimization methods capable of solving this problem efficiently. A new approach for parameter estimation of large scale models, called Cooperative Enhanced Scatter Search (CeSS), is presented. Its key feature is the cooperation between different programs ("threads") that run in parallel in different processors. Each thread implements a state of the art metaheuristic, the enhanced Scatter Search algorithm (eSS). Cooperation, meaning information sharing between threads, modifies the systemic properties of the algorithm and allows to speed up performance. Two parameter estimation problems involving models related with the central carbon metabolism of E. coli which include different regulatory levels (metabolic and transcriptional) are used as case studies. The performance and capabilities of the method are also evaluated using benchmark problems of large-scale global optimization, with excellent results. The cooperative CeSS strategy is a general purpose technique that can be applied to any model calibration problem. Its capability has been demonstrated by calibrating two large-scale models of different characteristics, improving the performance of previously existing methods in both cases. The cooperative metaheuristic presented here can be easily extended to incorporate other global and local search solvers and specific structural information for particular classes of problems.
A cooperative strategy for parameter estimation in large scale systems biology models
2012-01-01
Background Mathematical models play a key role in systems biology: they summarize the currently available knowledge in a way that allows to make experimentally verifiable predictions. Model calibration consists of finding the parameters that give the best fit to a set of experimental data, which entails minimizing a cost function that measures the goodness of this fit. Most mathematical models in systems biology present three characteristics which make this problem very difficult to solve: they are highly non-linear, they have a large number of parameters to be estimated, and the information content of the available experimental data is frequently scarce. Hence, there is a need for global optimization methods capable of solving this problem efficiently. Results A new approach for parameter estimation of large scale models, called Cooperative Enhanced Scatter Search (CeSS), is presented. Its key feature is the cooperation between different programs (“threads”) that run in parallel in different processors. Each thread implements a state of the art metaheuristic, the enhanced Scatter Search algorithm (eSS). Cooperation, meaning information sharing between threads, modifies the systemic properties of the algorithm and allows to speed up performance. Two parameter estimation problems involving models related with the central carbon metabolism of E. coli which include different regulatory levels (metabolic and transcriptional) are used as case studies. The performance and capabilities of the method are also evaluated using benchmark problems of large-scale global optimization, with excellent results. Conclusions The cooperative CeSS strategy is a general purpose technique that can be applied to any model calibration problem. Its capability has been demonstrated by calibrating two large-scale models of different characteristics, improving the performance of previously existing methods in both cases. The cooperative metaheuristic presented here can be easily extended to incorporate other global and local search solvers and specific structural information for particular classes of problems. PMID:22727112
NASA Astrophysics Data System (ADS)
Pujiastuti, E.; Waluya, B.; Mulyono
2018-03-01
There were many ways of solving the problem offered by the experts. The author combines various ways of solving the problem as a form of novelty. Among the learning model that was expected to support the growth of problem-solving skills was SAVI. The purpose, to obtain trace results from the analysis of the problem-solving ability of students in the Dual Integral material. The research method was a qualitative approach. Its activities include tests was filled with mathematical connections, observation, interviews, FGD, and triangulation. The results were: (1) some students were still experiencing difficulties in solving the problems. (2) The application of modification of SAVI learning model effective in supporting the growth of problem-solving abilities. (3) The strength of the students related to solving the problem, there were two students in the excellent category, there were three students in right classes and one student in the medium group.
Flexibility in Mathematics Problem Solving Based on Adversity Quotient
NASA Astrophysics Data System (ADS)
Dina, N. A.; Amin, S. M.; Masriyah
2018-01-01
Flexibility is an ability which is needed in problem solving. One of the ways in problem solving is influenced by Adversity Quotient (AQ). AQ is the power of facing difficulties. There are three categories of AQ namely climber, camper, and quitter. This research is a descriptive research using qualitative approach. The aim of this research is to describe flexibility in mathematics problem solving based on Adversity Quotient. The subjects of this research are climber student, camper student, and quitter student. This research was started by giving Adversity Response Profile (ARP) questioner continued by giving problem solving task and interviews. The validity of data measurement was using time triangulation. The results of this research shows that climber student uses two strategies in solving problem and doesn’t have difficulty. The camper student uses two strategies in solving problem but has difficulty to finish the second strategies. The quitter student uses one strategy in solving problem and has difficulty to finish it.
Analogy as a strategy for supporting complex problem solving under uncertainty.
Chan, Joel; Paletz, Susannah B F; Schunn, Christian D
2012-11-01
Complex problem solving in naturalistic environments is fraught with uncertainty, which has significant impacts on problem-solving behavior. Thus, theories of human problem solving should include accounts of the cognitive strategies people bring to bear to deal with uncertainty during problem solving. In this article, we present evidence that analogy is one such strategy. Using statistical analyses of the temporal dynamics between analogy and expressed uncertainty in the naturalistic problem-solving conversations among scientists on the Mars Rover Mission, we show that spikes in expressed uncertainty reliably predict analogy use (Study 1) and that expressed uncertainty reduces to baseline levels following analogy use (Study 2). In addition, in Study 3, we show with qualitative analyses that this relationship between uncertainty and analogy is not due to miscommunication-related uncertainty but, rather, is primarily concentrated on substantive problem-solving issues. Finally, we discuss a hypothesis about how analogy might serve as an uncertainty reduction strategy in naturalistic complex problem solving.
Interference thinking in constructing students’ knowledge to solve mathematical problems
NASA Astrophysics Data System (ADS)
Jayanti, W. E.; Usodo, B.; Subanti, S.
2018-04-01
This research aims to describe interference thinking in constructing students’ knowledge to solve mathematical problems. Interference thinking in solving problems occurs when students have two concepts that interfere with each other’s concept. Construction of problem-solving can be traced using Piaget’s assimilation and accommodation framework, helping to know the students’ thinking structures in solving the problems. The method of this research was a qualitative method with case research strategy. The data in this research involving problem-solving result and transcripts of interviews about students’ errors in solving the problem. The results of this research focus only on the student who experience proactive interference, where student in solving a problem using old information to interfere with the ability to recall new information. The student who experience interference thinking in constructing their knowledge occurs when the students’ thinking structures in the assimilation and accommodation process are incomplete. However, after being given reflection to the student, then the students’ thinking process has reached equilibrium condition even though the result obtained remains wrong.
Exploring quantum computing application to satellite data assimilation
NASA Astrophysics Data System (ADS)
Cheung, S.; Zhang, S. Q.
2015-12-01
This is an exploring work on potential application of quantum computing to a scientific data optimization problem. On classical computational platforms, the physical domain of a satellite data assimilation problem is represented by a discrete variable transform, and classical minimization algorithms are employed to find optimal solution of the analysis cost function. The computation becomes intensive and time-consuming when the problem involves large number of variables and data. The new quantum computer opens a very different approach both in conceptual programming and in hardware architecture for solving optimization problem. In order to explore if we can utilize the quantum computing machine architecture, we formulate a satellite data assimilation experimental case in the form of quadratic programming optimization problem. We find a transformation of the problem to map it into Quadratic Unconstrained Binary Optimization (QUBO) framework. Binary Wavelet Transform (BWT) will be applied to the data assimilation variables for its invertible decomposition and all calculations in BWT are performed by Boolean operations. The transformed problem will be experimented as to solve for a solution of QUBO instances defined on Chimera graphs of the quantum computer.
NASA Astrophysics Data System (ADS)
Vatankhah, Saeed; Renaut, Rosemary A.; Ardestani, Vahid E.
2018-04-01
We present a fast algorithm for the total variation regularization of the 3-D gravity inverse problem. Through imposition of the total variation regularization, subsurface structures presenting with sharp discontinuities are preserved better than when using a conventional minimum-structure inversion. The associated problem formulation for the regularization is nonlinear but can be solved using an iteratively reweighted least-squares algorithm. For small-scale problems the regularized least-squares problem at each iteration can be solved using the generalized singular value decomposition. This is not feasible for large-scale, or even moderate-scale, problems. Instead we introduce the use of a randomized generalized singular value decomposition in order to reduce the dimensions of the problem and provide an effective and efficient solution technique. For further efficiency an alternating direction algorithm is used to implement the total variation weighting operator within the iteratively reweighted least-squares algorithm. Presented results for synthetic examples demonstrate that the novel randomized decomposition provides good accuracy for reduced computational and memory demands as compared to use of classical approaches.
Insightful problem solving and emulation in brown capuchin monkeys.
Renner, Elizabeth; Abramo, Allison M; Karen Hambright, M; Phillips, Kimberley A
2017-05-01
We investigated problem solving abilities of capuchin monkeys via the "floating object problem," a task in which the subject must use creative problem solving to retrieve a favored food item from the bottom of a clear tube. Some great apes have solved this problem by adding water to raise the object to a level at which it can be easily grabbed. We presented seven capuchins with the task over eight trials (four "dry" and four "wet"). None of the subjects solved the task, indicating that no capuchin demonstrated insightful problem solving under these experimental conditions. We then investigated whether capuchins would emulate a solution to the task. Seven subjects observed a human model solve the problem by pouring water from a cup into the tube, which brought the object to the top of the tube, allowing the subject to retrieve it. Subjects were then allowed to interact freely with an unfilled tube containing the object in the presence of water and objects that could be used to solve the task. While most subjects were unable to solve the task after viewing a demonstrator solve it, one subject did so, but in a unique way. Our results are consistent with some previous results in great ape species and indicate that capuchins do not spontaneously solve the floating object problem via insight.
Path changing methods applied to the 4-D guidance of STOL aircraft.
DOT National Transportation Integrated Search
1971-11-01
Prior to the advent of large-scale commercial STOL service, some challenging navigation and guidance problems must be solved. Proposed terminal area operations may require that these aircraft be capable of accurately flying complex flight paths, and ...
Numerical Optimization Algorithms and Software for Systems Biology
DOE Office of Scientific and Technical Information (OSTI.GOV)
Saunders, Michael
2013-02-02
The basic aims of this work are: to develop reliable algorithms for solving optimization problems involving large stoi- chiometric matrices; to investigate cyclic dependency between metabolic and macromolecular biosynthetic networks; and to quantify the significance of thermodynamic constraints on prokaryotic metabolism.
ERIC Educational Resources Information Center
Coulter, Charles L.
1978-01-01
Discusses the scientific and financial aspects of large instrument sharing in natural science areas from the point of view of instrument centers, users and funding agencies. Concludes that effective actions, with well-defined goals, must be implemented to solve the problems of instrument sharing. (GA)
Solving Power Tool Problems in the School Shop
ERIC Educational Resources Information Center
Irvin, Daniel W.
1976-01-01
The school shop instructor is largely responsible for the preventive maintenance of power tools. These preventive measures primarily involve proper alignment, good lubrication, a reasonable maintenance program, and good operating procedures. Suggestions for maintenance of specific equipment is provided. (Author/BP)
Multi-GPU implementation of a VMAT treatment plan optimization algorithm
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tian, Zhen, E-mail: Zhen.Tian@UTSouthwestern.edu, E-mail: Xun.Jia@UTSouthwestern.edu, E-mail: Steve.Jiang@UTSouthwestern.edu; Folkerts, Michael; Tan, Jun
Purpose: Volumetric modulated arc therapy (VMAT) optimization is a computationally challenging problem due to its large data size, high degrees of freedom, and many hardware constraints. High-performance graphics processing units (GPUs) have been used to speed up the computations. However, GPU’s relatively small memory size cannot handle cases with a large dose-deposition coefficient (DDC) matrix in cases of, e.g., those with a large target size, multiple targets, multiple arcs, and/or small beamlet size. The main purpose of this paper is to report an implementation of a column-generation-based VMAT algorithm, previously developed in the authors’ group, on a multi-GPU platform tomore » solve the memory limitation problem. While the column-generation-based VMAT algorithm has been previously developed, the GPU implementation details have not been reported. Hence, another purpose is to present detailed techniques employed for GPU implementation. The authors also would like to utilize this particular problem as an example problem to study the feasibility of using a multi-GPU platform to solve large-scale problems in medical physics. Methods: The column-generation approach generates VMAT apertures sequentially by solving a pricing problem (PP) and a master problem (MP) iteratively. In the authors’ method, the sparse DDC matrix is first stored on a CPU in coordinate list format (COO). On the GPU side, this matrix is split into four submatrices according to beam angles, which are stored on four GPUs in compressed sparse row format. Computation of beamlet price, the first step in PP, is accomplished using multi-GPUs. A fast inter-GPU data transfer scheme is accomplished using peer-to-peer access. The remaining steps of PP and MP problems are implemented on CPU or a single GPU due to their modest problem scale and computational loads. Barzilai and Borwein algorithm with a subspace step scheme is adopted here to solve the MP problem. A head and neck (H and N) cancer case is then used to validate the authors’ method. The authors also compare their multi-GPU implementation with three different single GPU implementation strategies, i.e., truncating DDC matrix (S1), repeatedly transferring DDC matrix between CPU and GPU (S2), and porting computations involving DDC matrix to CPU (S3), in terms of both plan quality and computational efficiency. Two more H and N patient cases and three prostate cases are used to demonstrate the advantages of the authors’ method. Results: The authors’ multi-GPU implementation can finish the optimization process within ∼1 min for the H and N patient case. S1 leads to an inferior plan quality although its total time was 10 s shorter than the multi-GPU implementation due to the reduced matrix size. S2 and S3 yield the same plan quality as the multi-GPU implementation but take ∼4 and ∼6 min, respectively. High computational efficiency was consistently achieved for the other five patient cases tested, with VMAT plans of clinically acceptable quality obtained within 23–46 s. Conversely, to obtain clinically comparable or acceptable plans for all six of these VMAT cases that the authors have tested in this paper, the optimization time needed in a commercial TPS system on CPU was found to be in an order of several minutes. Conclusions: The results demonstrate that the multi-GPU implementation of the authors’ column-generation-based VMAT optimization can handle the large-scale VMAT optimization problem efficiently without sacrificing plan quality. The authors’ study may serve as an example to shed some light on other large-scale medical physics problems that require multi-GPU techniques.« less
Tenison, Caitlin; Fincham, Jon M; Anderson, John R
2014-02-01
This research explores how to determine when mathematical problems are solved by retrieval versus computation strategies. Past research has indicated that verbal reports, solution latencies, and neural imaging all provide imperfect indicators of this distinction. Participants in the current study solved mathematical problems involving two distinct problem types, called 'Pyramid' and 'Formula' problems. Participants were given extensive training solving 3 select Pyramid and 3 select Formula problems. Trained problems were highly practiced, whereas untrained problems were not. The distinction between untrained and trained problems was observed in the data. Untrained problems took longer to solve, more often used procedural strategies and showed a greater activation in the horizontal intraparietal sulcus (HIPS) when compared to trained problems. A classifier fit to the neural distinction between trained-untrained problems successfully predicted training within and between the two problem types. We employed this classifier to generate a prediction of strategy use. By combining evidence from the classifier, problem solving latencies, and retrospective reports, we predicted the strategy used to solve each problem in the scanner and gained unexpected insight into the distinction between different strategies. Copyright © 2013 Elsevier Ltd. All rights reserved.
Problem solving therapy - use and effectiveness in general practice.
Pierce, David
2012-09-01
Problem solving therapy (PST) is one of the focused psychological strategies supported by Medicare for use by appropriately trained general practitioners. This article reviews the evidence base for PST and its use in the general practice setting. Problem solving therapy involves patients learning or reactivating problem solving skills. These skills can then be applied to specific life problems associated with psychological and somatic symptoms. Problem solving therapy is suitable for use in general practice for patients experiencing common mental health conditions and has been shown to be as effective in the treatment of depression as antidepressants. Problem solving therapy involves a series of sequential stages. The clinician assists the patient to develop new empowering skills, and then supports them to work through the stages of therapy to determine and implement the solution selected by the patient. Many experienced GPs will identify their own existing problem solving skills. Learning about PST may involve refining and focusing these skills.
Collection of solved problems in physics
NASA Astrophysics Data System (ADS)
Koupilová, ZdeÅka; Mandíková, Dana; Snětinová, Marie
2017-01-01
To solve physics problems is a key ability which students should reach during their physics education. Ten years ago we started to develop a Collection of fully solved problems. The structure of problems' solutions is specially designed to substitute tutor's help during lesson and encourage students to solve at least some parts of a problem independently. Nowadays the database contains about 770 fully solved problems in physics in Czech, more than 100 problems in Polish and more than 140 problems in English. Other problems are still being translated. Except for physics problems, the Collection has also a mathematical part, which contains more than 300 fully solved problems in mathematics. This paper follows the presentation of the Collection of solved problems from previous years and introduces a new interface of the Collection, its enhanced functionality, new topics, newly created interface for teachers, user feedback and plans for future development. The database is placed at the website of the Department of Physics Education, Faculty of Mathematics and Physics, Charles University in Prague, the links are: http://reseneulohy.cz/fyzika (Czech version); http://www.physicstasks.eu/ (English version).
Pre-service mathematics teachers’ ability in solving well-structured problem
NASA Astrophysics Data System (ADS)
Paradesa, R.
2018-01-01
This study aimed to describe the mathematical problem-solving ability of undergraduate students of mathematics education in solving the well-structured problem. The type of this study was qualitative descriptive. The subjects in this study were 100 undergraduate students of Mathematics Education at one of the private universities in Palembang city. The data in this study was collected through two test items with essay form. The results of this study showed that, from the first problem, only 8% students can solve it, but do not check back again to validate the process. Based on a scoring rubric that follows Polya strategy, their answer satisfied 2 4 2 0 patterns. But, from the second problem, 45% students satisfied it. This is because the second problem imitated from the example that was given in learning process. The average score of undergraduate students mathematical problem-solving ability in solving well-structured problems showed 56.00 with standard deviation was 13.22. It means that, from 0 - 100 scale, undergraduate students mathematical problem-solving ability can be categorized low. From this result, the conclusion was undergraduate students of mathematics education in Palembang still have a problem in solving mathematics well-structured problem.
ERIC Educational Resources Information Center
Chen, Limin; Van Dooren, Wim; Chen, Qi; Verschaffel, Lieven
2011-01-01
In the present study, which is a part of a research project about realistic word problem solving and problem posing in Chinese elementary schools, a problem solving and a problem posing test were administered to 128 pre-service and in-service elementary school teachers from Tianjin City in China, wherein the teachers were asked to solve 3…
Abdollahi, Abbas; Abu Talib, Mansor; Carlbring, Per; Harvey, Richard; Yaacob, Siti Nor; Ismail, Zanariah
2016-06-01
This study was designed to examine the relationships between problem-solving skills, hardiness, and perceived stress and to test the moderating role of hardiness in the relationship between problem-solving skills and perceived stress among 500 undergraduates from Malaysian public universities. The analyses showed that undergraduates with poor problem-solving confidence, external personal control of emotion, and approach-avoidance style were more likely to report perceived stress. Hardiness moderated the relationships between problem-solving skills and perceived stress. These findings reinforce the importance of moderating role of hardiness as an influencing factor that explains how problem-solving skills affect perceived stress among undergraduates.
NASA Astrophysics Data System (ADS)
Jeon, Kyungmoon; Huffman, Douglas; Noh, Taehee
2005-10-01
This study investigated the effects of a thinking aloud pair problem solving (TAPPS) approach on students' chemistry problem-solving performance and verbal interactions. A total of 85 eleventh grade students from three classes in a Korean high school were randomly assigned to one of three groups; either individually using a problem-solving strategy, using a problem-solving strategy with TAPPS, or the control group. After instruction, students' problem-solving performance was examined. The results showed that students in both the individual and TAPPS groups performed better than those in the control group on recalling the related law and mathematical execution, while students in the TAPPS group performed better than those in the other groups on conceptual knowledge. To investigate the verbal behaviors using TAPPS, verbal behaviors of solvers and listeners were classified into 8 categories. Listeners' verbal behavior of "agreeing" and "pointing out", and solvers' verbal behavior of "modifying" were positively related with listeners' problem-solving performance. There was, however, a negative correlation between listeners' use of "point out" and solvers' problem-solving performance. The educational implications of this study are discussed.
Pedagogy and/or technology: Making difference in improving students' problem solving skills
NASA Astrophysics Data System (ADS)
Hrepic, Zdeslav; Lodder, Katherine; Shaw, Kimberly A.
2013-01-01
Pen input computers combined with interactive software may have substantial potential for promoting active instructional methodologies and for facilitating students' problem solving ability. An excellent example is a study in which introductory physics students improved retention, conceptual understanding and problem solving abilities when one of three weekly lectures was replaced with group problem solving sessions facilitated with Tablet PCs and DyKnow software [1,2]. The research goal of the present study was to isolate the effect of the methodology itself (using additional time to teach problem solving) from that of the involved technology. In Fall 2011 we compared the performance of students taking the same introductory physics lecture course while enrolled in two separate problem-solving sections. One section used pen-based computing to facilitate group problem solving while the other section used low-tech methods for one third of the semester (covering Kinematics), and then traded technologies for the middle third of the term (covering Dynamics). Analysis of quiz, exam and standardized pre-post test results indicated no significant difference in scores of the two groups. Combining this result with those of previous studies implies primacy of pedagogy (collaborative problem solving itself) over technology for student learning in problem solving recitations.
Working memory dysfunctions predict social problem solving skills in schizophrenia.
Huang, Jia; Tan, Shu-ping; Walsh, Sarah C; Spriggens, Lauren K; Neumann, David L; Shum, David H K; Chan, Raymond C K
2014-12-15
The current study aimed to examine the contribution of neurocognition and social cognition to components of social problem solving. Sixty-seven inpatients with schizophrenia and 31 healthy controls were administrated batteries of neurocognitive tests, emotion perception tests, and the Chinese Assessment of Interpersonal Problem Solving Skills (CAIPSS). MANOVAs were conducted to investigate the domains in which patients with schizophrenia showed impairments. Correlations were used to determine which impaired domains were associated with social problem solving, and multiple regression analyses were conducted to compare the relative contribution of neurocognitive and social cognitive functioning to components of social problem solving. Compared with healthy controls, patients with schizophrenia performed significantly worse in sustained attention, working memory, negative emotion, intention identification and all components of the CAIPSS. Specifically, sustained attention, working memory and negative emotion identification were found to correlate with social problem solving and 1-back accuracy significantly predicted the poor performance in social problem solving. Among the dysfunctions in schizophrenia, working memory contributed most to deficits in social problem solving in patients with schizophrenia. This finding provides support for targeting working memory in the development of future social problem solving rehabilitation interventions. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.
ERIC Educational Resources Information Center
Szetela, W.; Super, D.
A problem-solving program supplemented by calculators in one treatment group was conducted in 63 grade 7 classes with about 1350 students. Teachers were provided with problems correlated with textbooks, and instruction for teaching problem-solving strategies. School districts provided calculators and problem-solving materials. Pretest scores…
ERIC Educational Resources Information Center
Docktor, Jennifer L.; Dornfeld, Jay; Frodermann, Evan; Heller, Kenneth; Hsu, Leonardo; Jackson, Koblar Alan; Mason, Andrew; Ryan, Qing X.; Yang, Jie
2016-01-01
Problem solving is a complex process valuable in everyday life and crucial for learning in the STEM fields. To support the development of problem-solving skills it is important for researchers and curriculum developers to have practical tools that can measure the difference between novice and expert problem-solving performance in authentic…
Problem Solving: How Can We Help Students Overcome Cognitive Difficulties
ERIC Educational Resources Information Center
Cardellini, Liberato
2014-01-01
The traditional approach to teach problem solving usually consists in showing students the solutions of some example-problems and then in asking students to practice individually on solving a certain number of related problems. This approach does not ensure that students learn to solve problems and above all to think about the solution process in…
Step by Step: Biology Undergraduates' Problem-Solving Procedures during Multiple-Choice Assessment
ERIC Educational Resources Information Center
Prevost, Luanna B.; Lemons, Paula P.
2016-01-01
This study uses the theoretical framework of domain-specific problem solving to explore the procedures students use to solve multiple-choice problems about biology concepts. We designed several multiple-choice problems and administered them on four exams. We trained students to produce written descriptions of how they solved the problem, and this…
Analog Processor To Solve Optimization Problems
NASA Technical Reports Server (NTRS)
Duong, Tuan A.; Eberhardt, Silvio P.; Thakoor, Anil P.
1993-01-01
Proposed analog processor solves "traveling-salesman" problem, considered paradigm of global-optimization problems involving routing or allocation of resources. Includes electronic neural network and auxiliary circuitry based partly on concepts described in "Neural-Network Processor Would Allocate Resources" (NPO-17781) and "Neural Network Solves 'Traveling-Salesman' Problem" (NPO-17807). Processor based on highly parallel computing solves problem in significantly less time.
Problem Solving Appraisal of Delinquent Adolescents.
ERIC Educational Resources Information Center
Perez, Ruperto M.; And Others
The study investigated the following: (1) the relationship of problem solving appraisal to narcissistic vulnerability, locus of control, and depression; (2) the differences in problem solving appraisal, locus of control, and depression in first-time and repeat offenders; and (3) the prediction of problem solving appraisal by narcissistic…
Computer Programming: A Medium for Teaching Problem Solving.
ERIC Educational Resources Information Center
Casey, Patrick J.
1997-01-01
Argues that including computer programming in the curriculum as a medium for instruction is a feasible alternative for teaching problem solving. Discusses the nature of problem solving; the problem-solving elements of discovery, motivation, practical learning situations and flexibility which are inherent in programming; capabilities of computer…
Perceived Problem Solving, Stress, and Health among College Students
ERIC Educational Resources Information Center
Largo-Wight, Erin; Peterson, P. Michael; Chen, W. William
2005-01-01
Objective: To study the relationships among perceived problem solving, stress, and physical health. Methods: The Perceived Stress Questionnaire (PSQ), Personal Problem solving Inventory (PSI), and a stress-related physical health symptoms checklist were used to measure perceived stress, problem solving, and health among undergraduate college…
THE CURRENT STATUS OF RESEARCH AND THEORY IN HUMAN PROBLEM SOLVING.
ERIC Educational Resources Information Center
DAVIS, GARY A.
PROBLEM-SOLVING THEORIES IN THREE AREAS - TRADITIONAL (STIMULUS-RESPONSE) LEARNING, COGNITIVE-GESTALT APPROACHES, AND COMPUTER AND MATHEMATICAL MODELS - WERE SUMMARIZED. RECENT EMPIRICAL STUDIES (1960-65) ON PROBLEM SOLVING WERE CATEGORIZED ACCORDING TO TYPE OF BEHAVIOR ELICITED BY PARTICULAR PROBLEM-SOLVING TASKS. ANAGRAM,…
Developing Creativity through Collaborative Problem Solving
ERIC Educational Resources Information Center
Albert, Lillie R.; Kim, Rina
2013-01-01
This paper discusses an alternative approach for developing problem solving experiences for students. The major argument is that students can develop their creativity by engaging in collaborative problem solving activities in which they apply a variety of mathematical methods creatively to solve problems. The argument is supported by: considering…
The effects of expected reward on creative problem solving.
Cristofori, Irene; Salvi, Carola; Beeman, Mark; Grafman, Jordan
2018-06-12
Creative problem solving involves search processes, and it is known to be hard to motivate. Reward cues have been found to enhance performance across a range of tasks, even when cues are presented subliminally, without being consciously detected. It is uncertain whether motivational processes, such as reward, can influence problem solving. We tested the effect of supraliminal and subliminal reward on participant performance on problem solving that can be solved by deliberate analysis or by insight. Forty-one participants attempted to solve 100 compound remote associate problems. At the beginning of each problem, a potential reward cue (1 or 25 cents) was displayed, either subliminally (17 ms) or supraliminally (100 ms). Participants earned the displayed reward if they solved the problem correctly. Results showed that the higher subliminal reward increased the percentage of problems solved correctly overall. Second, we explored if subliminal rewards preferentially influenced solutions that were achieved via a sudden insight (mostly processed below awareness) or via a deliberate analysis. Participants solved more problems via insight following high subliminal reward when compared with low subliminal reward, and compared with high supraliminal reward, with no corresponding effect on analytic solving. Striatal dopamine (DA) is thought to influence motivation, reinforce behavior, and facilitate cognition. We speculate that subliminal rewards activate the striatal DA system, enhancing the kinds of automatic integrative processes that lead to more creative strategies for problem solving, without increasing the selectivity of attention, which could impede insight.
Find the Dimensions: Students Solving a Tiling Problem
ERIC Educational Resources Information Center
Obara, Samuel
2018-01-01
Students learn mathematics by solving problems. Mathematics textbooks are full of problems, and mathematics teachers use these problems to test students' understanding of mathematical concepts. This paper discusses how problem-solving skills can be fostered with a geometric tiling problem.
Penas, David R; González, Patricia; Egea, Jose A; Doallo, Ramón; Banga, Julio R
2017-01-21
The development of large-scale kinetic models is one of the current key issues in computational systems biology and bioinformatics. Here we consider the problem of parameter estimation in nonlinear dynamic models. Global optimization methods can be used to solve this type of problems but the associated computational cost is very large. Moreover, many of these methods need the tuning of a number of adjustable search parameters, requiring a number of initial exploratory runs and therefore further increasing the computation times. Here we present a novel parallel method, self-adaptive cooperative enhanced scatter search (saCeSS), to accelerate the solution of this class of problems. The method is based on the scatter search optimization metaheuristic and incorporates several key new mechanisms: (i) asynchronous cooperation between parallel processes, (ii) coarse and fine-grained parallelism, and (iii) self-tuning strategies. The performance and robustness of saCeSS is illustrated by solving a set of challenging parameter estimation problems, including medium and large-scale kinetic models of the bacterium E. coli, bakerés yeast S. cerevisiae, the vinegar fly D. melanogaster, Chinese Hamster Ovary cells, and a generic signal transduction network. The results consistently show that saCeSS is a robust and efficient method, allowing very significant reduction of computation times with respect to several previous state of the art methods (from days to minutes, in several cases) even when only a small number of processors is used. The new parallel cooperative method presented here allows the solution of medium and large scale parameter estimation problems in reasonable computation times and with small hardware requirements. Further, the method includes self-tuning mechanisms which facilitate its use by non-experts. We believe that this new method can play a key role in the development of large-scale and even whole-cell dynamic models.
Amirghasemi, Mehrdad; Zamani, Reza
2014-01-01
This paper presents an effective procedure for solving the job shop problem. Synergistically combining small and large neighborhood schemes, the procedure consists of four components, namely (i) a construction method for generating semi-active schedules by a forward-backward mechanism, (ii) a local search for manipulating a small neighborhood structure guided by a tabu list, (iii) a feedback-based mechanism for perturbing the solutions generated, and (iv) a very large-neighborhood local search guided by a forward-backward shifting bottleneck method. The combination of shifting bottleneck mechanism and tabu list is used as a means of the manipulation of neighborhood structures, and the perturbation mechanism employed diversifies the search. A feedback mechanism, called repeat-check, detects consequent repeats and ignites a perturbation when the total number of consecutive repeats for two identical makespan values reaches a given threshold. The results of extensive computational experiments on the benchmark instances indicate that the combination of these four components is synergetic, in the sense that they collectively make the procedure fast and robust.
Amesos2 and Belos: Direct and Iterative Solvers for Large Sparse Linear Systems
Bavier, Eric; Hoemmen, Mark; Rajamanickam, Sivasankaran; ...
2012-01-01
Solvers for large sparse linear systems come in two categories: direct and iterative. Amesos2, a package in the Trilinos software project, provides direct methods, and Belos, another Trilinos package, provides iterative methods. Amesos2 offers a common interface to many different sparse matrix factorization codes, and can handle any implementation of sparse matrices and vectors, via an easy-to-extend C++ traits interface. It can also factor matrices whose entries have arbitrary “Scalar” type, enabling extended-precision and mixed-precision algorithms. Belos includes many different iterative methods for solving large sparse linear systems and least-squares problems. Unlike competing iterative solver libraries, Belos completely decouples themore » algorithms from the implementations of the underlying linear algebra objects. This lets Belos exploit the latest hardware without changes to the code. Belos favors algorithms that solve higher-level problems, such as multiple simultaneous linear systems and sequences of related linear systems, faster than standard algorithms. The package also supports extended-precision and mixed-precision algorithms. Together, Amesos2 and Belos form a complete suite of sparse linear solvers.« less