Online adaptation and over-trial learning in macaque visuomotor control.
Braun, Daniel A; Aertsen, Ad; Paz, Rony; Vaadia, Eilon; Rotter, Stefan; Mehring, Carsten
2011-01-01
When faced with unpredictable environments, the human motor system has been shown to develop optimized adaptation strategies that allow for online adaptation during the control process. Such online adaptation is to be contrasted to slower over-trial learning that corresponds to a trial-by-trial update of the movement plan. Here we investigate the interplay of both processes, i.e., online adaptation and over-trial learning, in a visuomotor experiment performed by macaques. We show that simple non-adaptive control schemes fail to perform in this task, but that a previously suggested adaptive optimal feedback control model can explain the observed behavior. We also show that over-trial learning as seen in learning and aftereffect curves can be explained by learning in a radial basis function network. Our results suggest that both the process of over-trial learning and the process of online adaptation are crucial to understand visuomotor learning.
Online Adaptation and Over-Trial Learning in Macaque Visuomotor Control
Braun, Daniel A.; Aertsen, Ad; Paz, Rony; Vaadia, Eilon; Rotter, Stefan; Mehring, Carsten
2011-01-01
When faced with unpredictable environments, the human motor system has been shown to develop optimized adaptation strategies that allow for online adaptation during the control process. Such online adaptation is to be contrasted to slower over-trial learning that corresponds to a trial-by-trial update of the movement plan. Here we investigate the interplay of both processes, i.e., online adaptation and over-trial learning, in a visuomotor experiment performed by macaques. We show that simple non-adaptive control schemes fail to perform in this task, but that a previously suggested adaptive optimal feedback control model can explain the observed behavior. We also show that over-trial learning as seen in learning and aftereffect curves can be explained by learning in a radial basis function network. Our results suggest that both the process of over-trial learning and the process of online adaptation are crucial to understand visuomotor learning. PMID:21720526
Indirect learning control for nonlinear dynamical systems
NASA Technical Reports Server (NTRS)
Ryu, Yeong Soon; Longman, Richard W.
1993-01-01
In a previous paper, learning control algorithms were developed based on adaptive control ideas for linear time variant systems. The learning control methods were shown to have certain advantages over their adaptive control counterparts, such as the ability to produce zero tracking error in time varying systems, and the ability to eliminate repetitive disturbances. In recent years, certain adaptive control algorithms have been developed for multi-body dynamic systems such as robots, with global guaranteed convergence to zero tracking error for the nonlinear system euations. In this paper we study the relationship between such adaptive control methods designed for this specific class of nonlinear systems, and the learning control problem for such systems, seeking to converge to zero tracking error in following a specific command repeatedly, starting from the same initial conditions each time. The extension of these methods from the adaptive control problem to the learning control problem is seen to be trivial. The advantages and disadvantages of using learning control based on such adaptive control concepts for nonlinear systems, and the use of other currently available learning control algorithms are discussed.
Towards autonomous neuroprosthetic control using Hebbian reinforcement learning.
Mahmoudi, Babak; Pohlmeyer, Eric A; Prins, Noeline W; Geng, Shijia; Sanchez, Justin C
2013-12-01
Our goal was to design an adaptive neuroprosthetic controller that could learn the mapping from neural states to prosthetic actions and automatically adjust adaptation using only a binary evaluative feedback as a measure of desirability/undesirability of performance. Hebbian reinforcement learning (HRL) in a connectionist network was used for the design of the adaptive controller. The method combines the efficiency of supervised learning with the generality of reinforcement learning. The convergence properties of this approach were studied using both closed-loop control simulations and open-loop simulations that used primate neural data from robot-assisted reaching tasks. The HRL controller was able to perform classification and regression tasks using its episodic and sequential learning modes, respectively. In our experiments, the HRL controller quickly achieved convergence to an effective control policy, followed by robust performance. The controller also automatically stopped adapting the parameters after converging to a satisfactory control policy. Additionally, when the input neural vector was reorganized, the controller resumed adaptation to maintain performance. By estimating an evaluative feedback directly from the user, the HRL control algorithm may provide an efficient method for autonomous adaptation of neuroprosthetic systems. This method may enable the user to teach the controller the desired behavior using only a simple feedback signal.
Edwards, Ann L; Dawson, Michael R; Hebert, Jacqueline S; Sherstan, Craig; Sutton, Richard S; Chan, K Ming; Pilarski, Patrick M
2016-10-01
Myoelectric prostheses currently used by amputees can be difficult to control. Machine learning, and in particular learned predictions about user intent, could help to reduce the time and cognitive load required by amputees while operating their prosthetic device. The goal of this study was to compare two switching-based methods of controlling a myoelectric arm: non-adaptive (or conventional) control and adaptive control (involving real-time prediction learning). Case series study. We compared non-adaptive and adaptive control in two different experiments. In the first, one amputee and one non-amputee subject controlled a robotic arm to perform a simple task; in the second, three able-bodied subjects controlled a robotic arm to perform a more complex task. For both tasks, we calculated the mean time and total number of switches between robotic arm functions over three trials. Adaptive control significantly decreased the number of switches and total switching time for both tasks compared with the conventional control method. Real-time prediction learning was successfully used to improve the control interface of a myoelectric robotic arm during uninterrupted use by an amputee subject and able-bodied subjects. Adaptive control using real-time prediction learning has the potential to help decrease both the time and the cognitive load required by amputees in real-world functional situations when using myoelectric prostheses. © The International Society for Prosthetics and Orthotics 2015.
Rule-based mechanisms of learning for intelligent adaptive flight control
NASA Technical Reports Server (NTRS)
Handelman, David A.; Stengel, Robert F.
1990-01-01
How certain aspects of human learning can be used to characterize learning in intelligent adaptive control systems is investigated. Reflexive and declarative memory and learning are described. It is shown that model-based systems-theoretic adaptive control methods exhibit attributes of reflexive learning, whereas the problem-solving capabilities of knowledge-based systems of artificial intelligence are naturally suited for implementing declarative learning. Issues related to learning in knowledge-based control systems are addressed, with particular attention given to rule-based systems. A mechanism for real-time rule-based knowledge acquisition is suggested, and utilization of this mechanism within the context of failure diagnosis for fault-tolerant flight control is demonstrated.
Dynamic learning from adaptive neural network control of a class of nonaffine nonlinear systems.
Dai, Shi-Lu; Wang, Cong; Wang, Min
2014-01-01
This paper studies the problem of learning from adaptive neural network (NN) control of a class of nonaffine nonlinear systems in uncertain dynamic environments. In the control design process, a stable adaptive NN tracking control design technique is proposed for the nonaffine nonlinear systems with a mild assumption by combining a filtered tracking error with the implicit function theorem, input-to-state stability, and the small-gain theorem. The proposed stable control design technique not only overcomes the difficulty in controlling nonaffine nonlinear systems but also relaxes constraint conditions of the considered systems. In the learning process, the partial persistent excitation (PE) condition of radial basis function NNs is satisfied during tracking control to a recurrent reference trajectory. Under the PE condition and an appropriate state transformation, the proposed adaptive NN control is shown to be capable of acquiring knowledge on the implicit desired control input dynamics in the stable control process and of storing the learned knowledge in memory. Subsequently, an NN learning control design technique that effectively exploits the learned knowledge without re-adapting to the controller parameters is proposed to achieve closed-loop stability and improved control performance. Simulation studies are performed to demonstrate the effectiveness of the proposed design techniques.
Lessons Learned and Flight Results from the F15 Intelligent Flight Control System Project
NASA Technical Reports Server (NTRS)
Bosworth, John
2006-01-01
A viewgraph presentation on the lessons learned and flight results from the F15 Intelligent Flight Control System (IFCS) project is shown. The topics include: 1) F-15 IFCS Project Goals; 2) Motivation; 3) IFCS Approach; 4) NASA F-15 #837 Aircraft Description; 5) Flight Envelope; 6) Limited Authority System; 7) NN Floating Limiter; 8) Flight Experiment; 9) Adaptation Goals; 10) Handling Qualities Performance Metric; 11) Project Phases; 12) Indirect Adaptive Control Architecture; 13) Indirect Adaptive Experience and Lessons Learned; 14) Gen II Direct Adaptive Control Architecture; 15) Current Status; 16) Effect of Canard Multiplier; 17) Simulated Canard Failure Stab Open Loop; 18) Canard Multiplier Effect Closed Loop Freq. Resp.; 19) Simulated Canard Failure Stab Open Loop with Adaptation; 20) Canard Multiplier Effect Closed Loop with Adaptation; 21) Gen 2 NN Wts from Simulation; 22) Direct Adaptive Experience and Lessons Learned; and 23) Conclusions
Effectiveness of Adaptive Assessment versus Learner Control in a Multimedia Learning System
ERIC Educational Resources Information Center
Chen, Ching-Huei; Chang, Shu-Wei
2015-01-01
The purpose of this study was to explore the effectiveness of adaptive assessment versus learner control in a multimedia learning system designed to help secondary students learn science. Unlike other systems, this paper presents a workflow of adaptive assessment following instructional materials that better align with learners' cognitive…
NASA Technical Reports Server (NTRS)
Thau, F. E.; Montgomery, R. C.
1980-01-01
Techniques developed for the control of aircraft under changing operating conditions are used to develop a learning control system structure for a multi-configuration, flexible space vehicle. A configuration identification subsystem that is to be used with a learning algorithm and a memory and control process subsystem is developed. Adaptive gain adjustments can be achieved by this learning approach without prestoring of large blocks of parameter data and without dither signal inputs which will be suppressed during operations for which they are not compatible. The Space Shuttle Solar Electric Propulsion (SEP) experiment is used as a sample problem for the testing of adaptive/learning control system algorithms.
Adaptive categorization of ART networks in robot behavior learning using game-theoretic formulation.
Fung, Wai-keung; Liu, Yun-hui
2003-12-01
Adaptive Resonance Theory (ART) networks are employed in robot behavior learning. Two of the difficulties in online robot behavior learning, namely, (1) exponential memory increases with time, (2) difficulty for operators to specify learning tasks accuracy and control learning attention before learning. In order to remedy the aforementioned difficulties, an adaptive categorization mechanism is introduced in ART networks for perceptual and action patterns categorization in this paper. A game-theoretic formulation of adaptive categorization for ART networks is proposed for vigilance parameter adaptation for category size control on the categories formed. The proposed vigilance parameter update rule can help improving categorization performance in the aspect of category number stability and solve the problem of selecting initial vigilance parameter prior to pattern categorization in traditional ART networks. Behavior learning using physical robot is conducted to demonstrate the effectiveness of the proposed adaptive categorization mechanism in ART networks.
NASA Technical Reports Server (NTRS)
Jacklin, Stephen; Schumann, Johann; Gupta, Pramod; Richard, Michael; Guenther, Kurt; Soares, Fola
2005-01-01
Adaptive control technologies that incorporate learning algorithms have been proposed to enable automatic flight control and vehicle recovery, autonomous flight, and to maintain vehicle performance in the face of unknown, changing, or poorly defined operating environments. In order for adaptive control systems to be used in safety-critical aerospace applications, they must be proven to be highly safe and reliable. Rigorous methods for adaptive software verification and validation must be developed to ensure that control system software failures will not occur. Of central importance in this regard is the need to establish reliable methods that guarantee convergent learning, rapid convergence (learning) rate, and algorithm stability. This paper presents the major problems of adaptive control systems that use learning to improve performance. The paper then presents the major procedures and tools presently developed or currently being developed to enable the verification, validation, and ultimate certification of these adaptive control systems. These technologies include the application of automated program analysis methods, techniques to improve the learning process, analytical methods to verify stability, methods to automatically synthesize code, simulation and test methods, and tools to provide on-line software assurance.
Tsai, Jason Sheng-Hong; Du, Yan-Yi; Huang, Pei-Hsiang; Guo, Shu-Mei; Shieh, Leang-San; Chen, Yuhua
2011-07-01
In this paper, a digital redesign methodology of the iterative learning-based decentralized adaptive tracker is proposed to improve the dynamic performance of sampled-data linear large-scale control systems consisting of N interconnected multi-input multi-output subsystems, so that the system output will follow any trajectory which may not be presented by the analytic reference model initially. To overcome the interference of each sub-system and simplify the controller design, the proposed model reference decentralized adaptive control scheme constructs a decoupled well-designed reference model first. Then, according to the well-designed model, this paper develops a digital decentralized adaptive tracker based on the optimal analog control and prediction-based digital redesign technique for the sampled-data large-scale coupling system. In order to enhance the tracking performance of the digital tracker at specified sampling instants, we apply the iterative learning control (ILC) to train the control input via continual learning. As a result, the proposed iterative learning-based decentralized adaptive tracker not only has robust closed-loop decoupled property but also possesses good tracking performance at both transient and steady state. Besides, evolutionary programming is applied to search for a good learning gain to speed up the learning process of ILC. Copyright © 2011 ISA. Published by Elsevier Ltd. All rights reserved.
NASA Astrophysics Data System (ADS)
Wei, Caisheng; Luo, Jianjun; Dai, Honghua; Bian, Zilin; Yuan, Jianping
2018-05-01
In this paper, a novel learning-based adaptive attitude takeover control method is investigated for the postcapture space robot-target combination with guaranteed prescribed performance in the presence of unknown inertial properties and external disturbance. First, a new static prescribed performance controller is developed to guarantee that all the involved attitude tracking errors are uniformly ultimately bounded by quantitatively characterizing the transient and steady-state performance of the combination. Then, a learning-based supplementary adaptive strategy based on adaptive dynamic programming is introduced to improve the tracking performance of static controller in terms of robustness and adaptiveness only utilizing the input/output data of the combination. Compared with the existing works, the prominent advantage is that the unknown inertial properties are not required to identify in the development of learning-based adaptive control law, which dramatically decreases the complexity and difficulty of the relevant controller design. Moreover, the transient and steady-state performance is guaranteed a priori by designer-specialized performance functions without resorting to repeated regulations of the controller parameters. Finally, the three groups of illustrative examples are employed to verify the effectiveness of the proposed control method.
REVIEW: Internal models in sensorimotor integration: perspectives from adaptive control theory
NASA Astrophysics Data System (ADS)
Tin, Chung; Poon, Chi-Sang
2005-09-01
Internal models and adaptive controls are empirical and mathematical paradigms that have evolved separately to describe learning control processes in brain systems and engineering systems, respectively. This paper presents a comprehensive appraisal of the correlation between these paradigms with a view to forging a unified theoretical framework that may benefit both disciplines. It is suggested that the classic equilibrium-point theory of impedance control of arm movement is analogous to continuous gain-scheduling or high-gain adaptive control within or across movement trials, respectively, and that the recently proposed inverse internal model is akin to adaptive sliding control originally for robotic manipulator applications. Modular internal models' architecture for multiple motor tasks is a form of multi-model adaptive control. Stochastic methods, such as generalized predictive control, reinforcement learning, Bayesian learning and Hebbian feedback covariance learning, are reviewed and their possible relevance to motor control is discussed. Possible applicability of a Luenberger observer and an extended Kalman filter to state estimation problems—such as sensorimotor prediction or the resolution of vestibular sensory ambiguity—is also discussed. The important role played by vestibular system identification in postural control suggests an indirect adaptive control scheme whereby system states or parameters are explicitly estimated prior to the implementation of control. This interdisciplinary framework should facilitate the experimental elucidation of the mechanisms of internal models in sensorimotor systems and the reverse engineering of such neural mechanisms into novel brain-inspired adaptive control paradigms in future.
ERIC Educational Resources Information Center
Firth, Nola; Frydenberg, Erica; Greaves, Daryl
2008-01-01
This study explored the effect of a coping program and a teacher feedback intervention on perceived control and adaptive coping for 98 adolescent students who had specific learning disabilities. The coping program was modified to build personal control and to address the needs of students who have specific learning disabilities. The teacher…
Adaptive filter design using recurrent cerebellar model articulation controller.
Lin, Chih-Min; Chen, Li-Yang; Yeung, Daniel S
2010-07-01
A novel adaptive filter is proposed using a recurrent cerebellar-model-articulation-controller (CMAC). The proposed locally recurrent globally feedforward recurrent CMAC (RCMAC) has favorable properties of small size, good generalization, rapid learning, and dynamic response, thus it is more suitable for high-speed signal processing. To provide fast training, an efficient parameter learning algorithm based on the normalized gradient descent method is presented, in which the learning rates are on-line adapted. Then the Lyapunov function is utilized to derive the conditions of the adaptive learning rates, so the stability of the filtering error can be guaranteed. To demonstrate the performance of the proposed adaptive RCMAC filter, it is applied to a nonlinear channel equalization system and an adaptive noise cancelation system. The advantages of the proposed filter over other adaptive filters are verified through simulations.
Learning for intelligent mobile robots
NASA Astrophysics Data System (ADS)
Hall, Ernest L.; Liao, Xiaoqun; Alhaj Ali, Souma M.
2003-10-01
Unlike intelligent industrial robots which often work in a structured factory setting, intelligent mobile robots must often operate in an unstructured environment cluttered with obstacles and with many possible action paths. However, such machines have many potential applications in medicine, defense, industry and even the home that make their study important. Sensors such as vision are needed. However, in many applications some form of learning is also required. The purpose of this paper is to present a discussion of recent technical advances in learning for intelligent mobile robots. During the past 20 years, the use of intelligent industrial robots that are equipped not only with motion control systems but also with sensors such as cameras, laser scanners, or tactile sensors that permit adaptation to a changing environment has increased dramatically. However, relatively little has been done concerning learning. Adaptive and robust control permits one to achieve point to point and controlled path operation in a changing environment. This problem can be solved with a learning control. In the unstructured environment, the terrain and consequently the load on the robot"s motors are constantly changing. Learning the parameters of a proportional, integral and derivative controller (PID) and artificial neural network provides an adaptive and robust control. Learning may also be used for path following. Simulations that include learning may be conducted to see if a robot can learn its way through a cluttered array of obstacles. If a situation is performed repetitively, then learning can also be used in the actual application. To reach an even higher degree of autonomous operation, a new level of learning is required. Recently learning theories such as the adaptive critic have been proposed. In this type of learning a critic provides a grade to the controller of an action module such as a robot. The creative control process is used that is "beyond the adaptive critic." A mathematical model of the creative control process is presented that illustrates the use for mobile robots. Examples from a variety of intelligent mobile robot applications are also presented. The significance of this work is in providing a greater understanding of the applications of learning to mobile robots that could lead to many applications.
Fault-tolerant nonlinear adaptive flight control using sliding mode online learning.
Krüger, Thomas; Schnetter, Philipp; Placzek, Robin; Vörsmann, Peter
2012-08-01
An expanded nonlinear model inversion flight control strategy using sliding mode online learning for neural networks is presented. The proposed control strategy is implemented for a small unmanned aircraft system (UAS). This class of aircraft is very susceptible towards nonlinearities like atmospheric turbulence, model uncertainties and of course system failures. Therefore, these systems mark a sensible testbed to evaluate fault-tolerant, adaptive flight control strategies. Within this work the concept of feedback linearization is combined with feed forward neural networks to compensate for inversion errors and other nonlinear effects. Backpropagation-based adaption laws of the network weights are used for online training. Within these adaption laws the standard gradient descent backpropagation algorithm is augmented with the concept of sliding mode control (SMC). Implemented as a learning algorithm, this nonlinear control strategy treats the neural network as a controlled system and allows a stable, dynamic calculation of the learning rates. While considering the system's stability, this robust online learning method therefore offers a higher speed of convergence, especially in the presence of external disturbances. The SMC-based flight controller is tested and compared with the standard gradient descent backpropagation algorithm in the presence of system failures. Copyright © 2012 Elsevier Ltd. All rights reserved.
Applying Learning Analytics to Investigate Timed Release in Online Learning
ERIC Educational Resources Information Center
Martin, Florence; Whitmer, John C.
2016-01-01
Adaptive learning gives learners control of context, pace, and scope of their learning experience. This strategy can be implemented in online learning by using the "Adaptive Release" feature in learning management systems. The purpose of this study was to use learning analytics research methods to explore the extent to which the adaptive…
Adaptive and predictive control of a simulated robot arm.
Tolu, Silvia; Vanegas, Mauricio; Garrido, Jesús A; Luque, Niceto R; Ros, Eduardo
2013-06-01
In this work, a basic cerebellar neural layer and a machine learning engine are embedded in a recurrent loop which avoids dealing with the motor error or distal error problem. The presented approach learns the motor control based on available sensor error estimates (position, velocity, and acceleration) without explicitly knowing the motor errors. The paper focuses on how to decompose the input into different components in order to facilitate the learning process using an automatic incremental learning model (locally weighted projection regression (LWPR) algorithm). LWPR incrementally learns the forward model of the robot arm and provides the cerebellar module with optimal pre-processed signals. We present a recurrent adaptive control architecture in which an adaptive feedback (AF) controller guarantees a precise, compliant, and stable control during the manipulation of objects. Therefore, this approach efficiently integrates a bio-inspired module (cerebellar circuitry) with a machine learning component (LWPR). The cerebellar-LWPR synergy makes the robot adaptable to changing conditions. We evaluate how this scheme scales for robot-arms of a high number of degrees of freedom (DOFs) using a simulated model of a robot arm of the new generation of light weight robots (LWRs).
Adaptive critic learning techniques for engine torque and air-fuel ratio control.
Liu, Derong; Javaherian, Hossein; Kovalenko, Olesia; Huang, Ting
2008-08-01
A new approach for engine calibration and control is proposed. In this paper, we present our research results on the implementation of adaptive critic designs for self-learning control of automotive engines. A class of adaptive critic designs that can be classified as (model-free) action-dependent heuristic dynamic programming is used in this research project. The goals of the present learning control design for automotive engines include improved performance, reduced emissions, and maintained optimum performance under various operating conditions. Using the data from a test vehicle with a V8 engine, we developed a neural network model of the engine and neural network controllers based on the idea of approximate dynamic programming to achieve optimal control. We have developed and simulated self-learning neural network controllers for both engine torque (TRQ) and exhaust air-fuel ratio (AFR) control. The goal of TRQ control and AFR control is to track the commanded values. For both control problems, excellent neural network controller transient performance has been achieved.
A neural learning classifier system with self-adaptive constructivism for mobile robot control.
Hurst, Jacob; Bull, Larry
2006-01-01
For artificial entities to achieve true autonomy and display complex lifelike behavior, they will need to exploit appropriate adaptable learning algorithms. In this context adaptability implies flexibility guided by the environment at any given time and an open-ended ability to learn appropriate behaviors. This article examines the use of constructivism-inspired mechanisms within a neural learning classifier system architecture that exploits parameter self-adaptation as an approach to realize such behavior. The system uses a rule structure in which each rule is represented by an artificial neural network. It is shown that appropriate internal rule complexity emerges during learning at a rate controlled by the learner and that the structure indicates underlying features of the task. Results are presented in simulated mazes before moving to a mobile robot platform.
Neural robust stabilization via event-triggering mechanism and adaptive learning technique.
Wang, Ding; Liu, Derong
2018-06-01
The robust control synthesis of continuous-time nonlinear systems with uncertain term is investigated via event-triggering mechanism and adaptive critic learning technique. We mainly focus on combining the event-triggering mechanism with adaptive critic designs, so as to solve the nonlinear robust control problem. This can not only make better use of computation and communication resources, but also conduct controller design from the view of intelligent optimization. Through theoretical analysis, the nonlinear robust stabilization can be achieved by obtaining an event-triggered optimal control law of the nominal system with a newly defined cost function and a certain triggering condition. The adaptive critic technique is employed to facilitate the event-triggered control design, where a neural network is introduced as an approximator of the learning phase. The performance of the event-triggered robust control scheme is validated via simulation studies and comparisons. The present method extends the application domain of both event-triggered control and adaptive critic control to nonlinear systems possessing dynamical uncertainties. Copyright © 2018 Elsevier Ltd. All rights reserved.
Encoder-Decoder Optimization for Brain-Computer Interfaces
Merel, Josh; Pianto, Donald M.; Cunningham, John P.; Paninski, Liam
2015-01-01
Neuroprosthetic brain-computer interfaces are systems that decode neural activity into useful control signals for effectors, such as a cursor on a computer screen. It has long been recognized that both the user and decoding system can adapt to increase the accuracy of the end effector. Co-adaptation is the process whereby a user learns to control the system in conjunction with the decoder adapting to learn the user's neural patterns. We provide a mathematical framework for co-adaptation and relate co-adaptation to the joint optimization of the user's control scheme ("encoding model") and the decoding algorithm's parameters. When the assumptions of that framework are respected, co-adaptation cannot yield better performance than that obtainable by an optimal initial choice of fixed decoder, coupled with optimal user learning. For a specific case, we provide numerical methods to obtain such an optimized decoder. We demonstrate our approach in a model brain-computer interface system using an online prosthesis simulator, a simple human-in-the-loop pyschophysics setup which provides a non-invasive simulation of the BCI setting. These experiments support two claims: that users can learn encoders matched to fixed, optimal decoders and that, once learned, our approach yields expected performance advantages. PMID:26029919
Encoder-decoder optimization for brain-computer interfaces.
Merel, Josh; Pianto, Donald M; Cunningham, John P; Paninski, Liam
2015-06-01
Neuroprosthetic brain-computer interfaces are systems that decode neural activity into useful control signals for effectors, such as a cursor on a computer screen. It has long been recognized that both the user and decoding system can adapt to increase the accuracy of the end effector. Co-adaptation is the process whereby a user learns to control the system in conjunction with the decoder adapting to learn the user's neural patterns. We provide a mathematical framework for co-adaptation and relate co-adaptation to the joint optimization of the user's control scheme ("encoding model") and the decoding algorithm's parameters. When the assumptions of that framework are respected, co-adaptation cannot yield better performance than that obtainable by an optimal initial choice of fixed decoder, coupled with optimal user learning. For a specific case, we provide numerical methods to obtain such an optimized decoder. We demonstrate our approach in a model brain-computer interface system using an online prosthesis simulator, a simple human-in-the-loop pyschophysics setup which provides a non-invasive simulation of the BCI setting. These experiments support two claims: that users can learn encoders matched to fixed, optimal decoders and that, once learned, our approach yields expected performance advantages.
Rational metareasoning and the plasticity of cognitive control.
Lieder, Falk; Shenhav, Amitai; Musslick, Sebastian; Griffiths, Thomas L
2018-04-01
The human brain has the impressive capacity to adapt how it processes information to high-level goals. While it is known that these cognitive control skills are malleable and can be improved through training, the underlying plasticity mechanisms are not well understood. Here, we develop and evaluate a model of how people learn when to exert cognitive control, which controlled process to use, and how much effort to exert. We derive this model from a general theory according to which the function of cognitive control is to select and configure neural pathways so as to make optimal use of finite time and limited computational resources. The central idea of our Learned Value of Control model is that people use reinforcement learning to predict the value of candidate control signals of different types and intensities based on stimulus features. This model correctly predicts the learning and transfer effects underlying the adaptive control-demanding behavior observed in an experiment on visual attention and four experiments on interference control in Stroop and Flanker paradigms. Moreover, our model explained these findings significantly better than an associative learning model and a Win-Stay Lose-Shift model. Our findings elucidate how learning and experience might shape people's ability and propensity to adaptively control their minds and behavior. We conclude by predicting under which circumstances these learning mechanisms might lead to self-control failure.
Rational metareasoning and the plasticity of cognitive control
Shenhav, Amitai; Musslick, Sebastian; Griffiths, Thomas L.
2018-01-01
The human brain has the impressive capacity to adapt how it processes information to high-level goals. While it is known that these cognitive control skills are malleable and can be improved through training, the underlying plasticity mechanisms are not well understood. Here, we develop and evaluate a model of how people learn when to exert cognitive control, which controlled process to use, and how much effort to exert. We derive this model from a general theory according to which the function of cognitive control is to select and configure neural pathways so as to make optimal use of finite time and limited computational resources. The central idea of our Learned Value of Control model is that people use reinforcement learning to predict the value of candidate control signals of different types and intensities based on stimulus features. This model correctly predicts the learning and transfer effects underlying the adaptive control-demanding behavior observed in an experiment on visual attention and four experiments on interference control in Stroop and Flanker paradigms. Moreover, our model explained these findings significantly better than an associative learning model and a Win-Stay Lose-Shift model. Our findings elucidate how learning and experience might shape people’s ability and propensity to adaptively control their minds and behavior. We conclude by predicting under which circumstances these learning mechanisms might lead to self-control failure. PMID:29694347
Learning from adaptive neural dynamic surface control of strict-feedback systems.
Wang, Min; Wang, Cong
2015-06-01
Learning plays an essential role in autonomous control systems. However, how to achieve learning in the nonstationary environment for nonlinear systems is a challenging problem. In this paper, we present learning method for a class of n th-order strict-feedback systems by adaptive dynamic surface control (DSC) technology, which achieves the human-like ability of learning by doing and doing with learned knowledge. To achieve the learning, this paper first proposes stable adaptive DSC with auxiliary first-order filters, which ensures the boundedness of all the signals in the closed-loop system and the convergence of tracking errors in a finite time. With the help of DSC, the derivative of the filter output variable is used as the neural network (NN) input instead of traditional intermediate variables. As a result, the proposed adaptive DSC method reduces greatly the dimension of NN inputs, especially for high-order systems. After the stable DSC design, we decompose the stable closed-loop system into a series of linear time-varying perturbed subsystems. Using a recursive design, the recurrent property of NN input variables is easily verified since the complexity is overcome using DSC. Subsequently, the partial persistent excitation condition of the radial basis function NN is satisfied. By combining a state transformation, accurate approximations of the closed-loop system dynamics are recursively achieved in a local region along recurrent orbits. Then, the learning control method using the learned knowledge is proposed to achieve the closed-loop stability and the improved control performance. Simulation studies are performed to demonstrate the proposed scheme can not only reuse the learned knowledge to achieve the better control performance with the faster tracking convergence rate and the smaller tracking error but also greatly alleviate the computational burden because of reducing the number and complexity of NN input variables.
Beyond adaptive-critic creative learning for intelligent mobile robots
NASA Astrophysics Data System (ADS)
Liao, Xiaoqun; Cao, Ming; Hall, Ernest L.
2001-10-01
Intelligent industrial and mobile robots may be considered proven technology in structured environments. Teach programming and supervised learning methods permit solutions to a variety of applications. However, we believe that to extend the operation of these machines to more unstructured environments requires a new learning method. Both unsupervised learning and reinforcement learning are potential candidates for these new tasks. The adaptive critic method has been shown to provide useful approximations or even optimal control policies to non-linear systems. The purpose of this paper is to explore the use of new learning methods that goes beyond the adaptive critic method for unstructured environments. The adaptive critic is a form of reinforcement learning. A critic element provides only high level grading corrections to a cognition module that controls the action module. In the proposed system the critic's grades are modeled and forecasted, so that an anticipated set of sub-grades are available to the cognition model. The forecasting grades are interpolated and are available on the time scale needed by the action model. The success of the system is highly dependent on the accuracy of the forecasted grades and adaptability of the action module. Examples from the guidance of a mobile robot are provided to illustrate the method for simple line following and for the more complex navigation and control in an unstructured environment. The theory presented that is beyond the adaptive critic may be called creative theory. Creative theory is a form of learning that models the highest level of human learning - imagination. The application of the creative theory appears to not only be to mobile robots but also to many other forms of human endeavor such as educational learning and business forecasting. Reinforcement learning such as the adaptive critic may be applied to known problems to aid in the discovery of their solutions. The significance of creative theory is that it permits the discovery of the unknown problems, ones that are not yet recognized but may be critical to survival or success.
Adaptive critic autopilot design of bank-to-turn missiles using fuzzy basis function networks.
Lin, Chuan-Kai
2005-04-01
A new adaptive critic autopilot design for bank-to-turn missiles is presented. In this paper, the architecture of adaptive critic learning scheme contains a fuzzy-basis-function-network based associative search element (ASE), which is employed to approximate nonlinear and complex functions of bank-to-turn missiles, and an adaptive critic element (ACE) generating the reinforcement signal to tune the associative search element. In the design of the adaptive critic autopilot, the control law receives signals from a fixed gain controller, an ASE and an adaptive robust element, which can eliminate approximation errors and disturbances. Traditional adaptive critic reinforcement learning is the problem faced by an agent that must learn behavior through trial-and-error interactions with a dynamic environment, however, the proposed tuning algorithm can significantly shorten the learning time by online tuning all parameters of fuzzy basis functions and weights of ASE and ACE. Moreover, the weight updating law derived from the Lyapunov stability theory is capable of guaranteeing both tracking performance and stability. Computer simulation results confirm the effectiveness of the proposed adaptive critic autopilot.
NASA Technical Reports Server (NTRS)
Troudet, Terry; Merrill, Walter C.
1989-01-01
The ability of feed-forward neural net architectures to learn continuous-valued mappings in the presence of noise is demonstrated in relation to parameter identification and real-time adaptive control applications. Factors and parameters influencing the learning performance of such nets in the presence of noise are identified. Their effects are discussed through a computer simulation of the Back-Error-Propagation algorithm by taking the example of the cart-pole system controlled by a nonlinear control law. Adequate sampling of the state space is found to be essential for canceling the effect of the statistical fluctuations and allowing learning to take place.
Adaptive control of nonlinear system using online error minimum neural networks.
Jia, Chao; Li, Xiaoli; Wang, Kang; Ding, Dawei
2016-11-01
In this paper, a new learning algorithm named OEM-ELM (Online Error Minimized-ELM) is proposed based on ELM (Extreme Learning Machine) neural network algorithm and the spreading of its main structure. The core idea of this OEM-ELM algorithm is: online learning, evaluation of network performance, and increasing of the number of hidden nodes. It combines the advantages of OS-ELM and EM-ELM, which can improve the capability of identification and avoid the redundancy of networks. The adaptive control based on the proposed algorithm OEM-ELM is set up which has stronger adaptive capability to the change of environment. The adaptive control of chemical process Continuous Stirred Tank Reactor (CSTR) is also given for application. The simulation results show that the proposed algorithm with respect to the traditional ELM algorithm can avoid network redundancy and improve the control performance greatly. Copyright © 2016 ISA. Published by Elsevier Ltd. All rights reserved.
Adaptive Computerized Instruction.
ERIC Educational Resources Information Center
Ray, Roger D.; And Others
1995-01-01
Describes an artificially intelligent multimedia computerized instruction system capable of developing a conceptual image of what a student is learning while the student is learning it. It focuses on principles of learning and adaptive behavioral control systems theory upon which the system is designed and demonstrates multiple user modes.…
The lifecycle of e-learning course in the adaptive educational environment
NASA Astrophysics Data System (ADS)
Gustun, O. N.; Budaragin, N. V.
2017-01-01
In the article we have considered the lifecycle model of the e-learning course in the electronic educational environment. This model consists of three stages and nine phases. In order to implement the adaptive control of the learning process we have determined the actions which are necessary to undertake at different phases of the e-learning course lifecycle. The general characteristics of the SPACEL-technology is given for creating adaptive educational environments of the next generation.
Generalization in Adaptation to Stable and Unstable Dynamics
Kadiallah, Abdelhamid; Franklin, David W.; Burdet, Etienne
2012-01-01
Humans skillfully manipulate objects and tools despite the inherent instability. In order to succeed at these tasks, the sensorimotor control system must build an internal representation of both the force and mechanical impedance. As it is not practical to either learn or store motor commands for every possible future action, the sensorimotor control system generalizes a control strategy for a range of movements based on learning performed over a set of movements. Here, we introduce a computational model for this learning and generalization, which specifies how to learn feedforward muscle activity in a function of the state space. Specifically, by incorporating co-activation as a function of error into the feedback command, we are able to derive an algorithm from a gradient descent minimization of motion error and effort, subject to maintaining a stability margin. This algorithm can be used to learn to coordinate any of a variety of motor primitives such as force fields, muscle synergies, physical models or artificial neural networks. This model for human learning and generalization is able to adapt to both stable and unstable dynamics, and provides a controller for generating efficient adaptive motor behavior in robots. Simulation results exhibit predictions consistent with all experiments on learning of novel dynamics requiring adaptation of force and impedance, and enable us to re-examine some of the previous interpretations of experiments on generalization. PMID:23056191
Novel Hybrid Adaptive Controller for Manipulation in Complex Perturbation Environments
Smith, Alex M. C.; Yang, Chenguang; Ma, Hongbin; Culverhouse, Phil; Cangelosi, Angelo; Burdet, Etienne
2015-01-01
In this paper we present a hybrid control scheme, combining the advantages of task-space and joint-space control. The controller is based on a human-like adaptive design, which minimises both control effort and tracking error. Our novel hybrid adaptive controller has been tested in extensive simulations, in a scenario where a Baxter robot manipulator is affected by external disturbances in the form of interaction with the environment and tool-like end-effector perturbations. The results demonstrated improved performance in the hybrid controller over both of its component parts. In addition, we introduce a novel method for online adaptation of learning parameters, using the fuzzy control formalism to utilise expert knowledge from the experimenter. This mechanism of meta-learning induces further improvement in performance and avoids the need for tuning through trial testing. PMID:26029916
Enhancing Student Motivation and Learning within Adaptive Tutors
ERIC Educational Resources Information Center
Ostrow, Korinn S.
2015-01-01
My research is rooted in improving K-12 educational practice using motivational facets made possible through adaptive tutoring systems. In an attempt to isolate best practices within the science of learning, I conduct randomized controlled trials within ASSISTments, an online adaptive tutoring system that provides assistance and assessment to…
ERIC Educational Resources Information Center
Corbalan, Gemma; Kester, Liesbeth; van Merrienboer, Jeroen J. G.
2008-01-01
Complex skill acquisition by performing authentic learning tasks is constrained by limited working memory capacity [Baddeley, A. D. (1992). Working memory. "Science, 255", 556-559]. To prevent cognitive overload, task difficulty and support of each newly selected learning task can be adapted to the learner's competence level and perceived task…
Tutsoy, Onder; Barkana, Duygun Erol; Tugal, Harun
2018-05-01
In this paper, an adaptive controller is developed for discrete time linear systems that takes into account parametric uncertainty, internal-external non-parametric random uncertainties, and time varying control signal delay. Additionally, the proposed adaptive control is designed in such a way that it is utterly model free. Even though these properties are studied separately in the literature, they are not taken into account all together in adaptive control literature. The Q-function is used to estimate long-term performance of the proposed adaptive controller. Control policy is generated based on the long-term predicted value, and this policy searches an optimal stabilizing control signal for uncertain and unstable systems. The derived control law does not require an initial stabilizing control assumption as in the ones in the recent literature. Learning error, control signal convergence, minimized Q-function, and instantaneous reward are analyzed to demonstrate the stability and effectiveness of the proposed adaptive controller in a simulation environment. Finally, key insights on parameters convergence of the learning and control signals are provided. Copyright © 2018 ISA. Published by Elsevier Ltd. All rights reserved.
Jiang, Jiefeng; Beck, Jeffrey; Heller, Katherine; Egner, Tobias
2015-01-01
The anterior cingulate and lateral prefrontal cortices have been implicated in implementing context-appropriate attentional control, but the learning mechanisms underlying our ability to flexibly adapt the control settings to changing environments remain poorly understood. Here we show that human adjustments to varying control demands are captured by a reinforcement learner with a flexible, volatility-driven learning rate. Using model-based functional magnetic resonance imaging, we demonstrate that volatility of control demand is estimated by the anterior insula, which in turn optimizes the prediction of forthcoming demand in the caudate nucleus. The caudate's prediction of control demand subsequently guides the implementation of proactive and reactive attentional control in dorsal anterior cingulate and dorsolateral prefrontal cortices. These data enhance our understanding of the neuro-computational mechanisms of adaptive behaviour by connecting the classic cingulate-prefrontal cognitive control network to a subcortical control-learning mechanism that infers future demands by flexibly integrating remote and recent past experiences. PMID:26391305
Wei, Jianming; Zhang, Youan; Sun, Meimei; Geng, Baoliang
2017-09-01
This paper presents an adaptive iterative learning control scheme for a class of nonlinear systems with unknown time-varying delays and control direction preceded by unknown nonlinear backlash-like hysteresis. Boundary layer function is introduced to construct an auxiliary error variable, which relaxes the identical initial condition assumption of iterative learning control. For the controller design, integral Lyapunov function candidate is used, which avoids the possible singularity problem by introducing hyperbolic tangent funciton. After compensating for uncertainties with time-varying delays by combining appropriate Lyapunov-Krasovskii function with Young's inequality, an adaptive iterative learning control scheme is designed through neural approximation technique and Nussbaum function method. On the basis of the hyperbolic tangent function's characteristics, the system output is proved to converge to a small neighborhood of the desired trajectory by constructing Lyapunov-like composite energy function (CEF) in two cases, while keeping all the closed-loop signals bounded. Finally, a simulation example is presented to verify the effectiveness of the proposed approach. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
Flight Test Approach to Adaptive Control Research
NASA Technical Reports Server (NTRS)
Pavlock, Kate Maureen; Less, James L.; Larson, David Nils
2011-01-01
The National Aeronautics and Space Administration s Dryden Flight Research Center completed flight testing of adaptive controls research on a full-scale F-18 testbed. The validation of adaptive controls has the potential to enhance safety in the presence of adverse conditions such as structural damage or control surface failures. This paper describes the research interface architecture, risk mitigations, flight test approach and lessons learned of adaptive controls research.
Adaptive versus Learner Control in a Multiple Intelligence Learning Environment
ERIC Educational Resources Information Center
Kelly, Declan
2008-01-01
Within the field of technology enhanced learning, adaptive educational systems offer an advanced form of learning environment that attempts to meet the needs of different students. Such systems capture and represent, for each student, various characteristics such as knowledge and traits in an individual learner model. Subsequently, using the…
Shimansky, Yury P; Kang, Tao; He, Jiping
2004-02-01
A computational model of a learning system (LS) is described that acquires knowledge and skill necessary for optimal control of a multisegmental limb dynamics (controlled object or CO), starting from "knowing" only the dimensionality of the object's state space. It is based on an optimal control problem setup different from that of reinforcement learning. The LS solves the optimal control problem online while practicing the manipulation of CO. The system's functional architecture comprises several adaptive components, each of which incorporates a number of mapping functions approximated based on artificial neural nets. Besides the internal model of the CO's dynamics and adaptive controller that computes the control law, the LS includes a new type of internal model, the minimal cost (IM(mc)) of moving the controlled object between a pair of states. That internal model appears critical for the LS's capacity to develop an optimal movement trajectory. The IM(mc) interacts with the adaptive controller in a cooperative manner. The controller provides an initial approximation of an optimal control action, which is further optimized in real time based on the IM(mc). The IM(mc) in turn provides information for updating the controller. The LS's performance was tested on the task of center-out reaching to eight randomly selected targets with a 2DOF limb model. The LS reached an optimal level of performance in a few tens of trials. It also quickly adapted to movement perturbations produced by two different types of external force field. The results suggest that the proposed design of a self-optimized control system can serve as a basis for the modeling of motor learning that includes the formation and adaptive modification of the plan of a goal-directed movement.
Uncertainty, learning, and the optimal management of wildlife
Williams, B.K.
2001-01-01
Wildlife management is limited by uncontrolled and often unrecognized environmental variation, by limited capabilities to observe and control animal populations, and by a lack of understanding about the biological processes driving population dynamics. In this paper I describe a comprehensive framework for management that includes multiple models and likelihood values to account for structural uncertainty, along with stochastic factors to account for environmental variation, random sampling, and partial controllability. Adaptive optimization is developed in terms of the optimal control of incompletely understood populations, with the expected value of perfect information measuring the potential for improving control through learning. The framework for optimal adaptive control is generalized by including partial observability and non-adaptive, sample-based updating of model likelihoods. Passive adaptive management is derived as a special case of constrained adaptive optimization, representing a potentially efficient suboptimal alternative that nonetheless accounts for structural uncertainty.
Visually guided gait modifications for stepping over an obstacle: a bio-inspired approach.
Silva, Pedro; Matos, Vitor; Santos, Cristina P
2014-02-01
There is an increasing interest in conceiving robotic systems that are able to move and act in an unstructured and not predefined environment, for which autonomy and adaptability are crucial features. In nature, animals are autonomous biological systems, which often serve as bio-inspiration models, not only for their physical and mechanical properties, but also their control structures that enable adaptability and autonomy-for which learning is (at least) partially responsible. This work proposes a system which seeks to enable a quadruped robot to online learn to detect and to avoid stumbling on an obstacle in its path. The detection relies in a forward internal model that estimates the robot's perceptive information by exploring the locomotion repetitive nature. The system adapts the locomotion in order to place the robot optimally before attempting to step over the obstacle, avoiding any stumbling. Locomotion adaptation is achieved by changing control parameters of a central pattern generator (CPG)-based locomotion controller. The mechanism learns the necessary alterations to the stride length in order to adapt the locomotion by changing the required CPG parameter. Both learning tasks occur online and together define a sensorimotor map, which enables the robot to learn to step over the obstacle in its path. Simulation results show the feasibility of the proposed approach.
Miall, R Chris; Kitchen, Nick M; Nam, Se-Ho; Lefumat, Hannah; Renault, Alix G; Ørstavik, Kristin; Cole, Jonathan D; Sarlegna, Fabrice R
2018-05-19
It is uncertain how vision and proprioception contribute to adaptation of voluntary arm movements. In normal participants, adaptation to imposed forces is possible with or without vision, suggesting that proprioception is sufficient; in participants with proprioceptive loss (PL), adaptation is possible with visual feedback, suggesting that proprioception is unnecessary. In experiment 1 adaptation to, and retention of, perturbing forces were evaluated in three chronically deafferented participants. They made rapid reaching movements to move a cursor toward a visual target, and a planar robot arm applied orthogonal velocity-dependent forces. Trial-by-trial error correction was observed in all participants. Such adaptation has been characterized with a dual-rate model: a fast process that learns quickly, but retains poorly and a slow process that learns slowly and retains well. Experiment 2 showed that the PL participants had large individual differences in learning and retention rates compared to normal controls. Experiment 3 tested participants' perception of applied forces. With visual feedback, the PL participants could report the perturbation's direction as well as controls; without visual feedback, thresholds were elevated. Experiment 4 showed, in healthy participants, that force direction could be estimated from head motion, at levels close to the no-vision threshold for the PL participants. Our results show that proprioceptive loss influences perception, motor control and adaptation but that proprioception from the moving limb is not essential for adaptation to, or detection of, force fields. The differences in learning and retention seen between the three deafferented participants suggest that they achieve these tasks in idiosyncratic ways after proprioceptive loss, possibly integrating visual and vestibular information with individual cognitive strategies.
Reinforcement Learning with Orthonormal Basis Adaptation Based on Activity-Oriented Index Allocation
NASA Astrophysics Data System (ADS)
Satoh, Hideki
An orthonormal basis adaptation method for function approximation was developed and applied to reinforcement learning with multi-dimensional continuous state space. First, a basis used for linear function approximation of a control function is set to an orthonormal basis. Next, basis elements with small activities are replaced with other candidate elements as learning progresses. As this replacement is repeated, the number of basis elements with large activities increases. Example chaos control problems for multiple logistic maps were solved, demonstrating that the method for adapting an orthonormal basis can modify a basis while holding the orthonormality in accordance with changes in the environment to improve the performance of reinforcement learning and to eliminate the adverse effects of redundant noisy states.
The role of strategies in motor learning
Taylor, Jordan A.; Ivry, Richard B.
2015-01-01
There has been renewed interest in the role of strategies in sensorimotor learning. The combination of new behavioral methods and computational methods has begun to unravel the interaction between processes related to strategic control and processes related to motor adaptation. These processes may operate on very different error signals. Strategy learning is sensitive to goal-based performance error. In contrast, adaptation is sensitive to prediction errors between the desired and actual consequences of a planned movement. The former guides what the desired movement should be, whereas the latter guides how to implement the desired movement. Whereas traditional approaches have favored serial models in which an initial strategy-based phase gives way to more automatized forms of control, it now seems that strategic and adaptive processes operate with considerable independence throughout learning, although the relative weight given the two processes will shift with changes in performance. As such, skill acquisition involves the synergistic engagement of strategic and adaptive processes. PMID:22329960
Neural controller for adaptive movements with unforeseen payloads.
Kuperstein, M; Wang, J
1990-01-01
A theory and computer simulation of a neural controller that learns to move and position a link carrying an unforeseen payload accurately are presented. The neural controller learns adaptive dynamic control from its own experience. It does not use information about link mass, link length, or direction of gravity, and it uses only indirect uncalibrated information about payload and actuator limits. Its average positioning accuracy across a large range of payloads after learning is 3% of the positioning range. This neural controller can be used as a basis for coordinating any number of sensory inputs with limbs of any number of joints. The feedforward nature of control allows parallel implementation in real time across multiple joints.
Huang, Qi; Yang, Dapeng; Jiang, Li; Zhang, Huajie; Liu, Hong; Kotani, Kiyoshi
2017-01-01
Performance degradation will be caused by a variety of interfering factors for pattern recognition-based myoelectric control methods in the long term. This paper proposes an adaptive learning method with low computational cost to mitigate the effect in unsupervised adaptive learning scenarios. We presents a particle adaptive classifier (PAC), by constructing a particle adaptive learning strategy and universal incremental least square support vector classifier (LS-SVC). We compared PAC performance with incremental support vector classifier (ISVC) and non-adapting SVC (NSVC) in a long-term pattern recognition task in both unsupervised and supervised adaptive learning scenarios. Retraining time cost and recognition accuracy were compared by validating the classification performance on both simulated and realistic long-term EMG data. The classification results of realistic long-term EMG data showed that the PAC significantly decreased the performance degradation in unsupervised adaptive learning scenarios compared with NSVC (9.03% ± 2.23%, p < 0.05) and ISVC (13.38% ± 2.62%, p = 0.001), and reduced the retraining time cost compared with ISVC (2 ms per updating cycle vs. 50 ms per updating cycle). PMID:28608824
Huang, Qi; Yang, Dapeng; Jiang, Li; Zhang, Huajie; Liu, Hong; Kotani, Kiyoshi
2017-06-13
Performance degradation will be caused by a variety of interfering factors for pattern recognition-based myoelectric control methods in the long term. This paper proposes an adaptive learning method with low computational cost to mitigate the effect in unsupervised adaptive learning scenarios. We presents a particle adaptive classifier (PAC), by constructing a particle adaptive learning strategy and universal incremental least square support vector classifier (LS-SVC). We compared PAC performance with incremental support vector classifier (ISVC) and non-adapting SVC (NSVC) in a long-term pattern recognition task in both unsupervised and supervised adaptive learning scenarios. Retraining time cost and recognition accuracy were compared by validating the classification performance on both simulated and realistic long-term EMG data. The classification results of realistic long-term EMG data showed that the PAC significantly decreased the performance degradation in unsupervised adaptive learning scenarios compared with NSVC (9.03% ± 2.23%, p < 0.05) and ISVC (13.38% ± 2.62%, p = 0.001), and reduced the retraining time cost compared with ISVC (2 ms per updating cycle vs. 50 ms per updating cycle).
Hexacopter trajectory control using a neural network
NASA Astrophysics Data System (ADS)
Artale, V.; Collotta, M.; Pau, G.; Ricciardello, A.
2013-10-01
The modern flight control systems are complex due to their non-linear nature. In fact, modern aerospace vehicles are expected to have non-conventional flight envelopes and, then, they must guarantee a high level of robustness and adaptability in order to operate in uncertain environments. Neural Networks (NN), with real-time learning capability, for flight control can be used in applications with manned or unmanned aerial vehicles. Indeed, using proven lower level control algorithms with adaptive elements that exhibit long term learning could help in achieving better adaptation performance while performing aggressive maneuvers. In this paper we show a mathematical modeling and a Neural Network for a hexacopter dynamics in order to develop proper methods for stabilization and trajectory control.
Recent developments in learning control and system identification for robots and structures
NASA Technical Reports Server (NTRS)
Phan, M.; Juang, J.-N.; Longman, R. W.
1990-01-01
This paper reviews recent results in learning control and learning system identification, with particular emphasis on discrete-time formulation, and their relation to adaptive theory. Related continuous-time results are also discussed. Among the topics presented are proportional, derivative, and integral learning controllers, time-domain formulation of discrete learning algorithms. Newly developed techniques are described including the concept of the repetition domain, and the repetition domain formulation of learning control by linear feedback, model reference learning control, indirect learning control with parameter estimation, as well as related basic concepts, recursive and non-recursive methods for learning identification.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Aziz, H. M. Abdul; Zhu, Feng; Ukkusuri, Satish V.
Here, this research applies R-Markov Average Reward Technique based reinforcement learning (RL) algorithm, namely RMART, for vehicular signal control problem leveraging information sharing among signal controllers in connected vehicle environment. We implemented the algorithm in a network of 18 signalized intersections and compare the performance of RMART with fixed, adaptive, and variants of the RL schemes. Results show significant improvement in system performance for RMART algorithm with information sharing over both traditional fixed signal timing plans and real time adaptive control schemes. Additionally, the comparison with reinforcement learning algorithms including Q learning and SARSA indicate that RMART performs better atmore » higher congestion levels. Further, a multi-reward structure is proposed that dynamically adjusts the reward function with varying congestion states at the intersection. Finally, the results from test networks show significant reduction in emissions (CO, CO 2, NO x, VOC, PM 10) when RL algorithms are implemented compared to fixed signal timings and adaptive schemes.« less
Adaptive learning and control for MIMO system based on adaptive dynamic programming.
Fu, Jian; He, Haibo; Zhou, Xinmin
2011-07-01
Adaptive dynamic programming (ADP) is a promising research field for design of intelligent controllers, which can both learn on-the-fly and exhibit optimal behavior. Over the past decades, several generations of ADP design have been proposed in the literature, which have demonstrated many successful applications in various benchmarks and industrial applications. While many of the existing researches focus on multiple-inputs-single-output system with steepest descent search, in this paper we investigate a generalized multiple-input-multiple-output (GMIMO) ADP design for online learning and control, which is more applicable to a wide range of practical real-world applications. Furthermore, an improved weight-updating algorithm based on recursive Levenberg-Marquardt methods is presented and embodied in the GMIMO approach to improve its performance. Finally, we test the performance of this approach based on a practical complex system, namely, the learning and control of the tension and height of the looper system in a hot strip mill. Experimental results demonstrate that the proposed approach can achieve effective and robust performance.
Multi Car Elevator Control by using Learning Automaton
NASA Astrophysics Data System (ADS)
Shiraishi, Kazuaki; Hamagami, Tomoki; Hirata, Hironori
We study an adaptive control technique for multi car elevators (MCEs) by adopting learning automatons (LAs.) The MCE is a high performance and a near-future elevator system with multi shafts and multi cars. A strong point of the system is that realizing a large carrying capacity in small shaft area. However, since the operation is too complicated, realizing an efficient MCE control is difficult for top-down approaches. For example, “bunching up together" is one of the typical phenomenon in a simple traffic environment like the MCE. Furthermore, an adapting to varying environment in configuration requirement is a serious issue in a real elevator service. In order to resolve these issues, having an autonomous behavior is required to the control system of each car in MCE system, so that the learning automaton, as the solutions for this requirement, is supposed to be appropriate for the simple traffic control. First, we assign a stochastic automaton (SA) to each car control system. Then, each SA varies its stochastic behavior distributions for adapting to environment in which its policy is evaluated with each passenger waiting times. That is LA which learns the environment autonomously. Using the LA based control technique, the MCE operation efficiency is evaluated through simulation experiments. Results show the technique enables reducing waiting times efficiently, and we confirm the system can adapt to the dynamic environment.
An adaptive learning control system for large flexible structures
NASA Technical Reports Server (NTRS)
Thau, F. E.
1985-01-01
The objective of the research has been to study the design of adaptive/learning control systems for the control of large flexible structures. In the first activity an adaptive/learning control methodology for flexible space structures was investigated. The approach was based on using a modal model of the flexible structure dynamics and an output-error identification scheme to identify modal parameters. In the second activity, a least-squares identification scheme was proposed for estimating both modal parameters and modal-to-actuator and modal-to-sensor shape functions. The technique was applied to experimental data obtained from the NASA Langley beam experiment. In the third activity, a separable nonlinear least-squares approach was developed for estimating the number of excited modes, shape functions, modal parameters, and modal amplitude and velocity time functions for a flexible structure. In the final research activity, a dual-adaptive control strategy was developed for regulating the modal dynamics and identifying modal parameters of a flexible structure. A min-max approach was used for finding an input to provide modal parameter identification while not exceeding reasonable bounds on modal displacement.
On the role of model-based monitoring for adaptive planning under uncertainty
NASA Astrophysics Data System (ADS)
Raso, Luciano; Kwakkel, Jan; Timmermans, Jos; Haasnoot, Mariolijn
2016-04-01
Adaptive plans, designed to anticipate and respond to an unfolding uncertain future, have found a fertile application domain in the planning of deltas that are exposed to rapid socioeconomic development and climate change. Adaptive planning, under the moniker of adaptive delta management, is used in the Dutch Delta Program for developing a nation-wide plan to prepare for uncertain climate change and socio-economic developments. Scientifically, adaptive delta management relies heavily on Dynamic Adaptive Policy Pathways. Currently, in the Netherlands the focus is shifting towards implementing the adaptive delta plan. This shift is especially relevant because the efficacy of adaptive plans hinges on monitoring on-going developments and ensuring that actions are indeed taken if and when necessary. In the design of an effective monitoring system for an adaptive plan, three challenges have to be confronted: • Shadow of the past: The development of adaptive plans and the design of their monitoring system relies heavily on current knowledge of the system, and current beliefs about plausible future developments. A static monitoring system is therefore exposed to the exact same uncertainties one tries to address through adaptive planning. • Inhibition of learning: Recent applications of adaptive planning tend to overlook the importance of learning and new information, and fail to account for this explicitly in the design of adaptive plans. • Challenge of surprise: Adaptive policies are designed in light of the current foreseen uncertainties. However, developments that are not considered during the design phase as being plausible could still substantially affect the performance of adaptive policies. The shadow of the past, the inhibition of learning, and the challenge of surprise taken together suggest that there is a need for redesigning the concepts of monitoring and evaluation to support the implementation of adaptive plans. Innovations from control theory, triggered by the challenge of uncertainty in operational control, may offer solutions from which monitoring for adaptive planning can benefit. Specifically: (i) in control, observations are incorporated into the model through data assimilation, updating the present state, boundary conditions, and parameters based on new observations, diminishing the shadow of the past; (ii) adaptive control is a way to modify the characteristics of the internal model, incorporating new knowledge on the system, countervailing the inhibition of learning; and (iii) in closed-loop control, a continuous system update equips the controller with "inherent robustness", i.e. to capacity to adapts to new conditions even when these were not initially considered. We aim to explore how inherent robustness addresses the challenge of surprise. Innovations in model-based control might help to improve and adapt the models used to support adaptive delta management to new information (reducing uncertainty). Moreover, this would offer a starting point for using these models not only in the design of adaptive plans, but also as part of the monitoring. The proposed research requires multidisciplinary cooperation between control theory, the policy sciences, and integrated assessment modeling.
On Mixed Data and Event Driven Design for Adaptive-Critic-Based Nonlinear $H_{\\infty}$ Control.
Wang, Ding; Mu, Chaoxu; Liu, Derong; Ma, Hongwen
2018-04-01
In this paper, based on the adaptive critic learning technique, the control for a class of unknown nonlinear dynamic systems is investigated by adopting a mixed data and event driven design approach. The nonlinear control problem is formulated as a two-player zero-sum differential game and the adaptive critic method is employed to cope with the data-based optimization. The novelty lies in that the data driven learning identifier is combined with the event driven design formulation, in order to develop the adaptive critic controller, thereby accomplishing the nonlinear control. The event driven optimal control law and the time driven worst case disturbance law are approximated by constructing and tuning a critic neural network. Applying the event driven feedback control, the closed-loop system is built with stability analysis. Simulation studies are conducted to verify the theoretical results and illustrate the control performance. It is significant to observe that the present research provides a new avenue of integrating data-based control and event-triggering mechanism into establishing advanced adaptive critic systems.
Full-Scale Flight Research Testbeds: Adaptive and Intelligent Control
NASA Technical Reports Server (NTRS)
Pahle, Joe W.
2008-01-01
This viewgraph presentation describes the adaptive and intelligent control methods used for aircraft survival. The contents include: 1) Motivation for Adaptive Control; 2) Integrated Resilient Aircraft Control Project; 3) Full-scale Flight Assets in Use for IRAC; 4) NASA NF-15B Tail Number 837; 5) Gen II Direct Adaptive Control Architecture; 6) Limited Authority System; and 7) 837 Flight Experiments. A simulated destabilization failure analysis along with experience and lessons learned are also presented.
Enhanced Muscle Afferent Signals during Motor Learning in Humans.
Dimitriou, Michael
2016-04-25
Much has been revealed concerning human motor learning at the behavioral level [1, 2], but less is known about changes in the involved neural circuits and signals. By examining muscle spindle responses during a classic visuomotor adaptation task [3-6] performed by fully alert humans, I found substantial modulation of sensory afferent signals as a function of adaptation state. Specifically, spindle control was independent of concurrent muscle activity but was specific to movement direction (representing muscle lengthening versus shortening) and to different stages of learning. Increased spindle afferent responses to muscle stretch occurring early during learning reflected individual error size and were negatively related to subsequent antagonist activity (i.e., 60-80 ms thereafter). Relative increases in tonic afferent output early during learning were predictive of the subjects' adaptation rate. I also found that independent spindle control during sensory realignment (the "washout" stage) induced afferent signal "linearization" with respect to muscle length (i.e., signals were more tuned to hand position). The results demonstrate for the first time that motor learning also involves independent and state-related modulation of sensory mechanoreceptor signals. The current findings suggest that adaptive motor performance also relies on the independent control of sensors, not just of muscles. I propose that the "γ" motor system innervating spindles acts to facilitate the acquisition and extraction of task-relevant information at the early stages of sensorimotor adaptation. This designates a more active and targeted role for the human proprioceptive system during motor learning. Copyright © 2016 Elsevier Ltd. All rights reserved.
NASA Technical Reports Server (NTRS)
Jacklin, Stephen A.; Schumann, Johann; Guenther, Kurt; Bosworth, John
2006-01-01
Adaptive control technologies that incorporate learning algorithms have been proposed to enable autonomous flight control and to maintain vehicle performance in the face of unknown, changing, or poorly defined operating environments [1-2]. At the present time, however, it is unknown how adaptive algorithms can be routinely verified, validated, and certified for use in safety-critical applications. Rigorous methods for adaptive software verification end validation must be developed to ensure that. the control software functions as required and is highly safe and reliable. A large gap appears to exist between the point at which control system designers feel the verification process is complete, and when FAA certification officials agree it is complete. Certification of adaptive flight control software verification is complicated by the use of learning algorithms (e.g., neural networks) and degrees of system non-determinism. Of course, analytical efforts must be made in the verification process to place guarantees on learning algorithm stability, rate of convergence, and convergence accuracy. However, to satisfy FAA certification requirements, it must be demonstrated that the adaptive flight control system is also able to fail and still allow the aircraft to be flown safely or to land, while at the same time providing a means of crew notification of the (impending) failure. It was for this purpose that the NASA Ames Confidence Tool was developed [3]. This paper presents the Confidence Tool as a means of providing in-flight software assurance monitoring of an adaptive flight control system. The paper will present the data obtained from flight testing the tool on a specially modified F-15 aircraft designed to simulate loss of flight control faces.
Learning-Based Adaptive Optimal Tracking Control of Strict-Feedback Nonlinear Systems.
Gao, Weinan; Jiang, Zhong-Ping; Weinan Gao; Zhong-Ping Jiang; Gao, Weinan; Jiang, Zhong-Ping
2018-06-01
This paper proposes a novel data-driven control approach to address the problem of adaptive optimal tracking for a class of nonlinear systems taking the strict-feedback form. Adaptive dynamic programming (ADP) and nonlinear output regulation theories are integrated for the first time to compute an adaptive near-optimal tracker without any a priori knowledge of the system dynamics. Fundamentally different from adaptive optimal stabilization problems, the solution to a Hamilton-Jacobi-Bellman (HJB) equation, not necessarily a positive definite function, cannot be approximated through the existing iterative methods. This paper proposes a novel policy iteration technique for solving positive semidefinite HJB equations with rigorous convergence analysis. A two-phase data-driven learning method is developed and implemented online by ADP. The efficacy of the proposed adaptive optimal tracking control methodology is demonstrated via a Van der Pol oscillator with time-varying exogenous signals.
Fuzzy Adaptive Control for Intelligent Autonomous Space Exploration Problems
NASA Technical Reports Server (NTRS)
Esogbue, Augustine O.
1998-01-01
The principal objective of the research reported here is the re-design, analysis and optimization of our newly developed neural network fuzzy adaptive controller model for complex processes capable of learning fuzzy control rules using process data and improving its control through on-line adaption. The learned improvement is according to a performance objective function that provides evaluative feedback; this performance objective is broadly defined to meet long-range goals over time. Although fuzzy control had proven effective for complex, nonlinear, imprecisely-defined processes for which standard models and controls are either inefficient, impractical or cannot be derived, the state of the art prior to our work showed that procedures for deriving fuzzy control, however, were mostly ad hoc heuristics. The learning ability of neural networks was exploited to systematically derive fuzzy control and permit on-line adaption and in the process optimize control. The operation of neural networks integrates very naturally with fuzzy logic. The neural networks which were designed and tested using simulation software and simulated data, followed by realistic industrial data were reconfigured for application on several platforms as well as for the employment of improved algorithms. The statistical procedures of the learning process were investigated and evaluated with standard statistical procedures (such as ANOVA, graphical analysis of residuals, etc.). The computational advantage of dynamic programming-like methods of optimal control was used to permit on-line fuzzy adaptive control. Tests for the consistency, completeness and interaction of the control rules were applied. Comparisons to other methods and controllers were made so as to identify the major advantages of the resulting controller model. Several specific modifications and extensions were made to the original controller. Additional modifications and explorations have been proposed for further study. Some of these are in progress in our laboratory while others await additional support. All of these enhancements will improve the attractiveness of the controller as an effective tool for the on line control of an array of complex process environments.
Adaptive Critic Nonlinear Robust Control: A Survey.
Wang, Ding; He, Haibo; Liu, Derong
2017-10-01
Adaptive dynamic programming (ADP) and reinforcement learning are quite relevant to each other when performing intelligent optimization. They are both regarded as promising methods involving important components of evaluation and improvement, at the background of information technology, such as artificial intelligence, big data, and deep learning. Although great progresses have been achieved and surveyed when addressing nonlinear optimal control problems, the research on robustness of ADP-based control strategies under uncertain environment has not been fully summarized. Hence, this survey reviews the recent main results of adaptive-critic-based robust control design of continuous-time nonlinear systems. The ADP-based nonlinear optimal regulation is reviewed, followed by robust stabilization of nonlinear systems with matched uncertainties, guaranteed cost control design of unmatched plants, and decentralized stabilization of interconnected systems. Additionally, further comprehensive discussions are presented, including event-based robust control design, improvement of the critic learning rule, nonlinear H ∞ control design, and several notes on future perspectives. By applying the ADP-based optimal and robust control methods to a practical power system and an overhead crane plant, two typical examples are provided to verify the effectiveness of theoretical results. Overall, this survey is beneficial to promote the development of adaptive critic control methods with robustness guarantee and the construction of higher level intelligent systems.
Adaptive Inverse Control for Rotorcraft Vibration Reduction
NASA Technical Reports Server (NTRS)
Jacklin, Stephen A.
1985-01-01
This thesis extends the Least Mean Square (LMS) algorithm to solve the mult!ple-input, multiple-output problem of alleviating N/Rev (revolutions per minute by number of blades) helicopter fuselage vibration by means of adaptive inverse control. A frequency domain locally linear model is used to represent the transfer matrix relating the higher harmonic pitch control inputs to the harmonic vibration outputs to be controlled. By using the inverse matrix as the controller gain matrix, an adaptive inverse regulator is formed to alleviate the N/Rev vibration. The stability and rate of convergence properties of the extended LMS algorithm are discussed. It is shown that the stability ranges for the elements of the stability gain matrix are directly related to the eigenvalues of the vibration signal information matrix for the learning phase, but not for the control phase. The overall conclusion is that the LMS adaptive inverse control method can form a robust vibration control system, but will require some tuning of the input sensor gains, the stability gain matrix, and the amount of control relaxation to be used. The learning curve of the controller during the learning phase is shown to be quantitatively close to that predicted by averaging the learning curves of the normal modes. For higher order transfer matrices, a rough estimate of the inverse is needed to start the algorithm efficiently. The simulation results indicate that the factor which most influences LMS adaptive inverse control is the product of the control relaxation and the the stability gain matrix. A small stability gain matrix makes the controller less sensitive to relaxation selection, and permits faster and more stable vibration reduction, than by choosing the stability gain matrix large and the control relaxation term small. It is shown that the best selections of the stability gain matrix elements and the amount of control relaxation is basically a compromise between slow, stable convergence and fast convergence with increased possibility of unstable identification. In the simulation studies, the LMS adaptive inverse control algorithm is shown to be capable of adapting the inverse (controller) matrix to track changes in the flight conditions. The algorithm converges quickly for moderate disturbances, while taking longer for larger disturbances. Perfect knowledge of the inverse matrix is not required for good control of the N/Rev vibration. However it is shown that measurement noise will prevent the LMS adaptive inverse control technique from controlling the vibration, unless the signal averaging method presented is incorporated into the algorithm.
Visuomotor coordination and cortical connectivity of modular motor learning.
Burgos, Pablo I; Mariman, Juan J; Makeig, Scott; Rivera-Lillo, Gonzalo; Maldonado, Pedro E
2018-05-15
The ability to transfer sensorimotor skill components to new actions and the capacity to use skill components from whole actions are characteristic of the adaptability of the human sensorimotor system. However, behavioral evidence suggests complex limitations for transfer after combined or modular learning of motor adaptations. Also, to date, only behavioral analysis of the consequences of the modular learning has been reported, with little understanding of the sensorimotor mechanisms of control and the interaction between cortical areas. We programmed a video game with distorted kinematic and dynamic features to test the ability to combine sensorimotor skill components learned modularly (composition) and the capacity to use separate sensorimotor skill components learned in combination (decomposition). We examined motor performance, eye-hand coordination, and EEG connectivity. When tested for integrated learning, we found that combined practice initially performed better than separated practice, but differences disappeared after integrated practice. Separate learning promotes fewer anticipatory control mechanisms (depending more on feedback control), evidenced in a lower gaze leading behavior and in higher connectivity between visual and premotor domains, in comparison with the combined practice. The sensorimotor system can acquire motor modules in a separated or integrated manner. However, the system appears to require integrated practice to coordinate the adaptations with the skill learning and the networks involved in the integrated behavior. This integration seems to be related to the acquisition of anticipatory mechanism of control and with the decrement of feedback control. © 2018 Wiley Periodicals, Inc.
Integration of Online Parameter Identification and Neural Network for In-Flight Adaptive Control
NASA Technical Reports Server (NTRS)
Hageman, Jacob J.; Smith, Mark S.; Stachowiak, Susan
2003-01-01
An indirect adaptive system has been constructed for robust control of an aircraft with uncertain aerodynamic characteristics. This system consists of a multilayer perceptron pre-trained neural network, online stability and control derivative identification, a dynamic cell structure online learning neural network, and a model following control system based on the stochastic optimal feedforward and feedback technique. The pre-trained neural network and model following control system have been flight-tested, but the online parameter identification and online learning neural network are new additions used for in-flight adaptation of the control system model. A description of the modification and integration of these two stand-alone software packages into the complete system in preparation for initial flight tests is presented. Open-loop results using both simulation and flight data, as well as closed-loop performance of the complete system in a nonlinear, six-degree-of-freedom, flight validated simulation, are analyzed. Results show that this online learning system, in contrast to the nonlearning system, has the ability to adapt to changes in aerodynamic characteristics in a real-time, closed-loop, piloted simulation, resulting in improved flying qualities.
Flight Approach to Adaptive Control Research
NASA Technical Reports Server (NTRS)
Pavlock, Kate Maureen; Less, James L.; Larson, David Nils
2011-01-01
The National Aeronautics and Space Administration's Dryden Flight Research Center completed flight testing of adaptive controls research on a full-scale F-18 testbed. The testbed served as a full-scale vehicle to test and validate adaptive flight control research addressing technical challenges involved with reducing risk to enable safe flight in the presence of adverse conditions such as structural damage or control surface failures. This paper describes the research interface architecture, risk mitigations, flight test approach and lessons learned of adaptive controls research.
Closing the Certification Gaps in Adaptive Flight Control Software
NASA Technical Reports Server (NTRS)
Jacklin, Stephen A.
2008-01-01
Over the last five decades, extensive research has been performed to design and develop adaptive control systems for aerospace systems and other applications where the capability to change controller behavior at different operating conditions is highly desirable. Although adaptive flight control has been partially implemented through the use of gain-scheduled control, truly adaptive control systems using learning algorithms and on-line system identification methods have not seen commercial deployment. The reason is that the certification process for adaptive flight control software for use in national air space has not yet been decided. The purpose of this paper is to examine the gaps between the state-of-the-art methodologies used to certify conventional (i.e., non-adaptive) flight control system software and what will likely to be needed to satisfy FAA airworthiness requirements. These gaps include the lack of a certification plan or process guide, the need to develop verification and validation tools and methodologies to analyze adaptive controller stability and convergence, as well as the development of metrics to evaluate adaptive controller performance at off-nominal flight conditions. This paper presents the major certification gap areas, a description of the current state of the verification methodologies, and what further research efforts will likely be needed to close the gaps remaining in current certification practices. It is envisioned that closing the gap will require certain advances in simulation methods, comprehensive methods to determine learning algorithm stability and convergence rates, the development of performance metrics for adaptive controllers, the application of formal software assurance methods, the application of on-line software monitoring tools for adaptive controller health assessment, and the development of a certification case for adaptive system safety of flight.
Adaptive Fuzzy Bounded Control for Consensus of Multiple Strict-Feedback Nonlinear Systems.
Wang, Wei; Tong, Shaocheng
2018-02-01
This paper studies the adaptive fuzzy bounded control problem for leader-follower multiagent systems, where each follower is modeled by the uncertain nonlinear strict-feedback system. Combining the fuzzy approximation with the dynamic surface control, an adaptive fuzzy control scheme is developed to guarantee the output consensus of all agents under directed communication topologies. Different from the existing results, the bounds of the control inputs are known as a priori, and they can be determined by the feedback control gains. To realize smooth and fast learning, a predictor is introduced to estimate each error surface, and the corresponding predictor error is employed to learn the optimal fuzzy parameter vector. It is proved that the developed adaptive fuzzy control scheme guarantees the uniformly ultimate boundedness of the closed-loop systems, and the tracking error converges to a small neighborhood of the origin. The simulation results and comparisons are provided to show the validity of the control strategy presented in this paper.
MEMORY SYSTEMS STUDY. Annual Report No. 2, November 16, 1962 to November 15, 1963
DOE Office of Scientific and Technical Information (OSTI.GOV)
Peterson, G R; DeVries, R C; Melsa, J L
1964-10-31
S>Results of theoretical studies of learning control systems are presented. The need for definitions is discussed and definitions of successful, adaptive, and learning control systems are presented. The basic structural elements of learning control systems are discussed. The environmental characteristics of control situations in which learning may be applicable are discussed. Learning control systems are classified in accordance with the environmental situation in which they might operate. The structure and components suitable to various environmental situations are discussed. (auth)
ERIC Educational Resources Information Center
Reed, P.; Staytom, L.; Stott, S.; Truzoli, R.
2011-01-01
Background: This study investigated the relative ease of learning across four tasks suggested by an adaptation of Thomas's hierarchy of learning in children with Down syndrome, autism spectrum disorders and mental age-matched controls. Methods: Learning trials were carried out to investigate observational learning, instrumental learning, reversal…
Consensus-based distributed cooperative learning from closed-loop neural control systems.
Chen, Weisheng; Hua, Shaoyong; Zhang, Huaguang
2015-02-01
In this paper, the neural tracking problem is addressed for a group of uncertain nonlinear systems where the system structures are identical but the reference signals are different. This paper focuses on studying the learning capability of neural networks (NNs) during the control process. First, we propose a novel control scheme called distributed cooperative learning (DCL) control scheme, by establishing the communication topology among adaptive laws of NN weights to share their learned knowledge online. It is further proved that if the communication topology is undirected and connected, all estimated weights of NNs can converge to small neighborhoods around their optimal values over a domain consisting of the union of all state orbits. Second, as a corollary it is shown that the conclusion on the deterministic learning still holds in the decentralized adaptive neural control scheme where, however, the estimated weights of NNs just converge to small neighborhoods of the optimal values along their own state orbits. Thus, the learned controllers obtained by DCL scheme have the better generalization capability than ones obtained by decentralized learning method. A simulation example is provided to verify the effectiveness and advantages of the control schemes proposed in this paper.
Oyama, Shintaro; Shimoda, Shingo; Alnajjar, Fady S K; Iwatsuki, Katsuyuki; Hoshiyama, Minoru; Tanaka, Hirotaka; Hirata, Hitoshi
2016-01-01
Background: For mechanically reconstructing human biomechanical function, intuitive proportional control, and robustness to unexpected situations are required. Particularly, creating a functional hand prosthesis is a typical challenge in the reconstruction of lost biomechanical function. Nevertheless, currently available control algorithms are in the development phase. The most advanced algorithms for controlling multifunctional prosthesis are machine learning and pattern recognition of myoelectric signals. Despite the increase in computational speed, these methods cannot avoid the requirement of user consciousness and classified separation errors. "Tacit Learning System" is a simple but novel adaptive control strategy that can self-adapt its posture to environment changes. We introduced the strategy in the prosthesis rotation control to achieve compensatory reduction, as well as evaluated the system and its effects on the user. Methods: We conducted a non-randomized study involving eight prosthesis users to perform a bar relocation task with/without Tacit Learning System support. Hand piece and body motions were recorded continuously with goniometers, videos, and a motion-capture system. Findings: Reduction in the participants' upper extremity rotatory compensation motion was monitored during the relocation task in all participants. The estimated profile of total body energy consumption improved in five out of six participants. Interpretation: Our system rapidly accomplished nearly natural motion without unexpected errors. The Tacit Learning System not only adapts human motions but also enhances the human ability to adapt to the system quickly, while the system amplifies compensation generated by the residual limb. The concept can be extended to various situations for reconstructing lost functions that can be compensated.
Adapting human-machine interfaces to user performance.
Danziger, Zachary; Fishbach, Alon; Mussa-Ivaldi, Ferdinando A
2008-01-01
The goal of this study was to create and examine machine learning algorithms that adapt in a controlled and cadenced way to foster a harmonious learning environment between the user of a human-machine interface and the controlled device. In this experiment, subjects' high-dimensional finger motions remotely controlled the joint angles of a simulated planar 2-link arm, which was used to hit targets on a computer screen. Subjects were required to move the cursor at the endpoint of the simulated arm.
Adaptive nodes enrich nonlinear cooperative learning beyond traditional adaptation by links.
Sardi, Shira; Vardi, Roni; Goldental, Amir; Sheinin, Anton; Uzan, Herut; Kanter, Ido
2018-03-23
Physical models typically assume time-independent interactions, whereas neural networks and machine learning incorporate interactions that function as adjustable parameters. Here we demonstrate a new type of abundant cooperative nonlinear dynamics where learning is attributed solely to the nodes, instead of the network links which their number is significantly larger. The nodal, neuronal, fast adaptation follows its relative anisotropic (dendritic) input timings, as indicated experimentally, similarly to the slow learning mechanism currently attributed to the links, synapses. It represents a non-local learning rule, where effectively many incoming links to a node concurrently undergo the same adaptation. The network dynamics is now counterintuitively governed by the weak links, which previously were assumed to be insignificant. This cooperative nonlinear dynamic adaptation presents a self-controlled mechanism to prevent divergence or vanishing of the learning parameters, as opposed to learning by links, and also supports self-oscillations of the effective learning parameters. It hints on a hierarchical computational complexity of nodes, following their number of anisotropic inputs and opens new horizons for advanced deep learning algorithms and artificial intelligence based applications, as well as a new mechanism for enhanced and fast learning by neural networks.
NASA Technical Reports Server (NTRS)
Kopasakis, George
1997-01-01
Performance Seeking Control attempts to find the operating condition that will generate optimal performance and control the plant at that operating condition. In this paper a nonlinear multivariable Adaptive Performance Seeking Control (APSC) methodology will be developed and it will be demonstrated on a nonlinear system. The APSC is comprised of the Positive Gradient Control (PGC) and the Fuzzy Model Reference Learning Control (FMRLC). The PGC computes the positive gradients of the desired performance function with respect to the control inputs in order to drive the plant set points to the operating point that will produce optimal performance. The PGC approach will be derived in this paper. The feedback control of the plant is performed by the FMRLC. For the FMRLC, the conventional fuzzy model reference learning control methodology is utilized, with guidelines generated here for the effective tuning of the FMRLC controller.
Ong, M L; Ng, E Y K
2005-12-01
In the lower brain, body temperature is continually being regulated almost flawlessly despite huge fluctuations in ambient and physiological conditions that constantly threaten the well-being of the body. The underlying control problem defining thermal homeostasis is one of great enormity: Many systems and sub-systems are involved in temperature regulation and physiological processes are intrinsically complex and intertwined. Thus the defining control system has to take into account the complications of nonlinearities, system uncertainties, delayed feedback loops as well as internal and external disturbances. In this paper, we propose a self-tuning adaptive thermal controller based upon Hebbian feedback covariance learning where the system is to be regulated continually to best suit its environment. This hypothesis is supported in part by postulations of the presence of adaptive optimization behavior in biological systems of certain organisms which face limited resources vital for survival. We demonstrate the use of Hebbian feedback covariance learning as a possible self-adaptive controller in body temperature regulation. The model postulates an important role of Hebbian covariance adaptation as a means of reinforcement learning in the thermal controller. The passive system is based on a simplified 2-node core and shell representation of the body, where global responses are captured. Model predictions are consistent with observed thermoregulatory responses to conditions of exercise and rest, and heat and cold stress. An important implication of the model is that optimal physiological behaviors arising from self-tuning adaptive regulation in the thermal controller may be responsible for the departure from homeostasis in abnormal states, e.g., fever. This was previously unexplained using the conventional "set-point" control theory.
Braun, Daniel A.; Mehring, Carsten; Wolpert, Daniel M.
2010-01-01
‘Learning to learn’ phenomena have been widely investigated in cognition, perception and more recently also in action. During concept learning tasks, for example, it has been suggested that characteristic features are abstracted from a set of examples with the consequence that learning of similar tasks is facilitated—a process termed ‘learning to learn’. From a computational point of view such an extraction of invariants can be regarded as learning of an underlying structure. Here we review the evidence for structure learning as a ‘learning to learn’ mechanism, especially in sensorimotor control where the motor system has to adapt to variable environments. We review studies demonstrating that common features of variable environments are extracted during sensorimotor learning and exploited for efficient adaptation in novel tasks. We conclude that structure learning plays a fundamental role in skill learning and may underlie the unsurpassed flexibility and adaptability of the motor system. PMID:19720086
Stone, Amanda E; Roper, Jaimie A; Herman, Daniel C; Hass, Chris J
2018-05-01
Persons with anterior cruciate ligament reconstruction (ACLR) show deficits in gait and neuromuscular control following rehabilitation. This altered behavior extends to locomotor adaptation and learning, however the contributing factors to this observed behavior have yet to be investigated. The purpose of this study was to assess differences in locomotor adaptation and learning between ACLR and controls, and identify underlying contributors to motor adaptation in these individuals. Twenty ACLR individuals and 20 healthy controls (CON) agreed to participate in this study. Participants performed four cognitive and dexterity tasks (local version of Trail Making Test, reaction time test, electronic pursuit rotor test, and the Purdue pegboard). Three-dimensional kinematics were also collected while participants walked on a split-belt treadmill. ACLR individuals completed the local versions of Trails A and Trails B significantly faster than CON. During split-belt walking, ACLR individuals demonstrated smaller step length asymmetry during EARLY and LATE adaptation, smaller double support asymmetry during MID adaptation, and larger stance time asymmetry during DE-ADAPT compared with CON. ACLR individuals performed better during tasks that required visual attention and task switching and were less perturbed during split-belt walking compared to controls. Persons with ACLR may use different strategies than controls, cognitive or otherwise, to adapt locomotor patterns.
NASA Astrophysics Data System (ADS)
Rizki, H. T. N.; Frentika, D.; Wijaya, A.
2018-03-01
This study aims to explore junior high school students’ adaptive reasoning and the Van Hiele level of geometric thinking. The present study was a quasi-experiment with the non-equivalent control group design. The participants of the study were 34 seventh graders and 35 eighth graders in the experiment classes and 34 seventh graders and 34 eighth graders in the control classes. The students in the experiment classes learned geometry under the circumstances of a Knisley mathematical learning. The data were analyzed quantitatively by using inferential statistics. The results of data analysis show an improvement of adaptive reasoning skills both in the grade seven and grade eight. An improvement was also found for the Van Hiele level of geometric thinking. These results indicate the positive impact of Knisley learning model on students’ adaptive reasoning skills and Van Hiele level of geometric thinking.
A neural fuzzy controller learning by fuzzy error propagation
NASA Technical Reports Server (NTRS)
Nauck, Detlef; Kruse, Rudolf
1992-01-01
In this paper, we describe a procedure to integrate techniques for the adaptation of membership functions in a linguistic variable based fuzzy control environment by using neural network learning principles. This is an extension to our work. We solve this problem by defining a fuzzy error that is propagated back through the architecture of our fuzzy controller. According to this fuzzy error and the strength of its antecedent each fuzzy rule determines its amount of error. Depending on the current state of the controlled system and the control action derived from the conclusion, each rule tunes the membership functions of its antecedent and its conclusion. By this we get an unsupervised learning technique that enables a fuzzy controller to adapt to a control task by knowing just about the global state and the fuzzy error.
Hou, Runmin; Wang, Li; Gao, Qiang; Hou, Yuanglong; Wang, Chao
2017-09-01
This paper proposes a novel indirect adaptive fuzzy wavelet neural network (IAFWNN) to control the nonlinearity, wide variations in loads, time-variation and uncertain disturbance of the ac servo system. In the proposed approach, the self-recurrent wavelet neural network (SRWNN) is employed to construct an adaptive self-recurrent consequent part for each fuzzy rule of TSK fuzzy model. For the IAFWNN controller, the online learning algorithm is based on back propagation (BP) algorithm. Moreover, an improved particle swarm optimization (IPSO) is used to adapt the learning rate. The aid of an adaptive SRWNN identifier offers the real-time gradient information to the adaptive fuzzy wavelet neural controller to overcome the impact of parameter variations, load disturbances and other uncertainties effectively, and has a good dynamic. The asymptotical stability of the system is guaranteed by using the Lyapunov method. The result of the simulation and the prototype test prove that the proposed are effective and suitable. Copyright © 2017. Published by Elsevier Ltd.
de Rengervé, Antoine; Andry, Pierre; Gaussier, Philippe
2015-04-01
Imitation and learning from humans require an adequate sensorimotor controller to learn and encode behaviors. We present the Dynamic Muscle Perception-Action(DM-PerAc) model to control a multiple degrees-of-freedom (DOF) robot arm. In the original PerAc model, path-following or place-reaching behaviors correspond to the sensorimotor attractors resulting from the dynamics of learned sensorimotor associations. The DM-PerAc model, inspired by human muscles, permits one to combine impedance-like control with the capability of learning sensorimotor attraction basins. We detail a solution to learn incrementally online the DM-PerAc visuomotor controller. Postural attractors are learned by adapting the muscle activations in the model depending on movement errors. Visuomotor categories merging visual and proprioceptive signals are associated with these muscle activations. Thus, the visual and proprioceptive signals activate the motor action generating an attractor which satisfies both visual and proprioceptive constraints. This visuomotor controller can serve as a basis for imitative behaviors. In addition, the muscle activation patterns can define directions of movement instead of postural attractors. Such patterns can be used in state-action couples to generate trajectories like in the PerAc model. We discuss a possible extension of the DM-PerAc controller by adapting the Fukuyori's controller based on the Langevin's equation. This controller can serve not only to reach attractors which were not explicitly learned, but also to learn the state/action couples to define trajectories.
Biologically-inspired adaptive obstacle negotiation behavior of hexapod robots
Goldschmidt, Dennis; Wörgötter, Florentin; Manoonpong, Poramate
2014-01-01
Neurobiological studies have shown that insects are able to adapt leg movements and posture for obstacle negotiation in changing environments. Moreover, the distance to an obstacle where an insect begins to climb is found to be a major parameter for successful obstacle negotiation. Inspired by these findings, we present an adaptive neural control mechanism for obstacle negotiation behavior in hexapod robots. It combines locomotion control, backbone joint control, local leg reflexes, and neural learning. While the first three components generate locomotion including walking and climbing, the neural learning mechanism allows the robot to adapt its behavior for obstacle negotiation with respect to changing conditions, e.g., variable obstacle heights and different walking gaits. By successfully learning the association of an early, predictive signal (conditioned stimulus, CS) and a late, reflex signal (unconditioned stimulus, UCS), both provided by ultrasonic sensors at the front of the robot, the robot can autonomously find an appropriate distance from an obstacle to initiate climbing. The adaptive neural control was developed and tested first on a physical robot simulation, and was then successfully transferred to a real hexapod robot, called AMOS II. The results show that the robot can efficiently negotiate obstacles with a height up to 85% of the robot's leg length in simulation and 75% in a real environment. PMID:24523694
Adaptive Control Based Harvesting Strategy for a Predator-Prey Dynamical System.
Sen, Moitri; Simha, Ashutosh; Raha, Soumyendu
2018-04-23
This paper deals with designing a harvesting control strategy for a predator-prey dynamical system, with parametric uncertainties and exogenous disturbances. A feedback control law for the harvesting rate of the predator is formulated such that the population dynamics is asymptotically stabilized at a positive operating point, while maintaining a positive, steady state harvesting rate. The hierarchical block strict feedback structure of the dynamics is exploited in designing a backstepping control law, based on Lyapunov theory. In order to account for unknown parameters, an adaptive control strategy has been proposed in which the control law depends on an adaptive variable which tracks the unknown parameter. Further, a switching component has been incorporated to robustify the control performance against bounded disturbances. Proofs have been provided to show that the proposed adaptive control strategy ensures asymptotic stability of the dynamics at a desired operating point, as well as exact parameter learning in the disturbance-free case and learning with bounded error in the disturbance prone case. The dynamics, with uncertainty in the death rate of the predator, subjected to a bounded disturbance has been simulated with the proposed control strategy.
Learning algorithms for human-machine interfaces.
Danziger, Zachary; Fishbach, Alon; Mussa-Ivaldi, Ferdinando A
2009-05-01
The goal of this study is to create and examine machine learning algorithms that adapt in a controlled and cadenced way to foster a harmonious learning environment between the user and the controlled device. To evaluate these algorithms, we have developed a simple experimental framework. Subjects wear an instrumented data glove that records finger motions. The high-dimensional glove signals remotely control the joint angles of a simulated planar two-link arm on a computer screen, which is used to acquire targets. A machine learning algorithm was applied to adaptively change the transformation between finger motion and the simulated robot arm. This algorithm was either LMS gradient descent or the Moore-Penrose (MP) pseudoinverse transformation. Both algorithms modified the glove-to-joint angle map so as to reduce the endpoint errors measured in past performance. The MP group performed worse than the control group (subjects not exposed to any machine learning), while the LMS group outperformed the control subjects. However, the LMS subjects failed to achieve better generalization than the control subjects, and after extensive training converged to the same level of performance as the control subjects. These results highlight the limitations of coadaptive learning using only endpoint error reduction.
Learning Algorithms for Human–Machine Interfaces
Fishbach, Alon; Mussa-Ivaldi, Ferdinando A.
2012-01-01
The goal of this study is to create and examine machine learning algorithms that adapt in a controlled and cadenced way to foster a harmonious learning environment between the user and the controlled device. To evaluate these algorithms, we have developed a simple experimental framework. Subjects wear an instrumented data glove that records finger motions. The high-dimensional glove signals remotely control the joint angles of a simulated planar two-link arm on a computer screen, which is used to acquire targets. A machine learning algorithm was applied to adaptively change the transformation between finger motion and the simulated robot arm. This algorithm was either LMS gradient descent or the Moore–Penrose (MP) pseudoinverse transformation. Both algorithms modified the glove-to-joint angle map so as to reduce the endpoint errors measured in past performance. The MP group performed worse than the control group (subjects not exposed to any machine learning), while the LMS group outperformed the control subjects. However, the LMS subjects failed to achieve better generalization than the control subjects, and after extensive training converged to the same level of performance as the control subjects. These results highlight the limitations of coadaptive learning using only endpoint error reduction. PMID:19203886
Experimental setup for evaluating an adaptive user interface for teleoperation control
NASA Astrophysics Data System (ADS)
Wijayasinghe, Indika B.; Peetha, Srikanth; Abubakar, Shamsudeen; Saadatzi, Mohammad Nasser; Cremer, Sven; Popa, Dan O.
2017-05-01
A vital part of human interactions with a machine is the control interface, which single-handedly could define the user satisfaction and the efficiency of performing a task. This paper elaborates the implementation of an experimental setup to study an adaptive algorithm that can help the user better tele-operate the robot. The formulation of the adaptive interface and associate learning algorithms are general enough to apply when the mapping between the user controls and the robot actuators is complex and/or ambiguous. The method uses a genetic algorithm to find the optimal parameters that produce the input-output mapping for teleoperation control. In this paper, we describe the experimental setup and associated results that was used to validate the adaptive interface to a differential drive robot from two different input devices; a joystick, and a Myo gesture control armband. Results show that after the learning phase, the interface converges to an intuitive mapping that can help even inexperienced users drive the system to a goal location.
Pilarski, Patrick M; Dawson, Michael R; Degris, Thomas; Fahimi, Farbod; Carey, Jason P; Sutton, Richard S
2011-01-01
As a contribution toward the goal of adaptable, intelligent artificial limbs, this work introduces a continuous actor-critic reinforcement learning method for optimizing the control of multi-function myoelectric devices. Using a simulated upper-arm robotic prosthesis, we demonstrate how it is possible to derive successful limb controllers from myoelectric data using only a sparse human-delivered training signal, without requiring detailed knowledge about the task domain. This reinforcement-based machine learning framework is well suited for use by both patients and clinical staff, and may be easily adapted to different application domains and the needs of individual amputees. To our knowledge, this is the first my-oelectric control approach that facilitates the online learning of new amputee-specific motions based only on a one-dimensional (scalar) feedback signal provided by the user of the prosthesis. © 2011 IEEE
NASA Astrophysics Data System (ADS)
Billard, Aude
2000-10-01
This paper summarizes a number of experiments in biologically inspired robotics. The common feature to all experiments is the use of artificial neural networks as the building blocks for the controllers. The experiments speak in favor of using a connectionist approach for designing adaptive and flexible robot controllers, and for modeling neurological processes. I present 1) DRAMA, a novel connectionist architecture, which has general property for learning time series and extracting spatio-temporal regularities in multi-modal and highly noisy data; 2) Robota, a doll-shaped robot, which imitates and learns a proto-language; 3) an experiment in collective robotics, where a group of 4 to 15 Khepera robots learn dynamically the topography of an environment whose features change frequently; 4) an abstract, computational model of primate ability to learn by imitation; 5) a model for the control of locomotor gaits in a quadruped legged robot.
Flight Research into Simple Adaptive Control on the NASA FAST Aircraft
NASA Technical Reports Server (NTRS)
Hanson, Curtis E.
2011-01-01
A series of simple adaptive controllers with varying levels of complexity were designed, implemented and flight tested on the NASA Full-Scale Advanced Systems Testbed (FAST) aircraft. Lessons learned from the development and flight testing are presented.
Luo, Biao; Liu, Derong; Wu, Huai-Ning
2018-06-01
Reinforcement learning has proved to be a powerful tool to solve optimal control problems over the past few years. However, the data-based constrained optimal control problem of nonaffine nonlinear discrete-time systems has rarely been studied yet. To solve this problem, an adaptive optimal control approach is developed by using the value iteration-based Q-learning (VIQL) with the critic-only structure. Most of the existing constrained control methods require the use of a certain performance index and only suit for linear or affine nonlinear systems, which is unreasonable in practice. To overcome this problem, the system transformation is first introduced with the general performance index. Then, the constrained optimal control problem is converted to an unconstrained optimal control problem. By introducing the action-state value function, i.e., Q-function, the VIQL algorithm is proposed to learn the optimal Q-function of the data-based unconstrained optimal control problem. The convergence results of the VIQL algorithm are established with an easy-to-realize initial condition . To implement the VIQL algorithm, the critic-only structure is developed, where only one neural network is required to approximate the Q-function. The converged Q-function obtained from the critic-only VIQL method is employed to design the adaptive constrained optimal controller based on the gradient descent scheme. Finally, the effectiveness of the developed adaptive control method is tested on three examples with computer simulation.
NASA Astrophysics Data System (ADS)
Wang, Jing; Yang, Tianyu; Staskevich, Gennady; Abbe, Brian
2017-04-01
This paper studies the cooperative control problem for a class of multiagent dynamical systems with partially unknown nonlinear system dynamics. In particular, the control objective is to solve the state consensus problem for multiagent systems based on the minimisation of certain cost functions for individual agents. Under the assumption that there exist admissible cooperative controls for such class of multiagent systems, the formulated problem is solved through finding the optimal cooperative control using the approximate dynamic programming and reinforcement learning approach. With the aid of neural network parameterisation and online adaptive learning, our method renders a practically implementable approximately adaptive neural cooperative control for multiagent systems. Specifically, based on the Bellman's principle of optimality, the Hamilton-Jacobi-Bellman (HJB) equation for multiagent systems is first derived. We then propose an approximately adaptive policy iteration algorithm for multiagent cooperative control based on neural network approximation of the value functions. The convergence of the proposed algorithm is rigorously proved using the contraction mapping method. The simulation results are included to validate the effectiveness of the proposed algorithm.
A reductionist approach to the analysis of learning in brain-computer interfaces.
Danziger, Zachary
2014-04-01
The complexity and scale of brain-computer interface (BCI) studies limit our ability to investigate how humans learn to use BCI systems. It also limits our capacity to develop adaptive algorithms needed to assist users with their control. Adaptive algorithm development is forced offline and typically uses static data sets. But this is a poor substitute for the online, dynamic environment where algorithms are ultimately deployed and interact with an adapting user. This work evaluates a paradigm that simulates the control problem faced by human subjects when controlling a BCI, but which avoids the many complications associated with full-scale BCI studies. Biological learners can be studied in a reductionist way as they solve BCI-like control problems, and machine learning algorithms can be developed and tested in closed loop with the subjects before being translated to full BCIs. The method is to map 19 joint angles of the hand (representing neural signals) to the position of a 2D cursor which must be piloted to displayed targets (a typical BCI task). An investigation is presented on how closely the joint angle method emulates BCI systems; a novel learning algorithm is evaluated, and a performance difference between genders is discussed.
Decomposed fuzzy systems and their application in direct adaptive fuzzy control.
Hsueh, Yao-Chu; Su, Shun-Feng; Chen, Ming-Chang
2014-10-01
In this paper, a novel fuzzy structure termed as the decomposed fuzzy system (DFS) is proposed to act as the fuzzy approximator for adaptive fuzzy control systems. The proposed structure is to decompose each fuzzy variable into layers of fuzzy systems, and each layer is to characterize one traditional fuzzy set. Similar to forming fuzzy rules in traditional fuzzy systems, layers from different variables form the so-called component fuzzy systems. DFS is proposed to provide more adjustable parameters to facilitate possible adaptation in fuzzy rules, but without introducing a learning burden. It is because those component fuzzy systems are independent so that it can facilitate minimum distribution learning effects among component fuzzy systems. It can be seen from our experiments that even when the rule number increases, the learning time in terms of cycles is still almost constant. It can also be found that the function approximation capability and learning efficiency of the DFS are much better than that of the traditional fuzzy systems when employed in adaptive fuzzy control systems. Besides, in order to further reduce the computational burden, a simplified DFS is proposed in this paper to satisfy possible real time constraints required in many applications. From our simulation results, it can be seen that the simplified DFS can perform fairly with a more concise decomposition structure.
On the integration of reinforcement learning and approximate reasoning for control
NASA Technical Reports Server (NTRS)
Berenji, Hamid R.
1991-01-01
The author discusses the importance of strengthening the knowledge representation characteristic of reinforcement learning techniques using methods such as approximate reasoning. The ARIC (approximate reasoning-based intelligent control) architecture is an example of such a hybrid approach in which the fuzzy control rules are modified (fine-tuned) using reinforcement learning. ARIC also demonstrates that it is possible to start with an approximately correct control knowledge base and learn to refine this knowledge through further experience. On the other hand, techniques such as the TD (temporal difference) algorithm and Q-learning establish stronger theoretical foundations for their use in adaptive control and also in stability analysis of hybrid reinforcement learning and approximate reasoning-based controllers.
Wang, Shun-Yuan; Tseng, Chwan-Lu; Lin, Shou-Chuang; Chiu, Chun-Jung; Chou, Jen-Hsiang
2015-01-01
This paper presents the implementation of an adaptive supervisory sliding fuzzy cerebellar model articulation controller (FCMAC) in the speed sensorless vector control of an induction motor (IM) drive system. The proposed adaptive supervisory sliding FCMAC comprised a supervisory controller, integral sliding surface, and an adaptive FCMAC. The integral sliding surface was employed to eliminate steady-state errors and enhance the responsiveness of the system. The adaptive FCMAC incorporated an FCMAC with a compensating controller to perform a desired control action. The proposed controller was derived using the Lyapunov approach, which guarantees learning-error convergence. The implementation of three intelligent control schemes—the adaptive supervisory sliding FCMAC, adaptive sliding FCMAC, and adaptive sliding CMAC—were experimentally investigated under various conditions in a realistic sensorless vector-controlled IM drive system. The root mean square error (RMSE) was used as a performance index to evaluate the experimental results of each control scheme. The analysis results indicated that the proposed adaptive supervisory sliding FCMAC substantially improved the system performance compared with the other control schemes. PMID:25815450
Wang, Shun-Yuan; Tseng, Chwan-Lu; Lin, Shou-Chuang; Chiu, Chun-Jung; Chou, Jen-Hsiang
2015-03-25
This paper presents the implementation of an adaptive supervisory sliding fuzzy cerebellar model articulation controller (FCMAC) in the speed sensorless vector control of an induction motor (IM) drive system. The proposed adaptive supervisory sliding FCMAC comprised a supervisory controller, integral sliding surface, and an adaptive FCMAC. The integral sliding surface was employed to eliminate steady-state errors and enhance the responsiveness of the system. The adaptive FCMAC incorporated an FCMAC with a compensating controller to perform a desired control action. The proposed controller was derived using the Lyapunov approach, which guarantees learning-error convergence. The implementation of three intelligent control schemes--the adaptive supervisory sliding FCMAC, adaptive sliding FCMAC, and adaptive sliding CMAC--were experimentally investigated under various conditions in a realistic sensorless vector-controlled IM drive system. The root mean square error (RMSE) was used as a performance index to evaluate the experimental results of each control scheme. The analysis results indicated that the proposed adaptive supervisory sliding FCMAC substantially improved the system performance compared with the other control schemes.
Investigation of Drive-Reinforcement Learning and Application of Learning to Flight Control
1993-08-01
Attachment 1 138 Reprint of: Baird, L. (1991). Learning and Adaptive Hybrid Systems for Nonlinear Control, CSDL Report T-1099, M.S. Thesis , Department of...Aircraft, CSDL Report T-1127, S.M. Thesis , Department of Aeronautics and Astronautics, M.I.T. Attachment 3 351 . iprint of: Atkins, S. (1993...Incremental Synthesis of Optimal Control Laws Using Learning Algorithms, CSDL Report T-1181, S.M. Thesis , Department of Aeronautics and Astronautics, M.I.T
Control Systems with Normalized and Covariance Adaptation by Optimal Control Modification
NASA Technical Reports Server (NTRS)
Nguyen, Nhan T. (Inventor); Burken, John J. (Inventor); Hanson, Curtis E. (Inventor)
2016-01-01
Disclosed is a novel adaptive control method and system called optimal control modification with normalization and covariance adjustment. The invention addresses specifically to current challenges with adaptive control in these areas: 1) persistent excitation, 2) complex nonlinear input-output mapping, 3) large inputs and persistent learning, and 4) the lack of stability analysis tools for certification. The invention has been subject to many simulations and flight testing. The results substantiate the effectiveness of the invention and demonstrate the technical feasibility for use in modern aircraft flight control systems.
Methods of Adapting Digital Content for the Learning Process via Mobile Devices
ERIC Educational Resources Information Center
Lopez, J. L. Gimenez; Royo, T. Magal; Laborda, Jesus Garcia; Calvo, F. Garde
2009-01-01
This article analyses different methods of adapting digital content for its delivery via mobile devices taking into account two aspects which are a fundamental part of the learning process; on the one hand, functionality of the contents, and on the other, the actual controlled navigation requirements that the learner needs in order to acquire high…
The Role of Career Adaptabilities for Mid-Career Changers
ERIC Educational Resources Information Center
Brown, Alan; Bimrose, Jenny; Barnes, Sally-Anne; Hughes, Deirdre
2012-01-01
Career adaptability is mediated by personality factors and socio-psychological processes, with learning playing an important role. Using a five-fold career adapt-abilities competency framework (defined here as control, curiosity, commitment, confidence and concern), which was developed from the international quantitative study that is the focus of…
NASA Technical Reports Server (NTRS)
Padgett, Mary L. (Editor)
1993-01-01
The present conference discusses such neural networks (NN) related topics as their current development status, NN architectures, NN learning rules, NN optimization methods, NN temporal models, NN control methods, NN pattern recognition systems and applications, biological and biomedical applications of NNs, VLSI design techniques for NNs, NN systems simulation, fuzzy logic, and genetic algorithms. Attention is given to missileborne integrated NNs, adaptive-mixture NNs, implementable learning rules, an NN simulator for travelling salesman problem solutions, similarity-based forecasting, NN control of hypersonic aircraft takeoff, NN control of the Space Shuttle Arm, an adaptive NN robot manipulator controller, a synthetic approach to digital filtering, NNs for speech analysis, adaptive spline networks, an anticipatory fuzzy logic controller, and encoding operations for fuzzy associative memories.
A boundedness result for the direct heuristic dynamic programming.
Liu, Feng; Sun, Jian; Si, Jennie; Guo, Wentao; Mei, Shengwei
2012-08-01
Approximate/adaptive dynamic programming (ADP) has been studied extensively in recent years for its potential scalability to solve large state and control space problems, including those involving continuous states and continuous controls. The applicability of ADP algorithms, especially the adaptive critic designs has been demonstrated in several case studies. Direct heuristic dynamic programming (direct HDP) is one of the ADP algorithms inspired by the adaptive critic designs. It has been shown applicable to industrial scale, realistic and complex control problems. In this paper, we provide a uniformly ultimately boundedness (UUB) result for the direct HDP learning controller under mild and intuitive conditions. By using a Lyapunov approach we show that the estimation errors of the learning parameters or the weights in the action and critic networks remain UUB. This result provides a useful controller convergence guarantee for the first time for the direct HDP design. Copyright © 2012 Elsevier Ltd. All rights reserved.
Parameter learning for performance adaptation
NASA Technical Reports Server (NTRS)
Peek, Mark D.; Antsaklis, Panos J.
1990-01-01
A parameter learning method is introduced and used to broaden the region of operability of the adaptive control system of a flexible space antenna. The learning system guides the selection of control parameters in a process leading to optimal system performance. A grid search procedure is used to estimate an initial set of parameter values. The optimization search procedure uses a variation of the Hooke and Jeeves multidimensional search algorithm. The method is applicable to any system where performance depends on a number of adjustable parameters. A mathematical model is not necessary, as the learning system can be used whenever the performance can be measured via simulation or experiment. The results of two experiments, the transient regulation and the command following experiment, are presented.
Neural network-based model reference adaptive control system.
Patino, H D; Liu, D
2000-01-01
In this paper, an approach to model reference adaptive control based on neural networks is proposed and analyzed for a class of first-order continuous-time nonlinear dynamical systems. The controller structure can employ either a radial basis function network or a feedforward neural network to compensate adaptively the nonlinearities in the plant. A stable controller-parameter adjustment mechanism, which is determined using the Lyapunov theory, is constructed using a sigma-modification-type updating law. The evaluation of control error in terms of the neural network learning error is performed. That is, the control error converges asymptotically to a neighborhood of zero, whose size is evaluated and depends on the approximation error of the neural network. In the design and analysis of neural network-based control systems, it is important to take into account the neural network learning error and its influence on the control error of the plant. Simulation results showing the feasibility and performance of the proposed approach are given.
Probabilistic reversal learning is impaired in Parkinson's disease
Peterson, David A.; Elliott, Christian; Song, David D.; Makeig, Scott; Sejnowski, Terrence J.; Poizner, Howard
2009-01-01
In many everyday settings, the relationship between our choices and their potentially rewarding outcomes is probabilistic and dynamic. In addition, the difficulty of the choices can vary widely. Although a large body of theoretical and empirical evidence suggests that dopamine mediates rewarded learning, the influence of dopamine in probabilistic and dynamic rewarded learning remains unclear. We adapted a probabilistic rewarded learning task originally used to study firing rates of dopamine cells in primate substantia nigra pars compacta (Morris et al. 2006) for use as a reversal learning task with humans. We sought to investigate how the dopamine depletion in Parkinson's disease (PD) affects probabilistic reward learning and adaptation to a reversal in reward contingencies. Over the course of 256 trials subjects learned to choose the more favorable from among pairs of images with small or large differences in reward probabilities. During a subsequent otherwise identical reversal phase, the reward probability contingencies for the stimuli were reversed. Seventeen Parkinson's disease (PD) patients of mild to moderate severity were studied off of their dopaminergic medications and compared to 15 age-matched controls. Compared to controls, PD patients had distinct pre- and post-reversal deficiencies depending upon the difficulty of the choices they had to learn. The patients also exhibited compromised adaptability to the reversal. A computational model of the subjects’ trial-by-trial choices demonstrated that the adaptability was sensitive to the gain with which patients weighted pre-reversal feedback. Collectively, the results implicate the nigral dopaminergic system in learning to make choices in environments with probabilistic and dynamic reward contingencies. PMID:19628022
Algebraic and adaptive learning in neural control systems
NASA Astrophysics Data System (ADS)
Ferrari, Silvia
A systematic approach is developed for designing adaptive and reconfigurable nonlinear control systems that are applicable to plants modeled by ordinary differential equations. The nonlinear controller comprising a network of neural networks is taught using a two-phase learning procedure realized through novel techniques for initialization, on-line training, and adaptive critic design. A critical observation is that the gradients of the functions defined by the neural networks must equal corresponding linear gain matrices at chosen operating points. On-line training is based on a dual heuristic adaptive critic architecture that improves control for large, coupled motions by accounting for actual plant dynamics and nonlinear effects. An action network computes the optimal control law; a critic network predicts the derivative of the cost-to-go with respect to the state. Both networks are algebraically initialized based on prior knowledge of satisfactory pointwise linear controllers and continue to adapt on line during full-scale simulations of the plant. On-line training takes place sequentially over discrete periods of time and involves several numerical procedures. A backpropagating algorithm called Resilient Backpropagation is modified and successfully implemented to meet these objectives, without excessive computational expense. This adaptive controller is as conservative as the linear designs and as effective as a global nonlinear controller. The method is successfully implemented for the full-envelope control of a six-degree-of-freedom aircraft simulation. The results show that the on-line adaptation brings about improved performance with respect to the initialization phase during aircraft maneuvers that involve large-angle and coupled dynamics, and parameter variations.
Diminished Neural Adaptation during Implicit Learning in Autism
Schipul, Sarah E.; Just, Marcel Adam
2015-01-01
Neuroimaging studies have shown evidence of disrupted neural adaptation during learning in individuals with autism spectrum disorder (ASD) in several types of tasks, potentially stemming from frontal-posterior cortical underconnectivity (Schipul et al., 2012). The aim of the current study was to examine neural adaptations in an implicit learning task that entails participation of frontal and posterior regions. Sixteen high-functioning adults with ASD and sixteen neurotypical control participants were trained on and performed an implicit dot pattern prototype learning task in a functional magnetic resonance imaging (fMRI) session. During the preliminary exposure to the type of implicit prototype learning task later to be used in the scanner, the ASD participants took longer than the neurotypical group to learn the task, demonstrating altered implicit learning in ASD. After equating task structure learning, the two groups’ brain activation differed during their learning of a new prototype in the subsequent scanning session. The main findings indicated that neural adaptations in a distributed task network were reduced in the ASD group, relative to the neurotypical group, and were related to ASD symptom severity. Functional connectivity was reduced and did not change as much during learning for the ASD group, and was related to ASD symptom severity. These findings suggest that individuals with ASD show altered neural adaptations during learning, as seen in both activation and functional connectivity measures. This finding suggests why many real-world implicit learning situations may pose special challenges for ASD. PMID:26484826
Rapid feedback responses correlate with reach adaptation and properties of novel upper limb loads.
Cluff, Tyler; Scott, Stephen H
2013-10-02
A hallmark of voluntary motor control is the ability to adjust motor patterns for novel mechanical or visuomotor contexts. Recent work has also highlighted the importance of feedback for voluntary control, leading to the hypothesis that feedback responses should adapt when we learn new motor skills. We tested this prediction with a novel paradigm requiring that human subjects adapt to a viscous elbow load while reaching to three targets. Target 1 required combined shoulder and elbow motion, target 2 required only elbow motion, and target 3 (probe target) required shoulder but no elbow motion. This simple approach controlled muscle activity at the probe target before, during, and after the application of novel elbow loads. Our paradigm allowed us to perturb the elbow during reaching movements to the probe target and identify several key properties of adapted stretch responses. Adapted long-latency responses expressed (de-) adaptation similar to reaching errors observed when we introduced (removed) the elbow load. Moreover, reaching errors during learning correlated with changes in the long-latency response, showing subjects who adapted more to the elbow load displayed greater modulation of their stretch responses. These adapted responses were sensitive to the size and direction of the viscous training load. Our results highlight an important link between the adaptation of feedforward and feedback control and suggest a key part of motor adaptation is to adjust feedback responses to the requirements of novel motor skills.
Simulating closed- and open-loop voluntary movement: a nonlinear control-systems approach.
Davidson, Paul R; Jones, Richard D; Andreae, John H; Sirisena, Harsha R
2002-11-01
In many recent human motor control models, including feedback-error learning and adaptive model theory (AMT), feedback control is used to correct errors while an inverse model is simultaneously tuned to provide accurate feedforward control. This popular and appealing hypothesis, based on a combination of psychophysical observations and engineering considerations, predicts that once the tuning of the inverse model is complete the role of feedback control is limited to the correction of disturbances. This hypothesis was tested by looking at the open-loop behavior of the human motor system during adaptation. An experiment was carried out involving 20 normal adult subjects who learned a novel visuomotor relationship on a pursuit tracking task with a steering wheel for input. During learning, the response cursor was periodically blanked, removing all feedback about the external system (i.e., about the relationship between hand motion and response cursor motion). Open-loop behavior was not consistent with a progressive transfer from closed- to open-loop control. Our recently developed computational model of the brain--a novel nonlinear implementation of AMT--was able to reproduce the observed closed- and open-loop results. In contrast, other control-systems models exhibited only minimal feedback control following adaptation, leading to incorrect open-loop behavior. This is because our model continues to use feedback to control slow movements after adaptation is complete. This behavior enhances the internal stability of the inverse model. In summary, our computational model is currently the only motor control model able to accurately simulate the closed- and open-loop characteristics of the experimental response trajectories.
Evolutionary online behaviour learning and adaptation in real robots.
Silva, Fernando; Correia, Luís; Christensen, Anders Lyhne
2017-07-01
Online evolution of behavioural control on real robots is an open-ended approach to autonomous learning and adaptation: robots have the potential to automatically learn new tasks and to adapt to changes in environmental conditions, or to failures in sensors and/or actuators. However, studies have so far almost exclusively been carried out in simulation because evolution in real hardware has required several days or weeks to produce capable robots. In this article, we successfully evolve neural network-based controllers in real robotic hardware to solve two single-robot tasks and one collective robotics task. Controllers are evolved either from random solutions or from solutions pre-evolved in simulation. In all cases, capable solutions are found in a timely manner (1 h or less). Results show that more accurate simulations may lead to higher-performing controllers, and that completing the optimization process in real robots is meaningful, even if solutions found in simulation differ from solutions in reality. We furthermore demonstrate for the first time the adaptive capabilities of online evolution in real robotic hardware, including robots able to overcome faults injected in the motors of multiple units simultaneously, and to modify their behaviour in response to changes in the task requirements. We conclude by assessing the contribution of each algorithmic component on the performance of the underlying evolutionary algorithm.
An improved adaptive control for repetitive motion of robots
NASA Technical Reports Server (NTRS)
Pourboghrat, F.
1989-01-01
An adaptive control algorithm is proposed for a class of nonlinear systems, such as robotic manipulators, which is capable of improving its performance in repetitive motions. When the task is repeated, the error between the desired trajectory and that of the system is guaranteed to decrease. The design is based on the combination of a direct adaptive control and a learning process. This method does not require any knowledge of the dynamic parameters of the system.
ERIC Educational Resources Information Center
Kanar, Adam M.; Bell, Bradford S.
2013-01-01
Adaptive guidance is an instructional intervention that helps learners to make use of the control inherent in technology-based instruction. The present research investigated the interactive effects of guidance design (i.e., framing of guidance information) and individual differences (i.e., pretraining motivation and ability) on learning basic and…
Dynamic neural networks based on-line identification and control of high performance motor drives
NASA Technical Reports Server (NTRS)
Rubaai, Ahmed; Kotaru, Raj
1995-01-01
In the automated and high-tech industries of the future, there wil be a need for high performance motor drives both in the low-power range and in the high-power range. To meet very straight demands of tracking and regulation in the two quadrants of operation, advanced control technologies are of a considerable interest and need to be developed. In response a dynamics learning control architecture is developed with simultaneous on-line identification and control. the feature of the proposed approach, to efficiently combine the dual task of system identification (learning) and adaptive control of nonlinear motor drives into a single operation is presented. This approach, therefore, not only adapts to uncertainties of the dynamic parameters of the motor drives but also learns about their inherent nonlinearities. In fact, most of the neural networks based adaptive control approaches in use have an identification phase entirely separate from the control phase. Because these approaches separate the identification and control modes, it is not possible to cope with dynamic changes in a controlled process. Extensive simulation studies have been conducted and good performance was observed. The robustness characteristics of neuro-controllers to perform efficiently in a noisy environment is also demonstrated. With this initial success, the principal investigator believes that the proposed approach with the suggested neural structure can be used successfully for the control of high performance motor drives. Two identification and control topologies based on the model reference adaptive control technique are used in this present analysis. No prior knowledge of load dynamics is assumed in either topology while the second topology also assumes no knowledge of the motor parameters.
Globally Sustainable Management: A Dynamic Model of IHRM Learning and Control
ERIC Educational Resources Information Center
Takeda, Margaret B.; Helms, Marilyn M.
2010-01-01
Purpose: After a thorough literature review on multinational learning, it is apparent organizations "learn" when they capitalize on expatriate management, a "learning strategy" (international work teams, employee involvement and other human resource policies), technology transfer and political environment and cross-cultural adaptation. This…
On the Effectiveness of a Neural Network for Adaptive External Pacing.
ERIC Educational Resources Information Center
Montazemi, Ali R.; Wang, Feng
1995-01-01
Proposes a neural network model for an intelligent tutoring system featuring adaptive external control of student pacing. An experiment was conducted, and students using adaptive external pacing experienced improved mastery learning and increased motivation for time management. Contains 66 references. (JKP)
Machine-Learning Based Co-adaptive Calibration: A Perspective to Fight BCI Illiteracy
NASA Astrophysics Data System (ADS)
Vidaurre, Carmen; Sannelli, Claudia; Müller, Klaus-Robert; Blankertz, Benjamin
"BCI illiteracy" is one of the biggest problems and challenges in BCI research. It means that BCI control cannot be achieved by a non-negligible number of subjects (estimated 20% to 25%). There are two main causes for BCI illiteracy in BCI users: either no SMR idle rhythm is observed over motor areas, or this idle rhythm is not attenuated during motor imagery, resulting in a classification performance lower than 70% (criterion level) already for offline calibration data. In a previous work of the same authors, the concept of machine learning based co-adaptive calibration was introduced. This new type of calibration provided substantially improved performance for a variety of users. Here, we use a similar approach and investigate to what extent co-adapting learning enables substantial BCI control for completely novice users and those who suffered from BCI illiteracy before.
Bottiroli, Sara; Cavallini, Elena; Dunlosky, John; Vecchi, Tomaso; Hertzog, Christopher
2013-09-01
We investigated the benefits of strategy-adaptation training for promoting transfer effects. This learner-oriented approach--which directly encourages the learner to generalize strategic behavior to new tasks--helps older adults appraise new tasks and adapt trained strategies to them. In Experiment 1, older adults in a strategy-adaptation training group used 2 strategies (imagery and sentence generation) while practicing 2 tasks (list and associative learning); they were then instructed on how to do a simple task analysis to help them adapt the trained strategies for 2 different unpracticed tasks (place learning and text learning) that were discussed during training. Two additional criterion tasks (name-face associative learning and grocery-list learning) were never mentioned during training. Two other groups were included: A strategy training group (who received strategy training and transfer instructions but not strategy-adaptation training) and a waiting-list control group. Both training procedures enhanced older adults' performance on the trained tasks and those tasks that were discussed during training, but transfer was greatest after strategy-adaptation training. Experiment 2 found that strategy-adaptation training conducted via a manual that older adults used at home also promoted transfer. These findings demonstrate the importance of adopting a learner-oriented approach to promote transfer of strategy training. PsycINFO Database Record (c) 2013 APA, all rights reserved.
Applications of Adaptive Learning Controller to Synthetic Aperture Radar.
1985-02-01
FIGURE 37. Location of Two Sub- Phase Histories to be Utilized in Estimating Misfocus Coefficients A and C . . . A8 FIGURES 38.-94. ALC Learning Curves ...FIGURES (Concl uded) FIGURE 23. ALC Learning Curve .... .................. ... 45 .- " FIGURE 24. ALC Learning Curve ......... ................. 47 FIGURE...25. ALC Learning Curve .... .................. ... 48 FIGURE 26. ALC Learning Curve ....... .... ... .... 50 FIGURE 27. ALC Learning Curve
Efficient retrieval of landscape Hessian: Forced optimal covariance adaptive learning
NASA Astrophysics Data System (ADS)
Shir, Ofer M.; Roslund, Jonathan; Whitley, Darrell; Rabitz, Herschel
2014-06-01
Knowledge of the Hessian matrix at the landscape optimum of a controlled physical observable offers valuable information about the system robustness to control noise. The Hessian can also assist in physical landscape characterization, which is of particular interest in quantum system control experiments. The recently developed landscape theoretical analysis motivated the compilation of an automated method to learn the Hessian matrix about the global optimum without derivative measurements from noisy data. The current study introduces the forced optimal covariance adaptive learning (FOCAL) technique for this purpose. FOCAL relies on the covariance matrix adaptation evolution strategy (CMA-ES) that exploits covariance information amongst the control variables by means of principal component analysis. The FOCAL technique is designed to operate with experimental optimization, generally involving continuous high-dimensional search landscapes (≳30) with large Hessian condition numbers (≳104). This paper introduces the theoretical foundations of the inverse relationship between the covariance learned by the evolution strategy and the actual Hessian matrix of the landscape. FOCAL is presented and demonstrated to retrieve the Hessian matrix with high fidelity on both model landscapes and quantum control experiments, which are observed to possess nonseparable, nonquadratic search landscapes. The recovered Hessian forms were corroborated by physical knowledge of the systems. The implications of FOCAL extend beyond the investigated studies to potentially cover other physically motivated multivariate landscapes.
Shared internal models for feedforward and feedback control.
Wagner, Mark J; Smith, Maurice A
2008-10-15
A child often learns to ride a bicycle in the driveway, free of unforeseen obstacles. Yet when she first rides in the street, we hope that if a car suddenly pulls out in front of her, she will combine her innate goal of avoiding an accident with her learned knowledge of the bicycle, and steer away or brake. In general, when we train to perform a new motor task, our learning is most robust if it updates the rules of online error correction to reflect the rules and goals of the new task. Here we provide direct evidence that, after a new feedforward motor adaptation, motor feedback responses to unanticipated errors become precisely task appropriate, even when such errors were never experienced during training. To study this ability, we asked how, if at all, do online responses to occasional, unanticipated force pulses during reaching arm movements change after adapting to altered arm dynamics? Specifically, do they change in a task-appropriate manner? In our task, subjects learned novel velocity-dependent dynamics. However, occasional force-pulse perturbations produced unanticipated changes in velocity. Therefore, after adaptation, task-appropriate responses to unanticipated pulses should compensate corresponding changes in velocity-dependent dynamics. We found that after adaptation, pulse responses precisely compensated these changes, although they were never trained to do so. These results provide evidence for a smart feedback controller which automatically produces responses specific to the learned dynamics of the current task. To accomplish this, the neural processes underlying feedback control must (1) be capable of accurate real-time state prediction for velocity via a forward model and (2) have access to recently learned changes in internal models of limb dynamics.
Peternel, Luka; Noda, Tomoyuki; Petrič, Tadej; Ude, Aleš; Morimoto, Jun; Babič, Jan
2016-01-01
In this paper we propose an exoskeleton control method for adaptive learning of assistive joint torque profiles in periodic tasks. We use human muscle activity as feedback to adapt the assistive joint torque behaviour in a way that the muscle activity is minimised. The user can then relax while the exoskeleton takes over the task execution. If the task is altered and the existing assistive behaviour becomes inadequate, the exoskeleton gradually adapts to the new task execution so that the increased muscle activity caused by the new desired task can be reduced. The advantage of the proposed method is that it does not require biomechanical or dynamical models. Our proposed learning system uses Dynamical Movement Primitives (DMPs) as a trajectory generator and parameters of DMPs are modulated using Locally Weighted Regression. Then, the learning system is combined with adaptive oscillators that determine the phase and frequency of motion according to measured Electromyography (EMG) signals. We tested the method with real robot experiments where subjects wearing an elbow exoskeleton had to move an object of an unknown mass according to a predefined reference motion. We further evaluated the proposed approach on a whole-arm exoskeleton to show that it is able to adaptively derive assistive torques even for multiple-joint motion.
Peternel, Luka; Noda, Tomoyuki; Petrič, Tadej; Ude, Aleš; Morimoto, Jun; Babič, Jan
2016-01-01
In this paper we propose an exoskeleton control method for adaptive learning of assistive joint torque profiles in periodic tasks. We use human muscle activity as feedback to adapt the assistive joint torque behaviour in a way that the muscle activity is minimised. The user can then relax while the exoskeleton takes over the task execution. If the task is altered and the existing assistive behaviour becomes inadequate, the exoskeleton gradually adapts to the new task execution so that the increased muscle activity caused by the new desired task can be reduced. The advantage of the proposed method is that it does not require biomechanical or dynamical models. Our proposed learning system uses Dynamical Movement Primitives (DMPs) as a trajectory generator and parameters of DMPs are modulated using Locally Weighted Regression. Then, the learning system is combined with adaptive oscillators that determine the phase and frequency of motion according to measured Electromyography (EMG) signals. We tested the method with real robot experiments where subjects wearing an elbow exoskeleton had to move an object of an unknown mass according to a predefined reference motion. We further evaluated the proposed approach on a whole-arm exoskeleton to show that it is able to adaptively derive assistive torques even for multiple-joint motion. PMID:26881743
Adaptive management of rangeland systems
Allen, Craig R.; Angeler, David G.; Fontaine, Joseph J.; Garmestani, Ahjond S.; Hart, Noelle M.; Pope, Kevin L.; Twidwell, Dirac
2017-01-01
Adaptive management is an approach to natural resource management that uses structured learning to reduce uncertainties for the improvement of management over time. The origins of adaptive management are linked to ideas of resilience theory and complex systems. Rangeland management is particularly well suited for the application of adaptive management, having sufficient controllability and reducible uncertainties. Adaptive management applies the tools of structured decision making and requires monitoring, evaluation, and adjustment of management. Adaptive governance, involving sharing of power and knowledge among relevant stakeholders, is often required to address conflict situations. Natural resource laws and regulations can present a barrier to adaptive management when requirements for legal certainty are met with environmental uncertainty. However, adaptive management is possible, as illustrated by two cases presented in this chapter. Despite challenges and limitations, when applied appropriately adaptive management leads to improved management through structured learning, and rangeland management is an area in which adaptive management shows promise and should be further explored.
An application of adaptive learning to malfunction recovery
NASA Technical Reports Server (NTRS)
Cruz, R. E.
1986-01-01
A self-organizing controller is developed for a simplified two-dimensional aircraft model. The Controller learns how to pilot the aircraft through a navigational mission without exceeding pre-established position and velocity limits. The controller pilots the aircraft by activating one of eight directional actuators at all times. By continually monitoring the aircraft's position and velocity with respect to the mission, the controller progressively modifies its decision rules to improve the aircraft's performance. When the controller has learned how to pilot the aircraft, two actuators fail permanently. Despite this malfunction, the controller regains proficiency at its original task. The experimental results reported show the controller's capabilities for self-organizing control, learning, and malfunction recovery.
Thürer, Benjamin; Stein, Thorsten
2017-01-01
Intermanual transfer (motor memory generalization across arms) and motor memory interference (impairment of retest performance in consecutive motor learning) are well-investigated motor learning phenomena. However, the interplay of these phenomena remains elusive, i.e., whether intermanual interference occurs when two unimanual tasks are consecutively learned using different arms. Here, we examine intermanual interference when subjects consecutively adapt their right and left arm movements to novel dynamics. We considered two force field tasks A and B which were of the same structure but mirrored orientation (B = -A). The first test group (ABA-group) consecutively learned task A using their right arm and task B using their left arm before being retested for task A with their right arm. Another test group (AAA-group) learned only task A in the same right-left-right arm schedule. Control subjects learned task A using their right arm without intermediate left arm learning. All groups were able to adapt their right arm movements to force field A and both test groups showed significant intermanual transfer of this initial learning to the contralateral left arm of 21.9% (ABA-group) and 27.6% (AAA-group). Consecutively, both test groups adapted their left arm movements to force field B (ABA-group) or force field A (AAA-group). For the ABA-group, left arm learning caused significant intermanual interference of the initially learned right arm task (68.3% performance decrease). The performance decrease of the AAA-group (10.2%) did not differ from controls (15.5%). These findings suggest that motor control and learning of right and left arm movements involve partly similar neural networks or underlie a vital interhemispheric connectivity. Moreover, our results suggest a preferred internal task representation in extrinsic Cartesian-based coordinates rather than in intrinsic joint-based coordinates because interference was absent when learning was performed in extrinsically equivalent fashion (AAA-group) but interference occurred when learning was performed in intrinsically equivalent fashion (ABA-group). PMID:28459833
Stockinger, Christian; Thürer, Benjamin; Stein, Thorsten
2017-01-01
Intermanual transfer (motor memory generalization across arms) and motor memory interference (impairment of retest performance in consecutive motor learning) are well-investigated motor learning phenomena. However, the interplay of these phenomena remains elusive, i.e., whether intermanual interference occurs when two unimanual tasks are consecutively learned using different arms. Here, we examine intermanual interference when subjects consecutively adapt their right and left arm movements to novel dynamics. We considered two force field tasks A and B which were of the same structure but mirrored orientation (B = -A). The first test group (ABA-group) consecutively learned task A using their right arm and task B using their left arm before being retested for task A with their right arm. Another test group (AAA-group) learned only task A in the same right-left-right arm schedule. Control subjects learned task A using their right arm without intermediate left arm learning. All groups were able to adapt their right arm movements to force field A and both test groups showed significant intermanual transfer of this initial learning to the contralateral left arm of 21.9% (ABA-group) and 27.6% (AAA-group). Consecutively, both test groups adapted their left arm movements to force field B (ABA-group) or force field A (AAA-group). For the ABA-group, left arm learning caused significant intermanual interference of the initially learned right arm task (68.3% performance decrease). The performance decrease of the AAA-group (10.2%) did not differ from controls (15.5%). These findings suggest that motor control and learning of right and left arm movements involve partly similar neural networks or underlie a vital interhemispheric connectivity. Moreover, our results suggest a preferred internal task representation in extrinsic Cartesian-based coordinates rather than in intrinsic joint-based coordinates because interference was absent when learning was performed in extrinsically equivalent fashion (AAA-group) but interference occurred when learning was performed in intrinsically equivalent fashion (ABA-group).
The NASA F-15 Intelligent Flight Control Systems: Generation II
NASA Technical Reports Server (NTRS)
Buschbacher, Mark; Bosworth, John
2006-01-01
The Second Generation (Gen II) control system for the F-15 Intelligent Flight Control System (IFCS) program implements direct adaptive neural networks to demonstrate robust tolerance to faults and failures. The direct adaptive tracking controller integrates learning neural networks (NNs) with a dynamic inversion control law. The term direct adaptive is used because the error between the reference model and the aircraft response is being compensated or directly adapted to minimize error without regard to knowing the cause of the error. No parameter estimation is needed for this direct adaptive control system. In the Gen II design, the feedback errors are regulated with a proportional-plus-integral (PI) compensator. This basic compensator is augmented with an online NN that changes the system gains via an error-based adaptation law to improve aircraft performance at all times, including normal flight, system failures, mispredicted behavior, or changes in behavior resulting from damage.
Verification and Validation Challenges for Adaptive Flight Control of Complex Autonomous Systems
NASA Technical Reports Server (NTRS)
Nguyen, Nhan T.
2018-01-01
Autonomy of aerospace systems requires the ability for flight control systems to be able to adapt to complex uncertain dynamic environment. In spite of the five decades of research in adaptive control, the fact still remains that currently no adaptive control system has ever been deployed on any safety-critical or human-rated production systems such as passenger transport aircraft. The problem lies in the difficulty with the certification of adaptive control systems since existing certification methods cannot readily be used for nonlinear adaptive control systems. Research to address the notion of metrics for adaptive control began to appear in the recent years. These metrics, if accepted, could pave a path towards certification that would potentially lead to the adoption of adaptive control as a future control technology for safety-critical and human-rated production systems. Development of certifiable adaptive control systems represents a major challenge to overcome. Adaptive control systems with learning algorithms will never become part of the future unless it can be proven that they are highly safe and reliable. Rigorous methods for adaptive control software verification and validation must therefore be developed to ensure that adaptive control system software failures will not occur, to verify that the adaptive control system functions as required, to eliminate unintended functionality, and to demonstrate that certification requirements imposed by regulatory bodies such as the Federal Aviation Administration (FAA) can be satisfied. This presentation will discuss some of the technical issues with adaptive flight control and related V&V challenges.
Pohlmeyer, Eric A.; Mahmoudi, Babak; Geng, Shijia; Prins, Noeline W.; Sanchez, Justin C.
2014-01-01
Brain-machine interface (BMI) systems give users direct neural control of robotic, communication, or functional electrical stimulation systems. As BMI systems begin transitioning from laboratory settings into activities of daily living, an important goal is to develop neural decoding algorithms that can be calibrated with a minimal burden on the user, provide stable control for long periods of time, and can be responsive to fluctuations in the decoder’s neural input space (e.g. neurons appearing or being lost amongst electrode recordings). These are significant challenges for static neural decoding algorithms that assume stationary input/output relationships. Here we use an actor-critic reinforcement learning architecture to provide an adaptive BMI controller that can successfully adapt to dramatic neural reorganizations, can maintain its performance over long time periods, and which does not require the user to produce specific kinetic or kinematic activities to calibrate the BMI. Two marmoset monkeys used the Reinforcement Learning BMI (RLBMI) to successfully control a robotic arm during a two-target reaching task. The RLBMI was initialized using random initial conditions, and it quickly learned to control the robot from brain states using only a binary evaluative feedback regarding whether previously chosen robot actions were good or bad. The RLBMI was able to maintain control over the system throughout sessions spanning multiple weeks. Furthermore, the RLBMI was able to quickly adapt and maintain control of the robot despite dramatic perturbations to the neural inputs, including a series of tests in which the neuron input space was deliberately halved or doubled. PMID:24498055
Fast but fleeting: adaptive motor learning processes associated with aging and cognitive decline.
Trewartha, Kevin M; Garcia, Angeles; Wolpert, Daniel M; Flanagan, J Randall
2014-10-01
Motor learning has been shown to depend on multiple interacting learning processes. For example, learning to adapt when moving grasped objects with novel dynamics involves a fast process that adapts and decays quickly-and that has been linked to explicit memory-and a slower process that adapts and decays more gradually. Each process is characterized by a learning rate that controls how strongly motor memory is updated based on experienced errors and a retention factor determining the movement-to-movement decay in motor memory. Here we examined whether fast and slow motor learning processes involved in learning novel dynamics differ between younger and older adults. In addition, we investigated how age-related decline in explicit memory performance influences learning and retention parameters. Although the groups adapted equally well, they did so with markedly different underlying processes. Whereas the groups had similar fast processes, they had different slow processes. Specifically, the older adults exhibited decreased retention in their slow process compared with younger adults. Within the older group, who exhibited considerable variation in explicit memory performance, we found that poor explicit memory was associated with reduced retention in the fast process, as well as the slow process. These findings suggest that explicit memory resources are a determining factor in impairments in the both the fast and slow processes for motor learning but that aging effects on the slow process are independent of explicit memory declines. Copyright © 2014 the authors 0270-6474/14/3413411-11$15.00/0.
Startle reduces recall of a recently learned internal model.
Wright, Zachary; Patton, James L; Ravichandran, Venn
2011-01-01
Recent work has shown that preplanned motor programs are released early from subcortical areas by the using a startling acoustic stimulus (SAS). Our question is whether this response might also contain a recently learned internal model, which draws on experience to predict and compensate for expected perturbations in a feedforward manner. Studies of adaptation to robotic forces have shown some evidence of this, but were potentially confounded by cocontraction caused by startle. We performed a new adaptation experiment using a visually distorted field that could not be confounded by cocontraction. We found that in all subjects that exhibited startle, the startle stimulus (1) reduced performance of the recently learned task (2) reduced after-effect magnitudes. Because startle reduced but did not eliminate the recall of learned control, we suggest that multiple neural centers (cortical and subcortical) are involved in such learning and adaptation, which can impact training areas such as piloting, teleoperation, sports, and rehabilitation. © 2011 IEEE
Research on intelligent algorithm of electro - hydraulic servo control system
NASA Astrophysics Data System (ADS)
Wang, Yannian; Zhao, Yuhui; Liu, Chengtao
2017-09-01
In order to adapt the nonlinear characteristics of the electro-hydraulic servo control system and the influence of complex interference in the industrial field, using a fuzzy PID switching learning algorithm is proposed and a fuzzy PID switching learning controller is designed and applied in the electro-hydraulic servo controller. The designed controller not only combines the advantages of the fuzzy control and PID control, but also introduces the learning algorithm into the switching function, which makes the learning of the three parameters in the switching function can avoid the instability of the system during the switching between the fuzzy control and PID control algorithms. It also makes the switch between these two control algorithm more smoother than that of the conventional fuzzy PID.
Evolutionary online behaviour learning and adaptation in real robots
Correia, Luís; Christensen, Anders Lyhne
2017-01-01
Online evolution of behavioural control on real robots is an open-ended approach to autonomous learning and adaptation: robots have the potential to automatically learn new tasks and to adapt to changes in environmental conditions, or to failures in sensors and/or actuators. However, studies have so far almost exclusively been carried out in simulation because evolution in real hardware has required several days or weeks to produce capable robots. In this article, we successfully evolve neural network-based controllers in real robotic hardware to solve two single-robot tasks and one collective robotics task. Controllers are evolved either from random solutions or from solutions pre-evolved in simulation. In all cases, capable solutions are found in a timely manner (1 h or less). Results show that more accurate simulations may lead to higher-performing controllers, and that completing the optimization process in real robots is meaningful, even if solutions found in simulation differ from solutions in reality. We furthermore demonstrate for the first time the adaptive capabilities of online evolution in real robotic hardware, including robots able to overcome faults injected in the motors of multiple units simultaneously, and to modify their behaviour in response to changes in the task requirements. We conclude by assessing the contribution of each algorithmic component on the performance of the underlying evolutionary algorithm. PMID:28791130
Liu, Yan-Jun; Tang, Li; Tong, Shaocheng; Chen, C L Philip; Li, Dong-Juan
2015-01-01
Based on the neural network (NN) approximator, an online reinforcement learning algorithm is proposed for a class of affine multiple input and multiple output (MIMO) nonlinear discrete-time systems with unknown functions and disturbances. In the design procedure, two networks are provided where one is an action network to generate an optimal control signal and the other is a critic network to approximate the cost function. An optimal control signal and adaptation laws can be generated based on two NNs. In the previous approaches, the weights of critic and action networks are updated based on the gradient descent rule and the estimations of optimal weight vectors are directly adjusted in the design. Consequently, compared with the existing results, the main contributions of this paper are: 1) only two parameters are needed to be adjusted, and thus the number of the adaptation laws is smaller than the previous results and 2) the updating parameters do not depend on the number of the subsystems for MIMO systems and the tuning rules are replaced by adjusting the norms on optimal weight vectors in both action and critic networks. It is proven that the tracking errors, the adaptation laws, and the control inputs are uniformly bounded using Lyapunov analysis method. The simulation examples are employed to illustrate the effectiveness of the proposed algorithm.
Auditory-Perceptual Learning Improves Speech Motor Adaptation in Children
Shiller, Douglas M.; Rochon, Marie-Lyne
2015-01-01
Auditory feedback plays an important role in children’s speech development by providing the child with information about speech outcomes that is used to learn and fine-tune speech motor plans. The use of auditory feedback in speech motor learning has been extensively studied in adults by examining oral motor responses to manipulations of auditory feedback during speech production. Children are also capable of adapting speech motor patterns to perceived changes in auditory feedback, however it is not known whether their capacity for motor learning is limited by immature auditory-perceptual abilities. Here, the link between speech perceptual ability and the capacity for motor learning was explored in two groups of 5–7-year-old children who underwent a period of auditory perceptual training followed by tests of speech motor adaptation to altered auditory feedback. One group received perceptual training on a speech acoustic property relevant to the motor task while a control group received perceptual training on an irrelevant speech contrast. Learned perceptual improvements led to an enhancement in speech motor adaptation (proportional to the perceptual change) only for the experimental group. The results indicate that children’s ability to perceive relevant speech acoustic properties has a direct influence on their capacity for sensory-based speech motor adaptation. PMID:24842067
Grossberg, Stephen
2015-09-24
This article provides an overview of neural models of synaptic learning and memory whose expression in adaptive behavior depends critically on the circuits and systems in which the synapses are embedded. It reviews Adaptive Resonance Theory, or ART, models that use excitatory matching and match-based learning to achieve fast category learning and whose learned memories are dynamically stabilized by top-down expectations, attentional focusing, and memory search. ART clarifies mechanistic relationships between consciousness, learning, expectation, attention, resonance, and synchrony. ART models are embedded in ARTSCAN architectures that unify processes of invariant object category learning, recognition, spatial and object attention, predictive remapping, and eye movement search, and that clarify how conscious object vision and recognition may fail during perceptual crowding and parietal neglect. The generality of learned categories depends upon a vigilance process that is regulated by acetylcholine via the nucleus basalis. Vigilance can get stuck at too high or too low values, thereby causing learning problems in autism and medial temporal amnesia. Similar synaptic learning laws support qualitatively different behaviors: Invariant object category learning in the inferotemporal cortex; learning of grid cells and place cells in the entorhinal and hippocampal cortices during spatial navigation; and learning of time cells in the entorhinal-hippocampal system during adaptively timed conditioning, including trace conditioning. Spatial and temporal processes through the medial and lateral entorhinal-hippocampal system seem to be carried out with homologous circuit designs. Variations of a shared laminar neocortical circuit design have modeled 3D vision, speech perception, and cognitive working memory and learning. A complementary kind of inhibitory matching and mismatch learning controls movement. This article is part of a Special Issue entitled SI: Brain and Memory. Copyright © 2014 Elsevier B.V. All rights reserved.
Intelligent adaptive nonlinear flight control for a high performance aircraft with neural networks.
Savran, Aydogan; Tasaltin, Ramazan; Becerikli, Yasar
2006-04-01
This paper describes the development of a neural network (NN) based adaptive flight control system for a high performance aircraft. The main contribution of this work is that the proposed control system is able to compensate the system uncertainties, adapt to the changes in flight conditions, and accommodate the system failures. The underlying study can be considered in two phases. The objective of the first phase is to model the dynamic behavior of a nonlinear F-16 model using NNs. Therefore a NN-based adaptive identification model is developed for three angular rates of the aircraft. An on-line training procedure is developed to adapt the changes in the system dynamics and improve the identification accuracy. In this procedure, a first-in first-out stack is used to store a certain history of the input-output data. The training is performed over the whole data in the stack at every stage. To speed up the convergence rate and enhance the accuracy for achieving the on-line learning, the Levenberg-Marquardt optimization method with a trust region approach is adapted to train the NNs. The objective of the second phase is to develop intelligent flight controllers. A NN-based adaptive PID control scheme that is composed of an emulator NN, an estimator NN, and a discrete time PID controller is developed. The emulator NN is used to calculate the system Jacobian required to train the estimator NN. The estimator NN, which is trained on-line by propagating the output error through the emulator, is used to adjust the PID gains. The NN-based adaptive PID control system is applied to control three angular rates of the nonlinear F-16 model. The body-axis pitch, roll, and yaw rates are fed back via the PID controllers to the elevator, aileron, and rudder actuators, respectively. The resulting control system has learning, adaptation, and fault-tolerant abilities. It avoids the storage and interpolation requirements for the too many controller parameters of a typical flight control system. Performance of the control system is successfully tested by performing several six-degrees-of-freedom nonlinear simulations.
Adaptive, fast walking in a biped robot under neuronal control and learning.
Manoonpong, Poramate; Geng, Tao; Kulvicius, Tomas; Porr, Bernd; Wörgötter, Florentin
2007-07-01
Human walking is a dynamic, partly self-stabilizing process relying on the interaction of the biomechanical design with its neuronal control. The coordination of this process is a very difficult problem, and it has been suggested that it involves a hierarchy of levels, where the lower ones, e.g., interactions between muscles and the spinal cord, are largely autonomous, and where higher level control (e.g., cortical) arises only pointwise, as needed. This requires an architecture of several nested, sensori-motor loops where the walking process provides feedback signals to the walker's sensory systems, which can be used to coordinate its movements. To complicate the situation, at a maximal walking speed of more than four leg-lengths per second, the cycle period available to coordinate all these loops is rather short. In this study we present a planar biped robot, which uses the design principle of nested loops to combine the self-stabilizing properties of its biomechanical design with several levels of neuronal control. Specifically, we show how to adapt control by including online learning mechanisms based on simulated synaptic plasticity. This robot can walk with a high speed (>3.0 leg length/s), self-adapting to minor disturbances, and reacting in a robust way to abruptly induced gait changes. At the same time, it can learn walking on different terrains, requiring only few learning experiences. This study shows that the tight coupling of physical with neuronal control, guided by sensory feedback from the walking pattern itself, combined with synaptic learning may be a way forward to better understand and solve coordination problems in other complex motor tasks.
Network congestion control algorithm based on Actor-Critic reinforcement learning model
NASA Astrophysics Data System (ADS)
Xu, Tao; Gong, Lina; Zhang, Wei; Li, Xuhong; Wang, Xia; Pan, Wenwen
2018-04-01
Aiming at the network congestion control problem, a congestion control algorithm based on Actor-Critic reinforcement learning model is designed. Through the genetic algorithm in the congestion control strategy, the network congestion problems can be better found and prevented. According to Actor-Critic reinforcement learning, the simulation experiment of network congestion control algorithm is designed. The simulation experiments verify that the AQM controller can predict the dynamic characteristics of the network system. Moreover, the learning strategy is adopted to optimize the network performance, and the dropping probability of packets is adaptively adjusted so as to improve the network performance and avoid congestion. Based on the above finding, it is concluded that the network congestion control algorithm based on Actor-Critic reinforcement learning model can effectively avoid the occurrence of TCP network congestion.
Yuan, Chengzhi; Licht, Stephen; He, Haibo
2017-09-26
In this paper, a new concept of formation learning control is introduced to the field of formation control of multiple autonomous underwater vehicles (AUVs), which specifies a joint objective of distributed formation tracking control and learning/identification of nonlinear uncertain AUV dynamics. A novel two-layer distributed formation learning control scheme is proposed, which consists of an upper-layer distributed adaptive observer and a lower-layer decentralized deterministic learning controller. This new formation learning control scheme advances existing techniques in three important ways: 1) the multi-AUV system under consideration has heterogeneous nonlinear uncertain dynamics; 2) the formation learning control protocol can be designed and implemented by each local AUV agent in a fully distributed fashion without using any global information; and 3) in addition to the formation control performance, the distributed control protocol is also capable of accurately identifying the AUVs' heterogeneous nonlinear uncertain dynamics and utilizing experiences to improve formation control performance. Extensive simulations have been conducted to demonstrate the effectiveness of the proposed results.
NASA Astrophysics Data System (ADS)
Rigatos, Gerasimos
2016-12-01
It is shown that the model of the hypothalamic-pituitary-adrenal gland axis is a differentially flat one and this permits to transform it to the so-called linear canonical form. For the new description of the system's dynamics the transformed control inputs contain unknown terms which depend on the system's parameters. To identify these terms an adaptive fuzzy approximator is used in the control loop. Thus an adaptive fuzzy control scheme is implemented in which the unknown or unmodeled system dynamics is approximated by neurofuzzy networks and next this information is used by a feedback controller that makes the state variables (CRH - corticotropin releasing hormone, adenocortocotropic hormone - ACTH, cortisol) of the hypothalamic-pituitary-adrenal gland axis model converge to the desirable levels (setpoints). This adaptive control scheme is exclusively implemented with the use of output feedback, while the state vector elements which are not directly measured are estimated with the use of a state observer that operates in the control loop. The learning rate of the adaptive fuzzy system is suitably computed from Lyapunov analysis, so as to assure that both the learning procedure for the unknown system's parameters, the dynamics of the observer and the dynamics of the control loop will remain stable. The performed Lyapunov stability analysis depends on two Riccati equations, one associated with the feedback controller and one associated with the state observer. Finally, it is proven that for the control scheme that comprises the feedback controller, the state observer and the neurofuzzy approximator, an H-infinity tracking performance can be succeeded.
Modelling and Optimizing Mathematics Learning in Children
ERIC Educational Resources Information Center
Käser, Tanja; Busetto, Alberto Giovanni; Solenthaler, Barbara; Baschera, Gian-Marco; Kohn, Juliane; Kucian, Karin; von Aster, Michael; Gross, Markus
2013-01-01
This study introduces a student model and control algorithm, optimizing mathematics learning in children. The adaptive system is integrated into a computer-based training system for enhancing numerical cognition aimed at children with developmental dyscalculia or difficulties in learning mathematics. The student model consists of a dynamic…
Evolving fuzzy rules in a learning classifier system
NASA Technical Reports Server (NTRS)
Valenzuela-Rendon, Manuel
1993-01-01
The fuzzy classifier system (FCS) combines the ideas of fuzzy logic controllers (FLC's) and learning classifier systems (LCS's). It brings together the expressive powers of fuzzy logic as it has been applied in fuzzy controllers to express relations between continuous variables, and the ability of LCS's to evolve co-adapted sets of rules. The goal of the FCS is to develop a rule-based system capable of learning in a reinforcement regime, and that can potentially be used for process control.
Luo, Ying; Chen, Yangquan; Pi, Youguo
2010-10-01
Cogging effect which can be treated as a type of position-dependent periodic disturbance, is a serious disadvantage of the permanent magnetic synchronous motor (PMSM). In this paper, based on a simulation system model of PMSM position servo control, the cogging force, viscous friction, and applied load in the real PMSM control system are considered and presented. A dual high-order periodic adaptive learning compensation (DHO-PALC) method is proposed to minimize the cogging effect on the PMSM position and velocity servo system. In this DHO-PALC scheme, more than one previous periods stored information of both the composite tracking error and the estimate of the cogging force is used for the control law updating. Asymptotical stability proof with the proposed DHO-PALC scheme is presented. Simulation is implemented on the PMSM servo system model to illustrate the proposed method. When the constant speed reference is applied, the DHO-PALC can achieve a faster learning convergence speed than the first-order periodic adaptive learning compensation (FO-PALC). Moreover, when the designed reference signal changes periodically, the proposed DHO-PALC can obtain not only faster convergence speed, but also much smaller final error bound than the FO-PALC. Copyright © 2010 ISA. Published by Elsevier Ltd. All rights reserved.
Development of adaptive control applied to chaotic systems
NASA Astrophysics Data System (ADS)
Rhode, Martin Andreas
1997-12-01
Continuous-time derivative control and adaptive map-based recursive feedback control techniques are used to control chaos in a variety of systems and in situations that are of practical interest. The theoretical part of the research includes the review of fundamental concept of control theory in the context of its applications to deterministic chaotic systems, the development of a new adaptive algorithm to identify the linear system properties necessary for control, and the extension of the recursive proportional feedback control technique, RPF, to high dimensional systems. Chaos control was applied to models of a thermal pulsed combustor, electro-chemical dissolution and the hyperchaotic Rossler system. Important implications for combustion engineering were suggested by successful control of the model of the thermal pulsed combustor. The system was automatically tracked while maintaining control into regions of parameter and state space where no stable attractors exist. In a simulation of the electrochemical dissolution system, application of derivative control to stabilize a steady state, and adaptive RPF to stabilize a period one orbit, was demonstrated. The high dimensional adaptive control algorithm was applied in a simulation using the Rossler hyperchaotic system, where a period-two orbit with two unstable directions was stabilized and tracked over a wide range of a system parameter. In the experimental part, the electrochemical system was studied in parameter space, by scanning the applied potential and the frequency of the rotating copper disk. The automated control algorithm is demonstrated to be effective when applied to stabilize a period-one orbit in the experiment. We show the necessity of small random perturbations applied to the system in order to both learn the dynamics and control the system at the same time. The simultaneous learning and control capability is shown to be an important part of the active feedback control.
Jeste, Shafali S; Kirkham, Natasha; Senturk, Damla; Hasenstab, Kyle; Sugar, Catherine; Kupelian, Chloe; Baker, Elizabeth; Sanders, Andrew J; Shimizu, Christina; Norona, Amanda; Paparella, Tanya; Freeman, Stephanny F N; Johnson, Scott P
2015-01-01
Statistical learning is characterized by detection of regularities in one's environment without an awareness or intention to learn, and it may play a critical role in language and social behavior. Accordingly, in this study we investigated the electrophysiological correlates of visual statistical learning in young children with autism spectrum disorder (ASD) using an event-related potential shape learning paradigm, and we examined the relation between visual statistical learning and cognitive function. Compared to typically developing (TD) controls, the ASD group as a whole showed reduced evidence of learning as defined by N1 (early visual discrimination) and P300 (attention to novelty) components. Upon further analysis, in the ASD group there was a positive correlation between N1 amplitude difference and non-verbal IQ, and a positive correlation between P300 amplitude difference and adaptive social function. Children with ASD and a high non-verbal IQ and high adaptive social function demonstrated a distinctive pattern of learning. This is the first study to identify electrophysiological markers of visual statistical learning in children with ASD. Through this work we have demonstrated heterogeneity in statistical learning in ASD that maps onto non-verbal cognition and adaptive social function. © 2014 John Wiley & Sons Ltd.
Bio-inspired adaptive feedback error learning architecture for motor control.
Tolu, Silvia; Vanegas, Mauricio; Luque, Niceto R; Garrido, Jesús A; Ros, Eduardo
2012-10-01
This study proposes an adaptive control architecture based on an accurate regression method called Locally Weighted Projection Regression (LWPR) and on a bio-inspired module, such as a cerebellar-like engine. This hybrid architecture takes full advantage of the machine learning module (LWPR kernel) to abstract an optimized representation of the sensorimotor space while the cerebellar component integrates this to generate corrective terms in the framework of a control task. Furthermore, we illustrate how the use of a simple adaptive error feedback term allows to use the proposed architecture even in the absence of an accurate analytic reference model. The presented approach achieves an accurate control with low gain corrective terms (for compliant control schemes). We evaluate the contribution of the different components of the proposed scheme comparing the obtained performance with alternative approaches. Then, we show that the presented architecture can be used for accurate manipulation of different objects when their physical properties are not directly known by the controller. We evaluate how the scheme scales for simulated plants of high Degrees of Freedom (7-DOFs).
Learned Manipulation at Unconstrained Contacts Does Not Transfer across Hands
Fu, Qiushi; Choi, Jason Y.; Gordon, Andrew M.; Jesunathadas, Mark; Santello, Marco
2014-01-01
Recent studies about sensorimotor control of the human hand have focused on how dexterous manipulation is learned and generalized. Here we address this question by testing the extent to which learned manipulation can be transferred when the contralateral hand is used and/or object orientation is reversed. We asked subjects to use a precision grip to lift a grip device with an asymmetrical mass distribution while minimizing object roll during lifting by generating a compensatory torque. Subjects were allowed to grasp anywhere on the object’s vertical surfaces, and were therefore able to modulate both digit positions and forces. After every block of eight trials performed in one manipulation context (i.e., using the right hand and at a given object orientation), subjects had to lift the same object in the second context for one trial (transfer trial). Context changes were made by asking subjects to switch the hand used to lift the object and/or rotate the object 180° about a vertical axis. Therefore, three transfer conditions, hand switch (HS), object rotation (OR), and both hand switch and object rotation (HS+OR), were tested and compared with hand matched control groups who did not experience context changes. We found that subjects in all transfer conditions adapted digit positions across multiple transfer trials similar to the learning of control groups, regardless of different changes of contexts. Moreover, subjects in both HS and HS+OR group also adapted digit forces similar to the control group, suggesting independent learning of the left hand. In contrast, the OR group showed significant negative transfer of the compensatory torque due to an inability to adapt digit forces. Our results indicate that internal representations of dexterous manipulation tasks may be primarily built through the hand used for learning and cannot be transferred across hands. PMID:25233091
Indirect decentralized learning control
NASA Technical Reports Server (NTRS)
Longman, Richard W.; Lee, Soo C.; Phan, M.
1992-01-01
The new field of learning control develops controllers that learn to improve their performance at executing a given task, based on experience performing this specific task. In a previous work, the authors presented a theory of indirect learning control based on use of indirect adaptive control concepts employing simultaneous identification and control. This paper develops improved indirect learning control algorithms, and studies the use of such controllers in decentralized systems. The original motivation of the learning control field was learning in robots doing repetitive tasks such as on an assembly line. This paper starts with decentralized discrete time systems, and progresses to the robot application, modeling the robot as a time varying linear system in the neighborhood of the nominal trajectory, and using the usual robot controllers that are decentralized, treating each link as if it is independent of any coupling with other links. The basic result of the paper is to show that stability of the indirect learning controllers for all subsystems when the coupling between subsystems is turned off, assures convergence to zero tracking error of the decentralized indirect learning control of the coupled system, provided that the sample time in the digital learning controller is sufficiently short.
ERIC Educational Resources Information Center
Cook, David A.; Thompson, Warren G.; Thomas, Kris G.; Thomas, Matthew R.
2009-01-01
Background: Adaptation to learning styles has been proposed to enhance learning. Objective: We hypothesized that learners with sensing learning style would perform better using a problem-first instructional method while intuitive learners would do better using an information-first method. Design: Randomized, controlled, crossover trial. Setting:…
The congruency sequence effect 3.0: a critical test of conflict adaptation.
Duthoo, Wout; Abrahamse, Elger L; Braem, Senne; Boehler, C Nico; Notebaert, Wim
2014-01-01
Over the last two decades, the congruency sequence effect (CSE) -the finding of a reduced congruency effect following incongruent trials in conflict tasks- has played a central role in advancing research on cognitive control. According to the influential conflict-monitoring account, the CSE reflects adjustments in selective attention that enhance task focus when needed, often termed conflict adaptation. However, this dominant interpretation of the CSE has been called into question by several alternative accounts that stress the role of episodic memory processes: feature binding and (stimulus-response) contingency learning. To evaluate the notion of conflict adaptation in accounting for the CSE, we construed versions of three widely used experimental paradigms (the colour-word Stroop, picture-word Stroop and flanker task) that effectively control for feature binding and contingency learning. Results revealed that a CSE can emerge in all three tasks. This strongly suggests a contribution of attentional control to the CSE and highlights the potential of these unprecedentedly clean paradigms for further examining cognitive control.
The Congruency Sequence Effect 3.0: A Critical Test of Conflict Adaptation
Duthoo, Wout; Abrahamse, Elger L.; Braem, Senne; Boehler, C. Nico; Notebaert, Wim
2014-01-01
Over the last two decades, the congruency sequence effect (CSE) –the finding of a reduced congruency effect following incongruent trials in conflict tasks– has played a central role in advancing research on cognitive control. According to the influential conflict-monitoring account, the CSE reflects adjustments in selective attention that enhance task focus when needed, often termed conflict adaptation. However, this dominant interpretation of the CSE has been called into question by several alternative accounts that stress the role of episodic memory processes: feature binding and (stimulus-response) contingency learning. To evaluate the notion of conflict adaptation in accounting for the CSE, we construed versions of three widely used experimental paradigms (the colour-word Stroop, picture-word Stroop and flanker task) that effectively control for feature binding and contingency learning. Results revealed that a CSE can emerge in all three tasks. This strongly suggests a contribution of attentional control to the CSE and highlights the potential of these unprecedentedly clean paradigms for further examining cognitive control. PMID:25340396
Learning from ISS-modular adaptive NN control of nonlinear strict-feedback systems.
Wang, Cong; Wang, Min; Liu, Tengfei; Hill, David J
2012-10-01
This paper studies learning from adaptive neural control (ANC) for a class of nonlinear strict-feedback systems with unknown affine terms. To achieve the purpose of learning, a simple input-to-state stability (ISS) modular ANC method is first presented to ensure the boundedness of all the signals in the closed-loop system and the convergence of tracking errors in finite time. Subsequently, it is proven that learning with the proposed stable ISS-modular ANC can be achieved. The cascade structure and unknown affine terms of the considered systems make it very difficult to achieve learning using existing methods. To overcome these difficulties, the stable closed-loop system in the control process is decomposed into a series of linear time-varying (LTV) perturbed subsystems with the appropriate state transformation. Using a recursive design, the partial persistent excitation condition for the radial basis function neural network (NN) is established, which guarantees exponential stability of LTV perturbed subsystems. Consequently, accurate approximation of the closed-loop system dynamics is achieved in a local region along recurrent orbits of closed-loop signals, and learning is implemented during a closed-loop feedback control process. The learned knowledge is reused to achieve stability and an improved performance, thereby avoiding the tremendous repeated training process of NNs. Simulation studies are given to demonstrate the effectiveness of the proposed method.
Learning and tuning fuzzy logic controllers through reinforcements.
Berenji, H R; Khedkar, P
1992-01-01
A method for learning and tuning a fuzzy logic controller based on reinforcements from a dynamic system is presented. It is shown that: the generalized approximate-reasoning-based intelligent control (GARIC) architecture learns and tunes a fuzzy logic controller even when only weak reinforcement, such as a binary failure signal, is available; introduces a new conjunction operator in computing the rule strengths of fuzzy control rules; introduces a new localized mean of maximum (LMOM) method in combining the conclusions of several firing control rules; and learns to produce real-valued control actions. Learning is achieved by integrating fuzzy inference into a feedforward network, which can then adaptively improve performance by using gradient descent methods. The GARIC architecture is applied to a cart-pole balancing system and demonstrates significant improvements in terms of the speed of learning and robustness to changes in the dynamic system's parameters over previous schemes for cart-pole balancing.
Hassett, Thomas C; Hampton, Robert R
2017-05-01
Functionally distinct memory systems likely evolved in response to incompatible demands placed on learning by distinct environmental conditions. Working memory appears adapted, in part, for conditions that change frequently, making rapid acquisition and brief retention of information appropriate. In contrast, habits form gradually over many experiences, adapting organisms to contingencies of reinforcement that are stable over relatively long intervals. Serial reversal learning provides an opportunity to simultaneously examine the processes involved in adapting to rapidly changing and relatively stable contingencies. In serial reversal learning, selecting one of the two simultaneously presented stimuli is positively reinforced, while selection of the other is not. After a preference for the positive stimulus develops, the contingencies of reinforcement reverse. Naïve subjects adapt to such reversals gradually, perseverating in selection of the previously rewarded stimulus. Experts reverse rapidly according to a win-stay, lose-shift response pattern. We assessed whether a change in the relative control of choice by habit and working memory accounts for the development of serial reversal learning expertise. Across three experiments, we applied manipulations intended to attenuate the contribution of working memory but leave the contribution of habit intact. We contrasted performance following long and short intervals in Experiments 1 and 2, and we interposed a competing cognitive load between trials in Experiment 3. These manipulations slowed the acquisition of reversals in expert subjects, but not naïve subjects, indicating that serial reversal learning expertise is facilitated by a shift in the control of choice from passively acquired habit to actively maintained working memory.
Adaptive Optimization of Aircraft Engine Performance Using Neural Networks
NASA Technical Reports Server (NTRS)
Simon, Donald L.; Long, Theresa W.
1995-01-01
Preliminary results are presented on the development of an adaptive neural network based control algorithm to enhance aircraft engine performance. This work builds upon a previous National Aeronautics and Space Administration (NASA) effort known as Performance Seeking Control (PSC). PSC is an adaptive control algorithm which contains a model of the aircraft's propulsion system which is updated on-line to match the operation of the aircraft's actual propulsion system. Information from the on-line model is used to adapt the control system during flight to allow optimal operation of the aircraft's propulsion system (inlet, engine, and nozzle) to improve aircraft engine performance without compromising reliability or operability. Performance Seeking Control has been shown to yield reductions in fuel flow, increases in thrust, and reductions in engine fan turbine inlet temperature. The neural network based adaptive control, like PSC, will contain a model of the propulsion system which will be used to calculate optimal control commands on-line. Hopes are that it will be able to provide some additional benefits above and beyond those of PSC. The PSC algorithm is computationally intensive, it is valid only at near steady-state flight conditions, and it has no way to adapt or learn on-line. These issues are being addressed in the development of the optimal neural controller. Specialized neural network processing hardware is being developed to run the software, the algorithm will be valid at steady-state and transient conditions, and will take advantage of the on-line learning capability of neural networks. Future plans include testing the neural network software and hardware prototype against an aircraft engine simulation. In this paper, the proposed neural network software and hardware is described and preliminary neural network training results are presented.
Shih, Peter; Kaul, Brian C; Jagannathan, Sarangapani; Drallmeier, James A
2009-10-01
A novel reinforcement-learning-based output adaptive neural network (NN) controller, which is also referred to as the adaptive-critic NN controller, is developed to deliver the desired tracking performance for a class of nonlinear discrete-time systems expressed in nonstrict feedback form in the presence of bounded and unknown disturbances. The adaptive-critic NN controller consists of an observer, a critic, and two action NNs. The observer estimates the states and output, and the two action NNs provide virtual and actual control inputs to the nonlinear discrete-time system. The critic approximates a certain strategic utility function, and the action NNs minimize the strategic utility function and control inputs. All NN weights adapt online toward minimization of a performance index, utilizing the gradient-descent-based rule, in contrast with iteration-based adaptive-critic schemes. Lyapunov functions are used to show the stability of the closed-loop tracking error, weights, and observer estimates. Separation and certainty equivalence principles, persistency of excitation condition, and linearity in the unknown parameter assumption are not needed. Experimental results on a spark ignition (SI) engine operating lean at an equivalence ratio of 0.75 show a significant (25%) reduction in cyclic dispersion in heat release with control, while the average fuel input changes by less than 1% compared with the uncontrolled case. Consequently, oxides of nitrogen (NO(x)) drop by 30%, and unburned hydrocarbons drop by 16% with control. Overall, NO(x)'s are reduced by over 80% compared with stoichiometric levels.
D'Angelo, Maria C; Jiménez, Luis; Milliken, Bruce; Lupiáñez, Juan
2013-01-01
Individuals experience less interference from conflicting information following events that contain conflicting information. Recently, Jiménez, Lupiáñez, and Vaquero (2009) demonstrated that such adaptations to conflict occur even when the source of conflict arises from implicit knowledge of sequences. There is accumulating evidence that momentary changes in adaptations made in response to conflicting information are conflict-type specific (e.g., Funes, Lupiáñez, & Humphreys, 2010a), suggesting that there are multiple modes of control. The current study examined whether conflict-specific sequential congruency effects occur when the 2 sources of conflict are implicitly learned. Participants implicitly learned a motor sequence while simultaneously learning a perceptual sequence. In a first experiment, after learning the 2 orthogonal sequences, participants expressed knowledge of the 2 sequences independently of each other in a transfer phase. In Experiments 2 and 3, within each sequence, the presence of a single control trial disrupted the expression of this specific type of learning on the following trial. There was no evidence of cross-conflict modulations in the expression of sequence learning. The results suggest that the mechanisms involved in transient shifts in conflict-specific control, as reflected in sequential congruency effects, are also engaged when the source of conflict is implicit. (c) 2013 APA, all rights reserved.
Degraded expression of learned feedforward control in movements released by startle.
Wright, Zachary A; Carlsen, Anthony N; MacKinnon, Colum D; Patton, James L
2015-08-01
Recent work has shown that preplanned motor programs can be rapidly released via fast conducting pathways using a startling acoustic stimulus. Our question was whether the startle-elicited response might also release a recently learned internal model, which draws on experience to predict and compensate for expected perturbations in a feedforward manner. Our initial investigation using adaptation to robotically produced forces showed some evidence of this, but the results were potentially confounded by co-contraction caused by startle. In this study, we eliminated this confound by asking subjects to make reaching movements in the presence of a visual distortion. Results show that a startle stimulus (1) decreased performance of the recently learned task and (2) reduced after-effect magnitude. Since the recall of learned control was reduced, but not eliminated during startle trials, we suggest that multiple neural centers (cortical and subcortical) are involved in such learning and adaptation. These findings have implications for motor training in areas such as piloting, teleoperation, sports, and rehabilitation.
SVM-based tree-type neural networks as a critic in adaptive critic designs for control.
Deb, Alok Kanti; Jayadeva; Gopal, Madan; Chandra, Suresh
2007-07-01
In this paper, we use the approach of adaptive critic design (ACD) for control, specifically, the action-dependent heuristic dynamic programming (ADHDP) method. A least squares support vector machine (SVM) regressor has been used for generating the control actions, while an SVM-based tree-type neural network (NN) is used as the critic. After a failure occurs, the critic and action are retrained in tandem using the failure data. Failure data is binary classification data, where the number of failure states are very few as compared to the number of no-failure states. The difficulty of conventional multilayer feedforward NNs in learning this type of classification data has been overcome by using the SVM-based tree-type NN, which due to its feature to add neurons to learn misclassified data, has the capability to learn any binary classification data without a priori choice of the number of neurons or the structure of the network. The capability of the trained controller to handle unforeseen situations is demonstrated.
Effectiveness of a Learner-Directed Model for e-Learning
ERIC Educational Resources Information Center
Lee, Stella; Barker, Trevor; Kumar, Vivekanandan Suresh
2016-01-01
It is a hard task to strike a balance between extents of control a learner exercises and the amount of guidance, active or passive, afforded by the learning environment to guide, support, and motivate the learner. Adaptive systems strive to find the right balance in a spectrum that spans between self-control and system-guidance. They also concern…
Shih, Peter; Kaul, Brian C; Jagannathan, S; Drallmeier, James A
2008-08-01
A novel reinforcement-learning-based dual-control methodology adaptive neural network (NN) controller is developed to deliver a desired tracking performance for a class of complex feedback nonlinear discrete-time systems, which consists of a second-order nonlinear discrete-time system in nonstrict feedback form and an affine nonlinear discrete-time system, in the presence of bounded and unknown disturbances. For example, the exhaust gas recirculation (EGR) operation of a spark ignition (SI) engine is modeled by using such a complex nonlinear discrete-time system. A dual-controller approach is undertaken where primary adaptive critic NN controller is designed for the nonstrict feedback nonlinear discrete-time system whereas the secondary one for the affine nonlinear discrete-time system but the controllers together offer the desired performance. The primary adaptive critic NN controller includes an NN observer for estimating the states and output, an NN critic, and two action NNs for generating virtual control and actual control inputs for the nonstrict feedback nonlinear discrete-time system, whereas an additional critic NN and an action NN are included for the affine nonlinear discrete-time system by assuming the state availability. All NN weights adapt online towards minimization of a certain performance index, utilizing gradient-descent-based rule. Using Lyapunov theory, the uniformly ultimate boundedness (UUB) of the closed-loop tracking error, weight estimates, and observer estimates are shown. The adaptive critic NN controller performance is evaluated on an SI engine operating with high EGR levels where the controller objective is to reduce cyclic dispersion in heat release while minimizing fuel intake. Simulation and experimental results indicate that engine out emissions drop significantly at 20% EGR due to reduction in dispersion in heat release thus verifying the dual-control approach.
Creating a Reinforcement Learning Controller for Functional Electrical Stimulation of a Human Arm*
Thomas, Philip S.; Branicky, Michael; van den Bogert, Antonie; Jagodnik, Kathleen
2010-01-01
Clinical tests have shown that the dynamics of a human arm, controlled using Functional Electrical Stimulation (FES), can vary significantly between and during trials. In this paper, we study the application of Reinforcement Learning to create a controller that can adapt to these changing dynamics of a human arm. Development and tests were done in simulation using a two-dimensional arm model and Hill-based muscle dynamics. An actor-critic architecture is used with artificial neural networks for both the actor and the critic. We begin by training it using a Proportional Derivative (PD) controller as a supervisor. We then make clinically relevant changes to the dynamics of the arm and test the actor-critic’s ability to adapt without supervision in a reasonable number of episodes. PMID:22081795
Experimental Verification of Electric Drive Technologies Based on Artificial Intelligence Tools
NASA Technical Reports Server (NTRS)
Rubaai, Ahmed; Ricketts, Daniel; Kotaru, Raj; Thomas, Robert; Noga, Donald F. (Technical Monitor); Kankam, Mark D. (Technical Monitor)
2000-01-01
In this report, a fully integrated prototype of a flight servo control system is successfully developed and implemented using brushless dc motors. The control system is developed by the fuzzy logic theory, and implemented with a multilayer neural network. First, a neural network-based architecture is introduced for fuzzy logic control. The characteristic rules and their membership functions of fuzzy systems are represented as the processing nodes in the neural network structure. The network structure and the parameter learning are performed simultaneously and online in the fuzzy-neural network system. The structure learning is based on the partition of input space. The parameter learning is based on the supervised gradient decent method, using a delta adaptation law. Using experimental setup, the performance of the proposed control system is evaluated under various operating conditions. Test results are presented and discussed in the report. The proposed learning control system has several advantages, namely, simple structure and learning capability, robustness and high tracking performance and few nodes at hidden layers. In comparison with the PI controller, the proposed fuzzy-neural network system can yield a better dynamic performance with shorter settling time, and without overshoot. Experimental results have shown that the proposed control system is adaptive and robust in responding to a wide range of operating conditions. In summary, the goal of this study is to design and implement-advanced servosystems to actuate control surfaces for flight vehicles, namely, aircraft and helicopters, missiles and interceptors, and mini- and micro-air vehicles.
The remapping of space in motor learning and human-machine interfaces
Mussa-Ivaldi, F.A.; Danziger, Z.
2009-01-01
Studies of motor adaptation to patterns of deterministic forces have revealed the ability of the motor control system to form and use predictive representations of the environment. One of the most fundamental elements of our environment is space itself. This article focuses on the notion of Euclidean space as it applies to common sensory motor experiences. Starting from the assumption that we interact with the world through a system of neural signals, we observe that these signals are not inherently endowed with metric properties of the ordinary Euclidean space. The ability of the nervous system to represent these properties depends on adaptive mechanisms that reconstruct the Euclidean metric from signals that are not Euclidean. Gaining access to these mechanisms will reveal the process by which the nervous system handles novel sophisticated coordinate transformation tasks, thus highlighting possible avenues to create functional human-machine interfaces that can make that task much easier. A set of experiments is presented that demonstrate the ability of the sensory-motor system to reorganize coordination in novel geometrical environments. In these environments multiple degrees of freedom of body motions are used to control the coordinates of a point in a two-dimensional Euclidean space. We discuss how practice leads to the acquisition of the metric properties of the controlled space. Methods of machine learning based on the reduction of reaching errors are tested as a means to facilitate learning by adaptively changing he map from body motions to controlled device. We discuss the relevance of the results to the development of adaptive human machine interfaces and optimal control. PMID:19665553
Intelligent transportation systems for work zones : deployment benefits and lessons learned
DOT National Transportation Integrated Search
2000-12-01
This paper presents what has been learned in four principal areas of arterial management: 1) adaptive control strategies; 2) advanced traveler information systems; 3) automated enforcement; and 4) integration. The levels of deployment, benefits, depl...
NASA Astrophysics Data System (ADS)
Roslund, Jonathan; Shir, Ofer M.; Bäck, Thomas; Rabitz, Herschel
2009-10-01
Optimization of quantum systems by closed-loop adaptive pulse shaping offers a rich domain for the development and application of specialized evolutionary algorithms. Derandomized evolution strategies (DESs) are presented here as a robust class of optimizers for experimental quantum control. The combination of stochastic and quasi-local search embodied by these algorithms is especially amenable to the inherent topology of quantum control landscapes. Implementation of DES in the laboratory results in efficiency gains of up to ˜9 times that of the standard genetic algorithm, and thus is a promising tool for optimization of unstable or fragile systems. The statistical learning upon which these algorithms are predicated also provide the means for obtaining a control problem’s Hessian matrix with no additional experimental overhead. The forced optimal covariance adaptive learning (FOCAL) method is introduced to enable retrieval of the Hessian matrix, which can reveal information about the landscape’s local structure and dynamic mechanism. Exploitation of such algorithms in quantum control experiments should enhance their efficiency and provide additional fundamental insights.
A Robust Cooperated Control Method with Reinforcement Learning and Adaptive H∞ Control
NASA Astrophysics Data System (ADS)
Obayashi, Masanao; Uchiyama, Shogo; Kuremoto, Takashi; Kobayashi, Kunikazu
This study proposes a robust cooperated control method combining reinforcement learning with robust control to control the system. A remarkable characteristic of the reinforcement learning is that it doesn't require model formula, however, it doesn't guarantee the stability of the system. On the other hand, robust control system guarantees stability and robustness, however, it requires model formula. We employ both the actor-critic method which is a kind of reinforcement learning with minimal amount of computation to control continuous valued actions and the traditional robust control, that is, H∞ control. The proposed system was compared method with the conventional control method, that is, the actor-critic only used, through the computer simulation of controlling the angle and the position of a crane system, and the simulation result showed the effectiveness of the proposed method.
Intelligent robot trends and predictions for the .net future
NASA Astrophysics Data System (ADS)
Hall, Ernest L.
2001-10-01
An intelligent robot is a remarkably useful combination of a manipulator, sensors and controls. The use of these machines in factory automation can improve productivity, increase product quality and improve competitiveness. This paper presents a discussion of recent and future technical and economic trends. During the past twenty years the use of industrial robots that are equipped not only with precise motion control systems but also with sensors such as cameras, laser scanners, or tactile sensors that permit adaptation to a changing environment has increased dramatically. Intelligent robot products have been developed in many cases for factory automation and for some hospital and home applications. To reach an even higher degree of applications, the addition of learning may be required. Recently, learning theories such as the adaptive critic have been proposed. In this type of learning, a critic provides a grade to the controller of an action module such as a robot. The adaptive critic is a good model for human learning. In general, the critic may be considered to be the human with the teach pendant, plant manager, line supervisor, quality inspector or the consumer. If the ultimate critic is the consumer, then the quality inspector must model the consumer's decision-making process and use this model in the design and manufacturing operations. Can the adaptive critic be used to advance intelligent robots? Intelligent robots have historically taken decades to be developed and reduced to practice. Methods for speeding this development include technology such as rapid prototyping and product development and government, industry and university cooperation.
Video Game Adapts To Brain Waves
NASA Technical Reports Server (NTRS)
Pope, Alan T.; Bogart, Edward H.
1994-01-01
Electronic training system based on video game developed to help children afflicted with attention-deficit disorder (ADD) learn to prolong their attention spans. Uses combination of electroencephalography (EEG) and adaptive control to encourage attentiveness. Monitors trainee's brain-wave activity: if EEG signal indicates attention is waning, system increases difficulty of game, forcing trainee to devote more attention to it. Game designed to make trainees want to win and, in so doing, learn to pay attention for longer times.
Adaptation, Learning, and the Art of War: A Cybernetic Perspective
2014-05-14
William Ross Ashby and contemporary cybernetic thought, the study modeled the adaptive systems as control loops and the processes of adaptive systems...as a Markov process . Using this model , the study concluded that systems would return to the same relative equilibrium point, expressed in terms of...uncertain and ever-changing environment. Drawing from the works of William Ross Ashby and contemporary cybernetic thought, the study modeled the adaptive
Adaptivity in Agent-Based Routing for Data Networks
NASA Technical Reports Server (NTRS)
Wolpert, David H.; Kirshner, Sergey; Merz, Chris J.; Turner, Kagan
2000-01-01
Adaptivity, both of the individual agents and of the interaction structure among the agents, seems indispensable for scaling up multi-agent systems (MAS s) in noisy environments. One important consideration in designing adaptive agents is choosing their action spaces to be as amenable as possible to machine learning techniques, especially to reinforcement learning (RL) techniques. One important way to have the interaction structure connecting agents itself be adaptive is to have the intentions and/or actions of the agents be in the input spaces of the other agents, much as in Stackelberg games. We consider both kinds of adaptivity in the design of a MAS to control network packet routing. We demonstrate on the OPNET event-driven network simulator the perhaps surprising fact that simply changing the action space of the agents to be better suited to RL can result in very large improvements in their potential performance: at their best settings, our learning-amenable router agents achieve throughputs up to three and one half times better than that of the standard Bellman-Ford routing algorithm, even when the Bellman-Ford protocol traffic is maintained. We then demonstrate that much of that potential improvement can be realized by having the agents learn their settings when the agent interaction structure is itself adaptive.
Approximate Optimal Control as a Model for Motor Learning
ERIC Educational Resources Information Center
Berthier, Neil E.; Rosenstein, Michael T.; Barto, Andrew G.
2005-01-01
Current models of psychological development rely heavily on connectionist models that use supervised learning. These models adapt network weights when the network output does not match the target outputs computed by some agent. The authors present a model of motor learning in which the child uses exploration to discover appropriate ways of…
ERIC Educational Resources Information Center
Chang, Yoo Kyung
2010-01-01
Metacognition is widely studied for its influence on the effectiveness of learning. With Exploratory Computer-Based Learning Environments (ECBLE), metacognition is found to be especially important because these environments require adaptive metacognitive control by the learners due to their open-ended structure that allows for multiple learning…
Switching Adaptability in Human-Inspired Sidesteps: A Minimal Model.
Fujii, Keisuke; Yoshihara, Yuki; Tanabe, Hiroko; Yamamoto, Yuji
2017-01-01
Humans can adapt to abruptly changing situations by coordinating redundant components, even in bipedality. Conventional adaptability has been reproduced by various computational approaches, such as optimal control, neural oscillator, and reinforcement learning; however, the adaptability in bipedal locomotion necessary for biological and social activities, such as unpredicted direction change in chase-and-escape, is unknown due to the dynamically unstable multi-link closed-loop system. Here we propose a switching adaptation model for performing bipedal locomotion by improving autonomous distributed control, where autonomous actuators interact without central control and switch the roles for propulsion, balancing, and leg swing. Our switching mobility model achieved direction change at any time using only three actuators, although it showed higher motor costs than comparable models without direction change. Our method of evaluating such adaptation at any time should be utilized as a prerequisite for understanding universal motor control. The proposed algorithm may simply explain and predict the adaptation mechanism in human bipedality to coordinate the actuator functions within and between limbs.
Support patient search on pathology reports with interactive online learning based data extraction.
Zheng, Shuai; Lu, James J; Appin, Christina; Brat, Daniel; Wang, Fusheng
2015-01-01
Structural reporting enables semantic understanding and prompt retrieval of clinical findings about patients. While synoptic pathology reporting provides templates for data entries, information in pathology reports remains primarily in narrative free text form. Extracting data of interest from narrative pathology reports could significantly improve the representation of the information and enable complex structured queries. However, manual extraction is tedious and error-prone, and automated tools are often constructed with a fixed training dataset and not easily adaptable. Our goal is to extract data from pathology reports to support advanced patient search with a highly adaptable semi-automated data extraction system, which can adjust and self-improve by learning from a user's interaction with minimal human effort. We have developed an online machine learning based information extraction system called IDEAL-X. With its graphical user interface, the system's data extraction engine automatically annotates values for users to review upon loading each report text. The system analyzes users' corrections regarding these annotations with online machine learning, and incrementally enhances and refines the learning model as reports are processed. The system also takes advantage of customized controlled vocabularies, which can be adaptively refined during the online learning process to further assist the data extraction. As the accuracy of automatic annotation improves overtime, the effort of human annotation is gradually reduced. After all reports are processed, a built-in query engine can be applied to conveniently define queries based on extracted structured data. We have evaluated the system with a dataset of anatomic pathology reports from 50 patients. Extracted data elements include demographical data, diagnosis, genetic marker, and procedure. The system achieves F-1 scores of around 95% for the majority of tests. Extracting data from pathology reports could enable more accurate knowledge to support biomedical research and clinical diagnosis. IDEAL-X provides a bridge that takes advantage of online machine learning based data extraction and the knowledge from human's feedback. By combining iterative online learning and adaptive controlled vocabularies, IDEAL-X can deliver highly adaptive and accurate data extraction to support patient search.
Adaptive process control using fuzzy logic and genetic algorithms
NASA Technical Reports Server (NTRS)
Karr, C. L.
1993-01-01
Researchers at the U.S. Bureau of Mines have developed adaptive process control systems in which genetic algorithms (GA's) are used to augment fuzzy logic controllers (FLC's). GA's are search algorithms that rapidly locate near-optimum solutions to a wide spectrum of problems by modeling the search procedures of natural genetics. FLC's are rule based systems that efficiently manipulate a problem environment by modeling the 'rule-of-thumb' strategy used in human decision making. Together, GA's and FLC's possess the capabilities necessary to produce powerful, efficient, and robust adaptive control systems. To perform efficiently, such control systems require a control element to manipulate the problem environment, and a learning element to adjust to the changes in the problem environment. Details of an overall adaptive control system are discussed. A specific laboratory acid-base pH system is used to demonstrate the ideas presented.
Adaptive Process Control with Fuzzy Logic and Genetic Algorithms
NASA Technical Reports Server (NTRS)
Karr, C. L.
1993-01-01
Researchers at the U.S. Bureau of Mines have developed adaptive process control systems in which genetic algorithms (GA's) are used to augment fuzzy logic controllers (FLC's). GA's are search algorithms that rapidly locate near-optimum solutions to a wide spectrum of problems by modeling the search procedures of natural genetics. FLC's are rule based systems that efficiently manipulate a problem environment by modeling the 'rule-of-thumb' strategy used in human decision-making. Together, GA's and FLC's possess the capabilities necessary to produce powerful, efficient, and robust adaptive control systems. To perform efficiently, such control systems require a control element to manipulate the problem environment, an analysis element to recognize changes in the problem environment, and a learning element to adjust to the changes in the problem environment. Details of an overall adaptive control system are discussed. A specific laboratory acid-base pH system is used to demonstrate the ideas presented.
Genetic algorithms in adaptive fuzzy control
NASA Technical Reports Server (NTRS)
Karr, C. Lucas; Harper, Tony R.
1992-01-01
Researchers at the U.S. Bureau of Mines have developed adaptive process control systems in which genetic algorithms (GA's) are used to augment fuzzy logic controllers (FLC's). GA's are search algorithms that rapidly locate near-optimum solutions to a wide spectrum of problems by modeling the search procedures of natural genetics. FLC's are rule based systems that efficiently manipulate a problem environment by modeling the 'rule-of-thumb' strategy used in human decision making. Together, GA's and FLC's possess the capabilities necessary to produce powerful, efficient, and robust adaptive control systems. To perform efficiently, such control systems require a control element to manipulate the problem environment, an analysis element to recognize changes in the problem environment, and a learning element to adjust fuzzy membership functions in response to the changes in the problem environment. Details of an overall adaptive control system are discussed. A specific computer-simulated chemical system is used to demonstrate the ideas presented.
Pimashkin, Alexey; Gladkov, Arseniy; Mukhina, Irina; Kazantsev, Victor
2013-01-01
Learning in neuronal networks can be investigated using dissociated cultures on multielectrode arrays supplied with appropriate closed-loop stimulation. It was shown in previous studies that weakly respondent neurons on the electrodes can be trained to increase their evoked spiking rate within a predefined time window after the stimulus. Such neurons can be associated with weak synaptic connections in nearby culture network. The stimulation leads to the increase in the connectivity and in the response. However, it was not possible to perform the learning protocol for the neurons on electrodes with relatively strong synaptic inputs and responding at higher rates. We proposed an adaptive closed-loop stimulation protocol capable to achieve learning even for the highly respondent electrodes. It means that the culture network can reorganize appropriately its synaptic connectivity to generate a desired response. We introduced an adaptive reinforcement condition accounting for the response variability in control stimulation. It significantly enhanced the learning protocol to a large number of responding electrodes independently on its base response level. We also found that learning effect preserved after 4–6 h after training. PMID:23745105
Fuzzy Adaptive Decentralized Optimal Control for Strict Feedback Nonlinear Large-Scale Systems.
Sun, Kangkang; Sui, Shuai; Tong, Shaocheng
2018-04-01
This paper considers the optimal decentralized fuzzy adaptive control design problem for a class of interconnected large-scale nonlinear systems in strict feedback form and with unknown nonlinear functions. The fuzzy logic systems are introduced to learn the unknown dynamics and cost functions, respectively, and a state estimator is developed. By applying the state estimator and the backstepping recursive design algorithm, a decentralized feedforward controller is established. By using the backstepping decentralized feedforward control scheme, the considered interconnected large-scale nonlinear system in strict feedback form is changed into an equivalent affine large-scale nonlinear system. Subsequently, an optimal decentralized fuzzy adaptive control scheme is constructed. The whole optimal decentralized fuzzy adaptive controller is composed of a decentralized feedforward control and an optimal decentralized control. It is proved that the developed optimal decentralized controller can ensure that all the variables of the control system are uniformly ultimately bounded, and the cost functions are the smallest. Two simulation examples are provided to illustrate the validity of the developed optimal decentralized fuzzy adaptive control scheme.
Learning and tuning fuzzy logic controllers through reinforcements
NASA Technical Reports Server (NTRS)
Berenji, Hamid R.; Khedkar, Pratap
1992-01-01
A new method for learning and tuning a fuzzy logic controller based on reinforcements from a dynamic system is presented. In particular, our Generalized Approximate Reasoning-based Intelligent Control (GARIC) architecture: (1) learns and tunes a fuzzy logic controller even when only weak reinforcements, such as a binary failure signal, is available; (2) introduces a new conjunction operator in computing the rule strengths of fuzzy control rules; (3) introduces a new localized mean of maximum (LMOM) method in combining the conclusions of several firing control rules; and (4) learns to produce real-valued control actions. Learning is achieved by integrating fuzzy inference into a feedforward network, which can then adaptively improve performance by using gradient descent methods. We extend the AHC algorithm of Barto, Sutton, and Anderson to include the prior control knowledge of human operators. The GARIC architecture is applied to a cart-pole balancing system and has demonstrated significant improvements in terms of the speed of learning and robustness to changes in the dynamic system's parameters over previous schemes for cart-pole balancing.
Monitoring the Performance of a Neuro-Adaptive Controller
NASA Technical Reports Server (NTRS)
Schumann, Johann; Gupta, Pramod
2004-01-01
Traditional control has proven to be ineffective to deal with catastrophic changes or slow degradation of complex, highly nonlinear systems like aircraft or spacecraft, robotics, or flexible manufacturing systems. Control systems which can adapt toward changes in the plant have been proposed as they offer many advantages (e.g., better performance, controllability of aircraft despite of a damaged wing). In the last few years, use of neural networks in adaptive controllers (neuro-adaptive control) has been studied actively. Neural networks of various architectures have been used successfully for online learning adaptive controllers. In such a typical control architecture, the neural network receives as an input the current deviation between desired and actual plant behavior and, by on-line training, tries to minimize this discrepancy (e.g.; by producing a control augmentation signal). Even though neuro-adaptive controllers offer many advantages, they have not been used in mission- or safety-critical applications, because performance and safety guarantees cannot b e provided at development time-a major prerequisite for safety certification (e.g., by the FAA or NASA). Verification and Validation (V&V) of an adaptive controller requires the development of new analysis techniques which can demonstrate that the control system behaves safely under all operating conditions. Because of the requirement to adapt toward unforeseen changes during operation, i.e., in real time, design-time V&V is not sufficient.
Direct Adaptive Aircraft Control Using Dynamic Cell Structure Neural Networks
NASA Technical Reports Server (NTRS)
Jorgensen, Charles C.
1997-01-01
A Dynamic Cell Structure (DCS) Neural Network was developed which learns topology representing networks (TRNS) of F-15 aircraft aerodynamic stability and control derivatives. The network is integrated into a direct adaptive tracking controller. The combination produces a robust adaptive architecture capable of handling multiple accident and off- nominal flight scenarios. This paper describes the DCS network and modifications to the parameter estimation procedure. The work represents one step towards an integrated real-time reconfiguration control architecture for rapid prototyping of new aircraft designs. Performance was evaluated using three off-line benchmarks and on-line nonlinear Virtual Reality simulation. Flight control was evaluated under scenarios including differential stabilator lock, soft sensor failure, control and stability derivative variations, and air turbulence.
Electroencephalographic identifiers of motor adaptation learning
NASA Astrophysics Data System (ADS)
Özdenizci, Ozan; Yalçın, Mustafa; Erdoğan, Ahmetcan; Patoğlu, Volkan; Grosse-Wentrup, Moritz; Çetin, Müjdat
2017-08-01
Objective. Recent brain-computer interface (BCI) assisted stroke rehabilitation protocols tend to focus on sensorimotor activity of the brain. Relying on evidence claiming that a variety of brain rhythms beyond sensorimotor areas are related to the extent of motor deficits, we propose to identify neural correlates of motor learning beyond sensorimotor areas spatially and spectrally for further use in novel BCI-assisted neurorehabilitation settings. Approach. Electroencephalographic (EEG) data were recorded from healthy subjects participating in a physical force-field adaptation task involving reaching movements through a robotic handle. EEG activity recorded during rest prior to the experiment and during pre-trial movement preparation was used as features to predict motor adaptation learning performance across subjects. Main results. Subjects learned to perform straight movements under the force-field at different adaptation rates. Both resting-state and pre-trial EEG features were predictive of individual adaptation rates with relevance of a broad network of beta activity. Beyond sensorimotor regions, a parieto-occipital cortical component observed across subjects was involved strongly in predictions and a fronto-parietal cortical component showed significant decrease in pre-trial beta-powers for users with higher adaptation rates and increase in pre-trial beta-powers for users with lower adaptation rates. Significance. Including sensorimotor areas, a large-scale network of beta activity is presented as predictive of motor learning. Strength of resting-state parieto-occipital beta activity or pre-trial fronto-parietal beta activity can be considered in BCI-assisted stroke rehabilitation protocols with neurofeedback training or volitional control of neural activity for brain-robot interfaces to induce plasticity.
Pinzon-Morales, Ruben-Dario; Hirata, Yutaka
2014-01-01
To acquire and maintain precise movement controls over a lifespan, changes in the physical and physiological characteristics of muscles must be compensated for adaptively. The cerebellum plays a crucial role in such adaptation. Changes in muscle characteristics are not always symmetrical. For example, it is unlikely that muscles that bend and straighten a joint will change to the same degree. Thus, different (i.e., asymmetrical) adaptation is required for bending and straightening motions. To date, little is known about the role of the cerebellum in asymmetrical adaptation. Here, we investigate the cerebellar mechanisms required for asymmetrical adaptation using a bi-hemispheric cerebellar neuronal network model (biCNN). The bi-hemispheric structure is inspired by the observation that lesioning one hemisphere reduces motor performance asymmetrically. The biCNN model was constructed to run in real-time and used to control an unstable two-wheeled balancing robot. The load of the robot and its environment were modified to create asymmetrical perturbations. Plasticity at parallel fiber-Purkinje cell synapses in the biCNN model was driven by error signal in the climbing fiber (cf) input. This cf input was configured to increase and decrease its firing rate from its spontaneous firing rate (approximately 1 Hz) with sensory errors in the preferred and non-preferred direction of each hemisphere, as demonstrated in the monkey cerebellum. Our results showed that asymmetrical conditions were successfully handled by the biCNN model, in contrast to a single hemisphere model or a classical non-adaptive proportional and derivative controller. Further, the spontaneous activity of the cf, while relatively small, was critical for balancing the contribution of each cerebellar hemisphere to the overall motor command sent to the robot. Eliminating the spontaneous activity compromised the asymmetrical learning capabilities of the biCNN model. Thus, we conclude that a bi-hemispheric structure and adequate spontaneous activity of cf inputs are critical for cerebellar asymmetrical motor learning.
Pinzon-Morales, Ruben-Dario; Hirata, Yutaka
2014-01-01
To acquire and maintain precise movement controls over a lifespan, changes in the physical and physiological characteristics of muscles must be compensated for adaptively. The cerebellum plays a crucial role in such adaptation. Changes in muscle characteristics are not always symmetrical. For example, it is unlikely that muscles that bend and straighten a joint will change to the same degree. Thus, different (i.e., asymmetrical) adaptation is required for bending and straightening motions. To date, little is known about the role of the cerebellum in asymmetrical adaptation. Here, we investigate the cerebellar mechanisms required for asymmetrical adaptation using a bi-hemispheric cerebellar neuronal network model (biCNN). The bi-hemispheric structure is inspired by the observation that lesioning one hemisphere reduces motor performance asymmetrically. The biCNN model was constructed to run in real-time and used to control an unstable two-wheeled balancing robot. The load of the robot and its environment were modified to create asymmetrical perturbations. Plasticity at parallel fiber-Purkinje cell synapses in the biCNN model was driven by error signal in the climbing fiber (cf) input. This cf input was configured to increase and decrease its firing rate from its spontaneous firing rate (approximately 1 Hz) with sensory errors in the preferred and non-preferred direction of each hemisphere, as demonstrated in the monkey cerebellum. Our results showed that asymmetrical conditions were successfully handled by the biCNN model, in contrast to a single hemisphere model or a classical non-adaptive proportional and derivative controller. Further, the spontaneous activity of the cf, while relatively small, was critical for balancing the contribution of each cerebellar hemisphere to the overall motor command sent to the robot. Eliminating the spontaneous activity compromised the asymmetrical learning capabilities of the biCNN model. Thus, we conclude that a bi-hemispheric structure and adequate spontaneous activity of cf inputs are critical for cerebellar asymmetrical motor learning. PMID:25414644
Adaptive dynamic programming approach to experience-based systems identification and control.
Lendaris, George G
2009-01-01
Humans have the ability to make use of experience while selecting their control actions for distinct and changing situations, and their process speeds up and have enhanced effectiveness as more experience is gained. In contrast, current technological implementations slow down as more knowledge is stored. A novel way of employing Approximate (or Adaptive) Dynamic Programming (ADP) is described that shifts the underlying Adaptive Critic type of Reinforcement Learning method "up a level", away from designing individual (optimal) controllers to that of developing on-line algorithms that efficiently and effectively select designs from a repository of existing controller solutions (perhaps previously developed via application of ADP methods). The resulting approach is called Higher-Level Learning Algorithm. The approach and its rationale are described and some examples of its application are given. The notions of context and context discernment are important to understanding the human abilities noted above. These are first defined, in a manner appropriate to controls and system-identification, and as a foundation relating to the application arena, a historical view of the various phases during development of the controls field is given, organized by how the notion 'context' was, or was not, involved in each phase.
SU-D-BRB-05: Quantum Learning for Knowledge-Based Response-Adaptive Radiotherapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
El Naqa, I; Ten, R
Purpose: There is tremendous excitement in radiotherapy about applying data-driven methods to develop personalized clinical decisions for real-time response-based adaptation. However, classical statistical learning methods lack in terms of efficiency and ability to predict outcomes under conditions of uncertainty and incomplete information. Therefore, we are investigating physics-inspired machine learning approaches by utilizing quantum principles for developing a robust framework to dynamically adapt treatments to individual patient’s characteristics and optimize outcomes. Methods: We studied 88 liver SBRT patients with 35 on non-adaptive and 53 on adaptive protocols. Adaptation was based on liver function using a split-course of 3+2 fractions with amore » month break. The radiotherapy environment was modeled as a Markov decision process (MDP) of baseline and one month into treatment states. The patient environment was modeled by a 5-variable state represented by patient’s clinical and dosimetric covariates. For comparison of classical and quantum learning methods, decision-making to adapt at one month was considered. The MDP objective was defined by the complication-free tumor control (P{sup +}=TCPx(1-NTCP)). A simple regression model represented state-action mapping. Single bit in classical MDP and a qubit of 2-superimposed states in quantum MDP represented the decision actions. Classical decision selection was done using reinforcement Q-learning and quantum searching was performed using Grover’s algorithm, which applies uniform superposition over possible states and yields quadratic speed-up. Results: Classical/quantum MDPs suggested adaptation (probability amplitude ≥0.5) 79% of the time for splitcourses and 100% for continuous-courses. However, the classical MDP had an average adaptation probability of 0.5±0.22 while the quantum algorithm reached 0.76±0.28. In cases where adaptation failed, classical MDP yielded 0.31±0.26 average amplitude while the quantum approach averaged a more optimistic 0.57±0.4, but with high phase fluctuations. Conclusion: Our results demonstrate that quantum machine learning approaches provide a feasible and promising framework for real-time and sequential clinical decision-making in adaptive radiotherapy.« less
Telgen, Sebastian; Parvin, Darius; Diedrichsen, Jörn
2014-10-08
Motor learning tasks are often classified into adaptation tasks, which involve the recalibration of an existing control policy (the mapping that determines both feedforward and feedback commands), and skill-learning tasks, requiring the acquisition of new control policies. We show here that this distinction also applies to two different visuomotor transformations during reaching in humans: Mirror-reversal (left-right reversal over a mid-sagittal axis) of visual feedback versus rotation of visual feedback around the movement origin. During mirror-reversal learning, correct movement initiation (feedforward commands) and online corrections (feedback responses) were only generated at longer latencies. The earliest responses were directed into a nonmirrored direction, even after two training sessions. In contrast, for visual rotation learning, no dependency of directional error on reaction time emerged, and fast feedback responses to visual displacements of the cursor were immediately adapted. These results suggest that the motor system acquires a new control policy for mirror reversal, which initially requires extra processing time, while it recalibrates an existing control policy for visual rotations, exploiting established fast computational processes. Importantly, memory for visual rotation decayed between sessions, whereas memory for mirror reversals showed offline gains, leading to better performance at the beginning of the second session than in the end of the first. With shifts in time-accuracy tradeoff and offline gains, mirror-reversal learning shares common features with other skill-learning tasks. We suggest that different neuronal mechanisms underlie the recalibration of an existing versus acquisition of a new control policy and that offline gains between sessions are a characteristic of latter. Copyright © 2014 the authors 0270-6474/14/3413768-12$15.00/0.
An Approach to V&V of Embedded Adaptive Systems
NASA Technical Reports Server (NTRS)
Liu, Yan; Yerramalla, Sampath; Fuller, Edgar; Cukic, Bojan; Gururajan, Srikaruth
2004-01-01
Rigorous Verification and Validation (V&V) techniques are essential for high assurance systems. Lately, the performance of some of these systems is enhanced by embedded adaptive components in order to cope with environmental changes. Although the ability of adapting is appealing, it actually poses a problem in terms of V&V. Since uncertainties induced by environmental changes have a significant impact on system behavior, the applicability of conventional V&V techniques is limited. In safety-critical applications such as flight control system, the mechanisms of change must be observed, diagnosed, accommodated and well understood prior to deployment. In this paper, we propose a non-conventional V&V approach suitable for online adaptive systems. We apply our approach to an intelligent flight control system that employs a particular type of Neural Networks (NN) as the adaptive learning paradigm. Presented methodology consists of a novelty detection technique and online stability monitoring tools. The novelty detection technique is based on Support Vector Data Description that detects novel (abnormal) data patterns. The Online Stability Monitoring tools based on Lyapunov's Stability Theory detect unstable learning behavior in neural networks. Cases studies based on a high fidelity simulator of NASA's Intelligent Flight Control System demonstrate a successful application of the presented V&V methodology. ,
Selected Flight Test Results for Online Learning Neural Network-Based Flight Control System
NASA Technical Reports Server (NTRS)
Williams-Hayes, Peggy S.
2004-01-01
The NASA F-15 Intelligent Flight Control System project team developed a series of flight control concepts designed to demonstrate neural network-based adaptive controller benefits, with the objective to develop and flight-test control systems using neural network technology to optimize aircraft performance under nominal conditions and stabilize the aircraft under failure conditions. This report presents flight-test results for an adaptive controller using stability and control derivative values from an online learning neural network. A dynamic cell structure neural network is used in conjunction with a real-time parameter identification algorithm to estimate aerodynamic stability and control derivative increments to baseline aerodynamic derivatives in flight. This open-loop flight test set was performed in preparation for a future phase in which the learning neural network and parameter identification algorithm output would provide the flight controller with aerodynamic stability and control derivative updates in near real time. Two flight maneuvers are analyzed - pitch frequency sweep and automated flight-test maneuver designed to optimally excite the parameter identification algorithm in all axes. Frequency responses generated from flight data are compared to those obtained from nonlinear simulation runs. Flight data examination shows that addition of flight-identified aerodynamic derivative increments into the simulation improved aircraft pitch handling qualities.
Brosowsky, Nicholaus P; Crump, Matthew J C
2016-08-01
Recent work suggests that environmental cues associated with previous attentional control settings can rapidly and involuntarily adjust attentional priorities. The current study tests predictions from adaptive-learning and memory-based theories of contextual control about the role of intentions for setting attentional priorities. To extend the empirical boundaries of contextual control phenomena, and to determine whether theoretical principles of contextual control are generalizable we used a novel bi-dimensional stimulus sampling task. Subjects viewed briefly presented arrays of letters and colors presented above or below fixation, and identified specific stimuli according to a dimensional (letter or color) and positional cue. Location was predictive of the cued dimension, but not the position or identity. In contrast to previous findings, contextual control failed to develop through automatic, adaptive-learning processes. Instead, previous experience with intentionally changing attentional sampling priorities between different contexts was required for contextual control to develop. Copyright © 2016 Elsevier Inc. All rights reserved.
Self-teaching neural network learns difficult reactor control problem
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jouse, W.C.
1989-01-01
A self-teaching neural network used as an adaptive controller quickly learns to control an unstable reactor configuration. The network models the behavior of a human operator. It is trained by allowing it to operate the reactivity control impulsively. It is punished whenever either the power or fuel temperature stray outside technical limits. Using a simple paradigm, the network constructs an internal representation of the punishment and of the reactor system. The reactor is constrained to small power orbits.
Kong, Zehui; Liu, Teng
2017-01-01
To further improve the fuel economy of series hybrid electric tracked vehicles, a reinforcement learning (RL)-based real-time energy management strategy is developed in this paper. In order to utilize the statistical characteristics of online driving schedule effectively, a recursive algorithm for the transition probability matrix (TPM) of power-request is derived. The reinforcement learning (RL) is applied to calculate and update the control policy at regular time, adapting to the varying driving conditions. A facing-forward powertrain model is built in detail, including the engine-generator model, battery model and vehicle dynamical model. The robustness and adaptability of real-time energy management strategy are validated through the comparison with the stationary control strategy based on initial transition probability matrix (TPM) generated from a long naturalistic driving cycle in the simulation. Results indicate that proposed method has better fuel economy than stationary one and is more effective in real-time control. PMID:28671967
Kong, Zehui; Zou, Yuan; Liu, Teng
2017-01-01
To further improve the fuel economy of series hybrid electric tracked vehicles, a reinforcement learning (RL)-based real-time energy management strategy is developed in this paper. In order to utilize the statistical characteristics of online driving schedule effectively, a recursive algorithm for the transition probability matrix (TPM) of power-request is derived. The reinforcement learning (RL) is applied to calculate and update the control policy at regular time, adapting to the varying driving conditions. A facing-forward powertrain model is built in detail, including the engine-generator model, battery model and vehicle dynamical model. The robustness and adaptability of real-time energy management strategy are validated through the comparison with the stationary control strategy based on initial transition probability matrix (TPM) generated from a long naturalistic driving cycle in the simulation. Results indicate that proposed method has better fuel economy than stationary one and is more effective in real-time control.
Aircraft adaptive learning control
NASA Technical Reports Server (NTRS)
Lee, P. S. T.; Vanlandingham, H. F.
1979-01-01
The optimal control theory of stochastic linear systems is discussed in terms of the advantages of distributed-control systems, and the control of randomly-sampled systems. An optimal solution to longitudinal control is derived and applied to the F-8 DFBW aircraft. A randomly-sampled linear process model with additive process and noise is developed.
Fujiki, Soichiro; Aoi, Shinya; Funato, Tetsuro; Tomita, Nozomi; Senda, Kei; Tsuchiya, Kazuo
2015-01-01
Human walking behaviour adaptation strategies have previously been examined using split-belt treadmills, which have two parallel independently controlled belts. In such human split-belt treadmill walking, two types of adaptations have been identified: early and late. Early-type adaptations appear as rapid changes in interlimb and intralimb coordination activities when the belt speeds of the treadmill change between tied (same speed for both belts) and split-belt (different speeds for each belt) configurations. By contrast, late-type adaptations occur after the early-type adaptations as a gradual change and only involve interlimb coordination. Furthermore, interlimb coordination shows after-effects that are related to these adaptations. It has been suggested that these adaptations are governed primarily by the spinal cord and cerebellum, but the underlying mechanism remains unclear. Because various physiological findings suggest that foot contact timing is crucial to adaptive locomotion, this paper reports on the development of a two-layered control model for walking composed of spinal and cerebellar models, and on its use as the focus of our control model. The spinal model generates rhythmic motor commands using an oscillator network based on a central pattern generator and modulates the commands formulated in immediate response to foot contact, while the cerebellar model modifies motor commands through learning based on error information related to differences between the predicted and actual foot contact timings of each leg. We investigated adaptive behaviour and its mechanism by split-belt treadmill walking experiments using both computer simulations and an experimental bipedal robot. Our results showed that the robot exhibited rapid changes in interlimb and intralimb coordination that were similar to the early-type adaptations observed in humans. In addition, despite the lack of direct interlimb coordination control, gradual changes and after-effects in the interlimb coordination appeared in a manner that was similar to the late-type adaptations and after-effects observed in humans. The adaptation results of the robot were then evaluated in comparison with human split-belt treadmill walking, and the adaptation mechanism was clarified from a dynamic viewpoint. PMID:26289658
Fujiki, Soichiro; Aoi, Shinya; Funato, Tetsuro; Tomita, Nozomi; Senda, Kei; Tsuchiya, Kazuo
2015-09-06
Human walking behaviour adaptation strategies have previously been examined using split-belt treadmills, which have two parallel independently controlled belts. In such human split-belt treadmill walking, two types of adaptations have been identified: early and late. Early-type adaptations appear as rapid changes in interlimb and intralimb coordination activities when the belt speeds of the treadmill change between tied (same speed for both belts) and split-belt (different speeds for each belt) configurations. By contrast, late-type adaptations occur after the early-type adaptations as a gradual change and only involve interlimb coordination. Furthermore, interlimb coordination shows after-effects that are related to these adaptations. It has been suggested that these adaptations are governed primarily by the spinal cord and cerebellum, but the underlying mechanism remains unclear. Because various physiological findings suggest that foot contact timing is crucial to adaptive locomotion, this paper reports on the development of a two-layered control model for walking composed of spinal and cerebellar models, and on its use as the focus of our control model. The spinal model generates rhythmic motor commands using an oscillator network based on a central pattern generator and modulates the commands formulated in immediate response to foot contact, while the cerebellar model modifies motor commands through learning based on error information related to differences between the predicted and actual foot contact timings of each leg. We investigated adaptive behaviour and its mechanism by split-belt treadmill walking experiments using both computer simulations and an experimental bipedal robot. Our results showed that the robot exhibited rapid changes in interlimb and intralimb coordination that were similar to the early-type adaptations observed in humans. In addition, despite the lack of direct interlimb coordination control, gradual changes and after-effects in the interlimb coordination appeared in a manner that was similar to the late-type adaptations and after-effects observed in humans. The adaptation results of the robot were then evaluated in comparison with human split-belt treadmill walking, and the adaptation mechanism was clarified from a dynamic viewpoint. © 2015 The Authors.
Siu, Ho Chit; Arenas, Ana M; Sun, Tingxiao; Stirling, Leia A
2018-02-05
Upper-extremity exoskeletons have demonstrated potential as augmentative, assistive, and rehabilitative devices. Typical control of upper-extremity exoskeletons have relied on switches, force/torque sensors, and surface electromyography (sEMG), but these systems are usually reactionary, and/or rely on entirely hand-tuned parameters. sEMG-based systems may be able to provide anticipatory control, since they interface directly with muscle signals, but typically require expert placement of sensors on muscle bodies. We present an implementation of an adaptive sEMG-based exoskeleton controller that learns a mapping between muscle activation and the desired system state during interaction with a user, generating a personalized sEMG feature classifier to allow for anticipatory control. This system is robust to novice placement of sEMG sensors, as well as subdermal muscle shifts. We validate this method with 18 subjects using a thumb exoskeleton to complete a book-placement task. This learning-from-demonstration system for exoskeleton control allows for very short training times, as well as the potential for improvement in intent recognition over time, and adaptation to physiological changes in the user, such as those due to fatigue.
Arenas, Ana M.; Sun, Tingxiao
2018-01-01
Upper-extremity exoskeletons have demonstrated potential as augmentative, assistive, and rehabilitative devices. Typical control of upper-extremity exoskeletons have relied on switches, force/torque sensors, and surface electromyography (sEMG), but these systems are usually reactionary, and/or rely on entirely hand-tuned parameters. sEMG-based systems may be able to provide anticipatory control, since they interface directly with muscle signals, but typically require expert placement of sensors on muscle bodies. We present an implementation of an adaptive sEMG-based exoskeleton controller that learns a mapping between muscle activation and the desired system state during interaction with a user, generating a personalized sEMG feature classifier to allow for anticipatory control. This system is robust to novice placement of sEMG sensors, as well as subdermal muscle shifts. We validate this method with 18 subjects using a thumb exoskeleton to complete a book-placement task. This learning-from-demonstration system for exoskeleton control allows for very short training times, as well as the potential for improvement in intent recognition over time, and adaptation to physiological changes in the user, such as those due to fatigue. PMID:29401754
Brain-Machine Interface control of a robot arm using actor-critic rainforcement learning.
Pohlmeyer, Eric A; Mahmoudi, Babak; Geng, Shijia; Prins, Noeline; Sanchez, Justin C
2012-01-01
Here we demonstrate how a marmoset monkey can use a reinforcement learning (RL) Brain-Machine Interface (BMI) to effectively control the movements of a robot arm for a reaching task. In this work, an actor-critic RL algorithm used neural ensemble activity in the monkey's motor cortext to control the robot movements during a two-target decision task. This novel approach to decoding offers unique advantages for BMI control applications. Compared to supervised learning decoding methods, the actor-critic RL algorithm does not require an explicit set of training data to create a static control model, but rather it incrementally adapts the model parameters according to its current performance, in this case requiring only a very basic feedback signal. We show how this algorithm achieved high performance when mapping the monkey's neural states (94%) to robot actions, and only needed to experience a few trials before obtaining accurate real-time control of the robot arm. Since RL methods responsively adapt and adjust their parameters, they can provide a method to create BMIs that are robust against perturbations caused by changes in either the neural input space or the output actions they generate under different task requirements or goals.
Use of Frontal Lobe Hemodynamics as Reinforcement Signals to an Adaptive Controller
DiStasio, Marcello M.; Francis, Joseph T.
2013-01-01
Decision-making ability in the frontal lobe (among other brain structures) relies on the assignment of value to states of the animal and its environment. Then higher valued states can be pursued and lower (or negative) valued states avoided. The same principle forms the basis for computational reinforcement learning controllers, which have been fruitfully applied both as models of value estimation in the brain, and as artificial controllers in their own right. This work shows how state desirability signals decoded from frontal lobe hemodynamics, as measured with near-infrared spectroscopy (NIRS), can be applied as reinforcers to an adaptable artificial learning agent in order to guide its acquisition of skills. A set of experiments carried out on an alert macaque demonstrate that both oxy- and deoxyhemoglobin concentrations in the frontal lobe show differences in response to both primarily and secondarily desirable (versus undesirable) stimuli. This difference allows a NIRS signal classifier to serve successfully as a reinforcer for an adaptive controller performing a virtual tool-retrieval task. The agent's adaptability allows its performance to exceed the limits of the NIRS classifier decoding accuracy. We also show that decoding state desirabilities is more accurate when using relative concentrations of both oxyhemoglobin and deoxyhemoglobin, rather than either species alone. PMID:23894500
Reward and punishment enhance motor adaptation in stroke.
Quattrocchi, Graziella; Greenwood, Richard; Rothwell, John C; Galea, Joseph M; Bestmann, Sven
2017-09-01
The effects of motor learning, such as motor adaptation, in stroke rehabilitation are often transient, thus mandating approaches that enhance the amount of learning and retention. Previously, we showed in young individuals that reward and punishment feedback have dissociable effects on motor adaptation, with punishment improving adaptation and reward enhancing retention. If these findings were able to generalise to patients with stroke, they would provide a way to optimise motor learning in these patients. Therefore, we tested this in 45 patients with chronic stroke allocated in three groups. Patients performed reaching movements with their paretic arm with a robotic manipulandum. After training (day 1), day 2 involved adaptation to a novel force field. During the adaptation phase, patients received performance-based feedback according to the group they were allocated: reward, punishment or no feedback (neutral). On day 3, patients readapted to the force field but all groups now received neutral feedback. All patients adapted, with reward and punishment groups displaying greater adaptation and readaptation than the neutral group, irrespective of demographic, cognitive or functional differences. Remarkably, the reward and punishment groups adapted to similar degree as healthy controls. Finally, the reward group showed greater retention. This study provides, for the first time, evidence that reward and punishment can enhance motor adaptation in patients with stroke. Further research on reinforcement-based motor learning regimes is warranted to translate these promising results into clinical practice and improve motor rehabilitation outcomes in patients with stroke. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
Grounding cognitive control in associative learning.
Abrahamse, Elger; Braem, Senne; Notebaert, Wim; Verguts, Tom
2016-07-01
Cognitive control covers a broad range of cognitive functions, but its research and theories typically remain tied to a single domain. Here we outline and review an associative learning perspective on cognitive control in which control emerges from associative networks containing perceptual, motor, and goal representations. Our review identifies 3 trending research themes that are shared between the domains of conflict adaptation, task switching, response inhibition, and attentional control: Cognitive control is context-specific, can operate in the absence of awareness, and is modulated by reward. As these research themes can be envisaged as key characteristics of learning, we propose that their joint emergence across domains is not coincidental but rather reflects a (latent) growth of interest in learning-based control. Associative learning has the potential for providing broad-scaled integration to cognitive control theory, and offers a promising avenue for understanding cognitive control as a self-regulating system without postulating an ill-defined set of homunculi. We discuss novel predictions, theoretical implications, and immediate challenges that accompany an associative learning perspective on cognitive control. (PsycINFO Database Record (c) 2016 APA, all rights reserved).
ERIC Educational Resources Information Center
Wood, Milton E.
The purpose of the effort was to determine the benefits to be derived from the adaptive training technique of automatically adjusting task difficulty as a function of a student skill during early learning of a complex perceptual motor task. A digital computer provided the task dynamics, scoring, and adaptive control of a second-order, two-axis,…
Intelligent flight control systems
NASA Technical Reports Server (NTRS)
Stengel, Robert F.
1993-01-01
The capabilities of flight control systems can be enhanced by designing them to emulate functions of natural intelligence. Intelligent control functions fall in three categories. Declarative actions involve decision-making, providing models for system monitoring, goal planning, and system/scenario identification. Procedural actions concern skilled behavior and have parallels in guidance, navigation, and adaptation. Reflexive actions are spontaneous, inner-loop responses for control and estimation. Intelligent flight control systems learn knowledge of the aircraft and its mission and adapt to changes in the flight environment. Cognitive models form an efficient basis for integrating 'outer-loop/inner-loop' control functions and for developing robust parallel-processing algorithms.
Neurodevelopmental Status and Adaptive Behaviors in Preschool Children with Chronic Kidney Disease
ERIC Educational Resources Information Center
Duquette, Peter J.; Hooper, Stephen R.; Icard, Phil F.; Hower, Sarah J.; Mamak, Eva G.; Wetherington, Crista E.; Gipson, Debbie S.
2009-01-01
This study examines the early neurodevelopmental function of infants and preschool children who have chronic kidney disease (CKD). Fifteen patients with CKD are compared to a healthy control group using the "Mullen Scales of Early Learning" (MSEL) and the "Vineland Adaptive Behavior Scale" (VABS). Multivariate analysis reveals…
Probabilistic dual heuristic programming-based adaptive critic
NASA Astrophysics Data System (ADS)
Herzallah, Randa
2010-02-01
Adaptive critic (AC) methods have common roots as generalisations of dynamic programming for neural reinforcement learning approaches. Since they approximate the dynamic programming solutions, they are potentially suitable for learning in noisy, non-linear and non-stationary environments. In this study, a novel probabilistic dual heuristic programming (DHP)-based AC controller is proposed. Distinct to current approaches, the proposed probabilistic (DHP) AC method takes uncertainties of forward model and inverse controller into consideration. Therefore, it is suitable for deterministic and stochastic control problems characterised by functional uncertainty. Theoretical development of the proposed method is validated by analytically evaluating the correct value of the cost function which satisfies the Bellman equation in a linear quadratic control problem. The target value of the probabilistic critic network is then calculated and shown to be equal to the analytically derived correct value. Full derivation of the Riccati solution for this non-standard stochastic linear quadratic control problem is also provided. Moreover, the performance of the proposed probabilistic controller is demonstrated on linear and non-linear control examples.
Dusing, Stacey C; Thacker, Leroy R; Galloway, James C
2016-08-01
Infants born preterm are at increased risk of developmental disabilities, that may be attributed to their early experiences and ability to learn. The purpose of this paper was to evaluate the ability of infants born preterm to adapt their postural control to changing task demands. This study included 18 infants born at 32 weeks of gestation or less whose posture was compared in supine under 2 conditions, with and without a visual stimulus presented. The postural variability, measured with root mean squared displacement of the center of pressure, and postural complexity, measured with the approximate entropy of the center of pressure displacement were measured longitudinally from 2.5 to 5 months of age. The infants looked at the toys in midline for several months prior to adapting their postural variability in a manner similar to full term infants. Only after postural variability was reduced in both the caudal cephalic and medial lateral direction in the toy condition did the infants learn to reach for the toy. Postural complexity did not vary between conditions. These findings suggest that infants used a variety of strategies to control their posture. In contrast to research with infants born full term, the infants born preterm in this study did not identify the successful strategy of reducing movement of the center of pressure until months after showing interest in the toy. This delayed adaptation may impact the infants ability to learn over time. Copyright © 2016 Elsevier Inc. All rights reserved.
Dopamine Modulates Adaptive Prediction Error Coding in the Human Midbrain and Striatum.
Diederen, Kelly M J; Ziauddeen, Hisham; Vestergaard, Martin D; Spencer, Tom; Schultz, Wolfram; Fletcher, Paul C
2017-02-15
Learning to optimally predict rewards requires agents to account for fluctuations in reward value. Recent work suggests that individuals can efficiently learn about variable rewards through adaptation of the learning rate, and coding of prediction errors relative to reward variability. Such adaptive coding has been linked to midbrain dopamine neurons in nonhuman primates, and evidence in support for a similar role of the dopaminergic system in humans is emerging from fMRI data. Here, we sought to investigate the effect of dopaminergic perturbations on adaptive prediction error coding in humans, using a between-subject, placebo-controlled pharmacological fMRI study with a dopaminergic agonist (bromocriptine) and antagonist (sulpiride). Participants performed a previously validated task in which they predicted the magnitude of upcoming rewards drawn from distributions with varying SDs. After each prediction, participants received a reward, yielding trial-by-trial prediction errors. Under placebo, we replicated previous observations of adaptive coding in the midbrain and ventral striatum. Treatment with sulpiride attenuated adaptive coding in both midbrain and ventral striatum, and was associated with a decrease in performance, whereas bromocriptine did not have a significant impact. Although we observed no differential effect of SD on performance between the groups, computational modeling suggested decreased behavioral adaptation in the sulpiride group. These results suggest that normal dopaminergic function is critical for adaptive prediction error coding, a key property of the brain thought to facilitate efficient learning in variable environments. Crucially, these results also offer potential insights for understanding the impact of disrupted dopamine function in mental illness. SIGNIFICANCE STATEMENT To choose optimally, we have to learn what to expect. Humans dampen learning when there is a great deal of variability in reward outcome, and two brain regions that are modulated by the brain chemical dopamine are sensitive to reward variability. Here, we aimed to directly relate dopamine to learning about variable rewards, and the neural encoding of associated teaching signals. We perturbed dopamine in healthy individuals using dopaminergic medication and asked them to predict variable rewards while we made brain scans. Dopamine perturbations impaired learning and the neural encoding of reward variability, thus establishing a direct link between dopamine and adaptation to reward variability. These results aid our understanding of clinical conditions associated with dopaminergic dysfunction, such as psychosis. Copyright © 2017 Diederen et al.
Tandem internal models execute motor learning in the cerebellum.
Honda, Takeru; Nagao, Soichi; Hashimoto, Yuji; Ishikawa, Kinya; Yokota, Takanori; Mizusawa, Hidehiro; Ito, Masao
2018-06-25
In performing skillful movement, humans use predictions from internal models formed by repetition learning. However, the computational organization of internal models in the brain remains unknown. Here, we demonstrate that a computational architecture employing a tandem configuration of forward and inverse internal models enables efficient motor learning in the cerebellum. The model predicted learning adaptations observed in hand-reaching experiments in humans wearing a prism lens and explained the kinetic components of these behavioral adaptations. The tandem system also predicted a form of subliminal motor learning that was experimentally validated after training intentional misses of hand targets. Patients with cerebellar degeneration disease showed behavioral impairments consistent with tandemly arranged internal models. These findings validate computational tandemization of internal models in motor control and its potential uses in more complex forms of learning and cognition. Copyright © 2018 the Author(s). Published by PNAS.
Effects of practice schedule and task specificity on the adaptive process of motor learning.
Barros, João Augusto de Camargo; Tani, Go; Corrêa, Umberto Cesar
2017-10-01
This study investigated the effects of practice schedule and task specificity based on the perspective of adaptive process of motor learning. For this purpose, tasks with temporal and force control learning requirements were manipulated in experiments 1 and 2, respectively. Specifically, the task consisted of touching with the dominant hand the three sequential targets with specific movement time or force for each touch. Participants were children (N=120), both boys and girls, with an average age of 11.2years (SD=1.0). The design in both experiments involved four practice groups (constant, random, constant-random, and random-constant) and two phases (stabilisation and adaptation). The dependent variables included measures related to the task goal (accuracy and variability of error of the overall movement and force patterns) and movement pattern (macro- and microstructures). Results revealed a similar error of the overall patterns for all groups in both experiments and that they adapted themselves differently in terms of the macro- and microstructures of movement patterns. The study concludes that the effects of practice schedules on the adaptive process of motor learning were both general and specific to the task. That is, they were general to the task goal performance and specific regarding the movement pattern. Copyright © 2017 Elsevier B.V. All rights reserved.
A learning flight control system for the F8-DFBW aircraft. [Digital Fly-By-Wire
NASA Technical Reports Server (NTRS)
Montgomery, R. C.; Mekel, R.; Nachmias, S.
1978-01-01
This report contains a complete description of a learning control system designed for the F8-DFBW aircraft. The system is parameter-adaptive with the additional feature that it 'learns' the variation of the control system gains needed over the flight envelope. It, thus, generates and modifies its gain schedule when suitable data are available. The report emphasizes the novel learning features of the system: the forms of representation of the flight envelope and the process by which identified parameters are used to modify the gain schedule. It contains data taken during piloted real-time 6 degree-of-freedom simulations that were used to develop and evaluate the system.
An adaptive learning control system for aircraft
NASA Technical Reports Server (NTRS)
Mekel, R.; Nachmias, S.
1978-01-01
A learning control system and its utilization as a flight control system for F-8 Digital Fly-By-Wire (DFBW) research aircraft is studied. The system has the ability to adjust a gain schedule to account for changing plant characteristics and to improve its performance and the plant's performance in the course of its own operation. Three subsystems are detailed: (1) the information acquisition subsystem which identifies the plant's parameters at a given operating condition; (2) the learning algorithm subsystem which relates the identified parameters to predetermined analytical expressions describing the behavior of the parameters over a range of operating conditions; and (3) the memory and control process subsystem which consists of the collection of updated coefficients (memory) and the derived control laws. Simulation experiments indicate that the learning control system is effective in compensating for parameter variations caused by changes in flight conditions.
Deep reinforcement learning for automated radiation adaptation in lung cancer.
Tseng, Huan-Hsin; Luo, Yi; Cui, Sunan; Chien, Jen-Tzung; Ten Haken, Randall K; Naqa, Issam El
2017-12-01
To investigate deep reinforcement learning (DRL) based on historical treatment plans for developing automated radiation adaptation protocols for nonsmall cell lung cancer (NSCLC) patients that aim to maximize tumor local control at reduced rates of radiation pneumonitis grade 2 (RP2). In a retrospective population of 114 NSCLC patients who received radiotherapy, a three-component neural networks framework was developed for deep reinforcement learning (DRL) of dose fractionation adaptation. Large-scale patient characteristics included clinical, genetic, and imaging radiomics features in addition to tumor and lung dosimetric variables. First, a generative adversarial network (GAN) was employed to learn patient population characteristics necessary for DRL training from a relatively limited sample size. Second, a radiotherapy artificial environment (RAE) was reconstructed by a deep neural network (DNN) utilizing both original and synthetic data (by GAN) to estimate the transition probabilities for adaptation of personalized radiotherapy patients' treatment courses. Third, a deep Q-network (DQN) was applied to the RAE for choosing the optimal dose in a response-adapted treatment setting. This multicomponent reinforcement learning approach was benchmarked against real clinical decisions that were applied in an adaptive dose escalation clinical protocol. In which, 34 patients were treated based on avid PET signal in the tumor and constrained by a 17.2% normal tissue complication probability (NTCP) limit for RP2. The uncomplicated cure probability (P+) was used as a baseline reward function in the DRL. Taking our adaptive dose escalation protocol as a blueprint for the proposed DRL (GAN + RAE + DQN) architecture, we obtained an automated dose adaptation estimate for use at ∼2/3 of the way into the radiotherapy treatment course. By letting the DQN component freely control the estimated adaptive dose per fraction (ranging from 1-5 Gy), the DRL automatically favored dose escalation/de-escalation between 1.5 and 3.8 Gy, a range similar to that used in the clinical protocol. The same DQN yielded two patterns of dose escalation for the 34 test patients, but with different reward variants. First, using the baseline P+ reward function, individual adaptive fraction doses of the DQN had similar tendencies to the clinical data with an RMSE = 0.76 Gy; but adaptations suggested by the DQN were generally lower in magnitude (less aggressive). Second, by adjusting the P+ reward function with higher emphasis on mitigating local failure, better matching of doses between the DQN and the clinical protocol was achieved with an RMSE = 0.5 Gy. Moreover, the decisions selected by the DQN seemed to have better concordance with patients eventual outcomes. In comparison, the traditional temporal difference (TD) algorithm for reinforcement learning yielded an RMSE = 3.3 Gy due to numerical instabilities and lack of sufficient learning. We demonstrated that automated dose adaptation by DRL is a feasible and a promising approach for achieving similar results to those chosen by clinicians. The process may require customization of the reward function if individual cases were to be considered. However, development of this framework into a fully credible autonomous system for clinical decision support would require further validation on larger multi-institutional datasets. © 2017 American Association of Physicists in Medicine.
Development of a Countermeasure to Mitigate Postflight Locomotor Dysfunction
NASA Technical Reports Server (NTRS)
Bloomberg, J. J.; Mulavara, A. P.; Peters, B. T.; Cohen, H. S.; Richards, J. T.; Miller, C. A.; Brady, R.; Warren, L. E.; Ruttley, T. M.
2006-01-01
Astronauts returning from space flight experience locomotor dysfunction following their return to Earth. Our laboratory is currently developing a gait adaptability training program that is designed to facilitate recovery of locomotor function following a return to a gravitational environment. The training program exploits the ability of the sensorimotor system to generalize from exposure to multiple adaptive challenges during training so that the gait control system essentially learns to learn and therefore can reorganize more rapidly when faced with a novel adaptive challenge. Evidence for the potential efficacy of an adaptive generalization gait training program can be obtained from numerous studies in the motor learning literature which have demonstrated that systematically varying the conditions of training enhances the ability of the performer to learn and retain a novel motor task. These variable practice training approaches have been used in applied contexts to improve motor skills required in a number of different sports. The central nervous system (CNS) can produce voluntary movement in an almost infinite number of ways. For example, locomotion can be achieved with many different combinations of joint angles, muscle activation patterns and forces. The CNS can exploit these degrees of freedom to enhance motor response adaptability during periods of adaptive flux like that encountered during a change in gravitational environment. Ultimately, the functional goal of an adaptive generalization countermeasure is not necessarily to immediately return movement patterns back to normal. Rather the training regimen should facilitate the reorganization of available sensory and motor subsystems to achieve safe and effective locomotion as soon as possible after long duration space flight. Indeed, this approach has been proposed as a basic feature underlying effective neurological rehabilitation. We have previously confirmed that subjects participating in an adaptive generalization training program using a variety of visuomotor distortions and throwing as the dependent measure can learn to enhance their ability to adapt to a novel sensorimotor environment (Roller et al., 2001). Importantly, this increased adaptability was retained even one month after completion of the training period. Adaptive generalization has been observed in a variety of other tasks requiring sensorimotor transformations including manual control tasks and reaching (Bock et al., 2001, Seidler, 2003) and obstacle avoidance during walking (Lam and Dietz, 2004). Taken together, the evidence suggests that a training regimen exposing crewmembers to variation in locomotor conditions, with repeated transitions among states, may enhance their ability to learn how to reassemble appropriate locomotor patterns upon return from microgravity. We believe exposure to this type of training will extend crewmembers locomotor behavioral repertoires, facilitating the return of functional mobility after long duration space flight. In other words, our proposed training protocol will compel subjects to develop new behavioral solutions under varying sensorimotor demands. Over time subjects will learn to create appropriate locomotor solution more rapidly enabling acquisition of mobility sooner after long-duration space flight. A gait adaptability training program can be superimposed on nominal treadmill exercise activities thus ensuring that no additional crew time is required to perform this type of training regimen and that it can be implemented with current in-flight exercise systems available on the International Space Station.
Jia, Zi-Jun; Song, Yong-Duan
2017-06-01
This paper presents a new approach to construct neural adaptive control for uncertain nonaffine systems. By integrating locally weighted learning with barrier Lyapunov function (BLF), a novel control design method is presented to systematically address the two critical issues in neural network (NN) control field: one is how to fulfill the compact set precondition for NN approximation, and the other is how to use varying rather than a fixed NN structure to improve the functionality of NN control. A BLF is exploited to ensure the NN inputs to remain bounded during the entire system operation. To account for system nonlinearities, a neuron self-growing strategy is proposed to guide the process for adding new neurons to the system, resulting in a self-adjustable NN structure for better learning capabilities. It is shown that the number of neurons needed to accomplish the control task is finite, and better performance can be obtained with less number of neurons as compared with traditional methods. The salient feature of the proposed method also lies in the continuity of the control action everywhere. Furthermore, the resulting control action is smooth almost everywhere except for a few time instants at which new neurons are added. Numerical example illustrates the effectiveness of the proposed approach.
NASA Technical Reports Server (NTRS)
Troudet, Terry; Merrill, Walter C.
1990-01-01
The ability of feed-forward neural network architectures to learn continuous valued mappings in the presence of noise was demonstrated in relation to parameter identification and real-time adaptive control applications. An error function was introduced to help optimize parameter values such as number of training iterations, observation time, sampling rate, and scaling of the control signal. The learning performance depended essentially on the degree of embodiment of the control law in the training data set and on the degree of uniformity of the probability distribution function of the data that are presented to the net during sequence. When a control law was corrupted by noise, the fluctuations of the training data biased the probability distribution function of the training data sequence. Only if the noise contamination is minimized and the degree of embodiment of the control law is maximized, can a neural net develop a good representation of the mapping and be used as a neurocontroller. A multilayer net was trained with back-error-propagation to control a cart-pole system for linear and nonlinear control laws in the presence of data processing noise and measurement noise. The neurocontroller exhibited noise-filtering properties and was found to operate more smoothly than the teacher in the presence of measurement noise.
HyFIS: adaptive neuro-fuzzy inference systems and their application to nonlinear dynamical systems.
Kim, J; Kasabov, N
1999-11-01
This paper proposes an adaptive neuro-fuzzy system, HyFIS (Hybrid neural Fuzzy Inference System), for building and optimising fuzzy models. The proposed model introduces the learning power of neural networks to fuzzy logic systems and provides linguistic meaning to the connectionist architectures. Heuristic fuzzy logic rules and input-output fuzzy membership functions can be optimally tuned from training examples by a hybrid learning scheme comprised of two phases: rule generation phase from data; and rule tuning phase using error backpropagation learning scheme for a neural fuzzy system. To illustrate the performance and applicability of the proposed neuro-fuzzy hybrid model, extensive simulation studies of nonlinear complex dynamic systems are carried out. The proposed method can be applied to an on-line incremental adaptive learning for the prediction and control of nonlinear dynamical systems. Two benchmark case studies are used to demonstrate that the proposed HyFIS system is a superior neuro-fuzzy modelling technique.
Telerobotic control of a mobile coordinated robotic server, executive summary
NASA Technical Reports Server (NTRS)
Lee, Gordon
1993-01-01
This interim report continues with the research effort on advanced adaptive controls for space robotics systems. In particular, previous results developed by the principle investigator and his research team centered around fuzzy logic control (FLC) in which the lack of knowledge of the robotic system as well as the uncertainties of the environment are compensated for by a rule base structure which interacts with varying degrees of belief of control action using system measurements. An on-line adaptive algorithm was developed using a single parameter tuning scheme. In the effort presented, the methodology is further developed to include on-line scaling factor tuning and self-learning control as well as extended to the multi-input, multi-output (MIMO) case. Classical fuzzy logic control requires tuning input scale factors off-line through trial and error techniques. This is time-consuming and cannot adapt to new changes in the process. The new adaptive FLC includes a self-tuning scheme for choosing the scaling factors on-line. Further the rule base in classical FLC is usually produced by soliciting knowledge from human operators as to what is good control action for given circumstances. This usually requires full knowledge and experience of the process and operating conditions, which limits applicability. A self-learning scheme is developed which adaptively forms the rule base with very limited knowledge of the process. Finally, a MIMO method is presented employing optimization techniques. This is required for application to space robotics in which several degrees-of-freedom links are commonly used. Simulation examples are presented for terminal control - typical of robotic problems in which a desired terminal point is to be reached for each link. Future activities will be to implement the MIMO adaptive FLC on an INTEL microcontroller-based circuit and to test the algorithm on a robotic system at the Mars Mission Research Center at North Carolina State University.
Toward a Generative Model of the Teaching-Learning Process.
ERIC Educational Resources Information Center
McMullen, David W.
Until the rise of cognitive psychology, models of the teaching-learning process (TLP) stressed external rather than internal variables. Models remained general descriptions until control theory introduced explicit system analyses. Cybernetic models emphasize feedback and adaptivity but give little attention to creativity. Research on artificial…
Active Inference, homeostatic regulation and adaptive behavioural control
Pezzulo, Giovanni; Rigoli, Francesco; Friston, Karl
2015-01-01
We review a theory of homeostatic regulation and adaptive behavioural control within the Active Inference framework. Our aim is to connect two research streams that are usually considered independently; namely, Active Inference and associative learning theories of animal behaviour. The former uses a probabilistic (Bayesian) formulation of perception and action, while the latter calls on multiple (Pavlovian, habitual, goal-directed) processes for homeostatic and behavioural control. We offer a synthesis these classical processes and cast them as successive hierarchical contextualisations of sensorimotor constructs, using the generative models that underpin Active Inference. This dissolves any apparent mechanistic distinction between the optimization processes that mediate classical control or learning. Furthermore, we generalize the scope of Active Inference by emphasizing interoceptive inference and homeostatic regulation. The ensuing homeostatic (or allostatic) perspective provides an intuitive explanation for how priors act as drives or goals to enslave action, and emphasises the embodied nature of inference. PMID:26365173
Jansen-Osmann, Petra; Richter, Stefanie; Konczak, Jürgen; Kalveram, Karl-Theodor
2002-03-01
When humans perform goal-directed arm movements under the influence of an external damping force, they learn to adapt to these external dynamics. After removal of the external force field, they reveal kinematic aftereffects that are indicative of a neural controller that still compensates the no longer existing force. Such behavior suggests that the adult human nervous system uses a neural representation of inverse arm dynamics to control upper-extremity motion. Central to the notion of an inverse dynamic model (IDM) is that learning generalizes. Consequently, aftereffects should be observable even in untrained workspace regions. Adults have shown such behavior, but the ontogenetic development of this process remains unclear. This study examines the adaptive behavior of children and investigates whether learning a force field in one hemifield of the right arm workspace has an effect on force adaptation in the other hemifield. Thirty children (aged 6-10 years) and ten adults performed 30 degrees elbow flexion movements under two conditions of external damping (negative and null). We found that learning to compensate an external damping force transferred to the opposite hemifield, which indicates that a model of the limb dynamics rather than an association of visited space and experienced force was acquired. Aftereffects were more pronounced in the younger children and readaptation to a null-force condition was prolonged. This finding is consistent with the view that IDMs in children are imprecise neural representations of the actual arm dynamics. It indicates that the acquisition of IDMs is a developmental achievement and that the human motor system is inherently flexible enough to adapt to any novel force within the limits of the organism's biomechanics.
Conflict adaptation in patients diagnosed with schizophrenia.
Abrahamse, Elger; Ruitenberg, Marit; Boddewyn, Sarah; Oreel, Edith; de Schryver, Maarten; Morrens, Manuel; van Dijck, Jean-Philippe
2017-11-01
Cognitive control impairments may contribute strongly to the overall cognitive deficits observed in patients diagnosed with schizophrenia. In the current study we explore a specific cognitive control function referred to as conflict adaptation. Previous studies on conflict adaptation in schizophrenia showed equivocal results, and, moreover, were plagued by confounded research designs. Here we assessed for the first time conflict adaptation in schizophrenia with a design that avoided the major confounds of feature integration and stimulus-response contingency learning. Sixteen patients diagnosed with schizophrenia and sixteen healthy, matched controls performed a vocal Stroop task to determine the congruency sequence effect - a marker of conflict adaptation. A reliable congruency sequence effect was observed for both healthy controls and patients diagnosed with schizophrenia. These findings indicate that schizophrenia is not necessarily accompanied by impaired conflict adaptation. As schizophrenia has been related to abnormal functioning in core conflict adaptation areas such as anterior cingulate and dorsolateral prefrontal cortex, further research is required to better understand the precise impact of such abnormal brain functioning at the behavioral level. Copyright © 2017 Elsevier B.V. All rights reserved.
Garrido, Jesús A.; Luque, Niceto R.; D'Angelo, Egidio; Ros, Eduardo
2013-01-01
Adaptable gain regulation is at the core of the forward controller operation performed by the cerebro-cerebellar loops and it allows the intensity of motor acts to be finely tuned in a predictive manner. In order to learn and store information about body-object dynamics and to generate an internal model of movement, the cerebellum is thought to employ long-term synaptic plasticity. LTD at the PF-PC synapse has classically been assumed to subserve this function (Marr, 1969). However, this plasticity alone cannot account for the broad dynamic ranges and time scales of cerebellar adaptation. We therefore tested the role of plasticity distributed over multiple synaptic sites (Hansel et al., 2001; Gao et al., 2012) by generating an analog cerebellar model embedded into a control loop connected to a robotic simulator. The robot used a three-joint arm and performed repetitive fast manipulations with different masses along an 8-shape trajectory. In accordance with biological evidence, the cerebellum model was endowed with both LTD and LTP at the PF-PC, MF-DCN and PC-DCN synapses. This resulted in a network scheme whose effectiveness was extended considerably compared to one including just PF-PC synaptic plasticity. Indeed, the system including distributed plasticity reliably self-adapted to manipulate different masses and to learn the arm-object dynamics over a time course that included fast learning and consolidation, along the lines of what has been observed in behavioral tests. In particular, PF-PC plasticity operated as a time correlator between the actual input state and the system error, while MF-DCN and PC-DCN plasticity played a key role in generating the gain controller. This model suggests that distributed synaptic plasticity allows generation of the complex learning properties of the cerebellum. The incorporation of further plasticity mechanisms and of spiking signal processing will allow this concept to be extended in a more realistic computational scenario. PMID:24130518
Quality Control Barriers in Adapting "Metro-Centric" Education to Regional Needs
ERIC Educational Resources Information Center
Nagy, Judy; Robinson, Susan R.
2013-01-01
The massification and globalization of higher education, combined with the widespread adoption of processes underpinning accreditation and quality control of university programs, have tended to result in learning contexts that are increasingly narrowly conceived and tightly controlled. Underlying many quality control measures is a "one size…
NASA Astrophysics Data System (ADS)
Ji, Xuewu; He, Xiangkun; Lv, Chen; Liu, Yahui; Wu, Jian
2018-06-01
Modelling uncertainty, parameter variation and unknown external disturbance are the major concerns in the development of an advanced controller for vehicle stability at the limits of handling. Sliding mode control (SMC) method has proved to be robust against parameter variation and unknown external disturbance with satisfactory tracking performance. But modelling uncertainty, such as errors caused in model simplification, is inevitable in model-based controller design, resulting in lowered control quality. The adaptive radial basis function network (ARBFN) can effectively improve the control performance against large system uncertainty by learning to approximate arbitrary nonlinear functions and ensure the global asymptotic stability of the closed-loop system. In this paper, a novel vehicle dynamics stability control strategy is proposed using the adaptive radial basis function network sliding mode control (ARBFN-SMC) to learn system uncertainty and eliminate its adverse effects. This strategy adopts a hierarchical control structure which consists of reference model layer, yaw moment control layer, braking torque allocation layer and executive layer. Co-simulation using MATLAB/Simulink and AMESim is conducted on a verified 15-DOF nonlinear vehicle system model with the integrated-electro-hydraulic brake system (I-EHB) actuator in a Sine With Dwell manoeuvre. The simulation results show that ARBFN-SMC scheme exhibits superior stability and tracking performance in different running conditions compared with SMC scheme.
Indirect decentralized repetitive control
NASA Technical Reports Server (NTRS)
Lee, Soo Cheol; Longman, Richard W.
1993-01-01
Learning control refers to controllers that learn to improve their performance at executing a given task, based on experience performing this specific task. In a previous work, the authors presented a theory of indirect decentralized learning control based on use of indirect adaptive control concepts employing simultaneous identification and control. This paper extends these results to apply to the indirect repetitive control problem in which a periodic (i.e., repetitive) command is given to a control system. Decentralized indirect repetitive control algorithms are presented that have guaranteed convergence to zero tracking error under very general conditions. The original motivation of the repetitive control and learning control fields was learning in robots doing repetitive tasks such as on an assembly line. This paper starts with decentralized discrete time systems, and progresses to the robot application, modeling the robot as a time varying linear system in the neighborhood of the desired trajectory. Decentralized repetitive control is natural for this application because the feedback control for link rotations is normally implemented in a decentralized manner, treating each link as if it is independent of the other links.
Intelligent fault-tolerant controllers
NASA Technical Reports Server (NTRS)
Huang, Chien Y.
1987-01-01
A system with fault tolerant controls is one that can detect, isolate, and estimate failures and perform necessary control reconfiguration based on this new information. Artificial intelligence (AI) is concerned with semantic processing, and it has evolved to include the topics of expert systems and machine learning. This research represents an attempt to apply AI to fault tolerant controls, hence, the name intelligent fault tolerant control (IFTC). A generic solution to the problem is sought, providing a system based on logic in addition to analytical tools, and offering machine learning capabilities. The advantages are that redundant system specific algorithms are no longer needed, that reasonableness is used to quickly choose the correct control strategy, and that the system can adapt to new situations by learning about its effects on system dynamics.
Learning and tuning fuzzy logic controllers through reinforcements
NASA Technical Reports Server (NTRS)
Berenji, Hamid R.; Khedkar, Pratap
1992-01-01
This paper presents a new method for learning and tuning a fuzzy logic controller based on reinforcements from a dynamic system. In particular, our generalized approximate reasoning-based intelligent control (GARIC) architecture (1) learns and tunes a fuzzy logic controller even when only weak reinforcement, such as a binary failure signal, is available; (2) introduces a new conjunction operator in computing the rule strengths of fuzzy control rules; (3) introduces a new localized mean of maximum (LMOM) method in combining the conclusions of several firing control rules; and (4) learns to produce real-valued control actions. Learning is achieved by integrating fuzzy inference into a feedforward neural network, which can then adaptively improve performance by using gradient descent methods. We extend the AHC algorithm of Barto et al. (1983) to include the prior control knowledge of human operators. The GARIC architecture is applied to a cart-pole balancing system and demonstrates significant improvements in terms of the speed of learning and robustness to changes in the dynamic system's parameters over previous schemes for cart-pole balancing.
Adaptive Critic-based Neurofuzzy Controller for the Steam Generator Water Level
NASA Astrophysics Data System (ADS)
Fakhrazari, Amin; Boroushaki, Mehrdad
2008-06-01
In this paper, an adaptive critic-based neurofuzzy controller is presented for water level regulation of nuclear steam generators. The problem has been of great concern for many years as the steam generator is a highly nonlinear system showing inverse response dynamics especially at low operating power levels. Fuzzy critic-based learning is a reinforcement learning method based on dynamic programming. The only information available for the critic agent is the system feedback which is interpreted as the last action the controller has performed in the previous state. The signal produced by the critic agent is used alongside the backpropagation of error algorithm to tune online conclusion parts of the fuzzy inference rules. The critic agent here has a proportional-derivative structure and the fuzzy rule base has nine rules. The proposed controller shows satisfactory transient responses, disturbance rejection and robustness to model uncertainty. Its simple design procedure and structure, nominates it as one of the suitable controller designs for the steam generator water level control in nuclear power plant industry.
NASA Technical Reports Server (NTRS)
Bomben, Craig R.; Smolka, James W.; Bosworth, John T.; Silliams-Hayes, Peggy S.; Burken, John J.; Larson, Richard R.; Buschbacher, Mark J.; Maliska, Heather A.
2006-01-01
The Intelligent Flight Control System (IFCS) project at the NASA Dryden Flight Research Center, Edwards AFB, CA, has been investigating the use of neural network based adaptive control on a unique NF-15B test aircraft. The IFCS neural network is a software processor that stores measured aircraft response information to dynamically alter flight control gains. In 2006, the neural network was engaged and allowed to learn in real time to dynamically alter the aircraft handling qualities characteristics in the presence of actual aerodynamic failure conditions injected into the aircraft through the flight control system. The use of neural network and similar adaptive technologies in the design of highly fault and damage tolerant flight control systems shows promise in making future aircraft far more survivable than current technology allows. This paper will present the results of the IFCS flight test program conducted at the NASA Dryden Flight Research Center in 2006, with emphasis on challenges encountered and lessons learned.
Ensuring Success of Adaptive Control Research Through Project Lifecycle Risk Mitigation
NASA Technical Reports Server (NTRS)
Pavlock, Kate M.
2011-01-01
Lessons Learne: 1. Design-out unnecessary risk to prevent excessive mitigation management during flight. 2. Consider iterative checkouts to confirm or improve human factor characteristics. 3. Consider the total flight test profile to uncover unanticipated human-algorithm interactions. 4. Consider test card cadence as a metric to assess test readiness. 5. Full-scale flight test is critical to development, maturation, and acceptance of adaptive control laws for operational use.
The cerebellum and visual perceptual learning: evidence from a motion extrapolation task.
Deluca, Cristina; Golzar, Ashkan; Santandrea, Elisa; Lo Gerfo, Emanuele; Eštočinová, Jana; Moretto, Giuseppe; Fiaschi, Antonio; Panzeri, Marta; Mariotti, Caterina; Tinazzi, Michele; Chelazzi, Leonardo
2014-09-01
Visual perceptual learning is widely assumed to reflect plastic changes occurring along the cerebro-cortical visual pathways, including at the earliest stages of processing, though increasing evidence indicates that higher-level brain areas are also involved. Here we addressed the possibility that the cerebellum plays an important role in visual perceptual learning. Within the realm of motor control, the cerebellum supports learning of new skills and recalibration of motor commands when movement execution is consistently perturbed (adaptation). Growing evidence indicates that the cerebellum is also involved in cognition and mediates forms of cognitive learning. Therefore, the obvious question arises whether the cerebellum might play a similar role in learning and adaptation within the perceptual domain. We explored a possible deficit in visual perceptual learning (and adaptation) in patients with cerebellar damage using variants of a novel motion extrapolation, psychophysical paradigm. Compared to their age- and gender-matched controls, patients with focal damage to the posterior (but not the anterior) cerebellum showed strongly diminished learning, in terms of both rate and amount of improvement over time. Consistent with a double-dissociation pattern, patients with focal damage to the anterior cerebellum instead showed more severe clinical motor deficits, indicative of a distinct role of the anterior cerebellum in the motor domain. The collected evidence demonstrates that a pure form of slow-incremental visual perceptual learning is crucially dependent on the intact cerebellum, bearing the notion that the human cerebellum acts as a learning device for motor, cognitive and perceptual functions. We interpret the deficit in terms of an inability to fine-tune predictive models of the incoming flow of visual perceptual input over time. Moreover, our results suggest a strong dissociation between the role of different portions of the cerebellum in motor versus non-motor functions, with only the posterior lobe being responsible for learning in the perceptual domain. Copyright © 2014. Published by Elsevier Ltd.
Air-Breathing Hypersonic Vehicle Tracking Control Based on Adaptive Dynamic Programming.
Mu, Chaoxu; Ni, Zhen; Sun, Changyin; He, Haibo
2017-03-01
In this paper, we propose a data-driven supplementary control approach with adaptive learning capability for air-breathing hypersonic vehicle tracking control based on action-dependent heuristic dynamic programming (ADHDP). The control action is generated by the combination of sliding mode control (SMC) and the ADHDP controller to track the desired velocity and the desired altitude. In particular, the ADHDP controller observes the differences between the actual velocity/altitude and the desired velocity/altitude, and then provides a supplementary control action accordingly. The ADHDP controller does not rely on the accurate mathematical model function and is data driven. Meanwhile, it is capable to adjust its parameters online over time under various working conditions, which is very suitable for hypersonic vehicle system with parameter uncertainties and disturbances. We verify the adaptive supplementary control approach versus the traditional SMC in the cruising flight, and provide three simulation studies to illustrate the improved performance with the proposed approach.
Online learning control using adaptive critic designs with sparse kernel machines.
Xu, Xin; Hou, Zhongsheng; Lian, Chuanqiang; He, Haibo
2013-05-01
In the past decade, adaptive critic designs (ACDs), including heuristic dynamic programming (HDP), dual heuristic programming (DHP), and their action-dependent ones, have been widely studied to realize online learning control of dynamical systems. However, because neural networks with manually designed features are commonly used to deal with continuous state and action spaces, the generalization capability and learning efficiency of previous ACDs still need to be improved. In this paper, a novel framework of ACDs with sparse kernel machines is presented by integrating kernel methods into the critic of ACDs. To improve the generalization capability as well as the computational efficiency of kernel machines, a sparsification method based on the approximately linear dependence analysis is used. Using the sparse kernel machines, two kernel-based ACD algorithms, that is, kernel HDP (KHDP) and kernel DHP (KDHP), are proposed and their performance is analyzed both theoretically and empirically. Because of the representation learning and generalization capability of sparse kernel machines, KHDP and KDHP can obtain much better performance than previous HDP and DHP with manually designed neural networks. Simulation and experimental results of two nonlinear control problems, that is, a continuous-action inverted pendulum problem and a ball and plate control problem, demonstrate the effectiveness of the proposed kernel ACD methods.
Rauter, Georg; Sigrist, Roland; Riener, Robert; Wolf, Peter
2015-01-01
In literature, the effectiveness of haptics for motor learning is controversially discussed. Haptics is believed to be effective for motor learning in general; however, different types of haptic control enhance different movement aspects. Thus, in dependence on the movement aspects of interest, one type of haptic control may be effective whereas another one is not. Therefore, in the current work, it was investigated if and how different types of haptic controllers affect learning of spatial and temporal movement aspects. In particular, haptic controllers that enforce active participation of the participants were expected to improve spatial aspects. Only haptic controllers that provide feedback about the task's velocity profile were expected to improve temporal aspects. In a study on learning a complex trunk-arm rowing task, the effect of training with four different types of haptic control was investigated: position control, path control, adaptive path control, and reactive path control. A fifth group (control) trained with visual concurrent augmented feedback. As hypothesized, the position controller was most effective for learning of temporal movement aspects, while the path controller was most effective in teaching spatial movement aspects of the rowing task. Visual feedback was also effective for learning temporal and spatial movement aspects.
ERIC Educational Resources Information Center
altawallbeh, Manal; Thiam, Wun; alshourah, Sultan; Fong, Soon Fook
2015-01-01
The purpose of this study is to examine the moderating factors that effect on adaption e-learning among students in Jordanian universities. Two models of e-learning that are observed among adopting institutions are: E-learning as a supplement to traditional classroom model have been introduced by the respondents in this research. The paper takes a…
Learning in the tutorial group: a balance between individual freedom and institutional control.
McAllister, Anita; Aanstoot, Janna; Hammarström, Inger Lundeborg; Samuelsson, Christina; Johannesson, Eva; Sandström, Karin; Berglind, Ulrika
2014-01-01
The study investigates factors in problem-based learning tutorial groups which promote or inhibit learning. The informants were tutors and students from speech-language pathology and physiotherapy programmes. Semi-structured focus-group interviews and individual interviews were used. Results revealed three themes: Responsibility. Time and Support. Under responsibility, the delicate balance between individual and institutional responsibility and control was shown. Time included short and long-term perspectives on learning. Under support, supporting documents, activities and personnel resources were mentioned. In summary, an increased control by the program and tutors decreases student's motivation to assume responsibility for learning. Support in tutorial groups needs to adapt to student progression and to be well aligned to tutorial work to have the intended effect. A lifelong learning perspective may help students develop a meta-awareness regarding learning that could make tutorial work more meaningful.
Neural network applications in telecommunications
NASA Technical Reports Server (NTRS)
Alspector, Joshua
1994-01-01
Neural network capabilities include automatic and organized handling of complex information, quick adaptation to continuously changing environments, nonlinear modeling, and parallel implementation. This viewgraph presentation presents Bellcore work on applications, learning chip computational function, learning system block diagram, neural network equalization, broadband access control, calling-card fraud detection, software reliability prediction, and conclusions.
Effect of visuomotor-map uncertainty on visuomotor adaptation.
Saijo, Naoki; Gomi, Hiroaki
2012-03-01
Vision and proprioception contribute to generating hand movement. If a conflict between the visual and proprioceptive feedback of hand position is given, reaching movement is disturbed initially but recovers after training. Although previous studies have predominantly investigated the adaptive change in the motor output, it is unclear whether the contributions of visual and proprioceptive feedback controls to the reaching movement are modified by visuomotor adaptation. To investigate this, we focused on the change in proprioceptive feedback control associated with visuomotor adaptation. After the adaptation to gradually introduce visuomotor rotation, the hand reached the shifted position of the visual target to move the cursor to the visual target correctly. When the cursor feedback was occasionally eliminated (probe trial), the end point of the hand movement was biased in the visual-target direction, while the movement was initiated in the adapted direction, suggesting the incomplete adaptation of proprioceptive feedback control. Moreover, after the learning of uncertain visuomotor rotation, in which the rotation angle was randomly fluctuated on a trial-by-trial basis, the end-point bias in the probe trial increased, but the initial movement direction was not affected, suggesting a reduction in the adaptation level of proprioceptive feedback control. These results suggest that the change in the relative contribution of visual and proprioceptive feedback controls to the reaching movement in response to the visuomotor-map uncertainty is involved in visuomotor adaptation, whereas feedforward control might adapt in a manner different from that of the feedback control.
An extended ASLD trading system to enhance portfolio management.
Hung, Kei-Keung; Cheung, Yiu-Ming; Xu, Lei
2003-01-01
An adaptive supervised learning decision (ASLD) trading system has been presented by Xu and Cheung (1997) to optimize the expected returns of investment without considering risks. In this paper, we propose an extension of the ASLD system (EASLD), which combines the ASLD with a portfolio optimization scheme to take a balance between the expected returns and risks. This new system not only keeps the learning adaptability of the ASLD, but also dynamically controls the risk in pursuit of great profits by diversifying the capital to a time-varying portfolio of N assets. Consequently, it is shown that: 1) the EASLD system gives the investment risk much smaller than the ASLD one; and 2) more returns are gained through the EASLD system in comparison with the two individual portfolio optimization schemes that statically determine the portfolio weights without adaptive learning. We have justified these two issues by the experiments.
Development of a Countermeasure to Enhance Postflight Locomotor Adaptability
NASA Technical Reports Server (NTRS)
Bloomberg, Jacob J.
2006-01-01
Astronauts returning from space flight experience locomotor dysfunction following their return to Earth. Our laboratory is currently developing a gait adaptability training program that is designed to facilitate recovery of locomotor function following a return to a gravitational environment. The training program exploits the ability of the sensorimotor system to generalize from exposure to multiple adaptive challenges during training so that the gait control system essentially learns to learn and therefore can reorganize more rapidly when faced with a novel adaptive challenge. We have previously confirmed that subjects participating in adaptive generalization training programs using a variety of visuomotor distortions can enhance their ability to adapt to a novel sensorimotor environment. Importantly, this increased adaptability was retained even one month after completion of the training period. Adaptive generalization has been observed in a variety of other tasks requiring sensorimotor transformations including manual control tasks and reaching (Bock et al., 2001, Seidler, 2003) and obstacle avoidance during walking (Lam and Dietz, 2004). Taken together, the evidence suggests that a training regimen exposing crewmembers to variation in locomotor conditions, with repeated transitions among states, may enhance their ability to learn how to reassemble appropriate locomotor patterns upon return from microgravity. We believe exposure to this type of training will extend crewmembers locomotor behavioral repertoires, facilitating the return of functional mobility after long duration space flight. Our proposed training protocol will compel subjects to develop new behavioral solutions under varying sensorimotor demands. Over time subjects will learn to create appropriate locomotor solution more rapidly enabling acquisition of mobility sooner after long-duration space flight. Our laboratory is currently developing adaptive generalization training procedures and the associated flight hardware to implement such a training program during regular inflight treadmill operations. A visual display system will provide variation in visual flow patterns during treadmill exercise. Crewmembers will be exposed to a virtual scene that can translate and rotate in six-degrees-of freedom during their regular treadmill exercise period. Associated ground based studies are focused on determining optimal combinations of sensory manipulations (visual flow, body loading and support surface variation) and training schedules that will produce the greatest potential for adaptive flexibility in gait function during exposure to challenging and novel environments. An overview of our progress in these areas will be discussed during the presentation.
Observer-based distributed adaptive iterative learning control for linear multi-agent systems
NASA Astrophysics Data System (ADS)
Li, Jinsha; Liu, Sanyang; Li, Junmin
2017-10-01
This paper investigates the consensus problem for linear multi-agent systems from the viewpoint of two-dimensional systems when the state information of each agent is not available. Observer-based fully distributed adaptive iterative learning protocol is designed in this paper. A local observer is designed for each agent and it is shown that without using any global information about the communication graph, all agents achieve consensus perfectly for all undirected connected communication graph when the number of iterations tends to infinity. The Lyapunov-like energy function is employed to facilitate the learning protocol design and property analysis. Finally, simulation example is given to illustrate the theoretical analysis.
Adaptive Search through Constraint Violations
1990-01-01
procedural) knowledge? Different methodologies are used to investigate these questions: Psychological experiments, computer simulations, historical studies...learns control knowledge through adaptive search. Unlike most other psychological models of skill acquisition, HS is a model of analytical, or...Newzll, 1986; VanLehn, in press). Psychological models of skill acquisition employ different problem solving mechanisms (forward search, backward
Intelligent neural network and fuzzy logic control of industrial and power systems
NASA Astrophysics Data System (ADS)
Kuljaca, Ognjen
The main role played by neural network and fuzzy logic intelligent control algorithms today is to identify and compensate unknown nonlinear system dynamics. There are a number of methods developed, but often the stability analysis of neural network and fuzzy control systems was not provided. This work will meet those problems for the several algorithms. Some more complicated control algorithms included backstepping and adaptive critics will be designed. Nonlinear fuzzy control with nonadaptive fuzzy controllers is also analyzed. An experimental method for determining describing function of SISO fuzzy controller is given. The adaptive neural network tracking controller for an autonomous underwater vehicle is analyzed. A novel stability proof is provided. The implementation of the backstepping neural network controller for the coupled motor drives is described. Analysis and synthesis of adaptive critic neural network control is also provided in the work. Novel tuning laws for the system with action generating neural network and adaptive fuzzy critic are given. Stability proofs are derived for all those control methods. It is shown how these control algorithms and approaches can be used in practical engineering control. Stability proofs are given. Adaptive fuzzy logic control is analyzed. Simulation study is conducted to analyze the behavior of the adaptive fuzzy system on the different environment changes. A novel stability proof for adaptive fuzzy logic systems is given. Also, adaptive elastic fuzzy logic control architecture is described and analyzed. A novel membership function is used for elastic fuzzy logic system. The stability proof is proffered. Adaptive elastic fuzzy logic control is compared with the adaptive nonelastic fuzzy logic control. The work described in this dissertation serves as foundation on which analysis of particular representative industrial systems will be conducted. Also, it gives a good starting point for analysis of learning abilities of adaptive and neural network control systems, as well as for the analysis of the different algorithms such as elastic fuzzy systems.
Intelligent control of robotic arm/hand systems for the NASA EVA retriever using neural networks
NASA Technical Reports Server (NTRS)
Mclauchlan, Robert A.
1989-01-01
Adaptive/general learning algorithms using varying neural network models are considered for the intelligent control of robotic arm plus dextrous hand/manipulator systems. Results are summarized and discussed for the use of the Barto/Sutton/Anderson neuronlike, unsupervised learning controller as applied to the stabilization of an inverted pendulum on a cart system. Recommendations are made for the application of the controller and a kinematic analysis for trajectory planning to simple object retrieval (chase/approach and capture/grasp) scenarios in two dimensions.
Li, Cai; Lowe, Robert; Ziemke, Tom
2014-01-01
In this article, we propose an architecture of a bio-inspired controller that addresses the problem of learning different locomotion gaits for different robot morphologies. The modeling objective is split into two: baseline motion modeling and dynamics adaptation. Baseline motion modeling aims to achieve fundamental functions of a certain type of locomotion and dynamics adaptation provides a "reshaping" function for adapting the baseline motion to desired motion. Based on this assumption, a three-layer architecture is developed using central pattern generators (CPGs, a bio-inspired locomotor center for the baseline motion) and dynamic motor primitives (DMPs, a model with universal "reshaping" functions). In this article, we use this architecture with the actor-critic algorithms for finding a good "reshaping" function. In order to demonstrate the learning power of the actor-critic based architecture, we tested it on two experiments: (1) learning to crawl on a humanoid and, (2) learning to gallop on a puppy robot. Two types of actor-critic algorithms (policy search and policy gradient) are compared in order to evaluate the advantages and disadvantages of different actor-critic based learning algorithms for different morphologies. Finally, based on the analysis of the experimental results, a generic view/architecture for locomotion learning is discussed in the conclusion.
Li, Cai; Lowe, Robert; Ziemke, Tom
2014-01-01
In this article, we propose an architecture of a bio-inspired controller that addresses the problem of learning different locomotion gaits for different robot morphologies. The modeling objective is split into two: baseline motion modeling and dynamics adaptation. Baseline motion modeling aims to achieve fundamental functions of a certain type of locomotion and dynamics adaptation provides a “reshaping” function for adapting the baseline motion to desired motion. Based on this assumption, a three-layer architecture is developed using central pattern generators (CPGs, a bio-inspired locomotor center for the baseline motion) and dynamic motor primitives (DMPs, a model with universal “reshaping” functions). In this article, we use this architecture with the actor-critic algorithms for finding a good “reshaping” function. In order to demonstrate the learning power of the actor-critic based architecture, we tested it on two experiments: (1) learning to crawl on a humanoid and, (2) learning to gallop on a puppy robot. Two types of actor-critic algorithms (policy search and policy gradient) are compared in order to evaluate the advantages and disadvantages of different actor-critic based learning algorithms for different morphologies. Finally, based on the analysis of the experimental results, a generic view/architecture for locomotion learning is discussed in the conclusion. PMID:25324773
Masters of adaptation: learning in late life adjustments.
Roberson, Donald N
2005-01-01
The purpose of this research is to understand the relationship between human development in older adults and personal learning. Personal or self-directed learning (SDL) refers to a style of learning where the individual directs, controls, and evaluates what is learned. It may occur with formal classes, but most often takes place in non-formal situations. This study employed a descriptive qualitative design incorporating in-depth, semistructured interviews for data collection. The sample of 10 purposefully selected older adults from a rural area reflected diversity in gender, race, education, and employment. Data analysis was guided by the constant comparative method. The primary late life adjustments of these older adults were in response to having extra time, changes in family, and social and physical loss. This research also indicated that late life adjustments are a primary incentive for self-directed learning. The results of this study indicated that older adults become masters of adaptation through the use of self-directed learning activities.
Singh, Puneet; Jana, Sumitash; Ghosal, Ashitava; Murthy, Aditya
2016-12-13
The number of joints and muscles in a human arm is more than what is required for reaching to a desired point in 3D space. Although previous studies have emphasized how such redundancy and the associated flexibility may play an important role in path planning, control of noise, and optimization of motion, whether and how redundancy might promote motor learning has not been investigated. In this work, we quantify redundancy space and investigate its significance and effect on motor learning. We propose that a larger redundancy space leads to faster learning across subjects. We observed this pattern in subjects learning novel kinematics (visuomotor adaptation) and dynamics (force-field adaptation). Interestingly, we also observed differences in the redundancy space between the dominant hand and nondominant hand that explained differences in the learning of dynamics. Taken together, these results provide support for the hypothesis that redundancy aids in motor learning and that the redundant component of motor variability is not noise.
Exploration of joint redundancy but not task space variability facilitates supervised motor learning
Singh, Puneet; Jana, Sumitash; Ghosal, Ashitava; Murthy, Aditya
2016-01-01
The number of joints and muscles in a human arm is more than what is required for reaching to a desired point in 3D space. Although previous studies have emphasized how such redundancy and the associated flexibility may play an important role in path planning, control of noise, and optimization of motion, whether and how redundancy might promote motor learning has not been investigated. In this work, we quantify redundancy space and investigate its significance and effect on motor learning. We propose that a larger redundancy space leads to faster learning across subjects. We observed this pattern in subjects learning novel kinematics (visuomotor adaptation) and dynamics (force-field adaptation). Interestingly, we also observed differences in the redundancy space between the dominant hand and nondominant hand that explained differences in the learning of dynamics. Taken together, these results provide support for the hypothesis that redundancy aids in motor learning and that the redundant component of motor variability is not noise. PMID:27911808
Intelligent control based on fuzzy logic and neural net theory
NASA Technical Reports Server (NTRS)
Lee, Chuen-Chien
1991-01-01
In the conception and design of intelligent systems, one promising direction involves the use of fuzzy logic and neural network theory to enhance such systems' capability to learn from experience and adapt to changes in an environment of uncertainty and imprecision. Here, an intelligent control scheme is explored by integrating these multidisciplinary techniques. A self-learning system is proposed as an intelligent controller for dynamical processes, employing a control policy which evolves and improves automatically. One key component of the intelligent system is a fuzzy logic-based system which emulates human decision making behavior. It is shown that the system can solve a fairly difficult control learning problem. Simulation results demonstrate that improved learning performance can be achieved in relation to previously described systems employing bang-bang control. The proposed system is relatively insensitive to variations in the parameters of the system environment.
Learning-induced Dependence of Neuronal Activity in Primary Motor Cortex on Motor Task Condition.
Cai, X; Shimansky, Y; He, Jiping
2005-01-01
A brain-computer interface (BCI) system such as a cortically controlled robotic arm must have a capacity of adjusting its function to a specific environmental condition. We studied this capacity in non-human primates based on chronic multi-electrode recording from the primary motor cortex of a monkey during the animal's performance of a center-out 3D reaching task and adaptation to external force perturbations. The main condition-related feature of motor cortical activity observed before the onset of force perturbation was a phasic raise of activity immediately before the perturbation onset. This feature was observed during a series of perturbation trials, but were absent under no perturbations. After adaptation has been completed, it usually was taking the subject only one trial to recognize a change in the condition to switch the neuronal activity accordingly. These condition-dependent features of neuronal activity can be used by a BCI for recognizing a change in the environmental condition and making corresponding adjustments, which requires that the BCI-based control system possess such advanced properties of the neural motor control system as capacity to learn and adapt.
Zheng, Shiqi; Tang, Xiaoqi; Song, Bao; Lu, Shaowu; Ye, Bosheng
2013-07-01
In this paper, a stable adaptive PI control strategy based on the improved just-in-time learning (IJITL) technique is proposed for permanent magnet synchronous motor (PMSM) drive. Firstly, the traditional JITL technique is improved. The new IJITL technique has less computational burden and is more suitable for online identification of the PMSM drive system which is highly real-time compared to traditional JITL. In this way, the PMSM drive system is identified by IJITL technique, which provides information to an adaptive PI controller. Secondly, the adaptive PI controller is designed in discrete time domain which is composed of a PI controller and a supervisory controller. The PI controller is capable of automatically online tuning the control gains based on the gradient descent method and the supervisory controller is developed to eliminate the effect of the approximation error introduced by the PI controller upon the system stability in the Lyapunov sense. Finally, experimental results on the PMSM drive system show accurate identification and favorable tracking performance. Copyright © 2013 ISA. Published by Elsevier Ltd. All rights reserved.
Puzzles in modern biology. V. Why are genomes overwired?
Frank, Steven A
2017-01-01
Many factors affect eukaryotic gene expression. Transcription factors, histone codes, DNA folding, and noncoding RNA modulate expression. Those factors interact in large, broadly connected regulatory control networks. An engineer following classical principles of control theory would design a simpler regulatory network. Why are genomes overwired? Neutrality or enhanced robustness may lead to the accumulation of additional factors that complicate network architecture. Dynamics progresses like a ratchet. New factors get added. Genomes adapt to the additional complexity. The newly added factors can no longer be removed without significant loss of fitness. Alternatively, highly wired genomes may be more malleable. In large networks, most genomic variants tend to have a relatively small effect on gene expression and trait values. Many small effects lead to a smooth gradient, in which traits may change steadily with respect to underlying regulatory changes. A smooth gradient may provide a continuous path from a starting point up to the highest peak of performance. A potential path of increasing performance promotes adaptability and learning. Genomes gain by the inductive process of natural selection, a trial and error learning algorithm that discovers general solutions for adapting to environmental challenge. Similarly, deeply and densely connected computational networks gain by various inductive trial and error learning procedures, in which the networks learn to reduce the errors in sequential trials. Overwiring alters the geometry of induction by smoothing the gradient along the inductive pathways of improving performance. Those overwiring benefits for induction apply to both natural biological networks and artificial deep learning networks.
Modares, Hamidreza; Lewis, Frank L; Naghibi-Sistani, Mohammad-Bagher
2013-10-01
This paper presents an online policy iteration (PI) algorithm to learn the continuous-time optimal control solution for unknown constrained-input systems. The proposed PI algorithm is implemented on an actor-critic structure where two neural networks (NNs) are tuned online and simultaneously to generate the optimal bounded control policy. The requirement of complete knowledge of the system dynamics is obviated by employing a novel NN identifier in conjunction with the actor and critic NNs. It is shown how the identifier weights estimation error affects the convergence of the critic NN. A novel learning rule is developed to guarantee that the identifier weights converge to small neighborhoods of their ideal values exponentially fast. To provide an easy-to-check persistence of excitation condition, the experience replay technique is used. That is, recorded past experiences are used simultaneously with current data for the adaptation of the identifier weights. Stability of the whole system consisting of the actor, critic, system state, and system identifier is guaranteed while all three networks undergo adaptation. Convergence to a near-optimal control law is also shown. The effectiveness of the proposed method is illustrated with a simulation example.
Lefumat, Hannah Z.; Vercher, Jean-Louis; Miall, R. Chris; Cole, Jonathan; Buloup, Frank; Bringoux, Lionel; Bourdin, Christophe
2015-01-01
Humans can remarkably adapt their motor behavior to novel environmental conditions, yet it remains unclear which factors enable us to transfer what we have learned with one limb to the other. Here we tested the hypothesis that interlimb transfer of sensorimotor adaptation is determined by environmental conditions but also by individual characteristics. We specifically examined the adaptation of unconstrained reaching movements to a novel Coriolis, velocity-dependent force field. Right-handed subjects sat at the center of a rotating platform and performed forward reaching movements with the upper limb toward flashed visual targets in prerotation, per-rotation (i.e., adaptation), and postrotation tests. Here only the dominant arm was used during adaptation and interlimb transfer was assessed by comparing performance of the nondominant arm before and after dominant-arm adaptation. Vision and no-vision conditions did not significantly influence interlimb transfer of trajectory adaptation, which on average was significant but limited. We uncovered a substantial heterogeneity of interlimb transfer across subjects and found that interlimb transfer can be qualitatively and quantitatively predicted for each healthy young individual. A classifier showed that in our study, interlimb transfer could be predicted based on the subject's task performance, most notably motor variability during learning, and his or her laterality quotient. Positive correlations suggested that variability of motor performance and lateralization of arm movement control facilitate interlimb transfer. We further show that these individual characteristics can predict the presence and the magnitude of interlimb transfer of left-handers. Overall, this study suggests that individual characteristics shape the way the nervous system can generalize motor learning. PMID:26334018
Zhu, Feng; Aziz, H. M. Abdul; Qian, Xinwu; ...
2015-01-31
Our study develops a novel reinforcement learning algorithm for the challenging coordinated signal control problem. Traffic signals are modeled as intelligent agents interacting with the stochastic traffic environment. The model is built on the framework of coordinated reinforcement learning. The Junction Tree Algorithm (JTA) based reinforcement learning is proposed to obtain an exact inference of the best joint actions for all the coordinated intersections. Moreover, the algorithm is implemented and tested with a network containing 18 signalized intersections in VISSIM. Finally, our results show that the JTA based algorithm outperforms independent learning (Q-learning), real-time adaptive learning, and fixed timing plansmore » in terms of average delay, number of stops, and vehicular emissions at the network level.« less
Biohybrid Control of General Linear Systems Using the Adaptive Filter Model of Cerebellum.
Wilson, Emma D; Assaf, Tareq; Pearson, Martin J; Rossiter, Jonathan M; Dean, Paul; Anderson, Sean R; Porrill, John
2015-01-01
The adaptive filter model of the cerebellar microcircuit has been successfully applied to biological motor control problems, such as the vestibulo-ocular reflex (VOR), and to sensory processing problems, such as the adaptive cancelation of reafferent noise. It has also been successfully applied to problems in robotics, such as adaptive camera stabilization and sensor noise cancelation. In previous applications to inverse control problems, the algorithm was applied to the velocity control of a plant dominated by viscous and elastic elements. Naive application of the adaptive filter model to the displacement (as opposed to velocity) control of this plant results in unstable learning and control. To be more generally useful in engineering problems, it is essential to remove this restriction to enable the stable control of plants of any order. We address this problem here by developing a biohybrid model reference adaptive control (MRAC) scheme, which stabilizes the control algorithm for strictly proper plants. We evaluate the performance of this novel cerebellar-inspired algorithm with MRAC scheme in the experimental control of a dielectric electroactive polymer, a class of artificial muscle. The results show that the augmented cerebellar algorithm is able to accurately control the displacement response of the artificial muscle. The proposed solution not only greatly extends the practical applicability of the cerebellar-inspired algorithm, but may also shed light on cerebellar involvement in a wider range of biological control tasks.
Maity, Arnab; Hocht, Leonhard; Heise, Christian; Holzapfel, Florian
2018-01-01
A new efficient adaptive optimal control approach is presented in this paper based on the indirect model reference adaptive control (MRAC) architecture for improvement of adaptation and tracking performance of the uncertain system. The system accounts here for both matched and unmatched unknown uncertainties that can act as plant as well as input effectiveness failures or damages. For adaptation of the unknown parameters of these uncertainties, the frequency selective learning approach is used. Its idea is to compute a filtered expression of the system uncertainty using multiple filters based on online instantaneous information, which is used for augmentation of the update law. It is capable of adjusting a sudden change in system dynamics without depending on high adaptation gains and can satisfy exponential parameter error convergence under certain conditions in the presence of structured matched and unmatched uncertainties as well. Additionally, the controller of the MRAC system is designed using a new optimal control method. This method is a new linear quadratic regulator-based optimal control formulation for both output regulation and command tracking problems. It provides a closed-form control solution. The proposed overall approach is applied in a control of lateral dynamics of an unmanned aircraft problem to show its effectiveness.
Adaptive and perceptual learning technologies in medical education and training.
Kellman, Philip J
2013-10-01
Recent advances in the learning sciences offer remarkable potential to improve medical education and maximize the benefits of emerging medical technologies. This article describes 2 major innovation areas in the learning sciences that apply to simulation and other aspects of medical learning: Perceptual learning (PL) and adaptive learning technologies. PL technology offers, for the first time, systematic, computer-based methods for teaching pattern recognition, structural intuition, transfer, and fluency. Synergistic with PL are new adaptive learning technologies that optimize learning for each individual, embed objective assessment, and implement mastery criteria. The author describes the Adaptive Response-Time-based Sequencing (ARTS) system, which uses each learner's accuracy and speed in interactive learning to guide spacing, sequencing, and mastery. In recent efforts, these new technologies have been applied in medical learning contexts, including adaptive learning modules for initial medical diagnosis and perceptual/adaptive learning modules (PALMs) in dermatology, histology, and radiology. Results of all these efforts indicate the remarkable potential of perceptual and adaptive learning technologies, individually and in combination, to improve learning in a variety of medical domains. Reprint & Copyright © 2013 Association of Military Surgeons of the U.S.
Bryant, D P; Bryant, B R
1998-01-01
Cooperative learning (CL) is a common instructional arrangement that is used by classroom teachers to foster academic achievement and social acceptance of students with and without learning disabilities. Cooperative learning is appealing to classroom teachers because it can provide an opportunity for more instruction and feedback by peers than can be provided by teachers to individual students who require extra assistance. Recent studies suggest that students with LD may need adaptations during cooperative learning activities. The use of assistive technology adaptations may be necessary to help some students with LD compensate for their specific learning difficulties so that they can engage more readily in cooperative learning activities. A process for integrating technology adaptations into cooperative learning activities is discussed in terms of three components: selecting adaptations, monitoring the use of the adaptations during cooperative learning activities, and evaluating the adaptations' effectiveness. The article concludes with comments regarding barriers to and support systems for technology integration, technology and effective instructional practices, and the need to consider technology adaptations for students who have learning disabilities.
Habitual control of goal selection in humans
Cushman, Fiery; Morris, Adam
2015-01-01
Humans choose actions based on both habit and planning. Habitual control is computationally frugal but adapts slowly to novel circumstances, whereas planning is computationally expensive but can adapt swiftly. Current research emphasizes the competition between habits and plans for behavioral control, yet many complex tasks instead favor their integration. We consider a hierarchical architecture that exploits the computational efficiency of habitual control to select goals while preserving the flexibility of planning to achieve those goals. We formalize this mechanism in a reinforcement learning setting, illustrate its costs and benefits, and experimentally demonstrate its spontaneous application in a sequential decision-making task. PMID:26460050
2014-01-01
healthier and more prosperous. RAND is nonprofit, nonpartisan, and committed to the public interest. RAND’s publications do not necessarily reflect ...Toolkit 5 Continuous learning (related terms: adaptive learning, willingness to learn, active learning, metacognition ): Takes the necessary...when not under direct supervision; displays self -discipline and self -control; diligently checks work to ensure that all essential details have been
2015-01-01
reflect the opinions of its research clients and sponsors. Support RAND Make a tax-deductible charitable contribution at www.rand.org/giving...This Toolkit 5 Continuous learning (related terms: adaptive learning, willingness to learn, active learning, metacognition ): Takes the necessary...even when not under direct supervision; displays self -discipline and self -control; diligently checks work to ensure that all essential details have
2014-01-01
nonprofit, nonpartisan, and committed to the public interest. RAND’s publications do not necessarily reflect the opinions of its research clients...learning (related terms: adaptive learning, willingness to learn, active learning, metacognition ): Takes the necessary actions to develop and maintain...supervision; displays self -discipline and self -control; diligently checks work to ensure that all essential details have been considered; follows oral and
2015-01-01
RAND is nonprofit, nonpartisan, and committed to the public interest. RAND’s publications do not necessarily reflect the opinions of its research...Continuous learning (related terms: adaptive learning, willingness to learn, active learning, metacognition ): Takes the necessary actions to develop and...supervision; displays self -discipline and self -control; diligently checks work to ensure that all essential details have been considered; follows oral
Adaptive robotic control driven by a versatile spiking cerebellar network.
Casellato, Claudia; Antonietti, Alberto; Garrido, Jesus A; Carrillo, Richard R; Luque, Niceto R; Ros, Eduardo; Pedrocchi, Alessandra; D'Angelo, Egidio
2014-01-01
The cerebellum is involved in a large number of different neural processes, especially in associative learning and in fine motor control. To develop a comprehensive theory of sensorimotor learning and control, it is crucial to determine the neural basis of coding and plasticity embedded into the cerebellar neural circuit and how they are translated into behavioral outcomes in learning paradigms. Learning has to be inferred from the interaction of an embodied system with its real environment, and the same cerebellar principles derived from cell physiology have to be able to drive a variety of tasks of different nature, calling for complex timing and movement patterns. We have coupled a realistic cerebellar spiking neural network (SNN) with a real robot and challenged it in multiple diverse sensorimotor tasks. Encoding and decoding strategies based on neuronal firing rates were applied. Adaptive motor control protocols with acquisition and extinction phases have been designed and tested, including an associative Pavlovian task (Eye blinking classical conditioning), a vestibulo-ocular task and a perturbed arm reaching task operating in closed-loop. The SNN processed in real-time mossy fiber inputs as arbitrary contextual signals, irrespective of whether they conveyed a tone, a vestibular stimulus or the position of a limb. A bidirectional long-term plasticity rule implemented at parallel fibers-Purkinje cell synapses modulated the output activity in the deep cerebellar nuclei. In all tasks, the neurorobot learned to adjust timing and gain of the motor responses by tuning its output discharge. It succeeded in reproducing how human biological systems acquire, extinguish and express knowledge of a noisy and changing world. By varying stimuli and perturbations patterns, real-time control robustness and generalizability were validated. The implicit spiking dynamics of the cerebellar model fulfill timing, prediction and learning functions.
Different levels of learning interact to shape the congruency sequence effect.
Weissman, Daniel H; Hawks, Zoë W; Egner, Tobias
2016-04-01
The congruency effect in distracter interference tasks is often reduced after incongruent relative to congruent trials. Moreover, this congruency sequence effect (CSE) is influenced by learning related to concrete stimulus and response features as well as by learning related to abstract cognitive control processes. There is an ongoing debate, however, over whether interactions between these learning processes are best explained by an episodic retrieval account, an adaptation by binding account, or a cognitive efficiency account of the CSE. To make this distinction, we orthogonally manipulated the expression of these learning processes in a novel factorial design involving the prime-probe arrow task. In Experiment 1, these processes interacted in an over-additive fashion to influence CSE magnitude. In Experiment 2, we replicated this interaction while showing it was not driven by conditional differences in the size of the congruency effect. In Experiment 3, we ruled out an alternative account of this interaction as reflecting conditional differences in learning related to concrete stimulus and response features. These findings support an episodic retrieval account of the CSE, in which repeating a stimulus feature from the previous trial facilitates the retrieval and use of previous-trial control parameters, thereby boosting control in the current trial. In contrast, they do not fit with (a) an adaptation by binding account, in which CSE magnitude is directly related to the size of the congruency effect, or (b) a cognitive efficiency account, in which costly control processes are recruited only when behavioral adjustments cannot be mediated by low-level associative mechanisms. (c) 2016 APA, all rights reserved).
Adapting proportional myoelectric-controlled interfaces for prosthetic hands.
Pistohl, Tobias; Cipriani, Christian; Jackson, Andrew; Nazarpour, Kianoush
2013-01-01
Powered hand prostheses with many degrees of freedom are moving from research into the market for prosthetics. In order to make use of the prostheses' full functionality, it is essential to find efficient ways to control their multiple actuators. Human subjects can rapidly learn to employ electromyographic (EMG) activity of several hand and arm muscles to control the position of a cursor on a computer screen, even if the muscle-cursor map contradicts directions in which the muscles would act naturally. We investigated whether a similar control scheme, using signals from four hand muscles, could be adopted for real-time operation of a dexterous robotic hand. Despite different mapping strategies, learning to control the robotic hand over time was surprisingly similar to the learning of two-dimensional cursor control.
NASA Technical Reports Server (NTRS)
Warren, L. E.; Mulavara, A. P.; Peters, B. T.; Cohen, H. S.; Richards, J. T.; Miller, C. A.; Brady, R.; Ruttley, T. M.; Bloomberg, J. J.
2006-01-01
Space flight induces adaptive modification in sensorimotor function, allowing crewmembers to operate in the unique microgravity environment. This adaptive state, however, is inappropriate for a terrestrial environment. During a re-adaptation period upon their return to Earth, crewmembers experience alterations in sensorimotor function, causing various disturbances in perception, spatial orientation, posture, gait, and eye-head coordination. Following long duration space flight, sensorimotor dysfunction would prevent or extend the time required to make an emergency egress from the vehicle; compromising crew safety and mission objectives. We are investigating two types of motor learning that may interact with each other and influence a crewmember's ability to re-adapt to Earth's gravity environment. In strategic learning, crewmembers make rapid modifications in their motor control strategy emphasizing error reduction. This type of learning may be critical during the first minutes and hours after landing. In adaptive learning, long-term plastic transformations occur, involving morphological changes and synaptic modification. In recent literature these two behavioral components have been associated with separate brain structures that control the execution of motor strategies: the strategic component was linked to the posterior parietal cortex and the adaptive component was linked to the cerebellum (Pisella, et al. 2004). The goal of this paper was to demonstrate the relative contributions of the strategic and adaptive components to the re-adaptation process in locomotor control after long duration space flight missions on the International Space Station (ISS). The Functional Mobility Test (FMT) was developed to assess crewmember s ability to ambulate postflight from an operational and functional perspective. Sixteen crewmembers were tested preflight (3 sessions) and postflight (days 1, 2, 4, 7, 25) following a long duration space flight (approx 6 months) on the ISS. We have further analyzed the FMT data to characterize strategic and adaptive components during the postflight readaptation period. Crewmembers walked at a preferred pace through an obstacle course set up on a base of 10 cm thick medium density foam (Sunmate Foam, Dynamic Systems, Inc., Leicester, NC). The 6.0m X 4.0m course consisted of several pylons made of foam; a Styrofoam barrier 46.0cm high that crewmembers stepped over; and a portal constructed of two Styrofoam blocks, each 31cm high, with a horizontal bar covered by foam and suspended from the ceiling which was adjusted to the height of the crewmember s shoulder. The portal required crewmembers to bend at the waist and step over a barrier simultaneously. All obstacles were lightweight, soft and easily knocked over. Crewmembers were instructed to walk through the course as quickly and as safely as possible without touching any of the objects on the course. This task was performed three times in the clockwise direction and three times in the counterclockwise direction that was randomly chosen. The dependent measures for each trial were: time to complete the course (seconds) and the number of obstacles touched or knocked down. For each crewmember, the time to complete each FMT trial from postflight days 1, 2, 4, 7 and 25 were further analyzed. A single logarithmic curve using a least squares calculation was fit through these data to produce a single comprehensive curve (macro). This macro curve composed of data spanning 25 days, illustrates the re-adaptive learning function over the longer time scale term. Additionally, logarithmic curves were fit to the 6 data trials within each individual post flight test day to produce 5 separate daily curves. These micro curves, produced from data obtained over the course of minutes, illustrates the strategic learning function exhibited over a relative shorter time scale. The macro curve for all subjects exhibited adaptive motor learning patterns over the 25 day period. Howev, 9/16 crewmembers exhibited significant strategic motor learning patterns in their micro curves, as defined by m > 1 in the equation of the line y=m*LN(x) +b. These data indicate that postflight recovery in locomotor function involves both strategic and adaptive mechanisms. Future countermeasures will be designed to enhance both recovery processes.
Spontaneous neuronal activity as a self-organized critical phenomenon
NASA Astrophysics Data System (ADS)
de Arcangelis, L.; Herrmann, H. J.
2013-01-01
Neuronal avalanches are a novel mode of activity in neuronal networks, experimentally found in vitro and in vivo, and exhibit a robust critical behaviour. Avalanche activity can be modelled within the self-organized criticality framework, including threshold firing, refractory period and activity-dependent synaptic plasticity. The size and duration distributions confirm that the system acts in a critical state, whose scaling behaviour is very robust. Next, we discuss the temporal organization of neuronal avalanches. This is given by the alternation between states of high and low activity, named up and down states, leading to a balance between excitation and inhibition controlled by a single parameter. During these periods both the single neuron state and the network excitability level, keeping memory of past activity, are tuned by homeostatic mechanisms. Finally, we verify if a system with no characteristic response can ever learn in a controlled and reproducible way. Learning in the model occurs via plastic adaptation of synaptic strengths by a non-uniform negative feedback mechanism. Learning is a truly collective process and the learning dynamics exhibits universal features. Even complex rules can be learned provided that the plastic adaptation is sufficiently slow.
Intelligent control of an IPMC actuated manipulator using emotional learning-based controller
NASA Astrophysics Data System (ADS)
Shariati, Azadeh; Meghdari, Ali; Shariati, Parham
2008-08-01
In this research an intelligent emotional learning controller, Takagi- Sugeno- Kang (TSK) is applied to govern the dynamics of a novel Ionic-Polymer Metal Composite (IPMC) actuated manipulator. Ionic-Polymer Metal Composites are active actuators that show very large deformation in existence of low applied voltage. In this research, a new IPMC actuator is considered and applied to a 2-dof miniature manipulator. This manipulator is designed for miniature tasks. The control system consists of a set of neurofuzzy controller whose parameters are adapted according to the emotional learning rules, and a critic with task to assess the present situation resulted from the applied control action in terms of satisfactory achievement of the control goals and provides the emotional signal (the stress). The controller modifies its characteristics so that the critic's stress decreased.
Satterfield, Brieann C; Hinson, John M; Whitney, Paul; Schmidt, Michelle A; Wisor, Jonathan P; Van Dongen, Hans P A
2018-02-01
Adaptive decision making is profoundly impaired by total sleep deprivation (TSD). This suggests that TSD impacts fronto-striatal pathways involved in cognitive control, where dopamine is a key neuromodulator. In the prefrontal cortex (PFC), dopamine is catabolized by the enzyme catechol-O-methyltransferase (COMT). A functional polymorphism (Val158Met) influences COMT's enzymatic activity, resulting in markedly different levels of prefrontal dopamine. We investigated the effect of this polymorphism on adaptive decision making during TSD. Sixty-six healthy young adults participated in one of two in-laboratory studies. After a baseline day, subjects were randomized to either a TSD group (n = 32) with 38 h or 62 h of extended wakefulness or a well-rested control group (n = 34) with 10 h nighttime sleep opportunities. Subjects performed a go/no-go reversal learning (GNGr) task at well-rested baseline and again during TSD or equivalent control. During the task, subjects were required to learn stimulus-response relationships from accuracy feedback. The stimulus-response relationships were reversed halfway through the task, which required subjects to learn the new stimulus-response relationships from accuracy feedback. Performance on the GNGr task was quantified by discriminability (d') between go and no-go stimuli before and after the stimulus-response reversal. GNGr performance did not differ between COMT genotypes when subjects were well-rested. However, TSD exposed a significant vulnerability to adaptive decision making impairment in subjects with the Val allele. Our results indicate that sleep deprivation degrades cognitive control through a fronto-striatal, dopaminergic mechanism. Copyright © 2017 Elsevier Ltd. All rights reserved.
The Impact of an Automated Learning Component against a Traditional Lecturing Environment
ERIC Educational Resources Information Center
Maycock, Keith W.; Keating, J. G.
2017-01-01
This experimental study investigates the effect on the examination performance of a cohort of first-year undergraduate learners undertaking a Unified Modelling Language (UML) course using an adaptive learning system against a control group of learners undertaking the same UML course through a traditional lecturing environment. The adaptive…
Generating Adaptive Behaviour within a Memory-Prediction Framework
Rawlinson, David; Kowadlo, Gideon
2012-01-01
The Memory-Prediction Framework (MPF) and its Hierarchical-Temporal Memory implementation (HTM) have been widely applied to unsupervised learning problems, for both classification and prediction. To date, there has been no attempt to incorporate MPF/HTM in reinforcement learning or other adaptive systems; that is, to use knowledge embodied within the hierarchy to control a system, or to generate behaviour for an agent. This problem is interesting because the human neocortex is believed to play a vital role in the generation of behaviour, and the MPF is a model of the human neocortex. We propose some simple and biologically-plausible enhancements to the Memory-Prediction Framework. These cause it to explore and interact with an external world, while trying to maximize a continuous, time-varying reward function. All behaviour is generated and controlled within the MPF hierarchy. The hierarchy develops from a random initial configuration by interaction with the world and reinforcement learning only. Among other demonstrations, we show that a 2-node hierarchy can learn to successfully play “rocks, paper, scissors” against a predictable opponent. PMID:22272231
Study on application of adaptive fuzzy control and neural network in the automatic leveling system
NASA Astrophysics Data System (ADS)
Xu, Xiping; Zhao, Zizhao; Lan, Weiyong; Sha, Lei; Qian, Cheng
2015-04-01
This paper discusses the adaptive fuzzy control and neural network BP algorithm in large flat automatic leveling control system application. The purpose is to develop a measurement system with a flat quick leveling, Make the installation on the leveling system of measurement with tablet, to be able to achieve a level in precision measurement work quickly, improve the efficiency of the precision measurement. This paper focuses on the automatic leveling system analysis based on fuzzy controller, Use of the method of combining fuzzy controller and BP neural network, using BP algorithm improve the experience rules .Construct an adaptive fuzzy control system. Meanwhile the learning rate of the BP algorithm has also been run-rate adjusted to accelerate convergence. The simulation results show that the proposed control method can effectively improve the leveling precision of automatic leveling system and shorten the time of leveling.
Neger, Thordis M.; Rietveld, Toni; Janse, Esther
2014-01-01
Within a few sentences, listeners learn to understand severely degraded speech such as noise-vocoded speech. However, individuals vary in the amount of such perceptual learning and it is unclear what underlies these differences. The present study investigates whether perceptual learning in speech relates to statistical learning, as sensitivity to probabilistic information may aid identification of relevant cues in novel speech input. If statistical learning and perceptual learning (partly) draw on the same general mechanisms, then statistical learning in a non-auditory modality using non-linguistic sequences should predict adaptation to degraded speech. In the present study, 73 older adults (aged over 60 years) and 60 younger adults (aged between 18 and 30 years) performed a visual artificial grammar learning task and were presented with 60 meaningful noise-vocoded sentences in an auditory recall task. Within age groups, sentence recognition performance over exposure was analyzed as a function of statistical learning performance, and other variables that may predict learning (i.e., hearing, vocabulary, attention switching control, working memory, and processing speed). Younger and older adults showed similar amounts of perceptual learning, but only younger adults showed significant statistical learning. In older adults, improvement in understanding noise-vocoded speech was constrained by age. In younger adults, amount of adaptation was associated with lexical knowledge and with statistical learning ability. Thus, individual differences in general cognitive abilities explain listeners' variability in adapting to noise-vocoded speech. Results suggest that perceptual and statistical learning share mechanisms of implicit regularity detection, but that the ability to detect statistical regularities is impaired in older adults if visual sequences are presented quickly. PMID:25225475
Neger, Thordis M; Rietveld, Toni; Janse, Esther
2014-01-01
Within a few sentences, listeners learn to understand severely degraded speech such as noise-vocoded speech. However, individuals vary in the amount of such perceptual learning and it is unclear what underlies these differences. The present study investigates whether perceptual learning in speech relates to statistical learning, as sensitivity to probabilistic information may aid identification of relevant cues in novel speech input. If statistical learning and perceptual learning (partly) draw on the same general mechanisms, then statistical learning in a non-auditory modality using non-linguistic sequences should predict adaptation to degraded speech. In the present study, 73 older adults (aged over 60 years) and 60 younger adults (aged between 18 and 30 years) performed a visual artificial grammar learning task and were presented with 60 meaningful noise-vocoded sentences in an auditory recall task. Within age groups, sentence recognition performance over exposure was analyzed as a function of statistical learning performance, and other variables that may predict learning (i.e., hearing, vocabulary, attention switching control, working memory, and processing speed). Younger and older adults showed similar amounts of perceptual learning, but only younger adults showed significant statistical learning. In older adults, improvement in understanding noise-vocoded speech was constrained by age. In younger adults, amount of adaptation was associated with lexical knowledge and with statistical learning ability. Thus, individual differences in general cognitive abilities explain listeners' variability in adapting to noise-vocoded speech. Results suggest that perceptual and statistical learning share mechanisms of implicit regularity detection, but that the ability to detect statistical regularities is impaired in older adults if visual sequences are presented quickly.
Learning feedback and feedforward control in a mirror-reversed visual environment.
Kasuga, Shoko; Telgen, Sebastian; Ushiba, Junichi; Nozaki, Daichi; Diedrichsen, Jörn
2015-10-01
When we learn a novel task, the motor system needs to acquire both feedforward and feedback control. Currently, little is known about how the learning of these two mechanisms relate to each other. In the present study, we tested whether feedforward and feedback control need to be learned separately, or whether they are learned as common mechanism when a new control policy is acquired. Participants were trained to reach to two lateral and one central target in an environment with mirror (left-right)-reversed visual feedback. One group was allowed to make online movement corrections, whereas the other group only received visual information after the end of the movement. Learning of feedforward control was assessed by measuring the accuracy of the initial movement direction to lateral targets. Feedback control was measured in the responses to sudden visual perturbations of the cursor when reaching to the central target. Although feedforward control improved in both groups, it was significantly better when online corrections were not allowed. In contrast, feedback control only adaptively changed in participants who received online feedback and remained unchanged in the group without online corrections. Our findings suggest that when a new control policy is acquired, feedforward and feedback control are learned separately, and that there may be a trade-off in learning between feedback and feedforward controllers. Copyright © 2015 the American Physiological Society.
Learning feedback and feedforward control in a mirror-reversed visual environment
Kasuga, Shoko; Telgen, Sebastian; Ushiba, Junichi; Nozaki, Daichi
2015-01-01
When we learn a novel task, the motor system needs to acquire both feedforward and feedback control. Currently, little is known about how the learning of these two mechanisms relate to each other. In the present study, we tested whether feedforward and feedback control need to be learned separately, or whether they are learned as common mechanism when a new control policy is acquired. Participants were trained to reach to two lateral and one central target in an environment with mirror (left-right)-reversed visual feedback. One group was allowed to make online movement corrections, whereas the other group only received visual information after the end of the movement. Learning of feedforward control was assessed by measuring the accuracy of the initial movement direction to lateral targets. Feedback control was measured in the responses to sudden visual perturbations of the cursor when reaching to the central target. Although feedforward control improved in both groups, it was significantly better when online corrections were not allowed. In contrast, feedback control only adaptively changed in participants who received online feedback and remained unchanged in the group without online corrections. Our findings suggest that when a new control policy is acquired, feedforward and feedback control are learned separately, and that there may be a trade-off in learning between feedback and feedforward controllers. PMID:26245313
An integrated utility-based model of conflict evaluation and resolution in the Stroop task.
Chuderski, Adam; Smolen, Tomasz
2016-04-01
Cognitive control allows humans to direct and coordinate their thoughts and actions in a flexible way, in order to reach internal goals regardless of interference and distraction. The hallmark test used to examine cognitive control is the Stroop task, which elicits both the weakly learned but goal-relevant and the strongly learned but goal-irrelevant response tendencies, and requires people to follow the former while ignoring the latter. After reviewing the existing computational models of cognitive control in the Stroop task, its novel, integrated utility-based model is proposed. The model uses 3 crucial control mechanisms: response utility reinforcement learning, utility-based conflict evaluation using the Festinger formula for assessing the conflict level, and top-down adaptation of response utility in service of conflict resolution. Their complex, dynamic interaction led to replication of 18 experimental effects, being the largest data set explained to date by 1 Stroop model. The simulations cover the basic congruency effects (including the response latency distributions), performance dynamics and adaptation (including EEG indices of conflict), as well as the effects resulting from manipulations applied to stimulation and responding, which are yielded by the extant Stroop literature. (c) 2016 APA, all rights reserved).
Impedance learning for robotic contact tasks using natural actor-critic algorithm.
Kim, Byungchan; Park, Jooyoung; Park, Shinsuk; Kang, Sungchul
2010-04-01
Compared with their robotic counterparts, humans excel at various tasks by using their ability to adaptively modulate arm impedance parameters. This ability allows us to successfully perform contact tasks even in uncertain environments. This paper considers a learning strategy of motor skill for robotic contact tasks based on a human motor control theory and machine learning schemes. Our robot learning method employs impedance control based on the equilibrium point control theory and reinforcement learning to determine the impedance parameters for contact tasks. A recursive least-square filter-based episodic natural actor-critic algorithm is used to find the optimal impedance parameters. The effectiveness of the proposed method was tested through dynamic simulations of various contact tasks. The simulation results demonstrated that the proposed method optimizes the performance of the contact tasks in uncertain conditions of the environment.
The Neural Correlates of Implicit Sequence Learning in Schizophrenia
Marvel, Cherie L.; Turner, Beth M.; O’Leary, Daniel S.; Johnson, Hans J.; Pierson, Ronald K.; Boles Ponto, Laura L.; Andreasen, Nancy C.
2009-01-01
Twenty-seven schizophrenia spectrum patients and 25 healthy controls performed a probabilistic version of the serial reaction time task (SRT) that included sequence trials embedded within random trials. Patients showed diminished, yet measurable, sequence learning. Postexperimental analyses revealed that a group of patients performed above chance when generating short spans of the sequence. This high-generation group showed SRT learning that was similar in magnitude to that of controls. Their learning was evident from the very 1st block; however, unlike controls, learning did not develop further with continued testing. A subset of 12 patients and 11 controls performed the SRT in conjunction with positron emission tomography. High-generation performance, which corresponded to SRT learning in patients, correlated to activity in the premotor cortex and parahippocampus. These areas have been associated with stimulus-driven visuospatial processing. Taken together, these results suggest that a subset of patients who showed moderate success on the SRT used an explicit stimulus-driven strategy to process the sequential stimuli. This adaptive strategy facilitated sequence learning but may have interfered with conventional implicit learning of the overall stimulus pattern. PMID:17983290
The Study and Design of Adaptive Learning System Based on Fuzzy Set Theory
NASA Astrophysics Data System (ADS)
Jia, Bing; Zhong, Shaochun; Zheng, Tianyang; Liu, Zhiyong
Adaptive learning is an effective way to improve the learning outcomes, that is, the selection of learning content and presentation should be adapted to each learner's learning context, learning levels and learning ability. Adaptive Learning System (ALS) can provide effective support for adaptive learning. This paper proposes a new ALS based on fuzzy set theory. It can effectively estimate the learner's knowledge level by test according to learner's target. Then take the factors of learner's cognitive ability and preference into consideration to achieve self-organization and push plan of knowledge. This paper focuses on the design and implementation of domain model and user model in ALS. Experiments confirmed that the system providing adaptive content can effectively help learners to memory the content and improve their comprehension.
An Adaptive Scaffolding E-Learning System for Middle School Students' Physics Learning
ERIC Educational Resources Information Center
Chen, Ching-Huei
2014-01-01
This study presents a framework that utilizes cognitive and motivational aspects of learning to design an adaptive scaffolding e-learning system. It addresses scaffolding processes and conditions for designing adaptive scaffolds. The features and effectiveness of this adaptive scaffolding e-learning system are discussed and evaluated. An…
Investigating the Effect of an Adaptive Learning Intervention on Students' Learning
ERIC Educational Resources Information Center
Liu, Min; McKelroy, Emily; Corliss, Stephanie B.; Carrigan, Jamison
2017-01-01
Educators agree on the benefits of adaptive learning, but evidence-based research remains limited as the field of adaptive learning is still evolving within higher education. In this study, we investigated the impact of an adaptive learning intervention to provide remedial instruction in biology, chemistry, math, and information literacy to…
A Model for an Adaptive e-Learning Hypermedia System
ERIC Educational Resources Information Center
Mahnane, Lamia; Tayeb, Laskri Mohamed; Trigano, Philippe
2013-01-01
Recent years have shown increasing awareness for the importance of adaptivity in e-learning. Since the learning style of each learner is different. Adaptive e-learning hypermedia system (AEHS) must fit different learner's needs. A number of AEHS have been developed to support learning styles as a source for adaptation. However, these systems…
Conflict-driven adaptive control is enhanced by integral negative emotion on a short time scale.
Yang, Qian; Pourtois, Gilles
2018-02-05
Negative emotion influences cognitive control, and more specifically conflict adaptation. However, discrepant results have often been reported in the literature. In this study, we broke down negative emotion into integral and incidental components using a modern motivation-based framework, and assessed whether the former could change conflict adaptation. In the first experiment, we manipulated the duration of the inter-trial-interval (ITI) to assess the actual time-scale of this effect. Integral negative emotion was induced by using loss-related feedback contingent on task performance, and measured at the subjective and physiological levels. Results showed that conflict-driven adaptive control was enhanced when integral negative emotion was elicited, compared to a control condition without changes in defensive motivation. Importantly, this effect was only found when a short, as opposed to long ITI was used, suggesting that it had a short time scale. In the second experiment, we controlled for effects of feature repetition and contingency learning, and replicated an enhanced conflict adaptation effect when integral negative emotion was elicited and a short ITI was used. We interpret these new results against a standard cognitive control framework assuming that integral negative emotion amplifies specific control signals transiently, and in turn enhances conflict adaptation.
ICPL: Intelligent Cooperative Planning and Learning for Multi-agent Systems
2012-02-29
objective was to develop a new planning approach for teams!of multiple UAVs that tightly integrates learning and cooperative!control algorithms at... algorithms at multiple levels of the planning architecture. The research results enabled a team of mobile agents to learn to adapt and react to uncertainty in...expressive representation that incorporates feature conjunctions. Our algorithm is simple to implement, fast to execute, and can be combined with any
Towards Machine Learning of Motor Skills
NASA Astrophysics Data System (ADS)
Peters, Jan; Schaal, Stefan; Schölkopf, Bernhard
Autonomous robots that can adapt to novel situations has been a long standing vision of robotics, artificial intelligence, and cognitive sciences. Early approaches to this goal during the heydays of artificial intelligence research in the late 1980s, however, made it clear that an approach purely based on reasoning or human insights would not be able to model all the perceptuomotor tasks that a robot should fulfill. Instead, new hope was put in the growing wake of machine learning that promised fully adaptive control algorithms which learn both by observation and trial-and-error. However, to date, learning techniques have yet to fulfill this promise as only few methods manage to scale into the high-dimensional domains of manipulator robotics, or even the new upcoming trend of humanoid robotics, and usually scaling was only achieved in precisely pre-structured domains. In this paper, we investigate the ingredients for a general approach to motor skill learning in order to get one step closer towards human-like performance. For doing so, we study two major components for such an approach, i.e., firstly, a theoretically well-founded general approach to representing the required control structures for task representation and execution and, secondly, appropriate learning algorithms which can be applied in this setting.
NASA Astrophysics Data System (ADS)
Yang, Xiong; Liu, Derong; Wang, Ding
2014-03-01
In this paper, an adaptive reinforcement learning-based solution is developed for the infinite-horizon optimal control problem of constrained-input continuous-time nonlinear systems in the presence of nonlinearities with unknown structures. Two different types of neural networks (NNs) are employed to approximate the Hamilton-Jacobi-Bellman equation. That is, an recurrent NN is constructed to identify the unknown dynamical system, and two feedforward NNs are used as the actor and the critic to approximate the optimal control and the optimal cost, respectively. Based on this framework, the action NN and the critic NN are tuned simultaneously, without the requirement for the knowledge of system drift dynamics. Moreover, by using Lyapunov's direct method, the weights of the action NN and the critic NN are guaranteed to be uniformly ultimately bounded, while keeping the closed-loop system stable. To demonstrate the effectiveness of the present approach, simulation results are illustrated.
Towards Validation of an Adaptive Flight Control Simulation Using Statistical Emulation
NASA Technical Reports Server (NTRS)
He, Yuning; Lee, Herbert K. H.; Davies, Misty D.
2012-01-01
Traditional validation of flight control systems is based primarily upon empirical testing. Empirical testing is sufficient for simple systems in which a.) the behavior is approximately linear and b.) humans are in-the-loop and responsible for off-nominal flight regimes. A different possible concept of operation is to use adaptive flight control systems with online learning neural networks (OLNNs) in combination with a human pilot for off-nominal flight behavior (such as when a plane has been damaged). Validating these systems is difficult because the controller is changing during the flight in a nonlinear way, and because the pilot and the control system have the potential to co-adapt in adverse ways traditional empirical methods are unlikely to provide any guarantees in this case. Additionally, the time it takes to find unsafe regions within the flight envelope using empirical testing means that the time between adaptive controller design iterations is large. This paper describes a new concept for validating adaptive control systems using methods based on Bayesian statistics. This validation framework allows the analyst to build nonlinear models with modal behavior, and to have an uncertainty estimate for the difference between the behaviors of the model and system under test.
Butcher, Peter A; Ivry, Richard B; Kuo, Sheng-Han; Rydz, David; Krakauer, John W; Taylor, Jordan A
2017-09-01
Individuals with damage to the cerebellum perform poorly in sensorimotor adaptation paradigms. This deficit has been attributed to impairment in sensory prediction error-based updating of an internal forward model, a form of implicit learning. These individuals can, however, successfully counter a perturbation when instructed with an explicit aiming strategy. This successful use of an instructed aiming strategy presents a paradox: In adaptation tasks, why do individuals with cerebellar damage not come up with an aiming solution on their own to compensate for their implicit learning deficit? To explore this question, we employed a variant of a visuomotor rotation task in which, before executing a movement on each trial, the participants verbally reported their intended aiming location. Compared with healthy control participants, participants with spinocerebellar ataxia displayed impairments in both implicit learning and aiming. This was observed when the visuomotor rotation was introduced abruptly ( experiment 1 ) or gradually ( experiment 2 ). This dual deficit does not appear to be related to the increased movement variance associated with ataxia: Healthy undergraduates showed little change in implicit learning or aiming when their movement feedback was artificially manipulated to produce similar levels of variability ( experiment 3 ). Taken together the results indicate that a consequence of cerebellar dysfunction is not only impaired sensory prediction error-based learning but also a difficulty in developing and/or maintaining an aiming solution in response to a visuomotor perturbation. We suggest that this dual deficit can be explained by the cerebellum forming part of a network that learns and maintains action-outcome associations across trials. NEW & NOTEWORTHY Individuals with cerebellar pathology are impaired in sensorimotor adaptation. This deficit has been attributed to an impairment in error-based learning, specifically, from a deficit in using sensory prediction errors to update an internal model. Here we show that these individuals also have difficulty in discovering an aiming solution to overcome their adaptation deficit, suggesting a new role for the cerebellum in sensorimotor adaptation tasks. Copyright © 2017 the American Physiological Society.
Silverman, Arielle M; Pitonyak, Jennifer S; Nelson, Ian K; Matsuda, Patricia N; Kartin, Deborah; Molton, Ivan R
2018-05-01
To develop and test a novel impairment simulation activity to teach beginning rehabilitation students how people adapt to physical impairments. Masters of Occupational Therapy students (n = 14) and Doctor of Physical Therapy students (n = 18) completed the study during the first month of their program. Students were randomized to the experimental or control learning activity. Experimental students learned to perform simple tasks while simulating paraplegia and hemiplegia. Control students viewed videos of others completing tasks with these impairments. Before and after the learning activities, all students estimated average self-perceived health, life satisfaction, and depression ratings among people with paraplegia and hemiplegia. Experimental students increased their estimates of self-perceived health, and decreased their estimates of depression rates, among people with paraplegia and hemiplegia after the learning activity. The control activity had no effect on these estimates. Impairment simulation can be an effective way to teach rehabilitation students about the adaptations that people make to physical impairments. Positive impairment simulations should allow students to experience success in completing activities of daily living with impairments. Impairment simulation is complementary to other pedagogical methods, such as simulated clinical encounters using standardized patients. Implication of Rehabilitation It is important for rehabilitation students to learn how people live well with disabilities. Impairment simulations can improve students' assessments of quality of life with disabilities. To be beneficial, impairment simulations must include guided exposure to effective methods for completing daily tasks with disabilities.
Liu, Zongcheng; Dong, Xinmin; Xue, Jianping; Li, Hongbo; Chen, Yong
2016-09-01
This brief addresses the adaptive control problem for a class of pure-feedback systems with nonaffine functions possibly being nondifferentiable. Without using the mean value theorem, the difficulty of the control design for pure-feedback systems is overcome by modeling the nonaffine functions appropriately. With the help of neural network approximators, an adaptive neural controller is developed by combining the dynamic surface control (DSC) and minimal learning parameter (MLP) techniques. The key features of our approach are that, first, the restrictive assumptions on the partial derivative of nonaffine functions are removed, second, the DSC technique is used to avoid "the explosion of complexity" in the backstepping design, and the number of adaptive parameters is reduced significantly using the MLP technique, third, smooth robust compensators are employed to circumvent the influences of approximation errors and disturbances. Furthermore, it is proved that all the signals in the closed-loop system are semiglobal uniformly ultimately bounded. Finally, the simulation results are provided to demonstrate the effectiveness of the designed method.
Should the parameters of a BCI translation algorithm be continually adapted?
McFarland, Dennis J; Sarnacki, William A; Wolpaw, Jonathan R
2011-07-15
People with or without motor disabilities can learn to control sensorimotor rhythms (SMRs) recorded from the scalp to move a computer cursor in one or more dimensions or can use the P300 event-related potential as a control signal to make discrete selections. Data collected from individuals using an SMR-based or P300-based BCI were evaluated offline to estimate the impact on performance of continually adapting the parameters of the translation algorithm during BCI operation. The performance of the SMR-based BCI was enhanced by adaptive updating of the feature weights or adaptive normalization of the features. In contrast, P300 performance did not benefit from either of these procedures. Copyright © 2011 Elsevier B.V. All rights reserved.
Adaptive Modulation for DFIG and STATCOM With High-Voltage Direct Current Transmission.
Tang, Yufei; He, Haibo; Ni, Zhen; Wen, Jinyu; Huang, Tingwen
2016-08-01
This paper develops an adaptive modulation approach for power system control based on the approximate/adaptive dynamic programming method, namely, the goal representation heuristic dynamic programming (GrHDP). In particular, we focus on the fault recovery problem of a doubly fed induction generator (DFIG)-based wind farm and a static synchronous compensator (STATCOM) with high-voltage direct current (HVDC) transmission. In this design, the online GrHDP-based controller provides three adaptive supplementary control signals to the DFIG controller, STATCOM controller, and HVDC rectifier controller, respectively. The mechanism is to observe the system states and their derivatives and then provides supplementary control to the plant according to the utility function. With the GrHDP design, the controller can adaptively develop an internal goal representation signal according to the observed power system states, therefore, to achieve more effective learning and modulating. Our control approach is validated on a wind power integrated benchmark system with two areas connected by HVDC transmission lines. Compared with the classical direct HDP and proportional integral control, our GrHDP approach demonstrates the improved transient stability under system faults. Moreover, experiments under different system operating conditions with signal transmission delays are also carried out to further verify the effectiveness and robustness of the proposed approach.
ERIC Educational Resources Information Center
Nelson, Patricia L.
1986-01-01
Suggests studying New York's wetlands, both in the classroom and in the field, to illustrate ecological concepts of diversity, succession, and adaptation and to learn about their importance in controlling flooding, erosion, and pollution. (NEC)
NASA Astrophysics Data System (ADS)
Chen, Jiaxi; Li, Junmin
2018-02-01
In this paper, we investigate the perfect consensus problem for second-order linearly parameterised multi-agent systems (MAS) with imprecise communication topology structure. Takagi-Sugeno (T-S) fuzzy models are presented to describe the imprecise communication topology structure of leader-following MAS, and a distributed adaptive iterative learning control protocol is proposed with the dynamic of leader unknown to any of the agent. The proposed protocol guarantees that the follower agents can track the leader perfectly on [0,T] for the consensus problem. Under alignment condition, a sufficient condition of the consensus for closed-loop MAS is given based on Lyapunov stability theory. Finally, a numerical example and a multiple pendulum system are given to illustrate the effectiveness of the proposed algorithm.
Dai, Yun; Zhao, Lina; Xiao, Fei; Zhao, Haoxin; Bao, Hua; Zhou, Hong; Zhou, Yifeng; Zhang, Yudong
2015-02-10
An adaptive optics visual simulation combined with a perceptual learning (PL) system based on a 35-element bimorph deformable mirror (DM) was established. The larger stroke and smaller size of the bimorph DM made the system have larger aberration correction or superposition ability and be more compact. By simply modifying the control matrix or the reference matrix, select correction or superposition of aberrations was realized in real time similar to a conventional adaptive optics closed-loop correction. PL function was first integrated in addition to conventional adaptive optics visual simulation. PL training undertaken with high-order aberrations correction obviously improved the visual function of adult anisometropic amblyopia. The preliminary application of high-order aberrations correction with PL training on amblyopia treatment was being validated with a large scale population, which might have great potential in amblyopia treatment and visual performance maintenance.
Adaptive Learning Resources Sequencing in Educational Hypermedia Systems
ERIC Educational Resources Information Center
Karampiperis, Pythagoras; Sampson, Demetrios
2005-01-01
Adaptive learning resources selection and sequencing is recognized as among the most interesting research questions in adaptive educational hypermedia systems (AEHS). In order to adaptively select and sequence learning resources in AEHS, the definition of adaptation rules contained in the Adaptation Model, is required. Although, some efforts have…
Learning to push and learning to move: the adaptive control of contact forces
Casadio, Maura; Pressman, Assaf; Mussa-Ivaldi, Ferdinando A.
2015-01-01
To be successful at manipulating objects one needs to apply simultaneously well controlled movements and contact forces. We present a computational theory of how the brain may successfully generate a vast spectrum of interactive behaviors by combining two independent processes. One process is competent to control movements in free space and the other is competent to control contact forces against rigid constraints. Free space and rigid constraints are singularities at the boundaries of a continuum of mechanical impedance. Within this continuum, forces and motions occur in “compatible pairs” connected by the equations of Newtonian dynamics. The force applied to an object determines its motion. Conversely, inverse dynamics determine a unique force trajectory from a movement trajectory. In this perspective, we describe motor learning as a process leading to the discovery of compatible force/motion pairs. The learned compatible pairs constitute a local representation of the environment's mechanics. Experiments on force field adaptation have already provided us with evidence that the brain is able to predict and compensate the forces encountered when one is attempting to generate a motion. Here, we tested the theory in the dual case, i.e., when one attempts at applying a desired contact force against a simulated rigid surface. If the surface becomes unexpectedly compliant, the contact point moves as a function of the applied force and this causes the applied force to deviate from its desired value. We found that, through repeated attempts at generating the desired contact force, subjects discovered the unique compatible hand motion. When, after learning, the rigid contact was unexpectedly restored, subjects displayed after effects of learning, consistent with the concurrent operation of a motion control system and a force control system. Together, theory and experiment support a new and broader view of modularity in the coordinated control of forces and motions. PMID:26594163
Dong, Lu; Zhong, Xiangnan; Sun, Changyin; He, Haibo
2017-07-01
This paper presents the design of a novel adaptive event-triggered control method based on the heuristic dynamic programming (HDP) technique for nonlinear discrete-time systems with unknown system dynamics. In the proposed method, the control law is only updated when the event-triggered condition is violated. Compared with the periodic updates in the traditional adaptive dynamic programming (ADP) control, the proposed method can reduce the computation and transmission cost. An actor-critic framework is used to learn the optimal event-triggered control law and the value function. Furthermore, a model network is designed to estimate the system state vector. The main contribution of this paper is to design a new trigger threshold for discrete-time systems. A detailed Lyapunov stability analysis shows that our proposed event-triggered controller can asymptotically stabilize the discrete-time systems. Finally, we test our method on two different discrete-time systems, and the simulation results are included.
Quadrotor Control in the Presence of Unknown Mass Properties
NASA Astrophysics Data System (ADS)
Duivenvoorden, Rikky Ricardo Petrus Rufino
Quadrotor UAVs are popular due to their mechanical simplicity, as well as their capability to hover and vertically take-off and land. As applications diversify, quadrotors are increasingly required to operate under unknown mass properties, for example as a multirole sensor platform or for package delivery operations. The work presented here consists of the derivation of a generalized quadrotor dynamic model without the typical simplifying assumptions on the first and second moments of mass. The maximum payload capacity of a quadrotor in hover, and the observability of the unknown mass properties are discussed. A brief introduction of L1 adaptive control is provided, and three different L 1 adaptive controllers were designed for the Parrot AR.Drone quadrotor. Their tracking and disturbance rejection performance was compared to the baseline nonlinear controller in experiments. Finally, the results of the combination of L1 adaptive control with iterative learning control are presented, showing high performance trajectory tracking under uncertainty.
Rouhollahi, Korosh; Emadi Andani, Mehran; Karbassi, Seyed Mahdi; Izadi, Iman
2017-02-01
Deep brain stimulation (DBS) is an efficient therapy to control movement disorders of Parkinson's tremor. Stimulation of one area of basal ganglia (BG) by DBS with no feedback is the prevalent opinion. Reduction of additional stimulatory signal delivered to the brain is the advantage of using feedback. This results in reduction of side effects caused by the excessive stimulation intensity. In fact, the stimulatory intensity of controllers is decreased proportional to reduction of hand tremor. The objective of this study is to design a new controller structure to decrease three indicators: (i) the hand tremor; (ii) the level of delivered stimulation in disease condition; and (iii) the ratio of the level of delivered stimulation in health condition to disease condition. For this purpose, the authors offer a new closed-loop control structure to stimulate two areas of BG simultaneously. One area (STN: subthalamic nucleus) is stimulated by an adaptive controller with feedback error learning. The other area (GPi: globus pallidus internal) is stimulated by a partial state feedback (PSF) controller. Considering the three indicators, the results show that, stimulating two areas simultaneously leads to better performance compared with stimulating one area only. It is shown that both PSF and adaptive controllers are robust regarding system parameter uncertainties. In addition, a method is proposed to update the parameters of the BG model in real time. As a result, the parameters of the controllers can be updated based on the new parameters of the BG model.
OPUS One: An Intelligent Adaptive Learning Environment Using Artificial Intelligence Support
NASA Astrophysics Data System (ADS)
Pedrazzoli, Attilio
2010-06-01
AI based Tutoring and Learning Path Adaptation are well known concepts in e-Learning scenarios today and increasingly applied in modern learning environments. In order to gain more flexibility and to enhance existing e-learning platforms, the OPUS One LMS Extension package will enable a generic Intelligent Tutored Adaptive Learning Environment, based on a holistic Multidimensional Instructional Design Model (PENTHA ID Model), allowing AI based tutoring and adaptation functionality to existing Web-based e-learning systems. Relying on "real time" adapted profiles, it allows content- / course authors to apply a dynamic course design, supporting tutored, collaborative sessions and activities, as suggested by modern pedagogy. The concept presented combines a personalized level of surveillance, learning activity- and learning path adaptation suggestions to ensure the students learning motivation and learning success. The OPUS One concept allows to implement an advanced tutoring approach combining "expert based" e-tutoring with the more "personal" human tutoring function. It supplies the "Human Tutor" with precise, extended course activity data and "adaptation" suggestions based on predefined subject matter rules. The concept architecture is modular allowing a personalized platform configuration.
ERIC Educational Resources Information Center
Chang, Yi-Hsing; Chen, Yen-Yi; Chen, Nian-Shing; Lu, You-Te; Fang, Rong-Jyue
2016-01-01
This study designs and implements an adaptive learning management system based on Felder and Silverman's Learning Style Model and the Mashup technology. In this system, Felder and Silverman's Learning Style model is used to assess students' learning styles, in order to provide adaptive learning to leverage learners' learning preferences.…
Masia, Lorenzo; Frascarelli, Flaminia; Morasso, Pietro; Di Rosa, Giuseppe; Petrarca, Maurizio; Castelli, Enrico; Cappa, Paolo
2011-05-21
It is known that healthy adults can quickly adapt to a novel dynamic environment, generated by a robotic manipulandum as a structured disturbing force field. We suggest that it may be of clinical interest to evaluate to which extent this kind of motor learning capability is impaired in children affected by cerebal palsy. We adapted the protocol already used with adults, which employs a velocity dependant viscous field, and compared the performance of a group of subjects affected by Cerebral Palsy (CP group, 7 subjects) with a Control group of unimpaired age-matched children. The protocol included a familiarization phase (FA), during which no force was applied, a force field adaptation phase (CF), and a wash-out phase (WO) in which the field was removed. During the CF phase the field was shut down in a number of randomly selected "catch" trials, which were used in order to evaluate the "learning index" for each single subject and the two groups. Lateral deviation, speed and acceleration peaks and average speed were evaluated for each trajectory; a directional analysis was performed in order to inspect the role of the limb's inertial anisotropy in the different experimental phases. During the FA phase the movements of the CP subjects were more curved, displaying greater and variable directional error; over the course of the CF phase both groups showed a decreasing trend in the lateral error and an after-effect at the beginning of the wash-out, but the CP group had a non significant adaptation rate and a lower learning index, suggesting that CP subjects have reduced ability to learn to compensate external force. Moreover, a directional analysis of trajectories confirms that the control group is able to better predict the force field by tuning the kinematic features of the movements along different directions in order to account for the inertial anisotropy of arm. Spatial abnormalities in children affected by cerebral palsy may be related not only to disturbance in motor control signals generating weakness and spasticity, but also to an inefficient control strategy which is not based on a robust knowledge of the dynamical features of their upper limb. This lack of information could be related to the congenital nature of the brain damage and may contribute to a better delineation of therapeutic intervention.
A Telescopic Binary Learning Machine for Training Neural Networks.
Brunato, Mauro; Battiti, Roberto
2017-03-01
This paper proposes a new algorithm based on multiscale stochastic local search with binary representation for training neural networks [binary learning machine (BLM)]. We study the effects of neighborhood evaluation strategies, the effect of the number of bits per weight and that of the maximum weight range used for mapping binary strings to real values. Following this preliminary investigation, we propose a telescopic multiscale version of local search, where the number of bits is increased in an adaptive manner, leading to a faster search and to local minima of better quality. An analysis related to adapting the number of bits in a dynamic way is presented. The control on the number of bits, which happens in a natural manner in the proposed method, is effective to increase the generalization performance. The learning dynamics are discussed and validated on a highly nonlinear artificial problem and on real-world tasks in many application domains; BLM is finally applied to a problem requiring either feedforward or recurrent architectures for feedback control.
Flatness-based embedded adaptive fuzzy control of turbocharged diesel engines
NASA Astrophysics Data System (ADS)
Rigatos, Gerasimos; Siano, Pierluigi; Arsie, Ivan
2014-10-01
In this paper nonlinear embedded control for turbocharged Diesel engines is developed with the use of Differential flatness theory and adaptive fuzzy control. It is shown that the dynamic model of the turbocharged Diesel engine is differentially flat and admits dynamic feedback linearization. It is also shown that the dynamic model can be written in the linear Brunovsky canonical form for which a state feedback controller can be easily designed. To compensate for modeling errors and external disturbances an adaptive fuzzy control scheme is implemanted making use of the transformed dynamical system of the diesel engine that is obtained through the application of differential flatness theory. Since only the system's output is measurable the complete state vector has to be reconstructed with the use of a state observer. It is shown that a suitable learning law can be defined for neuro-fuzzy approximators, which are part of the controller, so as to preserve the closed-loop system stability. With the use of Lyapunov stability analysis it is proven that the proposed observer-based adaptive fuzzy control scheme results in H∞ tracking performance.
Active Inference, homeostatic regulation and adaptive behavioural control.
Pezzulo, Giovanni; Rigoli, Francesco; Friston, Karl
2015-11-01
We review a theory of homeostatic regulation and adaptive behavioural control within the Active Inference framework. Our aim is to connect two research streams that are usually considered independently; namely, Active Inference and associative learning theories of animal behaviour. The former uses a probabilistic (Bayesian) formulation of perception and action, while the latter calls on multiple (Pavlovian, habitual, goal-directed) processes for homeostatic and behavioural control. We offer a synthesis these classical processes and cast them as successive hierarchical contextualisations of sensorimotor constructs, using the generative models that underpin Active Inference. This dissolves any apparent mechanistic distinction between the optimization processes that mediate classical control or learning. Furthermore, we generalize the scope of Active Inference by emphasizing interoceptive inference and homeostatic regulation. The ensuing homeostatic (or allostatic) perspective provides an intuitive explanation for how priors act as drives or goals to enslave action, and emphasises the embodied nature of inference. Copyright © 2015 The Authors. Published by Elsevier Ltd.. All rights reserved.
Adaptable, Personalised E-Learning Incorporating Learning Styles
ERIC Educational Resources Information Center
Peter, Sophie E.; Bacon, Elizabeth; Dastbaz, Mohammad
2010-01-01
Purpose: The purpose of this paper is to discuss how learning styles and theories are currently used within personalised adaptable e-learning adaptive systems. This paper then aims to describe the e-learning platform iLearn and how this platform is designed to incorporate learning styles as part of the personalisation offered by the system.…
Genetic algorithms for adaptive real-time control in space systems
NASA Technical Reports Server (NTRS)
Vanderzijp, J.; Choudry, A.
1988-01-01
Genetic Algorithms that are used for learning as one way to control the combinational explosion associated with the generation of new rules are discussed. The Genetic Algorithm approach tends to work best when it can be applied to a domain independent knowledge representation. Applications to real time control in space systems are discussed.
Lefumat, Hannah Z; Vercher, Jean-Louis; Miall, R Chris; Cole, Jonathan; Buloup, Frank; Bringoux, Lionel; Bourdin, Christophe; Sarlegna, Fabrice R
2015-11-01
Humans can remarkably adapt their motor behavior to novel environmental conditions, yet it remains unclear which factors enable us to transfer what we have learned with one limb to the other. Here we tested the hypothesis that interlimb transfer of sensorimotor adaptation is determined by environmental conditions but also by individual characteristics. We specifically examined the adaptation of unconstrained reaching movements to a novel Coriolis, velocity-dependent force field. Right-handed subjects sat at the center of a rotating platform and performed forward reaching movements with the upper limb toward flashed visual targets in prerotation, per-rotation (i.e., adaptation), and postrotation tests. Here only the dominant arm was used during adaptation and interlimb transfer was assessed by comparing performance of the nondominant arm before and after dominant-arm adaptation. Vision and no-vision conditions did not significantly influence interlimb transfer of trajectory adaptation, which on average was significant but limited. We uncovered a substantial heterogeneity of interlimb transfer across subjects and found that interlimb transfer can be qualitatively and quantitatively predicted for each healthy young individual. A classifier showed that in our study, interlimb transfer could be predicted based on the subject's task performance, most notably motor variability during learning, and his or her laterality quotient. Positive correlations suggested that variability of motor performance and lateralization of arm movement control facilitate interlimb transfer. We further show that these individual characteristics can predict the presence and the magnitude of interlimb transfer of left-handers. Overall, this study suggests that individual characteristics shape the way the nervous system can generalize motor learning. Copyright © 2015 the American Physiological Society.
Distinctive neural processes during learning in autism.
Schipul, Sarah E; Williams, Diane L; Keller, Timothy A; Minshew, Nancy J; Just, Marcel Adam
2012-04-01
This functional magnetic resonance imaging study compared the neural activation patterns of 18 high-functioning individuals with autism and 18 IQ-matched neurotypical control participants as they learned to perform a social judgment task. Participants learned to identify liars among pairs of computer-animated avatars uttering the same sentence but with different facial and vocal expressions, namely those that have previously been associated with lying versus truth-telling. Despite showing a behavioral learning effect similar to the control group, the autism group did not show the same pattern of decreased activation in cortical association areas as they learned the task. Furthermore, the autism group showed a significantly smaller increase in interregion synchronization of activation (functional connectivity) with learning than did the control group. Finally, the autism group had decreased structural connectivity as measured by corpus callosum size, and this measure was reliably related to functional connectivity measures. The findings suggest that cortical underconnectivity in autism may constrain the ability of the brain to rapidly adapt during learning.
NASA Astrophysics Data System (ADS)
Pini, Giovanni; Tuci, Elio
2008-06-01
In biology/psychology, the capability of natural organisms to learn from the observation/interaction with conspecifics is referred to as social learning. Roboticists have recently developed an interest in social learning, since it might represent an effective strategy to enhance the adaptivity of a team of autonomous robots. In this study, we show that a methodological approach based on artifcial neural networks shaped by evolutionary computation techniques can be successfully employed to synthesise the individual and social learning mechanisms for robots required to learn a desired action (i.e. phototaxis or antiphototaxis).
Improving Adaptive Learning Technology through the Use of Response Times
ERIC Educational Resources Information Center
Mettler, Everett; Massey, Christine M.; Kellman, Philip J.
2011-01-01
Adaptive learning techniques have typically scheduled practice using learners' accuracy and item presentation history. We describe an adaptive learning system (Adaptive Response Time Based Sequencing--ARTS) that uses both accuracy and response time (RT) as direct inputs into sequencing. Response times are used to assess learning strength and…
Hiremath, Shivayogi V; Chen, Weidong; Wang, Wei; Foldes, Stephen; Yang, Ying; Tyler-Kabara, Elizabeth C; Collinger, Jennifer L; Boninger, Michael L
2015-01-01
A brain-computer interface (BCI) system transforms neural activity into control signals for external devices in real time. A BCI user needs to learn to generate specific cortical activity patterns to control external devices effectively. We call this process BCI learning, and it often requires significant effort and time. Therefore, it is important to study this process and develop novel and efficient approaches to accelerate BCI learning. This article reviews major approaches that have been used for BCI learning, including computer-assisted learning, co-adaptive learning, operant conditioning, and sensory feedback. We focus on BCIs based on electrocorticography and intracortical microelectrode arrays for restoring motor function. This article also explores the possibility of brain modulation techniques in promoting BCI learning, such as electrical cortical stimulation, transcranial magnetic stimulation, and optogenetics. Furthermore, as proposed by recent BCI studies, we suggest that BCI learning is in many ways analogous to motor and cognitive skill learning, and therefore skill learning should be a useful metaphor to model BCI learning.
Development of an Adaptive Learning System with Two Sources of Personalization Information
ERIC Educational Resources Information Center
Tseng, J. C. R.; Chu, H. C.; Hwang, G. J.; Tsai, C. C.
2008-01-01
Previous research of adaptive learning mainly focused on improving student learning achievements based only on single-source of personalization information, such as learning style, cognitive style or learning achievement. In this paper, an innovative adaptive learning approach is proposed by basing upon two main sources of personalization…
C3Conflict a Simulation Environment for Studying Teamwork in Command and Control
2011-06-01
the Sciences (pp. 173- 217). Amsterdam/New York: Rodopi. Kolb , D. A. (1984). Experiential Learning – Experience as a source of learning and...increases dramatically when the students can see a replay and discuss their collaboration. Kolb has expressed a generally accepted model of experiential ... learning ( Kolb , 1998). The model can be adapted for research and team training performed with computer-based simulations (Granlund, 2008). The main
Multi-Objective Reinforcement Learning-based Deep Neural Networks for Cognitive Space Communications
NASA Technical Reports Server (NTRS)
Ferreria, Paulo; Paffenroth, Randy; Wyglinski, Alexander M.; Hackett, Timothy; Bilen, Sven; Reinhart, Richard; Mortensen, Dale
2017-01-01
Future communication subsystems of space exploration missions can potentially benefit from software-defined radios (SDRs) controlled by machine learning algorithms. In this paper, we propose a novel hybrid radio resource allocation management control algorithm that integrates multi-objective reinforcement learning and deep artificial neural networks. The objective is to efficiently manage communications system resources by monitoring performance functions with common dependent variables that result in conflicting goals. The uncertainty in the performance of thousands of different possible combinations of radio parameters makes the trade-off between exploration and exploitation in reinforcement learning (RL) much more challenging for future critical space-based missions. Thus, the system should spend as little time as possible on exploring actions, and whenever it explores an action, it should perform at acceptable levels most of the time. The proposed approach enables on-line learning by interactions with the environment and restricts poor resource allocation performance through virtual environment exploration. Improvements in the multiobjective performance can be achieved via transmitter parameter adaptation on a packet-basis, with poorly predicted performance promptly resulting in rejected decisions. Simulations presented in this work considered the DVB-S2 standard adaptive transmitter parameters and additional ones expected to be present in future adaptive radio systems. Performance results are provided by analysis of the proposed hybrid algorithm when operating across a satellite communication channel from Earth to GEO orbit during clear sky conditions. The proposed approach constitutes part of the core cognitive engine proof-of-concept to be delivered to the NASA Glenn Research Center SCaN Testbed located onboard the International Space Station.
Multi-Objective Reinforcement Learning-Based Deep Neural Networks for Cognitive Space Communications
NASA Technical Reports Server (NTRS)
Ferreria, Paulo Victor R.; Paffenroth, Randy; Wyglinski, Alexander M.; Hackett, Timothy M.; Bilen, Sven G.; Reinhart, Richard C.; Mortensen, Dale J.
2017-01-01
Future communication subsystems of space exploration missions can potentially benefit from software-defined radios (SDRs) controlled by machine learning algorithms. In this paper, we propose a novel hybrid radio resource allocation management control algorithm that integrates multi-objective reinforcement learning and deep artificial neural networks. The objective is to efficiently manage communications system resources by monitoring performance functions with common dependent variables that result in conflicting goals. The uncertainty in the performance of thousands of different possible combinations of radio parameters makes the trade-off between exploration and exploitation in reinforcement learning (RL) much more challenging for future critical space-based missions. Thus, the system should spend as little time as possible on exploring actions, and whenever it explores an action, it should perform at acceptable levels most of the time. The proposed approach enables on-line learning by interactions with the environment and restricts poor resource allocation performance through virtual environment exploration. Improvements in the multiobjective performance can be achieved via transmitter parameter adaptation on a packet-basis, with poorly predicted performance promptly resulting in rejected decisions. Simulations presented in this work considered the DVB-S2 standard adaptive transmitter parameters and additional ones expected to be present in future adaptive radio systems. Performance results are provided by analysis of the proposed hybrid algorithm when operating across a satellite communication channel from Earth to GEO orbit during clear sky conditions. The proposed approach constitutes part of the core cognitive engine proof-of-concept to be delivered to the NASA Glenn Research Center SCaN Testbed located onboard the International Space Station.
Wong, Vincent; Smith, Ariella J; Hawkins, Nicholas J; Kumar, Rakesh K; Young, Noel; Kyaw, Merribel; Velan, Gary M
2015-10-01
Diagnostic imaging is under-represented in medical curricula globally. Adaptive tutorials, online intelligent tutoring systems that provide a personalized learning experience, have the potential to bridge this gap. However, there is limited evidence of their effectiveness for learning about diagnostic imaging. We performed a randomized mixed methods crossover trial to determine the impact of adaptive tutorials on perceived engagement and understanding of the appropriate use and interpretation of common diagnostic imaging investigations. Although concurrently engaged in disparate blocks of study, 99 volunteer medical students (from years 1-4 of the 6-year program) were randomly allocated to one of two groups. In the first arm of the trial on chest X-rays, one group received access to an adaptive tutorial, whereas the other received links to an existing peer-reviewed Web resource. These two groups crossed over in the second arm of the trial, which focused on computed tomography scans of the head, chest, and abdomen. At the conclusion of each arm of the trial, both groups completed an examination-style assessment, comprising questions both related and unrelated to the topics covered by the relevant adaptive tutorial. Online questionnaires were used to evaluate student perceptions of both learning resources. In both arms of the trial, the group using adaptive tutorials obtained significantly higher assessment scores than controls. This was because of higher assessment scores by senior students in the adaptive tutorial group when answering questions related to topics covered in those tutorials. Furthermore, students indicated significantly better engagement with adaptive tutorials than the Web resource and rated the tutorials as a significantly more valuable tool for learning. Medical students overwhelmingly accept adaptive tutorials for diagnostic imaging. The tutorials significantly improve the understanding of diagnostic imaging by senior students. Crown Copyright © 2015. Published by Elsevier Inc. All rights reserved.
Clipping in neurocontrol by adaptive dynamic programming.
Fairbank, Michael; Prokhorov, Danil; Alonso, Eduardo
2014-10-01
In adaptive dynamic programming, neurocontrol, and reinforcement learning, the objective is for an agent to learn to choose actions so as to minimize a total cost function. In this paper, we show that when discretized time is used to model the motion of the agent, it can be very important to do clipping on the motion of the agent in the final time step of the trajectory. By clipping, we mean that the final time step of the trajectory is to be truncated such that the agent stops exactly at the first terminal state reached, and no distance further. We demonstrate that when clipping is omitted, learning performance can fail to reach the optimum, and when clipping is done properly, learning performance can improve significantly. The clipping problem we describe affects algorithms that use explicit derivatives of the model functions of the environment to calculate a learning gradient. These include backpropagation through time for control and methods based on dual heuristic programming. However, the clipping problem does not significantly affect methods based on heuristic dynamic programming, temporal differences learning, or policy-gradient learning algorithms.
Coesmans, Michael; Röder, Christian H.; Smit, Albertine E.; Koekkoek, Sebastiaan K.E.; De Zeeuw, Chris I.; Frens, Maarten A.; van der Geest, Josef N.
2014-01-01
Background The notion that cerebellar deficits may underlie clinical symptoms in people with schizophrenia is tested by evaluating 2 forms of cerebellar learning in patients with recent-onset schizophrenia. A potential medication effect is evaluated by including patients with or without antipsychotics. Methods We assessed saccadic eye movement adaptation and eyeblink conditioning in men with recent-onset schizophrenia who were taking antipsychotic medication or who were antipsychotic-free and in age-matched controls. Results We included 39 men with schizophrenia (10 who were taking clozapine, 16 who were taking haloperidol and 13 who were antipsychotic-free) and 29 controls in our study. All participants showed significant saccadic adaptation. Adaptation strength did not differ between healthy controls and men with schizophrenia. The speed of saccade adaptation, however, was significantly lower in men with schizophrenia. They showed a significantly lower increase in the number of conditioned eyeblink responses. Over all experiments, no consistent effects of medication were observed. These outcomes did not correlate with age, years of education, psychopathology or dose of anti psychotics. Limitations As patients were not randomized for treatment, an influence of confounding variables associated with medication status cannot be excluded. Individual patients also varied along the schizophrenia spectrum despite the relative homogeneity with respect to onset of illness and short usage of medication. Finally, the relatively small number of participants may have concealed effects as a result of insufficient statistical power. Conclusion We found several cerebellar learning deficits in men with schizophrenia that we cannot attribute to the use of antipsychotics. Although this finding, combined with the fact that deficits are already present in patients with recent-onset schizophrenia, could suggest that cerebellar impairments are a trait deficit in people with schizophrenia. This should be confirmed in longitudinal studies. PMID:24083457
MRSA model of learning and adaptation: a qualitative study among the general public
2012-01-01
Background More people in the US now die from Methicillin Resistant Staphylococcus aureus (MRSA) infections than from HIV/AIDS. Often acquired in healthcare facilities or during healthcare procedures, the extremely high incidence of MRSA infections and the dangerously low levels of literacy regarding antibiotic resistance in the general public are on a collision course. Traditional medical approaches to infection control and the conventional attitude healthcare practitioners adopt toward public education are no longer adequate to avoid this collision. This study helps us understand how people acquire and process new information and then adapt behaviours based on learning. Methods Using constructivist theory, semi-structured face-to-face and phone interviews were conducted to gather pertinent data. This allowed participants to tell their stories so their experiences could deepen our understanding of this crucial health issue. Interview transcripts were analysed using grounded theory and sensitizing concepts. Results Our findings were classified into two main categories, each of which in turn included three subthemes. First, in the category of Learning, we identified how individuals used their Experiences with MRSA, to answer the questions: What was learned? and, How did learning occur? The second category, Adaptation gave us insights into Self-reliance, Reliance on others, and Reflections on the MRSA journey. Conclusions This study underscores the critical importance of educational programs for patients, and improved continuing education for healthcare providers. Five specific results of this study can reduce the vacuum that currently exists between the knowledge and information available to healthcare professionals, and how that information is conveyed to the public. These points include: 1) a common model of MRSA learning and adaptation; 2) the self-directed nature of adult learning; 3) the focus on general MRSA information, care and prevention, and antibiotic resistance; 4) the interconnected nature of adaptation; and, 5) the need for a consistent step by step plan to deal with MRSA provided at the time of diagnosis. PMID:22469420
Context-specific adjustment of cognitive control: Transfer of adaptive control sets.
Surrey, Caroline; Dreisbach, Gesine; Fischer, Rico
2017-11-01
Cognitive control protects processing of relevant information from interference by irrelevant information. The level of this processing selectivity can be flexibly adjusted to different control demands (e.g., frequency of conflict) associated with a certain context, leading to the formation of specific context-control associations. In the present study we investigated the robustness and transferability of the acquired context-control demands to new situations. In three experiments, we used a version of the context-specific proportion congruence (CSPC) paradigm, in which each context (e.g., location) is associated with a specific conflict frequency, determining high and low control demands. In a learning phase, associations between context and control demands were established. In a subsequent transfer block, stimulus-response mappings, whole task sets, or context-control demands changed. Results showed an impressive robustness of context-control associations, as context-specific adjustments of control from the learning phase were virtually unaffected by new stimuli and tasks in the transfer block. Only a change of the context-control demand eliminated the context-specific adjustment of control. These findings suggest that context-control associations that have proven to be adaptive in the past are continuously applied despite major changes in the task structure as long as the context-control associations remain the same.
Locomotor Dysfunction after Spaceflight: Characterization and Countermeasure Development
NASA Technical Reports Server (NTRS)
Mulavara, A. P.; Cohen, H. S.; Peters, B. T.; Miller, C. A.; Brady, R.; Bloomberg, Jacob J.
2007-01-01
Astronauts returning from space flight show disturbances in locomotor control manifested by changes in various sub-systems including head-trunk coordination, dynamic visual acuity, lower limb muscle activation patterning and kinematics (Glasauer, et al., 1995; Bloomberg, et al., 1997; McDonald, et al., 1996; 1997; Layne, et al., 1997; 1998, 2001, 2004; Newman, et al., 1997; Bloomberg and Mulavara, 2003). These post flight changes in locomotor performance, due to neural adaptation to the microgravity conditions of space flight, affect the ability of crewmembers especially after a long duration mission to egress their vehicle and perform extravehicular activities soon after landing on Earth or following a landing on the surface of the Moon or Mars. At present, no operational training intervention is available pre- or in- flight to mitigate post flight locomotor disturbances. Our laboratory is currently developing a gait adaptability training program that is designed to facilitate recovery of locomotor function following a return to a gravitational environment. The training program exploits the ability of the sensorimotor system to generalize from exposure to multiple adaptive challenges during training so that the gait control system essentially "learns to learn" and therefore can reorganize more rapidly when faced with a novel adaptive challenge. Ultimately, the functional goal of an adaptive generalization countermeasure is not necessarily to immediately return movement patterns back to "normal". Rather the training regimen should facilitate the reorganization of available sensorimotor sub-systems to achieve safe and effective locomotion as soon as possible after space flight. We have previously confirmed that subjects participating in adaptive generalization training programs, using a variety of visuomotor distortions and different motor tasks from throwing to negotiating an obstacle course as the dependent measure, can learn to enhance their ability to adapt to a novel sensorimotor environment (Roller et al., 2001; Cohen et al. 2005). Importantly, this increased adaptability is retained even one month after completion of the training period. Our laboratory is currently developing adaptive generalization training procedures and the associated flight hardware to implement such a training program, using variations of visual flow, subject loading, and treadmill speed; during regular in-flight treadmill operations.
Leading toward value: the role of strategic human resource management in health system adaptability.
Garman, Andrew N; Polavarapu, Nandakishor; Grady, Jane C; Canar, W Jeffrey
2013-01-01
Personnel costs typically account for 60% or more of total operating expenses in health systems, and as such become a necessary focus in most if not all substantive health reform adaptations. This study sought to assess whether strategic alignment of the human resource (HR) and learning functions was associated with greater adaptive capacity in U.S. health systems. Data were gathered using a survey that was distributed electronically to chief human resource officers from two U.S.-based associations. The survey included questions about organizational structure, strategic human resource management, strategic learning, and organizational response to health reform. Significant correlations were found between strategic alignment of HR and HR's involvement in responses related to cost control (r = 0.46, p < 0.01); quality improvement (r = 0.45, p < 0.01), and patient access (r = 0.39, p < 0.01). However, no significant relationships were found between strategic alignment of organizational learning and HR involvement with these responses. Results suggest that HR structure may affect an organization's capacity for adaptive response. Top-management teams in health systems should consider positioning HR as part of the core leadership team, with a reporting relationship that allows HR to maximally participate in formulating and implementing organizational adaptation.
Structure Learning in Bayesian Sensorimotor Integration
Genewein, Tim; Hez, Eduard; Razzaghpanah, Zeynab; Braun, Daniel A.
2015-01-01
Previous studies have shown that sensorimotor processing can often be described by Bayesian learning, in particular the integration of prior and feedback information depending on its degree of reliability. Here we test the hypothesis that the integration process itself can be tuned to the statistical structure of the environment. We exposed human participants to a reaching task in a three-dimensional virtual reality environment where we could displace the visual feedback of their hand position in a two dimensional plane. When introducing statistical structure between the two dimensions of the displacement, we found that over the course of several days participants adapted their feedback integration process in order to exploit this structure for performance improvement. In control experiments we found that this adaptation process critically depended on performance feedback and could not be induced by verbal instructions. Our results suggest that structural learning is an important meta-learning component of Bayesian sensorimotor integration. PMID:26305797
2015-10-01
higher effect sizes than others when comparing any intervention (e.g., computer trainers, human tutors, group learning) to a control . It is difficult... control number. PLEASE DO NOT RETURN YOUR FORM TO THE ABOVE ADDRESS. 1. REPORT DATE (DD-MM-YYYY) October 2015 2. REPORT TYPE Special Report 3...ABSTRACT While human tutoring and mentoring are common teaching tools, current US Army standards for training and education are group instruction and
Balboni, Giulia; Incognito, Oriana; Belacchi, Carmen; Bonichini, Sabrina; Cubelli, Roberto
2017-02-01
The evaluation of adaptive behavior is informative in children with attention-deficit/hyperactivity disorder (ADHD) or specific learning disorders (SLD). However, the few investigations available have focused only on the gross level of domains of adaptive behavior. To investigate which item subsets of the Vineland-II can discriminate children with ADHD or SLD from peers with typical development. Student's t-tests, ROC analysis, logistic regression, and linear discriminant function analysis were used to compare 24 children with ADHD, 61 elementary students with SLD, and controls matched on age, sex, school level attended, and both parents' education level. Several item subsets that address not only ADHD core symptoms, but also understanding in social context and development of interpersonal relationships, allowed discrimination of children with ADHD from controls. The combination of four item subsets (Listening and attending, Expressing complex ideas, Social communication, and Following instructions) classified children with ADHD with both sensitivity and specificity of 87.5%. Only Reading skills, Writing skills, and Time and dates discriminated children with SLD from controls. Evaluation of Vineland-II scores at the level of item content categories is a useful procedure for an efficient clinical description. Copyright © 2016 Elsevier Ltd. All rights reserved.
Links between motor control and classroom behaviors: Moderation by low birth weight
Razza, Rachel A.; Martin, Anne; Brooks-Gunn, Jeanne
2016-01-01
It is unclear from past research on effortful control whether one of its components, motor control, independently contributes to adaptive classroom behaviors. The goal of this study was to identify associations between early motor control, measured by the walk-a-line task at age 3, and teacher-reported learning-related behaviors (approaches to learning and attention problems) and behavior problems in kindergarten classrooms. Models tested whether children who were vulnerable to poorer learning behaviors and more behavior problems due to having been born low birth weight benefited more, less, or the same as other children from better motor control. Data were drawn from the national Fragile Families and Child-Wellbeing Study (n = 751). Regression models indicated that motor control was significantly associated with better approaches to learning and fewer behavior problems. Children who were low birth weight benefitted more than normal birth weight children from better motor control with respect to their approaches to learning, but equally with respect to behavior problems. Additionally, for low but not normal birth weight children, better motor control predicted fewer attention problems. These findings suggest that motor control follows a compensatory model of development for low birth weight children and classroom behaviors. PMID:27594776
Approach for Using Learner Satisfaction to Evaluate the Learning Adaptation Policy
ERIC Educational Resources Information Center
Jeghal, Adil; Oughdir, Lahcen; Tairi, Hamid; Radouane, Abdelhay
2016-01-01
The learning adaptation is a very important phase in a learning situation in human learning environments. This paper presents the authors' approach used to evaluate the effectiveness of learning adaptive systems. This approach is based on the analysis of learner satisfaction notices collected by a questionnaire on a learning situation; to analyze…
Bu, Xiangwei; Wu, Xiaoyan; Tian, Mingyan; Huang, Jiaqi; Zhang, Rui; Ma, Zhen
2015-09-01
In this paper, an adaptive neural controller is exploited for a constrained flexible air-breathing hypersonic vehicle (FAHV) based on high-order tracking differentiator (HTD). By utilizing functional decomposition methodology, the dynamic model is reasonably decomposed into the respective velocity subsystem and altitude subsystem. For the velocity subsystem, a dynamic inversion based neural controller is constructed. By introducing the HTD to adaptively estimate the newly defined states generated in the process of model transformation, a novel neural based altitude controller that is quite simpler than the ones derived from back-stepping is addressed based on the normal output-feedback form instead of the strict-feedback formulation. Based on minimal-learning parameter scheme, only two neural networks with two adaptive parameters are needed for neural approximation. Especially, a novel auxiliary system is explored to deal with the problem of control inputs constraints. Finally, simulation results are presented to test the effectiveness of the proposed control strategy in the presence of system uncertainties and actuators constraints. Copyright © 2015 ISA. Published by Elsevier Ltd. All rights reserved.
Adaptive Filter Design Using Type-2 Fuzzy Cerebellar Model Articulation Controller.
Lin, Chih-Min; Yang, Ming-Shu; Chao, Fei; Hu, Xiao-Min; Zhang, Jun
2016-10-01
This paper aims to propose an efficient network and applies it as an adaptive filter for the signal processing problems. An adaptive filter is proposed using a novel interval type-2 fuzzy cerebellar model articulation controller (T2FCMAC). The T2FCMAC realizes an interval type-2 fuzzy logic system based on the structure of the CMAC. Due to the better ability of handling uncertainties, type-2 fuzzy sets can solve some complicated problems with outstanding effectiveness than type-1 fuzzy sets. In addition, the Lyapunov function is utilized to derive the conditions of the adaptive learning rates, so that the convergence of the filtering error can be guaranteed. In order to demonstrate the performance of the proposed adaptive T2FCMAC filter, it is tested in signal processing applications, including a nonlinear channel equalization system, a time-varying channel equalization system, and an adaptive noise cancellation system. The advantages of the proposed filter over the other adaptive filters are verified through simulations.
Motor learning in childhood reveals distinct mechanisms for memory retention and re-learning.
Musselman, Kristin E; Roemmich, Ryan T; Garrett, Ben; Bastian, Amy J
2016-05-01
Adults can easily learn and access multiple versions of the same motor skill adapted for different conditions (e.g., walking in water, sand, snow). Following even a single session of adaptation, adults exhibit clear day-to-day retention and faster re-learning of the adapted pattern. Here, we studied the retention and re-learning of an adapted walking pattern in children aged 6-17 yr. We found that all children, regardless of age, showed adult-like patterns of retention of the adapted walking pattern. In contrast, children under 12 yr of age did not re-learn faster on the next day after washout had occurred-they behaved as if they had never adapted their walking before. Re-learning could be improved in younger children when the adaptation time on day 1 was increased to allow more practice at the plateau of the adapted pattern, but never to adult-like levels. These results show that the ability to store a separate, adapted version of the same general motor pattern does not fully develop until adolescence, and furthermore, that the mechanisms underlying the retention and rapid re-learning of adapted motor patterns are distinct. © 2016 Musselman et al.; Published by Cold Spring Harbor Laboratory Press.
Learners' Perceptions and Illusions of Adaptivity in Computer-Based Learning Environments
ERIC Educational Resources Information Center
Vandewaetere, Mieke; Vandercruysse, Sylke; Clarebout, Geraldine
2012-01-01
Research on computer-based adaptive learning environments has shown exemplary growth. Although the mechanisms of effective adaptive instruction are unraveled systematically, little is known about the relative effect of learners' perceptions of adaptivity in adaptive learning environments. As previous research has demonstrated that the learners'…
ERIC Educational Resources Information Center
Fasihuddin, Heba; Skinner, Geoff; Athauda, Rukshan
2017-01-01
Open learning represents a new form of online learning where courses are provided freely online for large numbers of learners. MOOCs are examples of this form of learning. The authors see an opportunity for personalising open learning environments by adapting to learners' learning styles and providing adaptive support to meet individual learner…
NASA Astrophysics Data System (ADS)
Lee, Michael; Freed, Adrian; Wessel, David
1992-08-01
In this report we present our tools for prototyping adaptive user interfaces in the context of real-time musical instrument control. Characteristic of most human communication is the simultaneous use of classified events and estimated parameters. We have integrated a neural network object into the MAX language to explore adaptive user interfaces that considers these facets of human communication. By placing the neural processing in the context of a flexible real-time musical programming environment, we can rapidly prototype experiments on applications of adaptive interfaces and learning systems to musical problems. We have trained networks to recognize gestures from a Mathews radio baton, Nintendo Power GloveTM, and MIDI keyboard gestural input devices. In one experiment, a network successfully extracted classification and attribute data from gestural contours transduced by a continuous space controller, suggesting their application in the interpretation of conducting gestures and musical instrument control. We discuss network architectures, low-level features extracted for the networks to operate on, training methods, and musical applications of adaptive techniques.
Model-Free Optimal Tracking Control via Critic-Only Q-Learning.
Luo, Biao; Liu, Derong; Huang, Tingwen; Wang, Ding
2016-10-01
Model-free control is an important and promising topic in control fields, which has attracted extensive attention in the past few years. In this paper, we aim to solve the model-free optimal tracking control problem of nonaffine nonlinear discrete-time systems. A critic-only Q-learning (CoQL) method is developed, which learns the optimal tracking control from real system data, and thus avoids solving the tracking Hamilton-Jacobi-Bellman equation. First, the Q-learning algorithm is proposed based on the augmented system, and its convergence is established. Using only one neural network for approximating the Q-function, the CoQL method is developed to implement the Q-learning algorithm. Furthermore, the convergence of the CoQL method is proved with the consideration of neural network approximation error. With the convergent Q-function obtained from the CoQL method, the adaptive optimal tracking control is designed based on the gradient descent scheme. Finally, the effectiveness of the developed CoQL method is demonstrated through simulation studies. The developed CoQL method learns with off-policy data and implements with a critic-only structure, thus it is easy to realize and overcome the inadequate exploration problem.
Wu, Howard G.
2013-01-01
The planning of goal-directed movements is highly adaptable; however, the basic mechanisms underlying this adaptability are not well understood. Even the features of movement that drive adaptation are hotly debated, with some studies suggesting remapping of goal locations and others suggesting remapping of the movement vectors leading to goal locations. However, several previous motor learning studies and the multiplicity of the neural coding underlying visually guided reaching movements stand in contrast to this either/or debate on the modes of motor planning and adaptation. Here we hypothesize that, during visuomotor learning, the target location and movement vector of trained movements are separately remapped, and we propose a novel computational model for how motor plans based on these remappings are combined during the control of visually guided reaching in humans. To test this hypothesis, we designed a set of experimental manipulations that effectively dissociated the effects of remapping goal location and movement vector by examining the transfer of visuomotor adaptation to untrained movements and movement sequences throughout the workspace. The results reveal that (1) motor adaptation differentially remaps goal locations and movement vectors, and (2) separate motor plans based on these features are effectively averaged during motor execution. We then show that, without any free parameters, the computational model we developed for combining movement-vector-based and goal-location-based planning predicts nearly 90% of the variance in novel movement sequences, even when multiple attributes are simultaneously adapted, demonstrating for the first time the ability to predict how motor adaptation affects movement sequence planning. PMID:23804099
Raza, Meher; Ivry, Richard B.
2016-01-01
In standard taxonomies, motor skills are typically treated as representative of implicit or procedural memory. We examined two emblematic tasks of implicit motor learning, sensorimotor adaptation and sequence learning, asking whether individual differences in learning are correlated between these tasks, as well as how individual differences within each task are related to different performance variables. As a prerequisite, it was essential to establish the reliability of learning measures for each task. Participants were tested twice on a visuomotor adaptation task and on a sequence learning task, either the serial reaction time task or the alternating reaction time task. Learning was evident in all tasks at the group level and reliable at the individual level in visuomotor adaptation and the alternating reaction time task but not in the serial reaction time task. Performance variability was predictive of learning in both domains, yet the relationship was in the opposite direction for adaptation and sequence learning. For the former, faster learning was associated with lower variability, consistent with models of sensorimotor adaptation in which learning rates are sensitive to noise. For the latter, greater learning was associated with higher variability and slower reaction times, factors that may facilitate the spread of activation required to form predictive, sequential associations. Interestingly, learning measures of the different tasks were not correlated. Together, these results oppose a shared process for implicit learning in sensorimotor adaptation and sequence learning and provide insight into the factors that account for individual differences in learning within each task domain. NEW & NOTEWORTHY We investigated individual differences in the ability to implicitly learn motor skills. As a prerequisite, we assessed whether individual differences were reliable across test sessions. We found that two commonly used tasks of implicit learning, visuomotor adaptation and the alternating serial reaction time task, exhibited good test-retest reliability in measures of learning and performance. However, the learning measures did not correlate between the two tasks, arguing against a shared process for implicit motor learning. PMID:27832611
How to Represent Adaptation in e-Learning with IMS Learning Design
ERIC Educational Resources Information Center
Burgos, Daniel; Tattersall, Colin; Koper, Rob
2007-01-01
Adaptation in e-learning has been an important research topic for the last few decades in computer-based education. In adaptivity the behaviour of the user triggers some actions in the system that guides the learning process. In adaptability, the user makes changes and takes decisions. Progressing from computer-based training and adaptive…
The neural correlates of implicit sequence learning in schizophrenia.
Marvel, Cherie L; Turner, Beth M; O'Leary, Daniel S; Johnson, Hans J; Pierson, Ronald K; Ponto, Laura L Boles; Andreasen, Nancy C
2007-11-01
Twenty-seven schizophrenia spectrum patients and 25 healthy controls performed a probabilistic version of the serial reaction time task (SRT) that included sequence trials embedded within random trials. Patients showed diminished, yet measurable, sequence learning. Postexperimental analyses revealed that a group of patients performed above chance when generating short spans of the sequence. This high-generation group showed SRT learning that was similar in magnitude to that of controls. Their learning was evident from the very 1st block; however, unlike controls, learning did not develop further with continued testing. A subset of 12 patients and 11 controls performed the SRT in conjunction with positron emission tomography. High-generation performance, which corresponded to SRT learning in patients, correlated to activity in the premotor cortex and parahippocampus. These areas have been associated with stimulus-driven visuospatial processing. Taken together, these results suggest that a subset of patients who showed moderate success on the SRT used an explicit stimulus-driven strategy to process the sequential stimuli. This adaptive strategy facilitated sequence learning but may have interfered with conventional implicit learning of the overall stimulus pattern. PsycINFO Database Record (c) 2007 APA, all rights reserved.
Motion Based Target Acquisition and Evaluation in an Adaptive Machine Vision System
1995-05-01
paths in facial recognition and learning. Annals of Neurology, 22, 41-45. Tolman, E.C. (1932) Purposive behavior in Animals and Men. New York: Appleton...Learned scan paths are the active processes of perception. Rizzo et al. (1987) studied the fixation patterns of two patients with impaired facial ... recognition and learning and found an increase in the randomness of the scan patterns compared to controls, indicating that the cortex was failing to direct
Is there evidence of learned helplessness in horses?
Hall, Carol; Goodwin, Deborah; Heleski, Camie; Randle, Hayley; Waran, Natalie
2008-01-01
Learned helplessness is a psychological condition whereby individuals learn that they have no control over unpleasant or harmful conditions, that their actions are futile, and that they are helpless. In a series of experiments in which dogs were exposed to inescapable shocks, this lack of control subsequently interfered with the ability to learn an avoidance task. There is evidence that both neural adaptations and behavioral despair occur in response to uncontrollable aversive experiences in rodents, although this has yet to be demonstrated in other species such as horses. However, certain traditional methods of horse training and some behavioral modification techniques--it has been suggested--may involve aversive conditions over which the horse has little or no control. When training and management procedures are repeatedly unpleasant for the horse and there is no clear association between behavior and outcome, this is likely to interfere with learning and performance-in addition to compromising welfare. This article reviews published literature and anecdotal evidence to explore the possibility that the phenomenon, learned helplessness, occurs in the horse.
Activation of the cerebellar cortex and the dentate nucleus in a prism adaptation fMRI study.
Küper, Michael; Wünnemann, Meret J S; Thürling, Markus; Stefanescu, Roxana M; Maderwald, Stefan; Elles, Hans G; Göricke, Sophia; Ladd, Mark E; Timmann, Dagmar
2014-04-01
During prism adaptation two types of learning processes can be distinguished. First, fast strategic motor control responses are predominant in the early course of prism adaptation to achieve rapid error correction within few trials. Second, slower spatial realignment occurs among the misaligned visual and proprioceptive sensorimotor coordinate system. The aim of the present ultra-highfield (7T) functional magnetic resonance imaging (fMRI) study was to explore cerebellar cortical and dentate nucleus activation during the course of prism adaptation in relation to a similar visuomotor task without prism exposure. Nineteen young healthy participants were included into the study. Recently developed normalization procedures were applied for the cerebellar cortex and the dentate nucleus. By means of subtraction analysis (early prism adaptation > visuomotor, early prism adaptation > late prism adaptation) we identified ipsilateral activation associated with strategic motor control responses within the posterior cerebellar cortex (lobules VIII and IX) and the ventro-caudal dentate nucleus. During the late phase of adaptation we observed pronounced activation of posterior parts of lobule VI, although subtraction analyses (late prism adaptation > visuomotor) remained negative. These results are in good accordance with the concept of a representation of non-motor functions, here strategic control, within the ventro-caudal dentate nucleus. Copyright © 2013 Wiley Periodicals, Inc.
NASA Astrophysics Data System (ADS)
Randler, Christoph; Kummer, Barbara; Wilhelm, Christian
2012-06-01
The aim of this study was to assess the outcome of a zoo visit in terms of learning and retention of knowledge concerning the adaptations and behavior of vertebrate species. Basis of the work was the concept of implementing zoo visits as an out-of-school setting for formal, curriculum based learning. Our theoretical framework centers on the self-determination theory, therefore, we used a group-based, hands-on learning environment. To address this questions, we used a treatment—control design (BACI) with different treatments and a control group. Pre-, post- and retention tests were applied. All treatments led to a substantial increase of learning and retention knowledge compared to the control group. Immediately after the zoo visit, the zoo-guide tour provided the highest scores, while after a delay of 6 weeks, the learner-centered environment combined with a teacher-guided summarizing scored best. We suggest incorporating the zoo as an out-of-school environment into formal school learning, and we propose different methods to improve learning in zoo settings.
Spatial features of synaptic adaptation affecting learning performance.
Berger, Damian L; de Arcangelis, Lucilla; Herrmann, Hans J
2017-09-08
Recent studies have proposed that the diffusion of messenger molecules, such as monoamines, can mediate the plastic adaptation of synapses in supervised learning of neural networks. Based on these findings we developed a model for neural learning, where the signal for plastic adaptation is assumed to propagate through the extracellular space. We investigate the conditions allowing learning of Boolean rules in a neural network. Even fully excitatory networks show very good learning performances. Moreover, the investigation of the plastic adaptation features optimizing the performance suggests that learning is very sensitive to the extent of the plastic adaptation and the spatial range of synaptic connections.
"In and Against" Lifelong Learning: Flexibility and the Corrosion of Character
ERIC Educational Resources Information Center
Crowther, Jim
2004-01-01
This paper argues against the dominant discourse of lifelong learning. It is primarily a mode of social control that acts as a new disciplinary technology to make people more compliant and adaptable for work in the era of flexible capitalism. Whilst the main reference point is trends in the UK, the argument has a wider resonance. Lifelong learning…
Enriching Adaptation in E-Learning Systems through a Situation-Aware Ontology Network
ERIC Educational Resources Information Center
Pernas, Ana Marilza; Diaz, Alicia; Motz, Regina; de Oliveira, Jose Palazzo Moreira
2012-01-01
Purpose: The broader adoption of the internet along with web-based systems has defined a new way of exchanging information. That advance added by the multiplication of mobile devices has required systems to be even more flexible and personalized. Maybe because of that, the traditional teaching-controlled learning style has given up space to a new…
Adaptive Diagrams: Handing Control over to the Learner to Manage Split-Attention Online
ERIC Educational Resources Information Center
Agostinho, Shirley; Tindall-Ford, Sharon; Roodenrys, Kylie
2013-01-01
Based on cognitive load theory, it is well known that when studying a diagram that includes explanatory text, optimal learning occurs when the text is physically positioned close to the diagram as it eliminates the need for learners to split their attention between the two sources of information. What is not known is the effect on learning when…
Understanding the Impact of Training on Performance
2014-05-01
learning argue that this method encourages metacognitive activity and self -regulation of learning; which, in turn, can aid the development of adaptable...Frese, M. (2005). Self -regulation in error management training: Emotion control and metacognition as mediators of performance effects. Journal...on aptitude by treatment interactions that indicate the effectiveness of training interventions can differ depending on trainee aptitude for self
Sensory motor remapping of space in human–machine interfaces
Mussa-Ivaldi, Ferdinando A.; Casadio, Maura; Danziger, Zachary C.; Mosier, Kristine M.; Scheidt, Robert A.
2012-01-01
Studies of adaptation to patterns of deterministic forces have revealed the ability of the motor control system to form and use predictive representations of the environment. These studies have also pointed out that adaptation to novel dynamics is aimed at preserving the trajectories of a controlled endpoint, either the hand of a subject or a transported object. We review some of these experiments and present more recent studies aimed at understanding how the motor system forms representations of the physical space in which actions take place. An extensive line of investigations in visual information processing has dealt with the issue of how the Euclidean properties of space are recovered from visual signals that do not appear to possess these properties. The same question is addressed here in the context of motor behavior and motor learning by observing how people remap hand gestures and body motions that control the state of an external device. We present some theoretical considerations and experimental evidence about the ability of the nervous system to create novel patterns of coordination that are consistent with the representation of extrapersonal space. We also discuss the perspective of endowing human–machine interfaces with learning algorithms that, combined with human learning, may facilitate the control of powered wheelchairs and other assistive devices. PMID:21741543
Selected Flight Test Results for Online Learning Neural Network-Based Flight Control System
NASA Technical Reports Server (NTRS)
Williams, Peggy S.
2004-01-01
The NASA F-15 Intelligent Flight Control System project team has developed a series of flight control concepts designed to demonstrate the benefits of a neural network-based adaptive controller. The objective of the team is to develop and flight-test control systems that use neural network technology to optimize the performance of the aircraft under nominal conditions as well as stabilize the aircraft under failure conditions. Failure conditions include locked or failed control surfaces as well as unforeseen damage that might occur to the aircraft in flight. This report presents flight-test results for an adaptive controller using stability and control derivative values from an online learning neural network. A dynamic cell structure neural network is used in conjunction with a real-time parameter identification algorithm to estimate aerodynamic stability and control derivative increments to the baseline aerodynamic derivatives in flight. This set of open-loop flight tests was performed in preparation for a future phase of flights in which the learning neural network and parameter identification algorithm output would provide the flight controller with aerodynamic stability and control derivative updates in near real time. Two flight maneuvers are analyzed a pitch frequency sweep and an automated flight-test maneuver designed to optimally excite the parameter identification algorithm in all axes. Frequency responses generated from flight data are compared to those obtained from nonlinear simulation runs. An examination of flight data shows that addition of the flight-identified aerodynamic derivative increments into the simulation improved the pitch handling qualities of the aircraft.
NASA Astrophysics Data System (ADS)
de Arcangelis, L.; Lombardi, F.; Herrmann, H. J.
2014-03-01
Spontaneous brain activity has been recently characterized by avalanche dynamics with critical features for systems in vitro and in vivo. In this contribution we present a review of experimental results on neuronal avalanches in cortex slices, together with numerical results from a neuronal model implementing several physiological properties of living neurons. Numerical data reproduce experimental results for avalanche statistics. The temporal organization of avalanches can be characterized by the distribution of waiting times between successive avalanches. Experimental measurements exhibit a non-monotonic behaviour, not usually found in other natural processes. Numerical simulations provide evidence that this behaviour is a consequence of the alternation between states of high and low activity, leading to a balance between excitation and inhibition controlled by a single parameter. During these periods both the single neuron state and the network excitability level, keeping memory of past activity, are tuned by homoeostatic mechanisms. Interestingly, the same homoeostatic balance is detected for neuronal activity at the scale of the whole brain. We finally review the learning abilities of this neuronal network. Learning occurs via plastic adaptation of synaptic strengths by a non-uniform negative feedback mechanism. The system is able to learn all the tested rules and the learning dynamics exhibits universal features as a function of the strength of plastic adaptation. Any rule could be learned provided that the plastic adaptation is sufficiently slow.
Neural learning of constrained nonlinear transformations
NASA Technical Reports Server (NTRS)
Barhen, Jacob; Gulati, Sandeep; Zak, Michail
1989-01-01
Two issues that are fundamental to developing autonomous intelligent robots, namely, rudimentary learning capability and dexterous manipulation, are examined. A powerful neural learning formalism is introduced for addressing a large class of nonlinear mapping problems, including redundant manipulator inverse kinematics, commonly encountered during the design of real-time adaptive control mechanisms. Artificial neural networks with terminal attractor dynamics are used. The rapid network convergence resulting from the infinite local stability of these attractors allows the development of fast neural learning algorithms. Approaches to manipulator inverse kinematics are reviewed, the neurodynamics model is discussed, and the neural learning algorithm is presented.
Unique characteristics of motor adaptation during walking in young children.
Musselman, Kristin E; Patrick, Susan K; Vasudevan, Erin V L; Bastian, Amy J; Yang, Jaynie F
2011-05-01
Children show precocious ability in the learning of languages; is this the case with motor learning? We used split-belt walking to probe motor adaptation (a form of motor learning) in children. Data from 27 children (ages 8-36 mo) were compared with those from 10 adults. Children walked with the treadmill belts at the same speed (tied belt), followed by walking with the belts moving at different speeds (split belt) for 8-10 min, followed again by tied-belt walking (postsplit). Initial asymmetries in temporal coordination (i.e., double support time) induced by split-belt walking were slowly reduced, with most children showing an aftereffect (i.e., asymmetry in the opposite direction to the initial) in the early postsplit period, indicative of learning. In contrast, asymmetries in spatial coordination (i.e., center of oscillation) persisted during split-belt walking and no aftereffect was seen. Step length, a measure of both spatial and temporal coordination, showed intermediate effects. The time course of learning in double support and step length was slower in children than in adults. Moreover, there was a significant negative correlation between the size of the initial asymmetry during early split-belt walking (called error) and the aftereffect for step length. Hence, children may have more difficulty learning when the errors are large. The findings further suggest that the mechanisms controlling temporal and spatial adaptation are different and mature at different times.
Bayesian nonparametric adaptive control using Gaussian processes.
Chowdhary, Girish; Kingravi, Hassan A; How, Jonathan P; Vela, Patricio A
2015-03-01
Most current model reference adaptive control (MRAC) methods rely on parametric adaptive elements, in which the number of parameters of the adaptive element are fixed a priori, often through expert judgment. An example of such an adaptive element is radial basis function networks (RBFNs), with RBF centers preallocated based on the expected operating domain. If the system operates outside of the expected operating domain, this adaptive element can become noneffective in capturing and canceling the uncertainty, thus rendering the adaptive controller only semiglobal in nature. This paper investigates a Gaussian process-based Bayesian MRAC architecture (GP-MRAC), which leverages the power and flexibility of GP Bayesian nonparametric models of uncertainty. The GP-MRAC does not require the centers to be preallocated, can inherently handle measurement noise, and enables MRAC to handle a broader set of uncertainties, including those that are defined as distributions over functions. We use stochastic stability arguments to show that GP-MRAC guarantees good closed-loop performance with no prior domain knowledge of the uncertainty. Online implementable GP inference methods are compared in numerical simulations against RBFN-MRAC with preallocated centers and are shown to provide better tracking and improved long-term learning.
Schema-based learning of adaptable and flexible prey-catching in anurans I. The basic architecture.
Corbacho, Fernando; Nishikawa, Kiisa C; Weerasuriya, Ananda; Liaw, Jim-Shih; Arbib, Michael A
2005-12-01
A motor action often involves the coordination of several motor synergies and requires flexible adjustment of the ongoing execution based on feedback signals. To elucidate the neural mechanisms underlying the construction and selection of motor synergies, we study prey-capture in anurans. Experimental data demonstrate the intricate interaction between different motor synergies, including the interplay of their afferent feedback signals (Weerasuriya 1991; Anderson and Nishikawa 1996). Such data provide insights for the general issues concerning two-way information flow between sensory centers, motor circuits and periphery in motor coordination. We show how different afferent feedback signals about the status of the different components of the motor apparatus play a critical role in motor control as well as in learning. This paper, along with its companion paper, extend the model by Liaw et al. (1994) by integrating a number of different motor pattern generators, different types of afferent feedback, as well as the corresponding control structure within an adaptive framework we call Schema-Based Learning. We develop a model of the different MPGs involved in prey-catching as a vehicle to investigate the following questions: What are the characteristic features of the activity of a single muscle? How can these features be controlled by the premotor circuit? What are the strategies employed to generate and synchronize motor synergies? What is the role of afferent feedback in shaping the activity of a MPG? How can several MPGs share the same underlying circuitry and yet give rise to different motor patterns under different input conditions? In the companion paper we also extend the model by incorporating learning components that give rise to more flexible, adaptable and robust behaviors. To show these aspects we incorporate studies on experiments on lesions and the learning processes that allow the animal to recover its proper functioning.
Policy improvement by a model-free Dyna architecture.
Hwang, Kao-Shing; Lo, Chia-Yue
2013-05-01
The objective of this paper is to accelerate the process of policy improvement in reinforcement learning. The proposed Dyna-style system combines two learning schemes, one of which utilizes a temporal difference method for direct learning; the other uses relative values for indirect learning in planning between two successive direct learning cycles. Instead of establishing a complicated world model, the approach introduces a simple predictor of average rewards to actor-critic architecture in the simulation (planning) mode. The relative value of a state, defined as the accumulated differences between immediate reward and average reward, is used to steer the improvement process in the right direction. The proposed learning scheme is applied to control a pendulum system for tracking a desired trajectory to demonstrate its adaptability and robustness. Through reinforcement signals from the environment, the system takes the appropriate action to drive an unknown dynamic to track desired outputs in few learning cycles. Comparisons are made between the proposed model-free method, a connectionist adaptive heuristic critic, and an advanced method of Dyna-Q learning in the experiments of labyrinth exploration. The proposed method outperforms its counterparts in terms of elapsed time and convergence rate.
NASA Astrophysics Data System (ADS)
Tryfonidis, Michail
It has been observed that during orbital spaceflight the absence of gravitation related sensory inputs causes incongruence between the expected and the actual sensory feedback resulting from voluntary movements. This incongruence results in a reinterpretation or neglect of gravity-induced sensory input signals. Over time, new internal models develop, gradually compensating for the loss of spatial reference. The study of adaptation of goal-directed movements is the main focus of this thesis. The hypothesis is that during the adaptive learning process the neural connections behave in ways that can be described by an adaptive control method. The investigation presented in this thesis includes two different sets of experiments. A series of dart throwing experiments took place onboard the space station Mir. Experiments also took place at the Biomechanics lab at MIT, where the subjects performed a series of continuous trajectory tracking movements while a planar robotic manipulandum exerted external torques on the subjects' moving arms. The experimental hypothesis for both experiments is that during the first few trials the subjects will perform poorly trying to follow a prescribed trajectory, or trying to hit a target. A theoretical framework is developed that is a modification of the sliding control method used in robotics. The new control framework is an attempt to explain the adaptive behavior of the subjects. Numerical simulations of the proposed framework are compared with experimental results and predictions from competitive models. The proposed control methodology extends the results of the sliding mode theory to human motor control. The resulting adaptive control model of the motor system is robust to external dynamics, even those of negative gain, uses only position and velocity feedback, and achieves bounded steady-state error without explicit knowledge of the system's nonlinearities. In addition, the experimental and modeling results demonstrate that visuomotor learning is important not only for error correction through internal model adaptation on ground or in microgravity, but also for the minimization of the total mean-square error in the presence of random variability. Thus human intelligent decision displays certain attributes that seem to conform to Bayesian statistical games. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge, MA 02139-4307. Ph. 617-253-5668; Fax 617-253-1690.)
Development of force adaptation during childhood.
Konczak, Jürgen; Jansen-Osmann, Petra; Kalveram, Karl-Theodor
2003-03-01
Humans learn to make reaching movements in novel dynamic environments by acquiring an internal motor model of their limb dynamics. Here, the authors investigated how 4- to 11-year-old children (N = 39) and adults (N = 7) adapted to changes in arm dynamics, and they examined whether those data support the view that the human brain acquires inverse dynamics models (IDM) during development. While external damping forces were applied, the children learned to perform goal-directed forearm flexion movements. After changes in damping, all children showed kinematic aftereffects indicative of a neural controller that still attempted to compensate the no longer existing damping force. With increasing age, the number of trials toward complete adaptation decreased. When damping was present, forearm paths were most perturbed and most variable in the youngest children but were improved in the older children. The findings indicate that the neural representations of limb dynamics are less precise in children and less stable in time than those of adults. Such controller instability might be a primary cause of the high kinematic variability observed in many motor tasks during childhood. Finally, the young children were not able to update those models at the same rate as the older children, who, in turn, adapted more slowly than adults. In conclusion, the ability to adapt to unknown forces is a developmental achievement. The present results are consistent with the view that the acquisition and modification of internal models of the limb dynamics form the basis of that adaptive process.
Liu, Xiaolin; Mosier, Kristine M; Mussa-Ivaldi, Ferdinando A; Casadio, Maura; Scheidt, Robert A
2011-01-01
We examined how people organize redundant kinematic control variables (finger joint configurations) while learning to make goal-directed movements of a virtual object (a cursor) within a low-dimensional task space (a computer screen). Subjects participated in three experiments performed on separate days. Learning progressed rapidly on day 1, resulting in reduced target capture error and increased cursor trajectory linearity. On days 2 and 3, one group of subjects adapted to a rotation of the nominal map, imposed either stepwise or randomly over trials. Another group experienced a scaling distortion. We report two findings. First, adaptation rates and memory-dependent motor command updating depended on distortion type. Stepwise application and removal of the rotation induced a marked increase in finger motion variability but scaling did not, suggesting that the rotation initiated a more exhaustive search through the space of viable finger motions to resolve the target capture task than did scaling. Indeed, subjects formed new coordination patterns in compensating the rotation but relied on patterns established during baseline practice to compensate the scaling. These findings support the idea that the brain compensates direction and extent errors separately and in computationally distinct ways, but are inconsistent with the idea that once a task is learned, command updating is limited to those degrees of freedom contributing to performance (thereby minimizing energetic or similar costs of control). Second, we report that subjects who learned a scaling while moving to just one target generalized more narrowly across directions than those who learned a rotation. This contrasts with results from whole-arm reaching studies, where a learned scaling generalizes more broadly across direction than rotation. Based on inverse- and forward-dynamics analyses of reaching with the arm, we propose the difference in results derives from extensive exposure in reaching with familiar arm dynamics versus the novelty of the manual task.
A New Approach for Constructing the Concept Map
ERIC Educational Resources Information Center
Tseng, Shian-Shyong; Sue, Pei-Chi; Su, Jun-Ming; Weng, Jui-Feng; Tsai, Wen-Nung
2007-01-01
In recent years, e-learning system has become more and more popular and many adaptive learning environments have been proposed to offer learners customized courses in accordance with their aptitudes and learning results. For achieving the adaptive learning, a predefined concept map of a course is often used to provide adaptive learning guidance…
Using Data to Understand How to Better Design Adaptive Learning
ERIC Educational Resources Information Center
Liu, Min; Kang, Jina; Zou, Wenting; Lee, Hyeyeon; Pan, Zilong; Corliss, Stephanie
2017-01-01
There is much enthusiasm in higher education about the benefits of adaptive learning and using big data to investigate learning processes to make data-informed educational decisions. The benefits of adaptive learning to achieve personalized learning are obvious. Yet, there lacks evidence-based research to understand how data such as user behavior…
A Context-Adaptive Teacher Training Model in a Ubiquitous Learning Environment
ERIC Educational Resources Information Center
Chen, Min; Chiang, Feng Kuang; Jiang, Ya Na; Yu, Sheng Quan
2017-01-01
In view of the discrepancies in teacher training and teaching practice, this paper put forward a context-adaptive teacher training model in a ubiquitous learning (u-learning) environment. The innovative model provides teachers of different subjects with adaptive and personalized learning content in a u-learning environment, implements intra- and…
Pattern Adaptation and Normalization Reweighting.
Westrick, Zachary M; Heeger, David J; Landy, Michael S
2016-09-21
Adaptation to an oriented stimulus changes both the gain and preferred orientation of neural responses in V1. Neurons tuned near the adapted orientation are suppressed, and their preferred orientations shift away from the adapter. We propose a model in which weights of divisive normalization are dynamically adjusted to homeostatically maintain response products between pairs of neurons. We demonstrate that this adjustment can be performed by a very simple learning rule. Simulations of this model closely match existing data from visual adaptation experiments. We consider several alternative models, including variants based on homeostatic maintenance of response correlations or covariance, as well as feedforward gain-control models with multiple layers, and we demonstrate that homeostatic maintenance of response products provides the best account of the physiological data. Adaptation is a phenomenon throughout the nervous system in which neural tuning properties change in response to changes in environmental statistics. We developed a model of adaptation that combines normalization (in which a neuron's gain is reduced by the summed responses of its neighbors) and Hebbian learning (in which synaptic strength, in this case divisive normalization, is increased by correlated firing). The model is shown to account for several properties of adaptation in primary visual cortex in response to changes in the statistics of contour orientation. Copyright © 2016 the authors 0270-6474/16/369805-12$15.00/0.
Adaptive neural network/expert system that learns fault diagnosis for different structures
NASA Astrophysics Data System (ADS)
Simon, Solomon H.
1992-08-01
Corporations need better real-time monitoring and control systems to improve productivity by watching quality and increasing production flexibility. The innovative technology to achieve this goal is evolving in the form artificial intelligence and neural networks applied to sensor processing, fusion, and interpretation. By using these advanced Al techniques, we can leverage existing systems and add value to conventional techniques. Neural networks and knowledge-based expert systems can be combined into intelligent sensor systems which provide real-time monitoring, control, evaluation, and fault diagnosis for production systems. Neural network-based intelligent sensor systems are more reliable because they can provide continuous, non-destructive monitoring and inspection. Use of neural networks can result in sensor fusion and the ability to model highly, non-linear systems. Improved models can provide a foundation for more accurate performance parameters and predictions. We discuss a research software/hardware prototype which integrates neural networks, expert systems, and sensor technologies and which can adapt across a variety of structures to perform fault diagnosis. The flexibility and adaptability of the prototype in learning two structures is presented. Potential applications are discussed.
NASA Technical Reports Server (NTRS)
Mulavara, A. P.; Wood, S. J.; Cohen, H. S.; Bloomberg, J. J.
2012-01-01
Exposure to the microgravity conditions of space flight induces adaptive modification in sensorimotor function allowing astronauts to operate in this unique environment. This adaptive state, however, is inappropriate for a 1-g environment. Consequently astronauts must spend time readapting to Earth s gravity following their return to Earth. During this readaptation period, alterations in sensorimotor function cause various disturbances in astronaut gait during postflight walking. They often rely more on vision for postural and gait stability and many report the need for greater cognitive supervision of motor actions that previous to space flight were fully automated. Over the last several years our laboratory has investigated postflight astronaut locomotion with the aim of better understanding how adaptive changes in underlying sensorimotor mechanisms contribute to postflight gait dysfunction. Exposure to the microgravity conditions of space flight induces adaptive modification in the control of vestibularly-mediated reflexive head movement during locomotion after space flight. Furthermore, during motor learning, adaptive transitions are composed of two main mechanisms: strategic and plastic. Strategic mechanisms represent immediate and transitory modifications in control to deal with changes in the prevailing environment that, if prolonged, induce plastic mechanisms designed to automate new behavioral responses. The goal of the present study was to examine the contributions of sensorimotor subsystems such as the vestibular and body load sensing (BLS) somatosensory influences on head movement control during locomotion after long-duration space flight. Further we present data on the two motor learning processes during readaptation of locomotor function after long-duration space flight.
NASA Astrophysics Data System (ADS)
Chen, Syuan-Yi; Gong, Sheng-Sian
2017-09-01
This study aims to develop an adaptive high-precision control system for controlling the speed of a vane-type air motor (VAM) pneumatic servo system. In practice, the rotor speed of a VAM depends on the input mass air flow, which can be controlled by the effective orifice area (EOA) of an electronic throttle valve (ETV). As the control variable of a second-order pneumatic system is the integral of the EOA, an observation-based adaptive dynamic sliding-mode control (ADSMC) system is proposed to derive the differential of the control variable, namely, the EOA control signal. In the ADSMC system, a proportional-integral-derivative fuzzy neural network (PIDFNN) observer is used to achieve an ideal dynamic sliding-mode control (DSMC), and a supervisor compensator is designed to eliminate the approximation error. As a result, the ADSMC incorporates the robustness of a DSMC and the online learning ability of a PIDFNN. To ensure the convergence of the tracking error, a Lyapunov-based analytical method is employed to obtain the adaptive algorithms required to tune the control parameters of the online ADSMC system. Finally, our experimental results demonstrate the precision and robustness of the ADSMC system for highly nonlinear and time-varying VAM pneumatic servo systems.
Heathcock, Jill C; Bhat, Anjana N; Lobo, Michele A; Galloway, James C
2005-01-01
Infants born preterm differ in their spontaneous kicking, as well as their learning and memory abilities in the mobile paradigm, compared with infants born full-term. In the mobile paradigm, a supine infant's ankle is tethered to a mobile so that leg kicks cause a proportional amount of mobile movement. The purpose of this study was to investigate the relative kicking frequency of the tethered (right) and nontethered (left) legs in these 2 groups of infants. Ten infants born full-term and 10 infants born preterm (<33 weeks gestational age, <2,500 g) and 10 comparison infants participated in the study. The relative kicking frequencies of the tethered and nontethered legs were analyzed during learning and short-term and long-term memory periods of the mobile paradigm. Infants born full-term showed an increase in the relative kicking frequency of the tethered leg during the learning period and the short-term memory period but not for the long-term memory period. Infants born preterm did not show a change in kicking pattern for learning or memory periods, and consistently kicked both legs in relatively equal amounts. Infants born full-term adapted their baseline kicking frequencies in a task-specific manner to move the mobile and then retained this adaptation for the short-term memory period. In contrast, infants born preterm showed no adaptation, suggesting a lack of purposeful leg control. This lack of control may reflect a general decrease in the ability of infants born preterm to use their limb movements to interact with their environment. As such, the mobile paradigm may be clinically useful in the early assessment and intervention of infants born preterm and at risk for future impairment.
Adaptive Neurotechnology for Making Neural Circuits Functional .
NASA Astrophysics Data System (ADS)
Jung, Ranu
2008-03-01
Two of the most important trends in recent technological developments are that technology is increasingly integrated with biological systems and that it is increasingly adaptive in its capabilities. Neuroprosthetic systems that provide lost sensorimotor function after a neural disability offer a platform to investigate this interplay between biological and engineered systems. Adaptive neurotechnology (hardware and software) could be designed to be biomimetic, guided by the physical and programmatic constraints observed in biological systems, and allow for real-time learning, stability, and error correction. An example will present biomimetic neural-network hardware that can be interfaced with the isolated spinal cord of a lower vertebrate to allow phase-locked real-time neural control. Another will present adaptive neural network control algorithms for functional electrical stimulation of the peripheral nervous system to provide desired movements of paralyzed limbs in rodents or people. Ultimately, the frontier lies in being able to utilize the adaptive neurotechnology to promote neuroplasticity in the living system on a long-time scale under co-adaptive conditions.
Learning to speciate: The biased learning of mate preferences promotes adaptive radiation
Gilman, R. Tucker; Kozak, Genevieve M.
2015-01-01
Bursts of rapid repeated speciation called adaptive radiations have generated much of Earth's biodiversity and fascinated biologists since Darwin, but we still do not know why some lineages radiate and others do not. Understanding what causes assortative mating to evolve rapidly and repeatedly in the same lineage is key to understanding adaptive radiation. Many species that have undergone adaptive radiations exhibit mate preference learning, where individuals acquire mate preferences by observing the phenotypes of other members of their populations. Mate preference learning can be biased if individuals also learn phenotypes to avoid in mates, and shift their preferences away from these avoided phenotypes. We used individual‐based computational simulations to study whether biased and unbiased mate preference learning promotes ecological speciation and adaptive radiation. We found that ecological speciation can be rapid and repeated when mate preferences are biased, but is inhibited when mate preferences are learned without bias. Our results suggest that biased mate preference learning may play an important role in generating animal biodiversity through adaptive radiation. PMID:26459795
Clustering of tethered satellite system simulation data by an adaptive neuro-fuzzy algorithm
NASA Technical Reports Server (NTRS)
Mitra, Sunanda; Pemmaraju, Surya
1992-01-01
Recent developments in neuro-fuzzy systems indicate that the concepts of adaptive pattern recognition, when used to identify appropriate control actions corresponding to clusters of patterns representing system states in dynamic nonlinear control systems, may result in innovative designs. A modular, unsupervised neural network architecture, in which fuzzy learning rules have been embedded is used for on-line identification of similar states. The architecture and control rules involved in Adaptive Fuzzy Leader Clustering (AFLC) allow this system to be incorporated in control systems for identification of system states corresponding to specific control actions. We have used this algorithm to cluster the simulation data of Tethered Satellite System (TSS) to estimate the range of delta voltages necessary to maintain the desired length rate of the tether. The AFLC algorithm is capable of on-line estimation of the appropriate control voltages from the corresponding length error and length rate error without a priori knowledge of their membership functions and familarity with the behavior of the Tethered Satellite System.
Self-organized adaptation of a simple neural circuit enables complex robot behaviour
NASA Astrophysics Data System (ADS)
Steingrube, Silke; Timme, Marc; Wörgötter, Florentin; Manoonpong, Poramate
2010-03-01
Controlling sensori-motor systems in higher animals or complex robots is a challenging combinatorial problem, because many sensory signals need to be simultaneously coordinated into a broad behavioural spectrum. To rapidly interact with the environment, this control needs to be fast and adaptive. Present robotic solutions operate with limited autonomy and are mostly restricted to few behavioural patterns. Here we introduce chaos control as a new strategy to generate complex behaviour of an autonomous robot. In the presented system, 18 sensors drive 18 motors by means of a simple neural control circuit, thereby generating 11 basic behavioural patterns (for example, orienting, taxis, self-protection and various gaits) and their combinations. The control signal quickly and reversibly adapts to new situations and also enables learning and synaptic long-term storage of behaviourally useful motor responses. Thus, such neural control provides a powerful yet simple way to self-organize versatile behaviours in autonomous agents with many degrees of freedom.
Functional Based Adaptive and Fuzzy Sliding Controller for Non-Autonomous Active Suspension System
NASA Astrophysics Data System (ADS)
Huang, Shiuh-Jer; Chen, Hung-Yi
In this paper, an adaptive sliding controller is developed for controlling a vehicle active suspension system. The functional approximation technique is employed to substitute the unknown non-autonomous functions of the suspension system and release the model-based requirement of sliding mode control algorithm. In order to improve the control performance and reduce the implementation problem, a fuzzy strategy with online learning ability is added to compensate the functional approximation error. The update laws of the functional approximation coefficients and the fuzzy tuning parameters are derived from the Lyapunov theorem to guarantee the system stability. The proposed controller is implemented on a quarter-car hydraulic actuating active suspension system test-rig. The experimental results show that the proposed controller suppresses the oscillation amplitude of the suspension system effectively.
Concept Based Approach for Adaptive Personalized Course Learning System
ERIC Educational Resources Information Center
Salahli, Mehmet Ali; Özdemir, Muzaffer; Yasar, Cumali
2013-01-01
One of the most important factors for improving the personalization aspects of learning systems is to enable adaptive properties to them. The aim of the adaptive personalized learning system is to offer the most appropriate learning path and learning materials to learners by taking into account their profiles. In this paper, a new approach to…
Designing a Semantic Bliki System to Support Different Types of Knowledge and Adaptive Learning
ERIC Educational Resources Information Center
Huang, Shiu-Li; Yang, Chia-Wei
2009-01-01
Though blogs and wikis have been used to support knowledge management and e-learning, existing blogs and wikis cannot support different types of knowledge and adaptive learning. A case in point, types of knowledge vary greatly in category and viewpoints. Additionally, adaptive learning is crucial to improving one's learning performance. This study…
ERIC Educational Resources Information Center
Yang, Tzu-Chi; Hwang, Gwo-Jen; Yang, Stephen Jen-Hwa
2013-01-01
In this study, an adaptive learning system is developed by taking multiple dimensions of personalized features into account. A personalized presentation module is proposed for developing adaptive learning systems based on the field dependent/independent cognitive style model and the eight dimensions of Felder-Silverman's learning style. An…
Lejeune, Caroline; Wansard, Murielle; Geurten, Marie; Meulemans, Thierry
2016-01-01
The aim of this study was to explore the differences in procedural learning abilities between children with DCD and typically developing children by investigating the steps that lead to skill automatization (i.e., the stages of fast learning, consolidation, and slow learning). Transfer of the skill to a new situation was also assessed. We tested 34 children aged 6-12 years with and without DCD on a perceptuomotor adaptation task, a form of procedural learning that is thought to involve the cerebellum and the basal ganglia (regions whose impairment has been associated with DCD) but also other brain areas including frontal regions. The results showed similar rates of learning, consolidation, and transfer in DCD and control children. However, the DCD children's performance remained slower than that of controls throughout the procedural task and they reached a lower asymptotic performance level; the difficulties observed at the outset did not diminish with practice.
A Collaborative 20 Questions Model for Target Search with Human-Machine Interaction
2013-05-01
optimal policies for entropy loss,” Journal of Applied Probability, vol. 49, pp. 114–136, 2012. [2] R. Castro and R. Nowak, “ Active learning and...vol. 10, pp. 223231, 1974. [8] R. Castro, Active Learning and Adaptive Sampling for Non- parametric Inference, Ph.D. thesis, Rice University, August...2007. [9] R. Castro and R. D. Nowak, “Upper and lower bounds for active learning ,” in 44th Annual Allerton Conference on Communica- tion, Control and Computing, 2006.
Neural networks for aircraft control
NASA Technical Reports Server (NTRS)
Linse, Dennis
1990-01-01
Current research in Artificial Neural Networks indicates that networks offer some potential advantages in adaptation and fault tolerance. This research is directed at determining the possible applicability of neural networks to aircraft control. The first application will be to aircraft trim. Neural network node characteristics, network topology and operation, neural network learning and example histories using neighboring optimal control with a neural net are discussed.
Stress attenuates the flexible updating of aversive value
Raio, Candace M.; Hartley, Catherine A.; Orederu, Temidayo A.; Li, Jian; Phelps, Elizabeth A.
2017-01-01
In a dynamic environment, sources of threat or safety can unexpectedly change, requiring the flexible updating of stimulus−outcome associations that promote adaptive behavior. However, aversive contexts in which we are required to update predictions of threat are often marked by stress. Acute stress is thought to reduce behavioral flexibility, yet its influence on the modulation of aversive value has not been well characterized. Given that stress exposure is a prominent risk factor for anxiety and trauma-related disorders marked by persistent, inflexible responses to threat, here we examined how acute stress affects the flexible updating of threat responses. Participants completed an aversive learning task, in which one stimulus was probabilistically associated with an electric shock, while the other stimulus signaled safety. A day later, participants underwent an acute stress or control manipulation before completing a reversal learning task during which the original stimulus−outcome contingencies switched. Skin conductance and neuroendocrine responses provided indices of sympathetic arousal and stress responses, respectively. Despite equivalent initial learning, stressed participants showed marked impairments in reversal learning relative to controls. Additionally, reversal learning deficits across participants were related to heightened levels of alpha-amylase, a marker of noradrenergic activity. Finally, fitting arousal data to a computational reinforcement learning model revealed that stress-induced reversal learning deficits emerged from stress-specific changes in the weight assigned to prediction error signals, disrupting the adaptive adjustment of learning rates. Our findings provide insight into how stress renders individuals less sensitive to changes in aversive reinforcement and have implications for understanding clinical conditions marked by stress-related psychopathology. PMID:28973957
Stark-Inbar, Alit; Raza, Meher; Taylor, Jordan A; Ivry, Richard B
2017-01-01
In standard taxonomies, motor skills are typically treated as representative of implicit or procedural memory. We examined two emblematic tasks of implicit motor learning, sensorimotor adaptation and sequence learning, asking whether individual differences in learning are correlated between these tasks, as well as how individual differences within each task are related to different performance variables. As a prerequisite, it was essential to establish the reliability of learning measures for each task. Participants were tested twice on a visuomotor adaptation task and on a sequence learning task, either the serial reaction time task or the alternating reaction time task. Learning was evident in all tasks at the group level and reliable at the individual level in visuomotor adaptation and the alternating reaction time task but not in the serial reaction time task. Performance variability was predictive of learning in both domains, yet the relationship was in the opposite direction for adaptation and sequence learning. For the former, faster learning was associated with lower variability, consistent with models of sensorimotor adaptation in which learning rates are sensitive to noise. For the latter, greater learning was associated with higher variability and slower reaction times, factors that may facilitate the spread of activation required to form predictive, sequential associations. Interestingly, learning measures of the different tasks were not correlated. Together, these results oppose a shared process for implicit learning in sensorimotor adaptation and sequence learning and provide insight into the factors that account for individual differences in learning within each task domain. We investigated individual differences in the ability to implicitly learn motor skills. As a prerequisite, we assessed whether individual differences were reliable across test sessions. We found that two commonly used tasks of implicit learning, visuomotor adaptation and the alternating serial reaction time task, exhibited good test-retest reliability in measures of learning and performance. However, the learning measures did not correlate between the two tasks, arguing against a shared process for implicit motor learning. Copyright © 2017 the American Physiological Society.
FMRI evidence of 'mirror' responses to geometric shapes.
Press, Clare; Catmur, Caroline; Cook, Richard; Widmann, Hannah; Heyes, Cecilia; Bird, Geoffrey
2012-01-01
Mirror neurons may be a genetic adaptation for social interaction. Alternatively, the associative hypothesis proposes that the development of mirror neurons is driven by sensorimotor learning, and that, given suitable experience, mirror neurons will respond to any stimulus. This hypothesis was tested using fMRI adaptation to index populations of cells with mirror properties. After sensorimotor training, where geometric shapes were paired with hand actions, BOLD response was measured while human participants experienced runs of events in which shape observation alternated with action execution or observation. Adaptation from shapes to action execution, and critically, observation, occurred in ventral premotor cortex (PMv) and inferior parietal lobule (IPL). Adaptation from shapes to execution indicates that neuronal populations responding to the shapes had motor properties, while adaptation to observation demonstrates that these populations had mirror properties. These results indicate that sensorimotor training induced populations of cells with mirror properties in PMv and IPL to respond to the observation of arbitrary shapes. They suggest that the mirror system has not been shaped by evolution to respond in a mirror fashion to biological actions; instead, its development is mediated by stimulus-general processes of learning within a system adapted for visuomotor control.
fMRI Evidence of ‘Mirror’ Responses to Geometric Shapes
Press, Clare; Catmur, Caroline; Cook, Richard; Widmann, Hannah; Heyes, Cecilia; Bird, Geoffrey
2012-01-01
Mirror neurons may be a genetic adaptation for social interaction [1]. Alternatively, the associative hypothesis [2], [3] proposes that the development of mirror neurons is driven by sensorimotor learning, and that, given suitable experience, mirror neurons will respond to any stimulus. This hypothesis was tested using fMRI adaptation to index populations of cells with mirror properties. After sensorimotor training, where geometric shapes were paired with hand actions, BOLD response was measured while human participants experienced runs of events in which shape observation alternated with action execution or observation. Adaptation from shapes to action execution, and critically, observation, occurred in ventral premotor cortex (PMv) and inferior parietal lobule (IPL). Adaptation from shapes to execution indicates that neuronal populations responding to the shapes had motor properties, while adaptation to observation demonstrates that these populations had mirror properties. These results indicate that sensorimotor training induced populations of cells with mirror properties in PMv and IPL to respond to the observation of arbitrary shapes. They suggest that the mirror system has not been shaped by evolution to respond in a mirror fashion to biological actions; instead, its development is mediated by stimulus-general processes of learning within a system adapted for visuomotor control. PMID:23251653
Neurocontrol and fuzzy logic: Connections and designs
NASA Technical Reports Server (NTRS)
Werbos, Paul J.
1991-01-01
Artificial neural networks (ANNs) and fuzzy logic are complementary technologies. ANNs extract information from systems to be learned or controlled, while fuzzy techniques mainly use verbal information from experts. Ideally, both sources of information should be combined. For example, one can learn rules in a hybrid fashion, and then calibrate them for better whole-system performance. ANNs offer universal approximation theorems, pedagogical advantages, very high-throughput hardware, and links to neurophysiology. Neurocontrol - the use of ANNs to directly control motors or actuators, etc. - uses five generalized designs, related to control theory, which can work on fuzzy logic systems as well as ANNs. These designs can copy what experts do instead of what they say, learn to track trajectories, generalize adaptive control, and maximize performance or minimize cost over time, even in noisy environments. Design tradeoffs and future directions are discussed throughout.
Reinforcement learning for a biped robot based on a CPG-actor-critic method.
Nakamura, Yutaka; Mori, Takeshi; Sato, Masa-aki; Ishii, Shin
2007-08-01
Animals' rhythmic movements, such as locomotion, are considered to be controlled by neural circuits called central pattern generators (CPGs), which generate oscillatory signals. Motivated by this biological mechanism, studies have been conducted on the rhythmic movements controlled by CPG. As an autonomous learning framework for a CPG controller, we propose in this article a reinforcement learning method we call the "CPG-actor-critic" method. This method introduces a new architecture to the actor, and its training is roughly based on a stochastic policy gradient algorithm presented recently. We apply this method to an automatic acquisition problem of control for a biped robot. Computer simulations show that training of the CPG can be successfully performed by our method, thus allowing the biped robot to not only walk stably but also adapt to environmental changes.
2010 Staff Organization for Optimum C2: A Private Sector Analysis
1998-02-13
control over business operations. Warfighting CINCs can benefit from the lessons learned in the private sector by adapting those lessons to future military... private sector analysis. Through the use of a networked command and control system and a "matrix" staff structure, the model consolidates the JFC staff
Fostering Self-Regulation in Training Complex Cognitive Tasks
ERIC Educational Resources Information Center
van Meeuwen, Ludo W.; Brand-Gruwel, Saskia; Kirschner, Paul A.; de Bock, Jeano J. P. R.; van Merriënboer, Jeroen J. G.
2018-01-01
In complex cognitive domains such as air traffic control, professionals must be able to adapt to and act upon continuing changes in a highly advanced technological work environment. To function optimally in such an environment, the controllers must be able to regulate their learning. Although these regulation skills should be part of their…
1985-07-01
learning ’, ’adaptations’, ’process’, and ’ abstraction ...hierarchy. Rather, he argues through a combina- tion of cybernetic epistemology, learning theory, and cognitive models, that patterned collective...systemic ability to perceive and respond to pattern and variance in the ’environment’. Each level of abstraction can be considered a ’meta’
Associative visual learning by tethered bees in a controlled visual environment.
Buatois, Alexis; Pichot, Cécile; Schultheiss, Patrick; Sandoz, Jean-Christophe; Lazzari, Claudio R; Chittka, Lars; Avarguès-Weber, Aurore; Giurfa, Martin
2017-10-10
Free-flying honeybees exhibit remarkable cognitive capacities but the neural underpinnings of these capacities cannot be studied in flying insects. Conversely, immobilized bees are accessible to neurobiological investigation but display poor visual learning. To overcome this limitation, we aimed at establishing a controlled visual environment in which tethered bees walking on a spherical treadmill learn to discriminate visual stimuli video projected in front of them. Freely flying bees trained to walk into a miniature Y-maze displaying these stimuli in a dark environment learned the visual discrimination efficiently when one of them (CS+) was paired with sucrose and the other with quinine solution (CS-). Adapting this discrimination to the treadmill paradigm with a tethered, walking bee was successful as bees exhibited robust discrimination and preferred the CS+ to the CS- after training. As learning was better in the maze, movement freedom, active vision and behavioral context might be important for visual learning. The nature of the punishment associated with the CS- also affects learning as quinine and distilled water enhanced the proportion of learners. Thus, visual learning is amenable to a controlled environment in which tethered bees learn visual stimuli, a result that is important for future neurobiological studies in virtual reality.
Development of Methodologies for IV and V of Neural Networks
NASA Technical Reports Server (NTRS)
Taylor, Brian; Darrah, Marjorie
2003-01-01
Non-deterministic systems often rely upon neural network (NN) technology to "lean" to manage flight systems under controlled conditions using carefully chosen training sets. How can these adaptive systems be certified to ensure that they will become increasingly efficient and behave appropriately in real-time situations? The bulk of Independent Verification and Validation (IV&V) research of non-deterministic software control systems such as Adaptive Flight Controllers (AFC's) addresses NNs in well-behaved and constrained environments such as simulations and strict process control. However, neither substantive research, nor effective IV&V techniques have been found to address AFC's learning in real-time and adapting to live flight conditions. Adaptive flight control systems offer good extensibility into commercial aviation as well as military aviation and transportation. Consequently, this area of IV&V represents an area of growing interest and urgency. ISR proposes to further the current body of knowledge to meet two objectives: Research the current IV&V methods and assess where these methods may be applied toward a methodology for the V&V of Neural Network; and identify effective methods for IV&V of NNs that learn in real-time, including developing a prototype test bed for IV&V of AFC's. Currently. no practical method exists. lSR will meet these objectives through the tasks identified and described below. First, ISR will conduct a literature review of current IV&V technology. TO do this, ISR will collect the existing body of research on IV&V of non-deterministic systems and neural network. ISR will also develop the framework for disseminating this information through specialized training. This effort will focus on developing NASA's capability to conduct IV&V of neural network systems and to provide training to meet the increasing need for IV&V expertise in such systems.
The effectiveness of a learning strategies program for university students.
Roces Montero, Cristina; Sierra Y Arizmendiarrieta, Beatriz
2017-11-01
University lecturers often complain about their students’ lack of learning strategies, but not many universities in Spain offer specific courses in this area. Studies on their effectiveness are also rare. This study presents the results of a Learning Strategies Course implemented at the School of Teacher Training and Education, University of Oviedo, Spain. A quasi-experimental design was used with an experi-mental (n = 60) and a control group (n = 57) of students on the Educational Psychology course. A Spanish adaptation of the Motivated Strategies for Learning Questionnaire (MSLQ): the CEAMR2 was used as a pre and post-test measure. Group A (EG) received training in learning strategies, while group B (CG) received no training. Post-test measures showed significant differences in five out of the ten learning strategies assessed: elaboration, organization, repetition, self-questioning and study space, and also an improvement in one out of the six motivational scales: control of learning beliefs. The results suggest that learning strategies courses with proven effectiveness should be offered to university students.
Kobza, Stefan; Ferrea, Stefano; Schnitzler, Alfons; Pollok, Bettina; Südmeyer, Martin; Bellebaum, Christian
2012-01-01
Feedback to both actively performed and observed behaviour allows adaptation of future actions. Positive feedback leads to increased activity of dopamine neurons in the substantia nigra, whereas dopamine neuron activity is decreased following negative feedback. Dopamine level reduction in unmedicated Parkinson's Disease patients has been shown to lead to a negative learning bias, i.e. enhanced learning from negative feedback. Recent findings suggest that the neural mechanisms of active and observational learning from feedback might differ, with the striatum playing a less prominent role in observational learning. Therefore, it was hypothesized that unmedicated Parkinson's Disease patients would show a negative learning bias only in active but not in observational learning. In a between-group design, 19 Parkinson's Disease patients and 40 healthy controls engaged in either an active or an observational probabilistic feedback-learning task. For both tasks, transfer phases aimed to assess the bias to learn better from positive or negative feedback. As expected, actively learning patients showed a negative learning bias, whereas controls learned better from positive feedback. In contrast, no difference between patients and controls emerged for observational learning, with both groups showing better learning from positive feedback. These findings add to neural models of reinforcement-learning by suggesting that dopamine-modulated input to the striatum plays a minor role in observational learning from feedback. Future research will have to elucidate the specific neural underpinnings of observational learning.
Vibration control of building structures using self-organizing and self-learning neural networks
NASA Astrophysics Data System (ADS)
Madan, Alok
2005-11-01
Past research in artificial intelligence establishes that artificial neural networks (ANN) are effective and efficient computational processors for performing a variety of tasks including pattern recognition, classification, associative recall, combinatorial problem solving, adaptive control, multi-sensor data fusion, noise filtering and data compression, modelling and forecasting. The paper presents a potentially feasible approach for training ANN in active control of earthquake-induced vibrations in building structures without the aid of teacher signals (i.e. target control forces). A counter-propagation neural network is trained to output the control forces that are required to reduce the structural vibrations in the absence of any feedback on the correctness of the output control forces (i.e. without any information on the errors in output activations of the network). The present study shows that, in principle, the counter-propagation network (CPN) can learn from the control environment to compute the required control forces without the supervision of a teacher (unsupervised learning). Simulated case studies are presented to demonstrate the feasibility of implementing the unsupervised learning approach in ANN for effective vibration control of structures under the influence of earthquake ground motions. The proposed learning methodology obviates the need for developing a mathematical model of structural dynamics or training a separate neural network to emulate the structural response for implementation in practice.
Luque, Niceto R.; Garrido, Jesús A.; Carrillo, Richard R.; D'Angelo, Egidio; Ros, Eduardo
2014-01-01
The cerebellum is known to play a critical role in learning relevant patterns of activity for adaptive motor control, but the underlying network mechanisms are only partly understood. The classical long-term synaptic plasticity between parallel fibers (PFs) and Purkinje cells (PCs), which is driven by the inferior olive (IO), can only account for limited aspects of learning. Recently, the role of additional forms of plasticity in the granular layer, molecular layer and deep cerebellar nuclei (DCN) has been considered. In particular, learning at DCN synapses allows for generalization, but convergence to a stable state requires hundreds of repetitions. In this paper we have explored the putative role of the IO-DCN connection by endowing it with adaptable weights and exploring its implications in a closed-loop robotic manipulation task. Our results show that IO-DCN plasticity accelerates convergence of learning by up to two orders of magnitude without conflicting with the generalization properties conferred by DCN plasticity. Thus, this model suggests that multiple distributed learning mechanisms provide a key for explaining the complex properties of procedural learning and open up new experimental questions for synaptic plasticity in the cerebellar network. PMID:25177290
The Basal Ganglia and Adaptive Motor Control
NASA Astrophysics Data System (ADS)
Graybiel, Ann M.; Aosaki, Toshihiko; Flaherty, Alice W.; Kimura, Minoru
1994-09-01
The basal ganglia are neural structures within the motor and cognitive control circuits in the mammalian forebrain and are interconnected with the neocortex by multiple loops. Dysfunction in these parallel loops caused by damage to the striatum results in major defects in voluntary movement, exemplified in Parkinson's disease and Huntington's disease. These parallel loops have a distributed modular architecture resembling local expert architectures of computational learning models. During sensorimotor learning, such distributed networks may be coordinated by widely spaced striatal interneurons that acquire response properties on the basis of experienced reward.
Adaptive Units of Learning and Educational Videogames
ERIC Educational Resources Information Center
Moreno-Ger, Pablo; Thomas, Pilar Sancho; Martinez-Ortiz, Ivan; Sierra, Jose Luis; Fernandez-Manjon, Baltasar
2007-01-01
In this paper, we propose three different ways of using IMS Learning Design to support online adaptive learning modules that include educational videogames. The first approach relies on IMS LD to support adaptation procedures where the educational games are considered as Learning Objects. These games can be included instead of traditional content…
Optimization of an organic memristor as an adaptive memory element
NASA Astrophysics Data System (ADS)
Berzina, Tatiana; Smerieri, Anteo; Bernabò, Marco; Pucci, Andrea; Ruggeri, Giacomo; Erokhin, Victor; Fontana, M. P.
2009-06-01
The combination of memory and signal handling characteristics of a memristor makes it a promising candidate for adaptive bioinspired information processing systems. This poses stringent requirements on the basic device, such as stability and reproducibility over a large number of training/learning cycles, and a large anisotropy in the fundamental control material parameter, in our case the electrical conductivity. In this work we report results on the improved performance of electrochemically controlled polymeric memristors, where optimization of a conducting polymer (polyaniline) in the active channel and better environmental control of fabrication methods led to a large increase both in the absolute values of the conductivity in the partially oxydized state of polyaniline and of the on-off conductivity ratio. These improvements are crucial for the application of the organic memristor to adaptive complex signal handling networks.
ERIC Educational Resources Information Center
Munoz-Organero, M.; Munoz-Merino, P. J.; Kloos, Carlos Delgado
2011-01-01
The use of technology in learning environments should be targeted at improving the learning outcome of the process. Several technology enhanced techniques can be used for maximizing the learning gain of particular students when having access to learning resources. One of them is content adaptation. Adapting content is especially important when…
Adaptivity in Game-Based Learning: A New Perspective on Story
NASA Astrophysics Data System (ADS)
Berger, Florian; Müller, Wolfgang
Game-based learning as a novel form of e-learning still has issues in fundamental questions, the lack of a general model for adaptivity being one of them. Since adaptive techniques in traditional e-learning applications bear close similarity to certain interactive storytelling approaches, we propose a new notion of story as the joining element of arbitraty learning paths.
Motor Learning in Childhood Reveals Distinct Mechanisms for Memory Retention and Re-Learning
ERIC Educational Resources Information Center
Musselman, Kristin E.; Roemmich, Ryan T.; Garrett, Ben; Bastian, Amy J.
2016-01-01
Adults can easily learn and access multiple versions of the same motor skill adapted for different conditions (e.g., walking in water, sand, snow). Following even a single session of adaptation, adults exhibit clear day-to-day retention and faster re-learning of the adapted pattern. Here, we studied the retention and re-learning of an adapted…
Enhancing Learning Performance and Adaptability for Complex Tasks
2005-03-30
development of active learning interventions and techniques that influence the focus and quality of learner regulatory activity (Kozlowski Toney et al...what are the effects of these goal representations on learning strategies, performance, and adaptability? Can active learning inductions, that influence...and mindful process - active learning - are generally associated with improved skill acquisition and adaptability for complex tasks (Smith et al
Towards adaptation in e-learning 2.0
NASA Astrophysics Data System (ADS)
Cristea, Alexandra I.; Ghali, Fawaz
2011-04-01
This paper presents several essential steps from an overall study on shaping new ways of learning and teaching, by using the synergetic merger of three different fields: Web 2.0, e-learning and adaptation (in particular, personalisation to the learner). These novel teaching and learning ways-the latter focus of this paper-are reflected in and finally adding to various versions of the My Online Teacher 2.0 adaptive system. In particular, this paper focuses on a study of how to more effectively use and combine the recommendation of peers and content adaptation to enhance the learning outcome in e-learning systems based on Web 2.0. In order to better isolate and examine the effects of peer recommendation and adaptive content presentation, we designed experiments inspecting collaboration between individuals based on recommendation of peers who have greater knowledge, and compare this to adaptive content recommendation, as well as to "simple" learning in a system with a minimum of Web 2.0 support. Overall, the results of adding peer recommendation and adaptive content presentation were encouraging, and are further discussed in detail in this paper.
NASA Astrophysics Data System (ADS)
Mulavara, Ajitkumar; Wood, Scott; Cohen, Helen; Bloomberg, Jacob
2012-07-01
Exposure to the microgravity conditions of space flight induces adaptive modification in sensorimotor function allowing astronauts to operate in this unique environment. This adaptive state, however, is inappropriate for a 1-g environment. Consequently astronauts must spend time readapting to Earth's gravity following their return to Earth. During this readaptation period, alterations in sensorimotor function cause various disturbances in astronaut gait during postflight walking. They often rely more on vision for postural and gait stability and many report the need for greater cognitive supervision of motor actions that previous to space flight were fully automated. Over the last several years our laboratory has investigated postflight astronaut locomotion with the aim of better understanding how adaptive changes in underlying sensorimotor mechanisms contribute to postflight gait dysfunction. Exposure to the microgravity conditions of space flight induces adaptive modification in the control of vestibularly-mediated reflexive head movement during locomotion after space flight. Furthermore, during motor learning, adaptive transitions are composed of two main mechanisms: strategic and plastic. Strategic mechanisms represent immediate and transitory modifications in control to deal with changes in the prevailing environment that, if prolonged, induce plastic mechanisms designed to automate new behavioral responses. The goal of the present study was to examine the contributions of sensorimotor subsystems such as the vestibular and body load sensing (BLS) somatosensory influences on head movement control during locomotion after long-duration space flight. Further we present data on the two motor learning processes during readaptation of locomotor function after long-duration space flight. Eighteen astronauts performed two tests of locomotion before and after 6 months of space flight: a treadmill walking test to examine vestibular reflexive mechanisms controlling head movement control and a functional mobility test to investigate overall functional locomotor ability. Postflight sessions were given on days 1, 2, 4, 7 after their return. Subjects walked on a treadmill driven at 1.8 m/s while performing a visual task. Motion data from head and trunk segmental motion data were obtained to calculate the angular head pitch (HP) movements during walking trials while subjects performed the visual task, to estimate the contributions of vestibular reflexive mechanisms in HP movements. Astronauts showed a heterogeneous response pattern of both increases and decreases in the amplitude of HP movement. We investigated the underlying mechanisms of this heterogeneity in postflight responses in head movement control by examining data obtained using the same experimental test paradigm on a vestibular clinical population (VC) and in normal subjects undergoing adaptation to acute body load support unloading. Results showed that exposure to unloaded locomotion caused a significant increase in HP movements, whereas in the VC patients the HP movements were significantly decreased. We infer that BLS-mediated somatosensory input centrally modulates vestibular input and can adaptively modify head-movement control during locomotion. Thus, space flight may cause a central adaptation of the converging vestibular and body load-sensing somatosensory systems. To investigate changes in functional mobility astronaut subjects walked at their preferred pace around an obstacle course consisting of several pylons and obstacles set up on a foam floor, which provided an unstable walking surface. Subjects were instructed to walk around the course as fast as possible without touching any of the objects on the course for a total of six individual trials per test session. One of the dependent measures was time to complete the course (TCC, sec). The learning rate over the six trials performed on preflight and the first day after landing (micro curve) was used to characterize the immediate compensatory strategic response. The learning rate over the six trials of the postflight test days (macro curve) was used to characterize the longer-term plastic response. Adaptation to space flight led to a 52% increase in TCC one day after landing. Recovery to pre-flight scores took an average of two weeks after landing. Subjects showed both strategic and plastic recovery patterns based on the slopes obtained from the micro and macro curves compared to preflight. A regression analysis revealed a significant correlation between the slope values of the macro and micro curves indicating a relationship between strategic and plastic recovery processes. Results showed that both strategic and plastic motor learning processes play a role in postflight restoration of functional mobility and showed a dynamic interplay between these two mechanisms during postflight recovery. These results suggest that gait adaptability training programs which are being developed to facilitate adaptive transition to planetary environments, coupled with low levels of electrical stimulation of the vestibular system, can be optimized to engage both strategic and plastic processes to facilitate rapid restoration of postflight functional mobility.
The Role of Motor Learning in Spatial Adaptation near a Tool
Brown, Liana E.; Doole, Robert; Malfait, Nicole
2011-01-01
Some visual-tactile (bimodal) cells have visual receptive fields (vRFs) that overlap and extend moderately beyond the skin of the hand. Neurophysiological evidence suggests, however, that a vRF will grow to encompass a hand-held tool following active tool use but not after passive holding. Why does active tool use, and not passive holding, lead to spatial adaptation near a tool? We asked whether spatial adaptation could be the result of motor or visual experience with the tool, and we distinguished between these alternatives by isolating motor from visual experience with the tool. Participants learned to use a novel, weighted tool. The active training group received both motor and visual experience with the tool, the passive training group received visual experience with the tool, but no motor experience, and finally, a no-training control group received neither visual nor motor experience using the tool. After training, we used a cueing paradigm to measure how quickly participants detected targets, varying whether the tool was placed near or far from the target display. Only the active training group detected targets more quickly when the tool was placed near, rather than far, from the target display. This effect of tool location was not present for either the passive-training or control groups. These results suggest that motor learning influences how visual space around the tool is represented. PMID:22174944
Adaptive quantum computation in changing environments using projective simulation
NASA Astrophysics Data System (ADS)
Tiersch, M.; Ganahl, E. J.; Briegel, H. J.
2015-08-01
Quantum information processing devices need to be robust and stable against external noise and internal imperfections to ensure correct operation. In a setting of measurement-based quantum computation, we explore how an intelligent agent endowed with a projective simulator can act as controller to adapt measurement directions to an external stray field of unknown magnitude in a fixed direction. We assess the agent’s learning behavior in static and time-varying fields and explore composition strategies in the projective simulator to improve the agent’s performance. We demonstrate the applicability by correcting for stray fields in a measurement-based algorithm for Grover’s search. Thereby, we lay out a path for adaptive controllers based on intelligent agents for quantum information tasks.
Adaptive quantum computation in changing environments using projective simulation
Tiersch, M.; Ganahl, E. J.; Briegel, H. J.
2015-01-01
Quantum information processing devices need to be robust and stable against external noise and internal imperfections to ensure correct operation. In a setting of measurement-based quantum computation, we explore how an intelligent agent endowed with a projective simulator can act as controller to adapt measurement directions to an external stray field of unknown magnitude in a fixed direction. We assess the agent’s learning behavior in static and time-varying fields and explore composition strategies in the projective simulator to improve the agent’s performance. We demonstrate the applicability by correcting for stray fields in a measurement-based algorithm for Grover’s search. Thereby, we lay out a path for adaptive controllers based on intelligent agents for quantum information tasks. PMID:26260263
Neuromodulation of reward-based learning and decision making in human aging
Eppinger, Ben; Hämmerer, Dorothea; Li, Shu-Chen
2013-01-01
In this paper, we review the current literature to highlight relations between age-associated declines in dopaminergic and serotonergic neuromodulation and adult age differences in adaptive goal-directed behavior. Specifically, we focus on evidence suggesting that deficits in neuromodulation contribute to older adults’ behavioral disadvantages in learning and decision making. These deficits are particularly pronounced when reward information is uncertain or the task context requires flexible adaptations to changing stimulus–reward contingencies. Moreover, emerging evidence points to age-related differences in the sensitivity to rewarding and aversive outcomes during learning and decision making if the acquisition of behavior critically depends on outcome processing. These age-related asymmetries in outcome valuation may be explained by age differences in the interplay of dopaminergic and serotonergic neuromodulation. This hypothesis is based on recent neurocomputational and psychopharmacological approaches, which suggest that dopamine and serotonin serve opponent roles in regulating the balance between approach behavior and inhibitory control. Studying adaptive regulation of behavior across the adult life span may shed new light on how the aging brain changes functionally in response to its diminishing resources. PMID:22023564
NASA Astrophysics Data System (ADS)
LaManna, Joseph C.; Sun, Xiaoyan; Ivy, Andre D.; Ward, Nicole L.
We have used a relatively simple model of hypoxia that triggers adaptive structural changes in the cerebral microvasculature to study the process of physiological angiogenesis. This model can be used to obtain mechanistic data for the processes that probably underlie the dynamic structural changes that occur in learning and the control of oxygen availability to the neurovascular unit. These mechanisms are broadly involved in a wide variety of pathophysiological processes. This is the vascular component to CNS functional plasticity, supporting learning and adaptation. The angiogenic process may wane with age, contributing to the decreasing ability to survive metabolic stress and the diminution of neuronal plasticity.
Yao, Bin; Han, Wei; Zeng, Lingxia; Guo, Xiong
2013-12-15
A retrospective nested case-control study was designed to explore whether freshman year mental health status and level of adaptation are predictors of Internet addiction. The study cohort was 977 college students at a university in northwest China. In the first college year, the students' mental health status and adaptation level were assessed using the Chinese College Student Mental Health Scale (CCSMHS) and the Chinese College Student Adjustment Scale (CCSAS). In the following 1-3 years, 62 Internet-addicted subjects were identified using Young's 8-item diagnostic questionnaire. Controls were matched for demographic characteristics. Using logistic regression analysis, freshman year mental health status, including factors such as somatization, anxiety, depression and self-contempt, and freshman year adaptive problems were found to be causal factors and predictors of Internet addiction. Freshman with features of depression, learning maladaptation and dissatisfaction could be an important target-intervention population for reducing Internet addiction. © 2013 Elsevier Ireland Ltd. All rights reserved.
Learner-Adaptive Educational Technology for Simulation in Healthcare: Foundations and Opportunities.
Lineberry, Matthew; Dev, Parvati; Lane, H Chad; Talbot, Thomas B
2018-06-01
Despite evidence that learners vary greatly in their learning needs, practical constraints tend to favor ''one-size-fits-all'' educational approaches, in simulation-based education as elsewhere. Adaptive educational technologies - devices and/or software applications that capture and analyze relevant data about learners to select and present individually tailored learning stimuli - are a promising aid in learners' and educators' efforts to provide learning experiences that meet individual needs. In this article, we summarize and build upon the 2017 Society for Simulation in Healthcare Research Summit panel discussion on adaptive learning. First, we consider the role of adaptivity in learning broadly. We then outline the basic functions that adaptive learning technologies must implement and the unique affordances and challenges of technology-based approaches for those functions, sharing an illustrative example from healthcare simulation. Finally, we consider future directions for accelerating research, development, and deployment of effective adaptive educational technology and techniques in healthcare simulation.
Fairbank, Michael; Li, Shuhui; Fu, Xingang; Alonso, Eduardo; Wunsch, Donald
2014-01-01
We present a recurrent neural-network (RNN) controller designed to solve the tracking problem for control systems. We demonstrate that a major difficulty in training any RNN is the problem of exploding gradients, and we propose a solution to this in the case of tracking problems, by introducing a stabilization matrix and by using carefully constrained context units. This solution allows us to achieve consistently lower training errors, and hence allows us to more easily introduce adaptive capabilities. The resulting RNN is one that has been trained off-line to be rapidly adaptive to changing plant conditions and changing tracking targets. The case study we use is a renewable-energy generator application; that of producing an efficient controller for a three-phase grid-connected converter. The controller we produce can cope with the random variation of system parameters and fluctuating grid voltages. It produces tracking control with almost instantaneous response to changing reference states, and virtually zero oscillation. This compares very favorably to the classical proportional integrator (PI) controllers, which we show produce a much slower response and settling time. In addition, the RNN we propose exhibits better learning stability and convergence properties, and can exhibit faster adaptation, than has been achieved with adaptive critic designs. Copyright © 2013 Elsevier Ltd. All rights reserved.
Shadmehr, Reza; Ohminami, Shinya; Tsutsumi, Ryosuke; Shirota, Yuichiro; Shimizu, Takahiro; Tanaka, Nobuyuki; Terao, Yasuo; Tsuji, Shoji; Ugawa, Yoshikazu; Uchimura, Motoaki; Inoue, Masato; Kitazawa, Shigeru
2015-01-01
Cerebellar damage can profoundly impair human motor adaptation. For example, if reaching movements are perturbed abruptly, cerebellar damage impairs the ability to learn from the perturbation-induced errors. Interestingly, if the perturbation is imposed gradually over many trials, people with cerebellar damage may exhibit improved adaptation. However, this result is controversial, since the differential effects of gradual vs. abrupt protocols have not been observed in all studies. To examine this question, we recruited patients with pure cerebellar ataxia due to cerebellar cortical atrophy (n = 13) and asked them to reach to a target while viewing the scene through wedge prisms. The prisms were computer controlled, making it possible to impose the full perturbation abruptly in one trial, or build up the perturbation gradually over many trials. To control visual feedback, we employed shutter glasses that removed visual feedback during the reach, allowing us to measure trial-by-trial learning from error (termed error-sensitivity), and trial-by-trial decay of motor memory (termed forgetting). We found that the patients benefited significantly from the gradual protocol, improving their performance with respect to the abrupt protocol by exhibiting smaller errors during the exposure block, and producing larger aftereffects during the postexposure block. Trial-by-trial analysis suggested that this improvement was due to increased error-sensitivity in the gradual protocol. Therefore, cerebellar patients exhibited an improved ability to learn from error if they experienced those errors gradually. This improvement coincided with increased error-sensitivity and was present in both groups of subjects, suggesting that control of error-sensitivity may be spared despite cerebellar damage. PMID:26311179
Peripheral Inflammation Undermines the Plasticity of the Isolated Spinal Cord
Huie, John R.; Grau, James W.
2009-01-01
Peripheral capsaicin treatment induces molecular changes that sensitize the responses of nociceptive neurons in the spinal dorsal horn. The current studies demonstrate that capsaicin also undermines the adaptive plasticity of the spinal cord, rendering the system incapable of learning a simple instrumental task. In these studies, male rats are transected at the second thoracic vertebra and are tested 24 to 48 hours later. During testing, subjects receive shock to one hindleg when it is extended (controllable stimulation). Rats quickly learn to maintain the leg in a flexed position. Rats that have been injected with capsaicin (1% or 3%) in the hindpaw fail to learn, even when tested on the leg contralateral to the injection. This learning deficit lasts at least 24 hours. Interestingly, training with controllable electrical stimulation prior to capsaicin administration protects the spinal cord against the maladaptive effects. Rats pretrained with controllable stimulation do not display a learning deficit or tactile allodynia. Moreover, controllable stimulation, combined with naltrexone, reverses the capsaicin-induced deficit. These data suggest that peripheral inflammation, accompanying spinal cord injuries, might have an adverse effect on recovery. PMID:18298266
Adaptive fuzzy controller for thermal comfort inside the air-conditioned automobile chamber
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tong, L.; Yu, B.; Chen, Z.
1999-07-01
In order to meet the passengers' demand for thermal comfort, the adaptive fuzzy logic control design methodology is applied for the automobile airconditioner system. In accordance with the theory of air flow and heat transfer, the air temperature field inside the airconditioned automobile chamber is simulated by a set of simplified half-empirical formula. Then, instead of PMV (Predicted Mean Vote) criterion, RIV (Real Individual Vote) criterion is adopted as the base of the control for passengers' thermal comfort. The proposed controller is applied to the air temperature regulation at the individual passenger position. The control procedure is based on partitioningmore » the state space of the system into cell-groups and fuzzily quantificating the state space into these cells. When the system model has some parameter perturbation, the controller can also adjust its control parameters to compensate for the perturbation and maintain the good performance. The learning procedure shows its ideal effect in both computer simulation and experiments. The final results demonstrate the ideal performance of this adaptive fuzzy controller.« less
The Influence of Student Characteristics on the Use of Adaptive E-Learning Material
ERIC Educational Resources Information Center
van Seters, J. R.; Ossevoort, M. A.; Tramper, J.; Goedhart, M. J.
2012-01-01
Adaptive e-learning materials can help teachers to educate heterogeneous student groups. This study provides empirical data about the way academic students differ in their learning when using adaptive e-learning materials. Ninety-four students participated in the study. We determined characteristics in a heterogeneous student group by collecting…
How Language Supports Adaptive Teaching through a Responsive Learning Culture
ERIC Educational Resources Information Center
Johnston, Peter; Dozier, Cheryl; Smit, Julie
2016-01-01
For students to learn optimally, teachers must design classrooms that are responsive to the full range of student development. The teacher must be adaptive, but so must each student and the learning culture itself. In other words, adaptive teaching means constructing a responsive learning culture that accommodates and even capitalizes on diversity…
Real-Time Global Nonlinear Aerodynamic Modeling for Learn-To-Fly
NASA Technical Reports Server (NTRS)
Morelli, Eugene A.
2016-01-01
Flight testing and modeling techniques were developed to accurately identify global nonlinear aerodynamic models for aircraft in real time. The techniques were developed and demonstrated during flight testing of a remotely-piloted subscale propeller-driven fixed-wing aircraft using flight test maneuvers designed to simulate a Learn-To-Fly scenario. Prediction testing was used to evaluate the quality of the global models identified in real time. The real-time global nonlinear aerodynamic modeling algorithm will be integrated and further tested with learning adaptive control and guidance for NASA Learn-To-Fly concept flight demonstrations.
Performance & Emotion--A Study on Adaptive E-Learning Based on Visual/Verbal Learning Styles
ERIC Educational Resources Information Center
Beckmann, Jennifer; Bertel, Sven; Zander, Steffi
2015-01-01
Adaptive e-Learning systems are able to adjust to a user's learning needs, usually by user modeling or tracking progress. Such learner-adaptive behavior has rapidly become a hot topic for e-Learning, furthered in part by the recent rapid increase in the use of MOOCs (Massive Open Online Courses). A lack of general, individual, and situational data…
Event-Based Robust Control for Uncertain Nonlinear Systems Using Adaptive Dynamic Programming.
Zhang, Qichao; Zhao, Dongbin; Wang, Ding
2018-01-01
In this paper, the robust control problem for a class of continuous-time nonlinear system with unmatched uncertainties is investigated using an event-based control method. First, the robust control problem is transformed into a corresponding optimal control problem with an augmented control and an appropriate cost function. Under the event-based mechanism, we prove that the solution of the optimal control problem can asymptotically stabilize the uncertain system with an adaptive triggering condition. That is, the designed event-based controller is robust to the original uncertain system. Note that the event-based controller is updated only when the triggering condition is satisfied, which can save the communication resources between the plant and the controller. Then, a single network adaptive dynamic programming structure with experience replay technique is constructed to approach the optimal control policies. The stability of the closed-loop system with the event-based control policy and the augmented control policy is analyzed using the Lyapunov approach. Furthermore, we prove that the minimal intersample time is bounded by a nonzero positive constant, which excludes Zeno behavior during the learning process. Finally, two simulation examples are provided to demonstrate the effectiveness of the proposed control scheme.
Krause, Mark A
2015-07-01
Inquiry into evolutionary adaptations has flourished since the modern synthesis of evolutionary biology. Comparative methods, genetic techniques, and various experimental and modeling approaches are used to test adaptive hypotheses. In psychology, the concept of adaptation is broadly applied and is central to comparative psychology and cognition. The concept of an adaptive specialization of learning is a proposed account for exceptions to general learning processes, as seen in studies of Pavlovian conditioning of taste aversions, sexual responses, and fear. The evidence generally consists of selective associations forming between biologically relevant conditioned and unconditioned stimuli, with conditioned responses differing in magnitude, persistence, or other measures relative to non-biologically relevant stimuli. Selective associations for biologically relevant stimuli may suggest adaptive specializations of learning, but do not necessarily confirm adaptive hypotheses as conceived of in evolutionary biology. Exceptions to general learning processes do not necessarily default to an adaptive specialization explanation, even if experimental results "make biological sense". This paper examines the degree to which hypotheses of adaptive specializations of learning in sexual and fear response systems have been tested using methodologies developed in evolutionary biology (e.g., comparative methods, quantitative and molecular genetics, survival experiments). A broader aim is to offer perspectives from evolutionary biology for testing adaptive hypotheses in psychological science.
Reinforcement learning or active inference?
Friston, Karl J; Daunizeau, Jean; Kiebel, Stefan J
2009-07-29
This paper questions the need for reinforcement learning or control theory when optimising behaviour. We show that it is fairly simple to teach an agent complicated and adaptive behaviours using a free-energy formulation of perception. In this formulation, agents adjust their internal states and sampling of the environment to minimize their free-energy. Such agents learn causal structure in the environment and sample it in an adaptive and self-supervised fashion. This results in behavioural policies that reproduce those optimised by reinforcement learning and dynamic programming. Critically, we do not need to invoke the notion of reward, value or utility. We illustrate these points by solving a benchmark problem in dynamic programming; namely the mountain-car problem, using active perception or inference under the free-energy principle. The ensuing proof-of-concept may be important because the free-energy formulation furnishes a unified account of both action and perception and may speak to a reappraisal of the role of dopamine in the brain.
Gómez-Moya, Rosinna; Díaz, Rosalinda; Fernandez-Ruiz, Juan
2016-04-01
Different processes are involved during visuomotor learning, including an error-based procedural and a strategy based cognitive mechanism. Our objective was to analyze if the changes in the adaptation or the aftereffect components of visuomotor learning measured across development, reflected different maturation rates of the aforementioned mechanisms. Ninety-five healthy children aged 4-12years and a group of young adults participated in a wedge prism and a dove prism throwing task, which laterally displace or horizontally reverse the visual field respectively. The results show that despite the age-related differences in motor control, all children groups adapted in the error-based wedge prisms condition. However, when removing the prism, small children showed a slower aftereffects extinction rate. On the strategy-based visual reversing task only the older children group reached adult-like levels. These results are consistent with the idea of different mechanisms with asynchronous maturation rates participating during visuomotor learning. Copyright © 2016 Elsevier B.V. All rights reserved.
Krakauer, John W.; Mazzoni, Pietro
2012-01-01
The public pays large sums of money to watch skilled motor performance. Notably, however, in recent decades motor skill learning (performance improvement beyond baseline levels) has received less experimental attention than motor adaptation (return to baseline performance in the setting of an external perturbation). Motor skill can be assessed at the levels of task success and movement quality, but the link between these levels remains poorly understood. We devised a motor skill task that required visually guided curved movements of the wrist without a perturbation, and we defined skill learning at the task level as a change in the speed–accuracy trade-off function (SAF). Practice in restricted speed ranges led to a global shift of the SAF. We asked how the SAF shift maps onto changes in trajectory kinematics, to establish a link between task-level performance and fine motor control. Although there were small changes in mean trajectory, improved performance largely consisted of reduction in trial-to-trial variability and increase in movement smoothness. We found evidence for improved feedback control, which could explain the reduction in variability but does not preclude other explanations such as an increased signal-to-noise ratio in cortical representations. Interestingly, submovement structure remained learning invariant. The global generalization of the SAF across a wide range of difficulty suggests that skill for this task is represented in a temporally scalable network. We propose that motor skill acquisition can be characterized as a slow reduction in movement variability, which is distinct from faster model-based learning that reduces systematic error in adaptation paradigms. PMID:22514286
Fostering cultural inclusiveness and learning in culturally mixed business classes.
Mak, Anita S; Daly, Anne; Barker, Michelle C
2014-01-01
Business educators have advocated that in order to build faculty's intercultural capability, it is vital to provide them with professional development in using intercultural training resources and with "community of practice" support in adapting such resources for enhancing their students' intercultural learning. This approach has been adopted in an Australian action research project titled "Internationalisation at Home" (IaH), which involved providing faculty with professional development adapted from an established intercultural training resource - the EXCELL (Excellence in Cultural Experiential Learning and Leadership) Program. In this paper, we present two case studies of the implementation of the IaH Project in business schools at the University of Canberra and at Griffith University. Lessons learned from the first study were incorporated in the design and evaluation of the second one. Faculty leaders will describe how they engage and support colleagues in adapting components of EXCELL to foster cultural inclusiveness and facilitate students' intercultural competence development. As part of project evaluation, we hypothesised that students who participated in IaH courses would report greater levels of (1) cultural inclusiveness in their educational environment, and (2) cultural learning development, compared with students who were not enrolled in IaH courses. Research participants in the Canberra case study comprised an intervention group of 140 business undergraduates enrolled in an IaH course, and a control group of 59 non-IaH undergraduates. At Griffith, participants were 211 first year management students in the intervention group and 84 students enrolled in a non-IaH first year course. In each case study, an end-of-semester survey showed that students who had completed courses with the IaH project intervention reported significantly greater levels of perceived cultural inclusiveness in multicultural classes, and of cultural learning development, than students in the control group. Faculty's reflections on project processes and outcomes further suggest that implementing strategic, structured active learning interventions such as in the IaH Project, could bring about more productive social interactions in multicultural classes and benefit domestic and especially international students. We will discuss implications of the findings for students' intercultural learning, faculty's needs for continual professional development, and the role of institutional support in intercultural competence development.
Bell, Adrian Viliami; Hernandez, Daniel
2017-03-01
Understanding the prevalence of adaptive culture in part requires understanding the dynamics of learning. Here we explore the adaptive value of social learning in groups and how formal social groups function as effective mediums of information exchange. We discuss the education literature on Cooperative Learning Groups (CLGs), which outlines the potential of group learning for enhancing learning outcomes. Four qualities appear essential for CLGs to enhance learning: (1) extended conversations, (2) regular interactions, (3) gathering of experts, and (4) incentives for sharing knowledge. We analyze these four qualities within the context of a small-scale agricultural society using data we collected in 2010 and 2012. Through an analysis of surveys, interviews, and observations in the Tongan islands, we describe the role CLGs likely plays in facilitating individuals' learning of adaptive information. Our analysis of group affiliation, membership, and topics of conversation suggest that the first three CLG qualities reflect conditions for adaptive learning in groups. We utilize ethnographic anecdotes to suggest the fourth quality is also conducive to adaptive group learning. Using an evolutionary model, we further explore the scope for CLGs outside the Tongan socioecological context. Model analysis shows that environmental volatility and migration rates among human groups mediate the scope for CLGs. We call for wider attention to how group structure facilitates learning in informal settings, which may be key to assessing the contribution of groups to the evolution of complex, adaptive culture.
Manoonpong, Poramate; Parlitz, Ulrich; Wörgötter, Florentin
2013-01-01
Living creatures, like walking animals, have found fascinating solutions for the problem of locomotion control. Their movements show the impression of elegance including versatile, energy-efficient, and adaptable locomotion. During the last few decades, roboticists have tried to imitate such natural properties with artificial legged locomotion systems by using different approaches including machine learning algorithms, classical engineering control techniques, and biologically-inspired control mechanisms. However, their levels of performance are still far from the natural ones. By contrast, animal locomotion mechanisms seem to largely depend not only on central mechanisms (central pattern generators, CPGs) and sensory feedback (afferent-based control) but also on internal forward models (efference copies). They are used to a different degree in different animals. Generally, CPGs organize basic rhythmic motions which are shaped by sensory feedback while internal models are used for sensory prediction and state estimations. According to this concept, we present here adaptive neural locomotion control consisting of a CPG mechanism with neuromodulation and local leg control mechanisms based on sensory feedback and adaptive neural forward models with efference copies. This neural closed-loop controller enables a walking machine to perform a multitude of different walking patterns including insect-like leg movements and gaits as well as energy-efficient locomotion. In addition, the forward models allow the machine to autonomously adapt its locomotion to deal with a change of terrain, losing of ground contact during stance phase, stepping on or hitting an obstacle during swing phase, leg damage, and even to promote cockroach-like climbing behavior. Thus, the results presented here show that the employed embodied neural closed-loop system can be a powerful way for developing robust and adaptable machines. PMID:23408775
Simon, Anja; Bock, Otmar
2015-01-01
A new 3-stage model based on neuroimaging evidence is proposed by Chein and Schneider (2012). Each stage is associated with different brain regions, and draws on cognitive abilities: the first stage on creativity, the second on selective attention, and the third on automatic processing. The purpose of the present study was to scrutinize the validity of this model for 1 popular learning paradigm, visuomotor adaptation. Participants completed tests for creativity, selective attention and automated processing before attending in a pointing task with adaptation to a 60° rotation of visual feedback. To examine the relationship between cognitive abilities and motor learning at different times of practice, associations between cognitive and adaptation scores were calculated repeatedly throughout adaptation. The authors found no benefit of high creativity for adaptive performance. High levels of selective attention were positively associated with early adaptation, but hardly with late adaptation and de-adaptation. High levels of automated execution were beneficial for late adaptation, but hardly for early and de-adaptation. From this we conclude that Chein and Schneider's first learning stage is difficult to confirm by research on visuomotor adaptation, and that the other 2 learning stages rather relate to workaround strategies than to actual adaptive recalibration.
A New Powered Lower Limb Prosthesis Control Framework Based on Adaptive Dynamic Programming.
Wen, Yue; Si, Jennie; Gao, Xiang; Huang, Stephanie; Huang, He Helen
2017-09-01
This brief presents a novel application of adaptive dynamic programming (ADP) for optimal adaptive control of powered lower limb prostheses, a type of wearable robots to assist the motor function of the limb amputees. Current control of these robotic devices typically relies on finite state impedance control (FS-IC), which lacks adaptability to the user's physical condition. As a result, joint impedance settings are often customized manually and heuristically in clinics, which greatly hinder the wide use of these advanced medical devices. This simulation study aimed at demonstrating the feasibility of ADP for automatic tuning of the twelve knee joint impedance parameters during a complete gait cycle to achieve balanced walking. Given that the accurate models of human walking dynamics are difficult to obtain, the model-free ADP control algorithms were considered. First, direct heuristic dynamic programming (dHDP) was applied to the control problem, and its performance was evaluated on OpenSim, an often-used dynamic walking simulator. For the comparison purposes, we selected another established ADP algorithm, the neural fitted Q with continuous action (NFQCA). In both cases, the ADP controllers learned to control the right knee joint and achieved balanced walking, but dHDP outperformed NFQCA in this application during a 200 gait cycle-based testing.
Adaptive control with an expert system based supervisory level. Thesis
NASA Technical Reports Server (NTRS)
Sullivan, Gerald A.
1991-01-01
Adaptive control is presently one of the methods available which may be used to control plants with poorly modelled dynamics or time varying dynamics. Although many variations of adaptive controllers exist, a common characteristic of all adaptive control schemes, is that input/output measurements from the plant are used to adjust a control law in an on-line fashion. Ideally the adjustment mechanism of the adaptive controller is able to learn enough about the dynamics of the plant from input/output measurements to effectively control the plant. In practice, problems such as measurement noise, controller saturation, and incorrect model order, to name a few, may prevent proper adjustment of the controller and poor performance or instability result. In this work we set out to avoid the inadequacies of procedurally implemented safety nets, by introducing a two level control scheme in which an expert system based 'supervisor' at the upper level provides all the safety net functions for an adaptive controller at the lower level. The expert system is based on a shell called IPEX, (Interactive Process EXpert), that we developed specifically for the diagnosis and treatment of dynamic systems. Some of the more important functions that the IPEX system provides are: (1) temporal reasoning; (2) planning of diagnostic activities; and (3) interactive diagnosis. Also, because knowledge and control logic are separate, the incorporation of new diagnostic and treatment knowledge is relatively simple. We note that the flexibility available in the system to express diagnostic and treatment knowledge, allows much greater functionality than could ever be reasonably expected from procedural implementations of safety nets. The remainder of this chapter is divided into three sections. In section 1.1 we give a detailed review of the literature in the area of supervisory systems for adaptive controllers. In particular, we describe the evolution of safety nets from simple ad hoc techniques, up to the use of expert systems for more advanced supervision capabilities.
ERIC Educational Resources Information Center
Schellings, Gonny L. M.; Broekkamp, Hein
2011-01-01
Self-regulated learning has been described as an adaptive process: students adapt their learning strategies for attaining different learning goals. In order to be adaptive, students must have a clear notion of what the task requirements consist of. Both trace data and questionnaire data indicate that students adapt study strategies in limited ways…
Towards Motivation-Based Adaptation of Difficulty in E-Learning Programs
ERIC Educational Resources Information Center
Endler, Anke; Rey, Gunter Daniel; Butz, Martin V.
2012-01-01
The objective of this study was to investigate if an e-learning environment may use measurements of the user's current motivation to adapt the level of task difficulty for more effective learning. In the reported study, motivation-based adaptation was applied randomly to collect a wide range of data for different adaptations in a variety of…
Adaptable Learning Pathway Generation with Ant Colony Optimization
ERIC Educational Resources Information Center
Wong, Lung-Hsiang; Looi, Chee-Kit
2009-01-01
One of the new major directions in research on web-based educational systems is the notion of adaptability: the educational system adapts itself to the learning profile, preferences and ability of the student. In this paper, we look into the issues of providing adaptability with respect to learning pathways. We explore the state of the art with…
Dynamic Cost Risk Assessment for Controlling the Cost of Naval Vessels
2008-04-23
for each individual RRA at the start of the project are depicted in Figures 2a and 2b, respectively. The PDFs are multimodal and cannot be...underestimates cost. 7 Probabilistic cost analysis A physician metaphor Adapted from Yacov Y. Haimes, NPS 2007 8 Dynamic cost risk management A physician... metaphor Adapted from Yacov Y. Haimes, NPS 2007 9 Sources of cost uncertainty Macroscopic analysis Economic, Materials & Labor, Learning rates
Simultaneous vibration control and energy harvesting using actor-critic based reinforcement learning
NASA Astrophysics Data System (ADS)
Loong, Cheng Ning; Chang, C. C.; Dimitrakopoulos, Elias G.
2018-03-01
Mitigating excessive vibration of civil engineering structures using various types of devices has been a conspicuous research topic in the past few decades. Some devices, such as electromagnetic transducers, which have a capability of exerting control forces while simultaneously harvesting energy, have been proposed recently. These devices make possible a self-regenerative system that can semi-actively mitigate structural vibration without the need of external energy. Integrating mechanical, electrical components, and control algorithms, these devices open up a new research domain that needs to be addressed. In this study, the feasibility of using an actor-critic based reinforcement learning control algorithm for simultaneous vibration control and energy harvesting for a civil engineering structure is investigated. The actor-critic based reinforcement learning control algorithm is a real-time, model-free adaptive technique that can adjust the controller parameters based on observations and reward signals without knowing the system characteristics. It is suitable for the control of a partially known nonlinear system with uncertain parameters. The feasibility of implementing this algorithm on a building structure equipped with an electromagnetic damper will be investigated in this study. Issues related to the modelling of learning algorithm, initialization and convergence will be presented and discussed.
Water in Fire Control--Basic Training Course.
ERIC Educational Resources Information Center
Forest Service (USDA), Washington, DC.
Prepared by a team of fire control officers, the training guide is designed to help fire crewmen learn the fundamentals of water use. The entire package can be used for a complete course or individual lessons and can be adapted to specific training needs. Throughout the guide, emphasis is placed on one primary training objective, performance in…
NASA Astrophysics Data System (ADS)
Marin, Nilo Eric
This study was conducted to determine if the use of the technology known as Classroom Performance System (CPS), specifically referred to as "Clickers", improves the learning gains of students enrolled in a biology course for science majors. CPS is one of a group of developing technologies adapted for providing feedback in the classroom using a learner-centered approach. It supports and facilitates discussion among students and between them and teachers, and provides for participation by passive students. Advocates, influenced by constructivist theories, claim increased academic achievement. In science teaching, the results have been mixed, but there is some evidence of improvements in conceptual understanding. The study employed a pretest-posttest, non-equivalent groups experimental design. The sample consisted of 226 participants in six sections of a college biology course at a large community college in South Florida with two instructors trained in the use of clickers. Each instructor randomly selected their sections into CPS (treatment) and non-CPS (control) groups. All participants filled out a survey that included demographic data at the beginning of the semester. The treatment group used clicker questions throughout, with discussions as necessary, whereas the control groups answered the same questions as quizzes, similarly engaging in discussion where necessary. The learning gains were assessed on a pre/post-test basis. The average learning gains, defined as the actual gain divided by the possible gain, were slightly better in the treatment group than in the control group, but the difference was statistically non-significant. An Analysis of Covariance (ANCOVA) statistic with pretest scores as the covariate was conducted to test for significant differences between the treatment and control groups on the posttest. A second ANCOVA was used to determine the significance of differences between the treatment and control groups on the posttest scores, after controlling for sex, GPA, academic status, experience with clickers, and instructional style. The results indicated a small increase in learning gains but these were not statistically significant. The data did not support an increase in learning based on the use of the CPS technology. This study adds to the body of research that questions whether CPS technology merits classroom adaptation.
Authoring Adaptive 3D Virtual Learning Environments
ERIC Educational Resources Information Center
Ewais, Ahmed; De Troyer, Olga
2014-01-01
The use of 3D and Virtual Reality is gaining interest in the context of academic discussions on E-learning technologies. However, the use of 3D for learning environments also has drawbacks. One way to overcome these drawbacks is by having an adaptive learning environment, i.e., an environment that dynamically adapts to the learner and the…
Jagannathan, Sarangapani; He, Pingan
2008-12-01
In this paper, a suite of adaptive neural network (NN) controllers is designed to deliver a desired tracking performance for the control of an unknown, second-order, nonlinear discrete-time system expressed in nonstrict feedback form. In the first approach, two feedforward NNs are employed in the controller with tracking error as the feedback variable whereas in the adaptive critic NN architecture, three feedforward NNs are used. In the adaptive critic architecture, two action NNs produce virtual and actual control inputs, respectively, whereas the third critic NN approximates certain strategic utility function and its output is employed for tuning action NN weights in order to attain the near-optimal control action. Both the NN control methods present a well-defined controller design and the noncausal problem in discrete-time backstepping design is avoided via NN approximation. A comparison between the controller methodologies is highlighted. The stability analysis of the closed-loop control schemes is demonstrated. The NN controller schemes do not require an offline learning phase and the NN weights can be initialized at zero or random. Results show that the performance of the proposed controller schemes is highly satisfactory while meeting the closed-loop stability.
An adaptive brain actuated system for augmenting rehabilitation
Roset, Scott A.; Gant, Katie; Prasad, Abhishek; Sanchez, Justin C.
2014-01-01
For people living with paralysis, restoration of hand function remains the top priority because it leads to independence and improvement in quality of life. In approaches to restore hand and arm function, a goal is to better engage voluntary control and counteract maladaptive brain reorganization that results from non-use. Standard rehabilitation augmented with developments from the study of brain-computer interfaces could provide a combined therapy approach for motor cortex rehabilitation and to alleviate motor impairments. In this paper, an adaptive brain-computer interface system intended for application to control a functional electrical stimulation (FES) device is developed as an experimental test bed for augmenting rehabilitation with a brain-computer interface. The system's performance is improved throughout rehabilitation by passive user feedback and reinforcement learning. By continuously adapting to the user's brain activity, similar adaptive systems could be used to support clinical brain-computer interface neurorehabilitation over multiple days. PMID:25565945
Effectiveness of a Multisystem Aquatic Therapy for Children with Autism Spectrum Disorders.
Caputo, Giovanni; Ippolito, Giovanni; Mazzotta, Marina; Sentenza, Luigi; Muzio, Mara Rosaria; Salzano, Sara; Conson, Massimiliano
2018-06-01
Aquatic therapy improves motor skills of persons with Autism Spectrum Disorders (ASD), but its usefulness for treating functional difficulties needs to be verified yet. We tested effectiveness of a multisystem aquatic therapy on behavioural, emotional, social and swimming skills of children with ASD. Multisystem aquatic therapy was divided in three phases (emotional adaptation, swimming adaptation and social integration) implemented in a 10-months-programme. At post-treatment, the aquatic therapy group showed significant improvements relative to controls on functional adaptation (Vineland Adaptive Behavior Scales), emotional response, adaptation to change and on activity level (Childhood Autism Rating Scale). Swimming skills learning was also demonstrated. Multisystem aquatic therapy is useful for ameliorating functional impairments of children with ASD, going well beyond a swimming training.
NASA Astrophysics Data System (ADS)
Kelley, Troy D.; McGhee, S.
2013-05-01
This paper describes the ongoing development of a robotic control architecture that inspired by computational cognitive architectures from the discipline of cognitive psychology. The Symbolic and Sub-Symbolic Robotics Intelligence Control System (SS-RICS) combines symbolic and sub-symbolic representations of knowledge into a unified control architecture. The new architecture leverages previous work in cognitive architectures, specifically the development of the Adaptive Character of Thought-Rational (ACT-R) and Soar. This paper details current work on learning from episodes or events. The use of episodic memory as a learning mechanism has, until recently, been largely ignored by computational cognitive architectures. This paper details work on metric level episodic memory streams and methods for translating episodes into abstract schemas. The presentation will include research on learning through novelty and self generated feedback mechanisms for autonomous systems.
Does Artificial Tutoring Foster Inquiry Based Learning?
ERIC Educational Resources Information Center
Schmoelz, Alexander; Swertz, Christian; Forstner, Alexandra; Barberi, Alessandro
2014-01-01
This contribution looks at the Intelligent Tutoring Interface for Technology Enhanced Learning, which integrates multistage-learning and inquiry-based learning in an adaptive e-learning system. Based on a common pedagogical ontology, adaptive e-learning systems can be enabled to recommend learning objects and activities, which follow inquiry-based…
Si, Wenjie; Dong, Xunde; Yang, Feifei
2018-03-01
This paper is concerned with the problem of decentralized adaptive backstepping state-feedback control for uncertain high-order large-scale stochastic nonlinear time-delay systems. For the control design of high-order large-scale nonlinear systems, only one adaptive parameter is constructed to overcome the over-parameterization, and neural networks are employed to cope with the difficulties raised by completely unknown system dynamics and stochastic disturbances. And then, the appropriate Lyapunov-Krasovskii functional and the property of hyperbolic tangent functions are used to deal with the unknown unmatched time-delay interactions of high-order large-scale systems for the first time. At last, on the basis of Lyapunov stability theory, the decentralized adaptive neural controller was developed, and it decreases the number of learning parameters. The actual controller can be designed so as to ensure that all the signals in the closed-loop system are semi-globally uniformly ultimately bounded (SGUUB) and the tracking error converges in the small neighborhood of zero. The simulation example is used to further show the validity of the design method. Copyright © 2018 Elsevier Ltd. All rights reserved.
The gravitational field and brain function.
Mei, L; Zhou, C D; Lan, J Q; Wang, Z G; Wu, W C; Xue, X M
1983-01-01
The frontal cortex is recognized as the highest adaptive control center of the human brain. The principle of the "frontalization" of human brain function offers new possibilities for brain research in space. There is evolutionary and experimental evidence indicating the validity of the principle, including it's role in nervous response to gravitational stimulation. The gravitational field is considered here as one of the more constant and comprehensive factors acting on brain evolution, which has undergone some successive crucial steps: "encephalization", "corticalization", "lateralization" and "frontalization". The dominating effects of electrical responses from the frontal cortex have been discovered 1) in experiments under gravitational stimulus; and 2) in processes potentially relating to gravitational adaptation, such as memory and learning, sensory information processing, motor programing, and brain state control. A brain research experiment during space flight is suggested to test the role of the frontal cortex in space adaptation and it's potentiality in brain control.
BCI Use and Its Relation to Adaptation in Cortical Networks.
Casimo, Kaitlyn; Weaver, Kurt E; Wander, Jeremiah; Ojemann, Jeffrey G
2017-10-01
Brain-computer interfaces (BCIs) carry great potential in the treatment of motor impairments. As a new motor output, BCIs interface with the native motor system, but acquisition of BCI proficiency requires a degree of learning to integrate this new function. In this review, we discuss how BCI designs often take advantage of the brain's motor system infrastructure as sources of command signals. We highlight a growing body of literature examining how this approach leads to changes in activity across cortex, including beyond motor regions, as a result of learning the new skill of BCI control. We discuss the previous research identifying patterns of neural activity associated with BCI skill acquisition and use that closely resembles those associated with learning traditional native motor tasks. We then discuss recent work in animals probing changes in connectivity of the BCI control site, which were linked to BCI skill acquisition, and use this as a foundation for our original work in humans. We present our novel work showing changes in resting state connectivity across cortex following the BCI learning process. We find substantial, heterogeneous changes in connectivity across regions and frequencies, including interactions that do not involve the BCI control site. We conclude from our review and original work that BCI skill acquisition may potentially lead to significant changes in evoked and resting state connectivity across multiple cortical regions. We recommend that future studies of BCIs look beyond motor regions to fully describe the cortical networks involved and long-term adaptations resulting from BCI skill acquisition.
Learning to Control Advanced Life Support Systems
NASA Technical Reports Server (NTRS)
Subramanian, Devika
2004-01-01
Advanced life support systems have many interacting processes and limited resources. Controlling and optimizing advanced life support systems presents unique challenges. In particular, advanced life support systems are nonlinear coupled dynamical systems and it is difficult for humans to take all interactions into account to design an effective control strategy. In this project. we developed several reinforcement learning controllers that actively explore the space of possible control strategies, guided by rewards from a user specified long term objective function. We evaluated these controllers using a discrete event simulation of an advanced life support system. This simulation, called BioSim, designed by Nasa scientists David Kortenkamp and Scott Bell has multiple, interacting life support modules including crew, food production, air revitalization, water recovery, solid waste incineration and power. They are implemented in a consumer/producer relationship in which certain modules produce resources that are consumed by other modules. Stores hold resources between modules. Control of this simulation is via adjusting flows of resources between modules and into/out of stores. We developed adaptive algorithms that control the flow of resources in BioSim. Our learning algorithms discovered several ingenious strategies for maximizing mission length by controlling the air and water recycling systems as well as crop planting schedules. By exploiting non-linearities in the overall system dynamics, the learned controllers easily out- performed controllers written by human experts. In sum, we accomplished three goals. We (1) developed foundations for learning models of coupled dynamical systems by active exploration of the state space, (2) developed and tested algorithms that learn to efficiently control air and water recycling processes as well as crop scheduling in Biosim, and (3) developed an understanding of the role machine learning in designing control systems for advanced life support.
Kobza, Stefan; Ferrea, Stefano; Schnitzler, Alfons; Pollok, Bettina
2012-01-01
Feedback to both actively performed and observed behaviour allows adaptation of future actions. Positive feedback leads to increased activity of dopamine neurons in the substantia nigra, whereas dopamine neuron activity is decreased following negative feedback. Dopamine level reduction in unmedicated Parkinson’s Disease patients has been shown to lead to a negative learning bias, i.e. enhanced learning from negative feedback. Recent findings suggest that the neural mechanisms of active and observational learning from feedback might differ, with the striatum playing a less prominent role in observational learning. Therefore, it was hypothesized that unmedicated Parkinson’s Disease patients would show a negative learning bias only in active but not in observational learning. In a between-group design, 19 Parkinson’s Disease patients and 40 healthy controls engaged in either an active or an observational probabilistic feedback-learning task. For both tasks, transfer phases aimed to assess the bias to learn better from positive or negative feedback. As expected, actively learning patients showed a negative learning bias, whereas controls learned better from positive feedback. In contrast, no difference between patients and controls emerged for observational learning, with both groups showing better learning from positive feedback. These findings add to neural models of reinforcement-learning by suggesting that dopamine-modulated input to the striatum plays a minor role in observational learning from feedback. Future research will have to elucidate the specific neural underpinnings of observational learning. PMID:23185586
Development of Adaptive Kanji Learning System for Mobile Phone
ERIC Educational Resources Information Center
Li, Mengmeng; Ogata, Hiroaki; Hou, Bin; Hashimoto, Satoshi; Liu, Yuqin; Uosaki, Noriko; Yano, Yoneo
2010-01-01
This paper describes an adaptive learning system based on mobile phone email to support the study of Japanese Kanji. In this study, the main emphasis is on using the adaptive learning to resolve one common problem of the mobile-based email or SMS language learning systems. To achieve this goal, the authors main efforts focus on three aspects:…
An adaptive deep Q-learning strategy for handwritten digit recognition.
Qiao, Junfei; Wang, Gongming; Li, Wenjing; Chen, Min
2018-02-22
Handwritten digits recognition is a challenging problem in recent years. Although many deep learning-based classification algorithms are studied for handwritten digits recognition, the recognition accuracy and running time still need to be further improved. In this paper, an adaptive deep Q-learning strategy is proposed to improve accuracy and shorten running time for handwritten digit recognition. The adaptive deep Q-learning strategy combines the feature-extracting capability of deep learning and the decision-making of reinforcement learning to form an adaptive Q-learning deep belief network (Q-ADBN). First, Q-ADBN extracts the features of original images using an adaptive deep auto-encoder (ADAE), and the extracted features are considered as the current states of Q-learning algorithm. Second, Q-ADBN receives Q-function (reward signal) during recognition of the current states, and the final handwritten digits recognition is implemented by maximizing the Q-function using Q-learning algorithm. Finally, experimental results from the well-known MNIST dataset show that the proposed Q-ADBN has a superiority to other similar methods in terms of accuracy and running time. Copyright © 2018 Elsevier Ltd. All rights reserved.
Compensation of significant parametric uncertainties using sliding mode online learning
NASA Astrophysics Data System (ADS)
Schnetter, Philipp; Kruger, Thomas
An augmented nonlinear inverse dynamics (NID) flight control strategy using sliding mode online learning for a small unmanned aircraft system (UAS) is presented. Because parameter identification for this class of aircraft often is not valid throughout the complete flight envelope, aerodynamic parameters used for model based control strategies may show significant deviations. For the concept of feedback linearization this leads to inversion errors that in combination with the distinctive susceptibility of small UAS towards atmospheric turbulence pose a demanding control task for these systems. In this work an adaptive flight control strategy using feedforward neural networks for counteracting such nonlinear effects is augmented with the concept of sliding mode control (SMC). SMC-learning is derived from variable structure theory. It considers a neural network and its training as a control problem. It is shown that by the dynamic calculation of the learning rates, stability can be guaranteed and thus increase the robustness against external disturbances and system failures. With the resulting higher speed of convergence a wide range of simultaneously occurring disturbances can be compensated. The SMC-based flight controller is tested and compared to the standard gradient descent (GD) backpropagation algorithm under the influence of significant model uncertainties and system failures.
Motor learning and consolidation: the case of visuomotor rotation.
Krakauer, John W
2009-01-01
Adaptation to visuomotor rotation is a particular form of motor learning distinct from force-field adaptation, sequence learning, and skill learning. Nevertheless, study of adaptation to visuomotor rotation has yielded a number of findings and principles that are likely of general importance to procedural learning and memory. First, rotation learning is implicit and appears to proceed through reduction in a visual prediction error generated by a forward model, such implicit adaptation occurs even when it is in conflict with an explicit task goal. Second, rotation learning is subject to different forms of interference: retrograde, anterograde through aftereffects, and contextual blocking of retrieval. Third, opposite rotations can be recalled within a short time interval without interference if implicit contextual cues (effector change) rather than explicit cues (color change) are used. Fourth, rotation learning consolidates both over time and with increased initial training (saturation learning).
Effect of Concurrent Visual Feedback Frequency on Postural Control Learning in Adolescents.
Marco-Ahulló, Adrià; Sánchez-Tormo, Alexis; García-Pérez, José A; Villarrasa-Sapiña, Israel; González, Luis M; García-Massó, Xavier
2018-04-13
The purpose was to find better augmented visual feedback frequency (100% or 67%) for learning a balance task in adolescents. Thirty subjects were divided randomly into a control group, and 100% and 67% feedback groups. The three groups performed pretest (3 trials), practice (12 trials), posttest (3 trials) and retention (3 trials, 24 hours later). The reduced feedback group showed lower RMS in the posttest than in the pretest (p = 0.04). The control and reduced feedback groups showed significant lower median frequency in the posttest than in the pretest (p < 0.05). Both feedback groups showed lower values in retention than in the pretest (p < 0.05). Even when the effect of feedback frequency could not be detected in motor learning, 67% of the feedback was recommended for motor adaptation.
MEAT: An Authoring Tool for Generating Adaptable Learning Resources
ERIC Educational Resources Information Center
Kuo, Yen-Hung; Huang, Yueh-Min
2009-01-01
Mobile learning (m-learning) is a new trend in the e-learning field. The learning services in m-learning environments are supported by fundamental functions, especially the content and assessment services, which need an authoring tool to rapidly generate adaptable learning resources. To fulfill the imperious demand, this study proposes an…
Alternating prism exposure causes dual adaptation and generalization to a novel displacement
NASA Technical Reports Server (NTRS)
Welch, Robert B.; Bridgeman, Bruce; Anand, Sulekha; Browman, Kaitlin E.
1993-01-01
In two experiments, we examined the hypothesis that repeatedly adapting and readapting to two mutually conflicting sensory environments fosters the development of a separate adaptation to each situation (dual adaptation) as well as an increased ability to adapt to a novel displacement (adaptive generalization). In the preliminary study, subjects alternated between adapting their visuomotor coordination to 30-diopter prismatic displacement and readapting to normal vision. Dual adaptation was observed by the end of 10 alternation cycles. However, an unconfounded test of adaptive generalization was prevented by an unexpected prism-adaptive shift in preexposure baselines for the dual-adapted subjects. In the primary experiment, the subjects adapted and readapted to opposite 15-diopter displacements for a total of 12 cycles. Both dual adaptation and adaptive generalization to a 30-diopter displacement were obtained. These findings may be understood in terms of serial reversal learning and 'learning to learn'.
Automatic learning rate adjustment for self-supervising autonomous robot control
NASA Technical Reports Server (NTRS)
Arras, Michael K.; Protzel, Peter W.; Palumbo, Daniel L.
1992-01-01
Described is an application in which an Artificial Neural Network (ANN) controls the positioning of a robot arm with five degrees of freedom by using visual feedback provided by two cameras. This application and the specific ANN model, local liner maps, are based on the work of Ritter, Martinetz, and Schulten. We extended their approach by generating a filtered, average positioning error from the continuous camera feedback and by coupling the learning rate to this error. When the network learns to position the arm, the positioning error decreases and so does the learning rate until the system stabilizes at a minimum error and learning rate. This abolishes the need for a predetermined cooling schedule. The automatic cooling procedure results in a closed loop control with no distinction between a learning phase and a production phase. If the positioning error suddenly starts to increase due to an internal failure such as a broken joint, or an environmental change such as a camera moving, the learning rate increases accordingly. Thus, learning is automatically activated and the network adapts to the new condition after which the error decreases again and learning is 'shut off'. The automatic cooling is therefore a prerequisite for the autonomy and the fault tolerance of the system.
NASA Technical Reports Server (NTRS)
Pavlock, Kate M.
2011-01-01
The National Aeronautics and Space Administration's Dryden Flight Research Center completed flight testing of adaptive controls research on the Full-Scale Advance Systems Testbed (FAST) in January of 2011. The research addressed technical challenges involved with reducing risk in an increasingly complex and dynamic national airspace. Specific challenges lie with the development of validated, multidisciplinary, integrated aircraft control design tools and techniques to enable safe flight in the presence of adverse conditions such as structural damage, control surface failures, or aerodynamic upsets. The testbed is an F-18 aircraft serving as a full-scale vehicle to test and validate adaptive flight control research and lends a significant confidence to the development, maturation, and acceptance process of incorporating adaptive control laws into follow-on research and the operational environment. The experimental systems integrated into FAST were designed to allow for flexible yet safe flight test evaluation and validation of modern adaptive control technologies and revolve around two major hardware upgrades: the modification of Production Support Flight Control Computers (PSFCC) and integration of two, fourth-generation Airborne Research Test Systems (ARTS). Post-hardware integration verification and validation provided the foundation for safe flight test of Nonlinear Dynamic Inversion and Model Reference Aircraft Control adaptive control law experiments. To ensure success of flight in terms of cost, schedule, and test results, emphasis on risk management was incorporated into early stages of design and flight test planning and continued through the execution of each flight test mission. Specific consideration was made to incorporate safety features within the hardware and software to alleviate user demands as well as into test processes and training to reduce human factor impacts to safe and successful flight test. This paper describes the research configuration, experiment functionality, overall risk mitigation, flight test approach and results, and lessons learned of adaptive controls research of the Full-Scale Advanced Systems Testbed.
Robot-assisted adaptive training: custom force fields for teaching movement patterns.
Patton, James L; Mussa-Ivaldi, Ferdinando A
2004-04-01
Based on recent studies of neuro-adaptive control, we tested a new iterative algorithm to generate custom training forces to "trick" subjects into altering their target-directed reaching movements to a prechosen movement as an after-effect of adaptation. The prechosen movement goal, a sinusoidal-shaped path from start to end point, was never explicitly conveyed to the subject. We hypothesized that the adaptation would cause an alteration in the feedforward command that would result in the prechosen movement. Our results showed that when forces were suddenly removed after a training period of 330 movements, trajectories were significantly shifted toward the prechosen movement. However, de-adaptation occurred (i.e., the after-effect "washed out") in the 50-75 movements that followed the removal of the training forces. A second experiment suppressed vision of hand location and found a detectable reduction in the washout of after-effects, suggesting that visual feedback of error critically influences learning. A final experiment demonstrated that after-effects were also present in the neighborhood of training--44% of original directional shift was seen in adjacent, unpracticed movement directions to targets that were 60 degrees different from the targets used for training. These results demonstrate the potential for these methods for teaching motor skills and for neuro-rehabilitation of brain-injured patients. This is a form of "implicit learning," because unlike explicit training methods, subjects learn movements with minimal instructions, no knowledge of, and little attention to the trajectory.
1981-01-01
errors more than twice that shown by the control group (32% versus 13%). The quantitative difference between training and control groups was further...could be enhanced (relative to control groups ) by means of brief training procedures. Educational implications of metaphor research were considered, with...instruction: It addresses itself to the kinds of research and instructional designs required for effective implementation of adaptive instruction ONR
The Future of Adaptive Learning: Does the Crowd Hold the Key?
ERIC Educational Resources Information Center
Heffernan, Neil T.; Ostrow, Korinn S.; Kelly, Kim; Selent, Douglas; Van Inwegen, Eric G.; Xiong, Xiaolu; Williams, Joseph Jay
2016-01-01
Due to substantial scientific and practical progress, learning technologies can effectively adapt to the characteristics and needs of students. This article considers how learning technologies can adapt over time by crowdsourcing contributions from teachers and students--explanations, feedback, and other pedagogical interactions. Considering the…
NASA Astrophysics Data System (ADS)
Bu, Xiangwei; Wu, Xiaoyan; He, Guangjun; Huang, Jiaqi
2016-03-01
This paper investigates the design of a novel adaptive neural controller for the longitudinal dynamics of a flexible air-breathing hypersonic vehicle with control input constraints. To reduce the complexity of controller design, the vehicle dynamics is decomposed into the velocity subsystem and the altitude subsystem, respectively. For each subsystem, only one neural network is utilized to approach the lumped unknown function. By employing a minimal-learning parameter method to estimate the norm of ideal weight vectors rather than their elements, there are only two adaptive parameters required for neural approximation. Thus, the computational burden is lower than the ones derived from neural back-stepping schemes. Specially, to deal with the control input constraints, additional systems are exploited to compensate the actuators. Lyapunov synthesis proves that all the closed-loop signals involved are uniformly ultimately bounded. Finally, simulation results show that the adopted compensation scheme can tackle actuator constraint effectively and moreover velocity and altitude can stably track their reference trajectories even when the physical limitations on control inputs are in effect.
NASA Astrophysics Data System (ADS)
Cui, Bing; Zhao, Chunhui; Ma, Tiedong; Feng, Chi
2017-02-01
In this paper, the cooperative adaptive consensus tracking problem for heterogeneous nonlinear multi-agent systems on directed graph is addressed. Each follower is modelled as a general nonlinear system with the unknown and nonidentical nonlinear dynamics, disturbances and actuator failures. Cooperative fault tolerant neural network tracking controllers with online adaptive learning features are proposed to guarantee that all agents synchronise to the trajectory of one leader with bounded adjustable synchronisation errors. With the help of linear quadratic regulator-based optimal design, a graph-dependent Lyapunov proof provides error bounds that depend on the graph topology, one virtual matrix and some design parameters. Of particular interest is that if the control gain is selected appropriately, the proposed control scheme can be implemented in a unified framework no matter whether there are faults or not. Furthermore, the fault detection and isolation are not needed to implement. Finally, a simulation is given to verify the effectiveness of the proposed method.
Hybrid Decompositional Verification for Discovering Failures in Adaptive Flight Control Systems
NASA Technical Reports Server (NTRS)
Thompson, Sarah; Davies, Misty D.; Gundy-Burlet, Karen
2010-01-01
Adaptive flight control systems hold tremendous promise for maintaining the safety of a damaged aircraft and its passengers. However, most currently proposed adaptive control methodologies rely on online learning neural networks (OLNNs), which necessarily have the property that the controller is changing during the flight. These changes tend to be highly nonlinear, and difficult or impossible to analyze using standard techniques. In this paper, we approach the problem with a variant of compositional verification. The overall system is broken into components. Undesirable behavior is fed backwards through the system. Components which can be solved using formal methods techniques explicitly for the ranges of safe and unsafe input bounds are treated as white box components. The remaining black box components are analyzed with heuristic techniques that try to predict a range of component inputs that may lead to unsafe behavior. The composition of these component inputs throughout the system leads to overall system test vectors that may elucidate the undesirable behavior
Dynamic Learner Profiling and Automatic Learner Classification for Adaptive E-Learning Environment
ERIC Educational Resources Information Center
Premlatha, K. R.; Dharani, B.; Geetha, T. V.
2016-01-01
E-learning allows learners individually to learn "anywhere, anytime" and offers immediate access to specific information. However, learners have different behaviors, learning styles, attitudes, and aptitudes, which affect their learning process, and therefore learning environments need to adapt according to these differences, so as to…
The Framework of Intervention Engine Based on Learning Analytics
ERIC Educational Resources Information Center
Sahin, Muhittin; Yurdugül, Halil
2017-01-01
Learning analytics primarily deals with the optimization of learning environments and the ultimate goal of learning analytics is to improve learning and teaching efficiency. Studies on learning analytics seem to have been made in the form of adaptation engine and intervention engine. Adaptation engine studies are quite widespread, but intervention…
Liu, Jing-Ying; Liu, Yan-Hui; Yang, Ji-Peng
2014-01-01
The aim of this study was to explore the relationships among study engagement, learning adaptability, and time management disposition in a sample of Chinese baccalaureate nursing students. A convenient sample of 467 baccalaureate nursing students was surveyed in two universities in Tianjin, China. Students completed a questionnaire that included their demographic information, Chinese Utrecht Work Engagement Scale-Student Questionnaire, Learning Adaptability Scale, and Adolescence Time Management Disposition Scale. One-way analysis of variance tests were used to assess the relationship between certain characteristics of baccalaureate nursing students. Pearson correlation was performed to test the correlation among study engagement, learning adaptability, and time management disposition. Hierarchical linear regression analyses were performed to explore the mediating role of time management disposition. The results revealed that study engagement (F = 7.20, P < .01) and learning adaptability (F = 4.41, P < .01) differed across grade groups. Learning adaptability (r = 0.382, P < .01) and time management disposition (r = 0.741, P < .01) were positively related with study engagement. Time management disposition had a partially mediating effect on the relationship between study engagement and learning adaptability. The findings implicate that educators should not only promote interventions to increase engagement of baccalaureate nursing students but also focus on development, investment in adaptability, and time management. Copyright © 2014 Elsevier Inc. All rights reserved.
Adaptive effort investment in cognitive and physical tasks: a neurocomputational model
Verguts, Tom; Vassena, Eliana; Silvetti, Massimo
2015-01-01
Despite its importance in everyday life, the computational nature of effort investment remains poorly understood. We propose an effort model obtained from optimality considerations, and a neurocomputational approximation to the optimal model. Both are couched in the framework of reinforcement learning. It is shown that choosing when or when not to exert effort can be adaptively learned, depending on rewards, costs, and task difficulty. In the neurocomputational model, the limbic loop comprising anterior cingulate cortex (ACC) and ventral striatum in the basal ganglia allocates effort to cortical stimulus-action pathways whenever this is valuable. We demonstrate that the model approximates optimality. Next, we consider two hallmark effects from the cognitive control literature, namely proportion congruency and sequential congruency effects. It is shown that the model exerts both proactive and reactive cognitive control. Then, we simulate two physical effort tasks. In line with empirical work, impairing the model's dopaminergic pathway leads to apathetic behavior. Thus, we conceptually unify the exertion of cognitive and physical effort, studied across a variety of literatures (e.g., motivation and cognitive control) and animal species. PMID:25805978
2017-02-08
Georgia Tech Research Corporation 505 Tenth Street NW Atlanta, GA 30332 -0420 ABSTRACT Final Report: MURI: Neuro-Inspired Adaptive Perception and...Conquer Strategy for Optimal Trajectory Planning via Mixed-Integer Programming, IEEE Transactions on Robotics, (12 2015): 0. doi: 10.1109/TRO...Learning Day, Microsoft Corporation , Cambridge, MA, May 18, 2015. (c) Presentations 09/06/2015 09/08/2015 125 131 Ali Borji, Dicky N. Sihite, Laurent Itti
Ludolph, Nicolas; Giese, Martin A; Ilg, Winfried
2017-10-16
There is increasing evidence that sensorimotor learning under real-life conditions relies on a composition of several learning processes. Nevertheless, most studies examine learning behaviour in relation to one specific learning mechanism. In this study, we examined the interaction between reward-based skill acquisition and motor adaptation to changes of object dynamics. Thirty healthy subjects, split into two groups, acquired the skill of balancing a pole on a cart in virtual reality. In one group, we gradually increased the gravity, making the task easier in the beginning and more difficult towards the end. In the second group, subjects had to acquire the skill on the maximum, most difficult gravity level. We hypothesized that the gradual increase in gravity during skill acquisition supports learning despite the necessary adjustments to changes in cart-pole dynamics. We found that the gradual group benefits from the slow increment, although overall improvement was interrupted by the changes in gravity and resulting system dynamics, which caused short-term degradations in performance and timing of actions. In conclusion, our results deliver evidence for an interaction of reward-based skill acquisition and motor adaptation processes, which indicates the importance of both processes for the development of optimized skill acquisition schedules.
The interference effects of non-rotated versus counter-rotated trials in visuomotor adaptation.
Hinder, Mark R; Walk, Laura; Woolley, Daniel G; Riek, Stephan; Carson, Richard G
2007-07-01
An isometric torque-production task was used to investigate interference and retention in adaptation to multiple visuomotor environments. Subjects produced isometric flexion-extension and pronation-supination elbow torques to move a cursor to acquire targets as quickly as possible. Adaptation to a 30 degrees counter-clockwise (CCW) rotation (task A), was followed by a period of rest (control), trials with no rotation (task B0), or trials with a 60 degrees clockwise (CW) rotation (task B60). For all groups, retention of task A was assessed 5 h later. With initial training, all groups reduced the angular deviation of cursor paths early in the movements, indicating feedforward adaptation. For the control group, performance at commencement of the retest was significantly better than that at the beginning of the initial learning. For the B0 group, performance in the retest of task A was not dissimilar to that at the start of the initial learning, while for the B60 group retest performance in task A was markedly worse than initially observed. Our results indicate that close juxtaposition of two visuomotor environments precludes improved retest performance in the initial environment. Data for the B60 group, specifically larger angular errors upon retest compared with initial exposures, are consistent with the presence of anterograde interference. Furthermore, full interference occurred even when the visuomotor environment encountered in the second task was not rotated (B0). This latter novel result differs from those obtained for force field learning, where interference does not occur when task B does not impose perturbing forces, i.e., when B consists of a null field (Brashers-Krug et al., Nature 382:252-255, 1996). The results are consistent with recent proposals suggesting different interference mechanisms for visuomotor (kinematic) compared to force field (dynamic) adaptations, and have implications for the use of washout trials when studying interference between multiple visuomotor environments.
Koller, Jeffrey R; Jacobs, Daniel A; Ferris, Daniel P; Remy, C David
2015-11-04
Robotic ankle exoskeletons can provide assistance to users and reduce metabolic power during walking. Our research group has investigated the use of proportional myoelectric control for controlling robotic ankle exoskeletons. Previously, these controllers have relied on a constant gain to map user's muscle activity to actuation control signals. A constant gain may act as a constraint on the user, so we designed a controller that dynamically adapts the gain to the user's myoelectric amplitude. We hypothesized that an adaptive gain proportional myoelectric controller would reduce metabolic energy expenditure compared to walking with the ankle exoskeleton unpowered because users could choose their preferred control gain. We tested eight healthy subjects walking with the adaptive gain proportional myoelectric controller with bilateral ankle exoskeletons. The adaptive gain was updated each stride such that on average the user's peak muscle activity was mapped to maximal power output of the exoskeleton. All subjects participated in three identical training sessions where they walked on a treadmill for 50 minutes (30 minutes of which the exoskeleton was powered) at 1.2 ms(-1). We calculated and analyzed metabolic energy consumption, muscle recruitment, inverse kinematics, inverse dynamics, and exoskeleton mechanics. Using our controller, subjects achieved a metabolic reduction similar to that seen in previous work in about a third of the training time. The resulting controller gain was lower than that seen in previous work (β=1.50±0.14 versus a constant β=2). The adapted gain allowed users more total ankle joint power than that of unassisted walking, increasing ankle power in exchange for a decrease in hip power. Our findings indicate that humans prefer to walk with greater ankle mechanical power output than their unassisted gait when provided with an ankle exoskeleton using an adaptive controller. This suggests that robotic assistance from an exoskeleton can allow humans to adopt gait patterns different from their normal choices for locomotion. In our specific experiment, subjects increased ankle power and decreased hip power to walk with a reduction in metabolic cost. Future exoskeleton devices that rely on proportional myolectric control are likely to demonstrate improved performance by including an adaptive gain.
Flow Navigation by Smart Microswimmers via Reinforcement Learning
NASA Astrophysics Data System (ADS)
Colabrese, Simona; Gustavsson, Kristian; Celani, Antonio; Biferale, Luca
2017-04-01
Smart active particles can acquire some limited knowledge of the fluid environment from simple mechanical cues and exert a control on their preferred steering direction. Their goal is to learn the best way to navigate by exploiting the underlying flow whenever possible. As an example, we focus our attention on smart gravitactic swimmers. These are active particles whose task is to reach the highest altitude within some time horizon, given the constraints enforced by fluid mechanics. By means of numerical experiments, we show that swimmers indeed learn nearly optimal strategies just by experience. A reinforcement learning algorithm allows particles to learn effective strategies even in difficult situations when, in the absence of control, they would end up being trapped by flow structures. These strategies are highly nontrivial and cannot be easily guessed in advance. This Letter illustrates the potential of reinforcement learning algorithms to model adaptive behavior in complex flows and paves the way towards the engineering of smart microswimmers that solve difficult navigation problems.
Embodying a cognitive model in a mobile robot
NASA Astrophysics Data System (ADS)
Benjamin, D. Paul; Lyons, Damian; Lonsdale, Deryle
2006-10-01
The ADAPT project is a collaboration of researchers in robotics, linguistics and artificial intelligence at three universities to create a cognitive architecture specifically designed to be embodied in a mobile robot. There are major respects in which existing cognitive architectures are inadequate for robot cognition. In particular, they lack support for true concurrency and for active perception. ADAPT addresses these deficiencies by modeling the world as a network of concurrent schemas, and modeling perception as problem solving. Schemas are represented using the RS (Robot Schemas) language, and are activated by spreading activation. RS provides a powerful language for distributed control of concurrent processes. Also, The formal semantics of RS provides the basis for the semantics of ADAPT's use of natural language. We have implemented the RS language in Soar, a mature cognitive architecture originally developed at CMU and used at a number of universities and companies. Soar's subgoaling and learning capabilities enable ADAPT to manage the complexity of its environment and to learn new schemas from experience. We describe the issues faced in developing an embodied cognitive architecture, and our implementation choices.
ERIC Educational Resources Information Center
Debevc, Matjaž; Stjepanovic, Zoran; Holzinger, Andreas
2014-01-01
Web-based and adapted e-learning materials provide alternative methods of learning to those used in a traditional classroom. Within the study described in this article, deaf and hard of hearing people used an adaptive e-learning environment to improve their computer literacy. This environment included streaming video with sign language interpreter…
Online EEG-Based Workload Adaptation of an Arithmetic Learning Environment.
Walter, Carina; Rosenstiel, Wolfgang; Bogdan, Martin; Gerjets, Peter; Spüler, Martin
2017-01-01
In this paper, we demonstrate a closed-loop EEG-based learning environment, that adapts instructional learning material online, to improve learning success in students during arithmetic learning. The amount of cognitive workload during learning is crucial for successful learning and should be held in the optimal range for each learner. Based on EEG data from 10 subjects, we created a prediction model that estimates the learner's workload to obtain an unobtrusive workload measure. Furthermore, we developed an interactive learning environment that uses the prediction model to estimate the learner's workload online based on the EEG data and adapt the difficulty of the learning material to keep the learner's workload in an optimal range. The EEG-based learning environment was used by 13 subjects to learn arithmetic addition in the octal number system, leading to a significant learning effect. The results suggest that it is feasible to use EEG as an unobtrusive measure of cognitive workload to adapt the learning content. Further it demonstrates that a promptly workload prediction is possible using a generalized prediction model without the need for a user-specific calibration.
Trial-to-trial adaptation in control of arm reaching and standing posture
Pienciak-Siewert, Alison; Horan, Dylan P.
2016-01-01
Classical theories of motor learning hypothesize that adaptation is driven by sensorimotor error; this is supported by studies of arm and eye movements that have shown that trial-to-trial adaptation increases with error. Studies of postural control have shown that anticipatory postural adjustments increase with the magnitude of a perturbation. However, differences in adaptation have been observed between the two modalities, possibly due to either the inherent instability or sensory uncertainty in standing posture. Therefore, we hypothesized that trial-to-trial adaptation in posture should be driven by error, similar to what is observed in arm reaching, but the nature of the relationship between error and adaptation may differ. Here we investigated trial-to-trial adaptation of arm reaching and postural control concurrently; subjects made reaching movements in a novel dynamic environment of varying strengths, while standing and holding the handle of a force-generating robotic arm. We found that error and adaptation increased with perturbation strength in both arm and posture. Furthermore, in both modalities, adaptation showed a significant correlation with error magnitude. Our results indicate that adaptation scales proportionally with error in the arm and near proportionally in posture. In posture only, adaptation was not sensitive to small error sizes, which were similar in size to errors experienced in unperturbed baseline movements due to inherent variability. This finding may be explained as an effect of uncertainty about the source of small errors. Our findings suggest that in rehabilitation, postural error size should be considered relative to the magnitude of inherent movement variability. PMID:27683888
Trial-to-trial adaptation in control of arm reaching and standing posture.
Pienciak-Siewert, Alison; Horan, Dylan P; Ahmed, Alaa A
2016-12-01
Classical theories of motor learning hypothesize that adaptation is driven by sensorimotor error; this is supported by studies of arm and eye movements that have shown that trial-to-trial adaptation increases with error. Studies of postural control have shown that anticipatory postural adjustments increase with the magnitude of a perturbation. However, differences in adaptation have been observed between the two modalities, possibly due to either the inherent instability or sensory uncertainty in standing posture. Therefore, we hypothesized that trial-to-trial adaptation in posture should be driven by error, similar to what is observed in arm reaching, but the nature of the relationship between error and adaptation may differ. Here we investigated trial-to-trial adaptation of arm reaching and postural control concurrently; subjects made reaching movements in a novel dynamic environment of varying strengths, while standing and holding the handle of a force-generating robotic arm. We found that error and adaptation increased with perturbation strength in both arm and posture. Furthermore, in both modalities, adaptation showed a significant correlation with error magnitude. Our results indicate that adaptation scales proportionally with error in the arm and near proportionally in posture. In posture only, adaptation was not sensitive to small error sizes, which were similar in size to errors experienced in unperturbed baseline movements due to inherent variability. This finding may be explained as an effect of uncertainty about the source of small errors. Our findings suggest that in rehabilitation, postural error size should be considered relative to the magnitude of inherent movement variability. Copyright © 2016 the American Physiological Society.
Online Learning of Genetic Network Programming and its Application to Prisoner’s Dilemma Game
NASA Astrophysics Data System (ADS)
Mabu, Shingo; Hirasawa, Kotaro; Hu, Jinglu; Murata, Junichi
A new evolutionary model with the network structure named Genetic Network Programming (GNP) has been proposed recently. GNP, that is, an expansion of GA and GP, represents solutions as a network structure and evolves it by using “offline learning (selection, mutation, crossover)”. GNP can memorize the past action sequences in the network flow, so it can deal with Partially Observable Markov Decision Process (POMDP) well. In this paper, in order to improve the ability of GNP, Q learning (an off-policy TD control algorithm) that is one of the famous online methods is introduced for online learning of GNP. Q learning is suitable for GNP because (1) in reinforcement learning, the rewards an agent will get in the future can be estimated, (2) TD control doesn’t need much memory and can learn quickly, and (3) off-policy is suitable in order to search for an optimal solution independently of the policy. Finally, in the simulations, online learning of GNP is applied to a player for “Prisoner’s dilemma game” and its ability for online adaptation is confirmed.
Impairment of probabilistic reward-based learning in schizophrenia.
Weiler, Julia A; Bellebaum, Christian; Brüne, Martin; Juckel, Georg; Daum, Irene
2009-09-01
Recent models assume that some symptoms of schizophrenia originate from defective reward processing mechanisms. Understanding the precise nature of reward-based learning impairments might thus make an important contribution to the understanding of schizophrenia and the development of treatment strategies. The present study investigated several features of probabilistic reward-based stimulus association learning, namely the acquisition of initial contingencies, reversal learning, generalization abilities, and the effects of reward magnitude. Compared to healthy controls, individuals with schizophrenia exhibited attenuated overall performance during acquisition, whereas learning rates across blocks were similar to the rates of controls. On the group level, persons with schizophrenia were, however, unable to learn the reversal of the initial reward contingencies. Exploratory analysis of only the subgroup of individuals with schizophrenia who showed significant learning during acquisition yielded deficits in reversal learning with low reward magnitudes only. There was further evidence of a mild generalization impairment of the persons with schizophrenia in an acquired equivalence task. In summary, although there was evidence of intact basic processing of reward magnitudes, individuals with schizophrenia were impaired at using this feedback for the adaptive guidance of behavior.
Investigating Work and Learning through Complex Adaptive Organisations
ERIC Educational Resources Information Center
Lizier, Amanda Louise
2017-01-01
Purpose: The purpose of this paper is to outline an empirical study of how professionals experience work and learning in complex adaptive organisations. The study uses a complex adaptive systems approach, which forms the basis of a specifically developed conceptual framework for explaining professionals' experiences of work and learning.…
ERIC Educational Resources Information Center
Kontoghiorghes, Constantine; Awbre, Susan M.; Feurig, Pamela L.
2005-01-01
The main purpose of this exploratory study was to examine the relationship between certain learning organization characteristics and change adaptation, innovation, and bottom-line organizational performance. The following learning organization characteristics were found to be the strongest predictors of rapid change adaptation, quick product or…
Evolution of social learning does not explain the origin of human cumulative culture.
Enquist, Magnus; Ghirlanda, Stefano
2007-05-07
Because culture requires transmission of information between individuals, thinking about the origin of culture has mainly focused on the genetic evolution of abilities for social learning. Current theory considers how social learning affects the adaptiveness of a single cultural trait, yet human culture consists of the accumulation of very many traits. Here we introduce a new modeling strategy that tracks the adaptive value of many cultural traits, showing that genetic evolution favors only limited social learning owing to the accumulation of maladaptive as well as adaptive culture. We further show that culture can be adaptive, and refined social learning can evolve, if individuals can identify and discard maladaptive culture. This suggests that the evolution of such "adaptive filtering" mechanisms may have been crucial for the birth of human culture.
Hebbian Learning of Cognitive Control: Dealing with Specific and Nonspecific Adaptation
ERIC Educational Resources Information Center
Verguts, Tom; Notebaert, Wim
2008-01-01
The conflict monitoring model of M. M. Botvinick, T. S. Braver, D. M. Barch, C. S. Carter, and J. D. Cohen (2001) triggered several research programs investigating various aspects of cognitive control. One problematic aspect of the Botvinick et al. model is that there is no clear account of how the cognitive system knows where to intervene when…
A biologically inspired meta-control navigation system for the Psikharpax rat robot.
Caluwaerts, K; Staffa, M; N'Guyen, S; Grand, C; Dollé, L; Favre-Félix, A; Girard, B; Khamassi, M
2012-06-01
A biologically inspired navigation system for the mobile rat-like robot named Psikharpax is presented, allowing for self-localization and autonomous navigation in an initially unknown environment. The ability of parts of the model (e.g. the strategy selection mechanism) to reproduce rat behavioral data in various maze tasks has been validated before in simulations. But the capacity of the model to work on a real robot platform had not been tested. This paper presents our work on the implementation on the Psikharpax robot of two independent navigation strategies (a place-based planning strategy and a cue-guided taxon strategy) and a strategy selection meta-controller. We show how our robot can memorize which was the optimal strategy in each situation, by means of a reinforcement learning algorithm. Moreover, a context detector enables the controller to quickly adapt to changes in the environment-recognized as new contexts-and to restore previously acquired strategy preferences when a previously experienced context is recognized. This produces adaptivity closer to rat behavioral performance and constitutes a computational proposition of the role of the rat prefrontal cortex in strategy shifting. Moreover, such a brain-inspired meta-controller may provide an advancement for learning architectures in robotics.
Unsupervised learning in general connectionist systems.
Dente, J A; Mendes, R Vilela
1996-01-01
There is a common framework in which different connectionist systems may be treated in a unified way. The general system in which they may all be mapped is a network which, in addition to the connection strengths, has an adaptive node parameter controlling the output intensity. In this paper we generalize two neural network learning schemes to networks with node parameters. In generalized Hebbian learning we find improvements to the convergence rate for small eigenvalues in principal component analysis. For competitive learning the use of node parameters also seems useful in that, by emphasizing or de-emphasizing the dominance of winning neurons, either improved robustness or discrimination is obtained.
NASA Technical Reports Server (NTRS)
Rodriguez, Guillermo (Editor)
1990-01-01
Various papers on intelligent control and adaptive systems are presented. Individual topics addressed include: control architecture for a Mars walking vehicle, representation for error detection and recovery in robot task plans, real-time operating system for robots, execution monitoring of a mobile robot system, statistical mechanics models for motion and force planning, global kinematics for manipulator planning and control, exploration of unknown mechanical assemblies through manipulation, low-level representations for robot vision, harmonic functions for robot path construction, simulation of dual behavior of an autonomous system. Also discussed are: control framework for hand-arm coordination, neural network approach to multivehicle navigation, electronic neural networks for global optimization, neural network for L1 norm linear regression, planning for assembly with robot hands, neural networks in dynamical systems, control design with iterative learning, improved fuzzy process control of spacecraft autonomous rendezvous using a genetic algorithm.
ERIC Educational Resources Information Center
Thalmann, Stefan
2014-01-01
Personalised e-Learning represents a major step-change from the one-size-fits-all approach of traditional learning platforms to a more customised and interactive provision of learning materials. Adaptive learning can support the learning process by tailoring learning materials to individual needs. However, this requires the initial preparation of…