Modular Track System For Positioning Mobile Robots
NASA Technical Reports Server (NTRS)
Miller, Jeff
1995-01-01
Conceptual system for positioning mobile robotic manipulators on large main structure includes modular tracks and ancillary structures assembled easily along with main structure. System, called "tracked robotic location system" (TROLS), originally intended for application to platforms in outer space, but TROLS concept might also prove useful on Earth; for example, to position robots in factories and warehouses. T-cross-section rail keeps mobile robot on track. Bar codes mark locations along track. Each robot equipped with bar-code-recognizing circuitry so it quickly finds way to assigned location.
Laser-based pedestrian tracking in outdoor environments by multiple mobile robots.
Ozaki, Masataka; Kakimuma, Kei; Hashimoto, Masafumi; Takahashi, Kazuhiko
2012-10-29
This paper presents an outdoors laser-based pedestrian tracking system using a group of mobile robots located near each other. Each robot detects pedestrians from its own laser scan image using an occupancy-grid-based method, and the robot tracks the detected pedestrians via Kalman filtering and global-nearest-neighbor (GNN)-based data association. The tracking data is broadcast to multiple robots through intercommunication and is combined using the covariance intersection (CI) method. For pedestrian tracking, each robot identifies its own posture using real-time-kinematic GPS (RTK-GPS) and laser scan matching. Using our cooperative tracking method, all the robots share the tracking data with each other; hence, individual robots can always recognize pedestrians that are invisible to any other robot. The simulation and experimental results show that cooperating tracking provides the tracking performance better than conventional individual tracking does. Our tracking system functions in a decentralized manner without any central server, and therefore, this provides a degree of scalability and robustness that cannot be achieved by conventional centralized architectures.
Hybrid position and orientation tracking for a passive rehabilitation table-top robot.
Wojewoda, K K; Culmer, P R; Gallagher, J F; Jackson, A E; Levesley, M C
2017-07-01
This paper presents a real time hybrid 2D position and orientation tracking system developed for an upper limb rehabilitation robot. Designed to work on a table-top, the robot is to enable home-based upper-limb rehabilitative exercise for stroke patients. Estimates of the robot's position are computed by fusing data from two tracking systems, each utilizing a different sensor type: laser optical sensors and a webcam. Two laser optical sensors are mounted on the underside of the robot and track the relative motion of the robot with respect to the surface on which it is placed. The webcam is positioned directly above the workspace, mounted on a fixed stand, and tracks the robot's position with respect to a fixed coordinate system. The optical sensors sample the position data at a higher frequency than the webcam, and a position and orientation fusion scheme is proposed to fuse the data from the two tracking systems. The proposed fusion scheme is validated through an experimental set-up whereby the rehabilitation robot is moved by a humanoid robotic arm replicating previously recorded movements of a stroke patient. The results prove that the presented hybrid position tracking system can track the position and orientation with greater accuracy than the webcam or optical sensors alone. The results also confirm that the developed system is capable of tracking recovery trends during rehabilitation therapy.
Laser-Based Pedestrian Tracking in Outdoor Environments by Multiple Mobile Robots
Ozaki, Masataka; Kakimuma, Kei; Hashimoto, Masafumi; Takahashi, Kazuhiko
2012-01-01
This paper presents an outdoors laser-based pedestrian tracking system using a group of mobile robots located near each other. Each robot detects pedestrians from its own laser scan image using an occupancy-grid-based method, and the robot tracks the detected pedestrians via Kalman filtering and global-nearest-neighbor (GNN)-based data association. The tracking data is broadcast to multiple robots through intercommunication and is combined using the covariance intersection (CI) method. For pedestrian tracking, each robot identifies its own posture using real-time-kinematic GPS (RTK-GPS) and laser scan matching. Using our cooperative tracking method, all the robots share the tracking data with each other; hence, individual robots can always recognize pedestrians that are invisible to any other robot. The simulation and experimental results show that cooperating tracking provides the tracking performance better than conventional individual tracking does. Our tracking system functions in a decentralized manner without any central server, and therefore, this provides a degree of scalability and robustness that cannot be achieved by conventional centralized architectures. PMID:23202171
Human-like object tracking and gaze estimation with PKD android
Wijayasinghe, Indika B.; Miller, Haylie L.; Das, Sumit K; Bugnariu, Nicoleta L.; Popa, Dan O.
2018-01-01
As the use of robots increases for tasks that require human-robot interactions, it is vital that robots exhibit and understand human-like cues for effective communication. In this paper, we describe the implementation of object tracking capability on Philip K. Dick (PKD) android and a gaze tracking algorithm, both of which further robot capabilities with regard to human communication. PKD's ability to track objects with human-like head postures is achieved with visual feedback from a Kinect system and an eye camera. The goal of object tracking with human-like gestures is twofold : to facilitate better human-robot interactions and to enable PKD as a human gaze emulator for future studies. The gaze tracking system employs a mobile eye tracking system (ETG; SensoMotoric Instruments) and a motion capture system (Cortex; Motion Analysis Corp.) for tracking the head orientations. Objects to be tracked are displayed by a virtual reality system, the Computer Assisted Rehabilitation Environment (CAREN; MotekForce Link). The gaze tracking algorithm converts eye tracking data and head orientations to gaze information facilitating two objectives: to evaluate the performance of the object tracking system for PKD and to use the gaze information to predict the intentions of the user, enabling the robot to understand physical cues by humans. PMID:29416193
Human-like object tracking and gaze estimation with PKD android
NASA Astrophysics Data System (ADS)
Wijayasinghe, Indika B.; Miller, Haylie L.; Das, Sumit K.; Bugnariu, Nicoleta L.; Popa, Dan O.
2016-05-01
As the use of robots increases for tasks that require human-robot interactions, it is vital that robots exhibit and understand human-like cues for effective communication. In this paper, we describe the implementation of object tracking capability on Philip K. Dick (PKD) android and a gaze tracking algorithm, both of which further robot capabilities with regard to human communication. PKD's ability to track objects with human-like head postures is achieved with visual feedback from a Kinect system and an eye camera. The goal of object tracking with human-like gestures is twofold: to facilitate better human-robot interactions and to enable PKD as a human gaze emulator for future studies. The gaze tracking system employs a mobile eye tracking system (ETG; SensoMotoric Instruments) and a motion capture system (Cortex; Motion Analysis Corp.) for tracking the head orientations. Objects to be tracked are displayed by a virtual reality system, the Computer Assisted Rehabilitation Environment (CAREN; MotekForce Link). The gaze tracking algorithm converts eye tracking data and head orientations to gaze information facilitating two objectives: to evaluate the performance of the object tracking system for PKD and to use the gaze information to predict the intentions of the user, enabling the robot to understand physical cues by humans.
Development and validation of a low-cost mobile robotics testbed
NASA Astrophysics Data System (ADS)
Johnson, Michael; Hayes, Martin J.
2012-03-01
This paper considers the design, construction and validation of a low-cost experimental robotic testbed, which allows for the localisation and tracking of multiple robotic agents in real time. The testbed system is suitable for research and education in a range of different mobile robotic applications, for validating theoretical as well as practical research work in the field of digital control, mobile robotics, graphical programming and video tracking systems. It provides a reconfigurable floor space for mobile robotic agents to operate within, while tracking the position of multiple agents in real-time using the overhead vision system. The overall system provides a highly cost-effective solution to the topical problem of providing students with practical robotics experience within severe budget constraints. Several problems encountered in the design and development of the mobile robotic testbed and associated tracking system, such as radial lens distortion and the selection of robot identifier templates are clearly addressed. The testbed performance is quantified and several experiments involving LEGO Mindstorm NXT and Merlin System MiaBot robots are discussed.
2017-06-01
implement human following on a mobile robot in an indoor environment . B. FUTURE WORK Future work that could be conducted in the realm of this thesis...FEASIBILITY OF CONDUCTING HUMAN TRACKING AND FOLLOWING IN AN INDOOR ENVIRONMENT USING A MICROSOFT KINECT AND THE ROBOT OPERATING SYSTEM by...FEASIBILITY OF CONDUCTING HUMAN TRACKING AND FOLLOWING IN AN INDOOR ENVIRONMENT USING A MICROSOFT KINECT AND THE ROBOT OPERATING SYSTEM 5. FUNDING NUMBERS
SU-G-JeP3-08: Robotic System for Ultrasound Tracking in Radiation Therapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kuhlemann, I; Graduate School for Computing in Medicine and Life Sciences, University of Luebeck; Jauer, P
Purpose: For safe and accurate real-time tracking of tumors for IGRT using 4D ultrasound, it is necessary to make use of novel, high-end force-sensitive lightweight robots designed for human-machine interaction. Such a robot will be integrated into an existing robotized ultrasound system for non-invasive 4D live tracking, using a newly developed real-time control and communication framework. Methods: The new KUKA LWR iiwa robot is used for robotized ultrasound real-time tumor tracking. Besides more precise probe contact pressure detection, this robot provides an additional 7th link, enhancing the dexterity of the kinematic and the mounted transducer. Several integrated, certified safety featuresmore » create a safe environment for the patients during treatment. However, to remotely control the robot for the ultrasound application, a real-time control and communication framework has to be developed. Based on a client/server concept, client-side control commands are received and processed by a central server unit and are implemented by a client module running directly on the robot’s controller. Several special functionalities for robotized ultrasound applications are integrated and the robot can now be used for real-time control of the image quality by adjusting the transducer position, and contact pressure. The framework was evaluated looking at overall real-time capability for communication and processing of three different standard commands. Results: Due to inherent, certified safety modules, the new robot ensures a safe environment for patients during tumor tracking. Furthermore, the developed framework shows overall real-time capability with a maximum average latency of 3.6 ms (Minimum 2.5 ms; 5000 trials). Conclusion: The novel KUKA LBR iiwa robot will advance the current robotized ultrasound tracking system with important features. With the developed framework, it is now possible to remotely control this robot and use it for robotized ultrasound tracking applications, including image quality control and target tracking.« less
Trajectory planning and optimal tracking for an industrial mobile robot
NASA Astrophysics Data System (ADS)
Hu, Huosheng; Brady, J. Michael; Probert, Penelope J.
1994-02-01
This paper introduces a unified approach to trajectory planning and tracking for an industrial mobile robot subject to non-holonomic constraints. We show (1) how a smooth trajectory is generated that takes into account the constraints from the dynamic environment and the robot kinematics; and (2) how a general predictive controller works to provide optimal tracking capability for nonlinear systems. The tracking performance of the proposed guidance system is analyzed by simulation.
A novel traveling wave piezoelectric actuated tracked mobile robot utilizing friction effect
NASA Astrophysics Data System (ADS)
Wang, Liang; Shu, Chengyou; Jin, Jiamei; Zhang, Jianhui
2017-03-01
A novel traveling wave piezoelectric-actuated tracked mobile robot with potential application to robotic rovers was proposed and investigated in this study. The proposed tracked mobile robot is composed of a parallelogram-frame-structure piezoelectric transducer with four rings and a metal track. Utilizing the converse piezoelectric and friction effects, traveling waves were propagated in the rings and then the metal track was actuated by the piezoelectric transducer. Compared with traditional tracked mechanisms, the proposed tracked mobile robot has a simpler and more compact structure without lubricant, which eliminates the problem of lubricant volatilization and deflation, thus, it could be operated in the vacuum environment. Dynamic characteristics were simulated and measured to reveal the mechanism of actuating track of the piezoelectric transducer. Experimental investigations of the traveling wave piezoelectric-actuated tracked mobile robot were then carried out, and the results indicated that the robot prototype with a pair of exciting voltages of 460 Vpp is able to achieve a maximum velocity of 57 mm s-1 moving on the foam plate and possesses the obstacle crossing capability with a maximum height of 27 mm. The proposed tracked mobile robot exhibits potential to be the driving system of robotic rovers.
Real Time Target Tracking Using Dedicated Vision Hardware
NASA Astrophysics Data System (ADS)
Kambies, Keith; Walsh, Peter
1988-03-01
This paper describes a real-time vision target tracking system developed by Adaptive Automation, Inc. and delivered to NASA's Launch Equipment Test Facility, Kennedy Space Center, Florida. The target tracking system is part of the Robotic Application Development Laboratory (RADL) which was designed to provide NASA with a general purpose robotic research and development test bed for the integration of robot and sensor systems. One of the first RADL system applications is the closing of a position control loop around a six-axis articulated arm industrial robot using a camera and dedicated vision processor as the input sensor so that the robot can locate and track a moving target. The vision system is inside of the loop closure of the robot tracking system, therefore, tight throughput and latency constraints are imposed on the vision system that can only be met with specialized hardware and a concurrent approach to the processing algorithms. State of the art VME based vision boards capable of processing the image at frame rates were used with a real-time, multi-tasking operating system to achieve the performance required. This paper describes the high speed vision based tracking task, the system throughput requirements, the use of dedicated vision hardware architecture, and the implementation design details. Important to the overall philosophy of the complete system was the hierarchical and modular approach applied to all aspects of the system, hardware and software alike, so there is special emphasis placed on this topic in the paper.
A Kinect-Based Real-Time Compressive Tracking Prototype System for Amphibious Spherical Robots
Pan, Shaowu; Shi, Liwei; Guo, Shuxiang
2015-01-01
A visual tracking system is essential as a basis for visual servoing, autonomous navigation, path planning, robot-human interaction and other robotic functions. To execute various tasks in diverse and ever-changing environments, a mobile robot requires high levels of robustness, precision, environmental adaptability and real-time performance of the visual tracking system. In keeping with the application characteristics of our amphibious spherical robot, which was proposed for flexible and economical underwater exploration in 2012, an improved RGB-D visual tracking algorithm is proposed and implemented. Given the limited power source and computational capabilities of mobile robots, compressive tracking (CT), which is the effective and efficient algorithm that was proposed in 2012, was selected as the basis of the proposed algorithm to process colour images. A Kalman filter with a second-order motion model was implemented to predict the state of the target and select candidate patches or samples for the CT tracker. In addition, a variance ratio features shift (VR-V) tracker with a Kalman estimation mechanism was used to process depth images. Using a feedback strategy, the depth tracking results were used to assist the CT tracker in updating classifier parameters at an adaptive rate. In this way, most of the deficiencies of CT, including drift and poor robustness to occlusion and high-speed target motion, were partly solved. To evaluate the proposed algorithm, a Microsoft Kinect sensor, which combines colour and infrared depth cameras, was adopted for use in a prototype of the robotic tracking system. The experimental results with various image sequences demonstrated the effectiveness, robustness and real-time performance of the tracking system. PMID:25856331
A Kinect-based real-time compressive tracking prototype system for amphibious spherical robots.
Pan, Shaowu; Shi, Liwei; Guo, Shuxiang
2015-04-08
A visual tracking system is essential as a basis for visual servoing, autonomous navigation, path planning, robot-human interaction and other robotic functions. To execute various tasks in diverse and ever-changing environments, a mobile robot requires high levels of robustness, precision, environmental adaptability and real-time performance of the visual tracking system. In keeping with the application characteristics of our amphibious spherical robot, which was proposed for flexible and economical underwater exploration in 2012, an improved RGB-D visual tracking algorithm is proposed and implemented. Given the limited power source and computational capabilities of mobile robots, compressive tracking (CT), which is the effective and efficient algorithm that was proposed in 2012, was selected as the basis of the proposed algorithm to process colour images. A Kalman filter with a second-order motion model was implemented to predict the state of the target and select candidate patches or samples for the CT tracker. In addition, a variance ratio features shift (VR-V) tracker with a Kalman estimation mechanism was used to process depth images. Using a feedback strategy, the depth tracking results were used to assist the CT tracker in updating classifier parameters at an adaptive rate. In this way, most of the deficiencies of CT, including drift and poor robustness to occlusion and high-speed target motion, were partly solved. To evaluate the proposed algorithm, a Microsoft Kinect sensor, which combines colour and infrared depth cameras, was adopted for use in a prototype of the robotic tracking system. The experimental results with various image sequences demonstrated the effectiveness, robustness and real-time performance of the tracking system.
Decentralised consensus-based formation tracking of multiple differential drive robots
NASA Astrophysics Data System (ADS)
Chu, Xing; Peng, Zhaoxia; Wen, Guoguang; Rahmani, Ahmed
2017-11-01
This article investigates the control problem for formation tracking of multiple nonholonomic robots under distributed manner which means each robot only needs local information exchange. A class of general state and input transform is introduced to convert the formation-tracking issue of multi-robot systems into the consensus-like problem with time-varying reference. The distributed observer-based protocol with nonlinear dynamics is developed for each robot to achieve the consensus tracking of the new system, which namely means a group of nonholonomic mobile robots can form the desired formation configuration with its centroid moving along the predefined reference trajectory. The finite-time stability of observer and control law is analysed rigorously by using the Lyapunov direct method, algebraic graph theory and matrix analysis. Numerical examples are finally provided to illustrate the effectiveness of the theory results proposed in this paper.
A soft biomimetic tongue: model reconstruction and motion tracking
NASA Astrophysics Data System (ADS)
Lu, Xuanming; Xu, Weiliang; Li, Xiaoning
2016-04-01
A bioinspired robotic tongue which is actuated by a network of compressed air is proposed for the purpose of mimicking the movements of human tongue. It can be applied in the fields such as medical science and food engineering. The robotic tongue is made of two kinds of silicone rubber Ecoflex 0030 and PDMS with the shape simplified from real human tongue. In order to characterize the robotic tongue, a series of experiments were carried out. Laser scan was applied to reconstruct the static model of robotic tongue when it was under pressurization. After each scan, the robotic tongue was scattered into dense points in the same 3D coordinate system and the coordinates of each point were recorded. Motion tracking system (OptiTrack) was used to track and record the whole process of deformation dynamically during the loading and unloading phase. In the experiments, five types of deformation were achieved including roll-up, roll-down, elongation, groove and twist. Utilizing the discrete points generated by laser scan, the accurate parameterized outline of robotic tongue under different pressure was obtained, which could help demonstrate the static characteristic of robotic tongue. The precise deformation process under one pressure was acquired through the OptiTrack system which contains a series of digital cameras, markers on the robotic tongue and a set of hardware and software for data processing. By means of tracking and recording different process of deformation under different pressure, the dynamic characteristic of robotic tongue could be achieved.
Robot trajectory tracking with self-tuning predicted control
NASA Technical Reports Server (NTRS)
Cui, Xianzhong; Shin, Kang G.
1988-01-01
A controller that combines self-tuning prediction and control is proposed for robot trajectory tracking. The controller has two feedback loops: one is used to minimize the prediction error, and the other is designed to make the system output track the set point input. Because the velocity and position along the desired trajectory are given and the future output of the system is predictable, a feedforward loop can be designed for robot trajectory tracking with self-tuning predicted control (STPC). Parameters are estimated online to account for the model uncertainty and the time-varying property of the system. The authors describe the principle of STPC, analyze the system performance, and discuss the simplification of the robot dynamic equations. To demonstrate its utility and power, the controller is simulated for a Stanford arm.
Shang, Weijian; Su, Hao; Li, Gang; Fischer, Gregory S.
2014-01-01
This paper presents a surgical master-slave tele-operation system for percutaneous interventional procedures under continuous magnetic resonance imaging (MRI) guidance. This system consists of a piezoelectrically actuated slave robot for needle placement with integrated fiber optic force sensor utilizing Fabry-Perot interferometry (FPI) sensing principle. The sensor flexure is optimized and embedded to the slave robot for measuring needle insertion force. A novel, compact opto-mechanical FPI sensor interface is integrated into an MRI robot control system. By leveraging the complementary features of pneumatic and piezoelectric actuation, a pneumatically actuated haptic master robot is also developed to render force associated with needle placement interventions to the clinician. An aluminum load cell is implemented and calibrated to close the impedance control loop of the master robot. A force-position control algorithm is developed to control the hybrid actuated system. Teleoperated needle insertion is demonstrated under live MR imaging, where the slave robot resides in the scanner bore and the user manipulates the master beside the patient outside the bore. Force and position tracking results of the master-slave robot are demonstrated to validate the tracking performance of the integrated system. It has a position tracking error of 0.318mm and sine wave force tracking error of 2.227N. PMID:25126446
Shang, Weijian; Su, Hao; Li, Gang; Fischer, Gregory S
2013-01-01
This paper presents a surgical master-slave tele-operation system for percutaneous interventional procedures under continuous magnetic resonance imaging (MRI) guidance. This system consists of a piezoelectrically actuated slave robot for needle placement with integrated fiber optic force sensor utilizing Fabry-Perot interferometry (FPI) sensing principle. The sensor flexure is optimized and embedded to the slave robot for measuring needle insertion force. A novel, compact opto-mechanical FPI sensor interface is integrated into an MRI robot control system. By leveraging the complementary features of pneumatic and piezoelectric actuation, a pneumatically actuated haptic master robot is also developed to render force associated with needle placement interventions to the clinician. An aluminum load cell is implemented and calibrated to close the impedance control loop of the master robot. A force-position control algorithm is developed to control the hybrid actuated system. Teleoperated needle insertion is demonstrated under live MR imaging, where the slave robot resides in the scanner bore and the user manipulates the master beside the patient outside the bore. Force and position tracking results of the master-slave robot are demonstrated to validate the tracking performance of the integrated system. It has a position tracking error of 0.318mm and sine wave force tracking error of 2.227N.
DOE Office of Scientific and Technical Information (OSTI.GOV)
2005-03-30
The Robotic Follow Algorithm enables allows any robotic vehicle to follow a moving target while reactively choosing a route around nearby obstacles. The robotic follow behavior can be used with different camera systems and can be used with thermal or visual tracking as well as other tracking methods such as radio frequency tags.
Robotics virtual rail system and method
Bruemmer, David J [Idaho Falls, ID; Few, Douglas A [Idaho Falls, ID; Walton, Miles C [Idaho Falls, ID
2011-07-05
A virtual track or rail system and method is described for execution by a robot. A user, through a user interface, generates a desired path comprised of at least one segment representative of the virtual track for the robot. Start and end points are assigned to the desired path and velocities are also associated with each of the at least one segment of the desired path. A waypoint file is generated including positions along the virtual track representing the desired path with the positions beginning from the start point to the end point including the velocities of each of the at least one segment. The waypoint file is sent to the robot for traversing along the virtual track.
Daluja, Sachin; Golenberg, Lavie; Cao, Alex; Pandya, Abhilash K; Auner, Gregory W; Klein, Michael D
2009-01-01
Robotic surgery has gradually gained acceptance due to its numerous advantages such as tremor filtration, increased dexterity and motion scaling. There remains, however, a significant scope for improvement, especially in the areas of surgeon-robot interface and autonomous procedures. Previous studies have attempted to identify factors affecting a surgeon's performance in a master-slave robotic system by tracking hand movements. These studies relied on conventional optical or magnetic tracking systems, making their use impracticable in the operating room. This study concentrated on building an intrinsic movement capture platform using microcontroller based hardware wired to a surgical robot. Software was developed to enable tracking and analysis of hand movements while surgical tasks were performed. Movement capture was applied towards automated movements of the robotic instruments. By emulating control signals, recorded surgical movements were replayed by the robot's end-effectors. Though this work uses a surgical robot as the platform, the ideas and concepts put forward are applicable to telerobotic systems in general.
NASA Technical Reports Server (NTRS)
Trube, Matthew J.; Hyslop, Andrew M.; Carignan, Craig R.; Easley, Joseph W.
2012-01-01
A hardware-in-the-loop ground system was developed for simulating a robotic servicer spacecraft tracking a target satellite at short range. A relative navigation sensor package "Argon" is mounted on the end-effector of a Fanuc 430 manipulator, which functions as the base platform of the robotic spacecraft servicer. Machine vision algorithms estimate the pose of the target spacecraft, mounted on a Rotopod R-2000 platform, relay the solution to a simulation of the servicer spacecraft running in "Freespace", which performs guidance, navigation and control functions, integrates dynamics, and issues motion commands to a Fanuc platform controller so that it tracks the simulated servicer spacecraft. Results will be reviewed for several satellite motion scenarios at different ranges. Key words: robotics, satellite, servicing, guidance, navigation, tracking, control, docking.
Hu, Xiaoling; Tong, K Y; Li, R; Chen, M; Xue, J J; Ho, S K; Chen, P N
2010-01-01
Functional electrical stimulation (FES) and rehabilitation robots are techniques used to assist in post-stroke rehabilitation. However, FES and rehabilitation robots are still separate systems currently; and their combined training effects on persons after experiencing a stroke have not been well studied yet. In this work, a new combined FES-robot system driven by user's voluntary intention was developed for wrist joint training after stroke. The performance of the FES-robot assisted wrist tracking was evaluated on five subjects with chronic stroke. With simultaneous assistance from both the FES and robot parts of the system, the motion accuracy was improved and excessive activation in elbow flexor was reduced during wrist tracking.
Direct manipulation of tool-like masters for controlling a master-slave surgical robotic system.
Zhang, Linan; Zhou, Ningxin; Wang, Shuxin
2014-12-01
Robotic-assisted minimally invasive surgery (MIS) can benefit both patients and surgeons. However, the learning curve for robotically assisted procedures can be long and the total system costs are high. Therefore, there is considerable interest in new methods and lower cost controllers for a surgical robotic system. In this study, a knife-master and a forceps-master, shaped similarly to a surgical knife and forceps, were developed as input devices for control of a master-slave surgical robotic system. In addition, a safety strategy was developed to eliminate the master-slave orientation difference and stabilize the surgical system. Master-slave tracking experiments and a ring-and-bar experiment showed that the safety tracking strategy could ensure that the robot system moved stably without any tremor in the tracking motion. Subjects could manipulate the surgical tool to achieve the master-slave operation with less training compared to a mechanical master. Direct manipulation of the small, light and low-cost surgical tools to control a robotic system is a possible operating mode. Surgeons can operate the robotic system in their own familiar way, without long training. The main potential safety issues can be solved by the proposed safety control strategy. Copyright © 2013 John Wiley & Sons, Ltd.
Baigzadehnoe, Barmak; Rahmani, Zahra; Khosravi, Alireza; Rezaie, Behrooz
2017-09-01
In this paper, the position and force tracking control problem of cooperative robot manipulator system handling a common rigid object with unknown dynamical models and unknown external disturbances is investigated. The universal approximation properties of fuzzy logic systems are employed to estimate the unknown system dynamics. On the other hand, by defining new state variables based on the integral and differential of position and orientation errors of the grasped object, the error system of coordinated robot manipulators is constructed. Subsequently by defining the appropriate change of coordinates and using the backstepping design strategy, an adaptive fuzzy backstepping position tracking control scheme is proposed for multi-robot manipulator systems. By utilizing the properties of internal forces, extra terms are also added to the control signals to consider the force tracking problem. Moreover, it is shown that the proposed adaptive fuzzy backstepping position/force control approach ensures all the signals of the closed loop system uniformly ultimately bounded and tracking errors of both positions and forces can converge to small desired values by proper selection of the design parameters. Finally, the theoretic achievements are tested on the two three-link planar robot manipulators cooperatively handling a common object to illustrate the effectiveness of the proposed approach. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
A tracked robot with novel bio-inspired passive "legs".
Sun, Bo; Jing, Xingjian
2017-01-01
For track-based robots, an important aspect is the suppression design, which determines the trafficability and comfort of the whole system. The trafficability limits the robot's working capability, and the riding comfort limits the robot's working effectiveness, especially with some sensitive instruments mounted on or operated. To these aims, a track-based robot equipped with a novel passive bio-inspired suspension is designed and studied systematically in this paper. Animal or insects have very special leg or limb structures which are good for motion control and adaptable to different environments. Inspired by this, a new track-based robot is designed with novel "legs" for connecting the loading wheels to the robot body. Each leg is designed with passive structures and can achieve very high loading capacity but low dynamic stiffness such that the robot can move on rough ground similar to a multi-leg animal or insect. Therefore, the trafficability and riding comfort can be significantly improved without losing loading capacity. The new track-based robot can be well applied to various engineering tasks for providing a stable moving platform of high mobility, better trafficability and excellent loading capacity.
Visual Detection and Tracking System for a Spherical Amphibious Robot
Guo, Shuxiang; Pan, Shaowu; Shi, Liwei; Guo, Ping; He, Yanlin; Tang, Kun
2017-01-01
With the goal of supporting close-range observation tasks of a spherical amphibious robot, such as ecological observations and intelligent surveillance, a moving target detection and tracking system was designed and implemented in this study. Given the restrictions presented by the amphibious environment and the small-sized spherical amphibious robot, an industrial camera and vision algorithms using adaptive appearance models were adopted to construct the proposed system. To handle the problem of light scattering and absorption in the underwater environment, the multi-scale retinex with color restoration algorithm was used for image enhancement. Given the environmental disturbances in practical amphibious scenarios, the Gaussian mixture model was used to detect moving targets entering the field of view of the robot. A fast compressive tracker with a Kalman prediction mechanism was used to track the specified target. Considering the limited load space and the unique mechanical structure of the robot, the proposed vision system was fabricated with a low power system-on-chip using an asymmetric and heterogeneous computing architecture. Experimental results confirmed the validity and high efficiency of the proposed system. The design presented in this paper is able to meet future demands of spherical amphibious robots in biological monitoring and multi-robot cooperation. PMID:28420134
Visual Detection and Tracking System for a Spherical Amphibious Robot.
Guo, Shuxiang; Pan, Shaowu; Shi, Liwei; Guo, Ping; He, Yanlin; Tang, Kun
2017-04-15
With the goal of supporting close-range observation tasks of a spherical amphibious robot, such as ecological observations and intelligent surveillance, a moving target detection and tracking system was designed and implemented in this study. Given the restrictions presented by the amphibious environment and the small-sized spherical amphibious robot, an industrial camera and vision algorithms using adaptive appearance models were adopted to construct the proposed system. To handle the problem of light scattering and absorption in the underwater environment, the multi-scale retinex with color restoration algorithm was used for image enhancement. Given the environmental disturbances in practical amphibious scenarios, the Gaussian mixture model was used to detect moving targets entering the field of view of the robot. A fast compressive tracker with a Kalman prediction mechanism was used to track the specified target. Considering the limited load space and the unique mechanical structure of the robot, the proposed vision system was fabricated with a low power system-on-chip using an asymmetric and heterogeneous computing architecture. Experimental results confirmed the validity and high efficiency of the proposed system. The design presented in this paper is able to meet future demands of spherical amphibious robots in biological monitoring and multi-robot cooperation.
Verification hybrid control of a wheeled mobile robot and manipulator
NASA Astrophysics Data System (ADS)
Muszynska, Magdalena; Burghardt, Andrzej; Kurc, Krzysztof; Szybicki, Dariusz
2016-04-01
In this article, innovative approaches to realization of the wheeled mobile robots and manipulator tracking are presented. Conceptions include application of the neural-fuzzy systems to compensation of the controlled system's nonlinearities in the tracking control task. Proposed control algorithms work on-line, contain structure, that adapt to the changeable work conditions of the controlled systems, and do not require the preliminary learning. The algorithm was verification on the real object which was a Scorbot - ER 4pc robotic manipulator and a Pioneer - 2DX mobile robot.
Imamoglu, Nevrez; Dorronzoro, Enrique; Wei, Zhixuan; Shi, Huangjun; Sekine, Masashi; González, José; Gu, Dongyun; Chen, Weidong; Yu, Wenwei
2014-01-01
Our research is focused on the development of an at-home health care biomonitoring mobile robot for the people in demand. Main task of the robot is to detect and track a designated subject while recognizing his/her activity for analysis and to provide warning in an emergency. In order to push forward the system towards its real application, in this study, we tested the robustness of the robot system with several major environment changes, control parameter changes, and subject variation. First, an improved color tracker was analyzed to find out the limitations and constraints of the robot visual tracking considering the suitable illumination values and tracking distance intervals. Then, regarding subject safety and continuous robot based subject tracking, various control parameters were tested on different layouts in a room. Finally, the main objective of the system is to find out walking activities for different patterns for further analysis. Therefore, we proposed a fast, simple, and person specific new activity recognition model by making full use of localization information, which is robust to partial occlusion. The proposed activity recognition algorithm was tested on different walking patterns with different subjects, and the results showed high recognition accuracy.
Imamoglu, Nevrez; Dorronzoro, Enrique; Wei, Zhixuan; Shi, Huangjun; González, José; Gu, Dongyun; Yu, Wenwei
2014-01-01
Our research is focused on the development of an at-home health care biomonitoring mobile robot for the people in demand. Main task of the robot is to detect and track a designated subject while recognizing his/her activity for analysis and to provide warning in an emergency. In order to push forward the system towards its real application, in this study, we tested the robustness of the robot system with several major environment changes, control parameter changes, and subject variation. First, an improved color tracker was analyzed to find out the limitations and constraints of the robot visual tracking considering the suitable illumination values and tracking distance intervals. Then, regarding subject safety and continuous robot based subject tracking, various control parameters were tested on different layouts in a room. Finally, the main objective of the system is to find out walking activities for different patterns for further analysis. Therefore, we proposed a fast, simple, and person specific new activity recognition model by making full use of localization information, which is robust to partial occlusion. The proposed activity recognition algorithm was tested on different walking patterns with different subjects, and the results showed high recognition accuracy. PMID:25587560
NASA Technical Reports Server (NTRS)
Wilcox, Brian H.; Tso, Kam S.; Litwin, Todd E.; Hayati, Samad A.; Bon, Bruce B.
1991-01-01
Experimental robotic system semiautomatically grasps rotating object, stops rotation, and pulls object to rest in fixture. Based on combination of advanced techniques for sensing and control, constructed to test concepts for robotic recapture of spinning artificial satellites. Potential terrestrial applications for technology developed with help of system includes tracking and grasping of industrial parts on conveyor belts, tracking of vehicles and animals, and soft grasping of moving objects in general.
NASA Technical Reports Server (NTRS)
Tonkay, Gregory
1990-01-01
The following separate topics are addressed: (1) improving a robotic tracking system; and (2) providing insights into orbiter position calibration for radiator inspection. The objective of the tracking system project was to provide the capability to track moving targets more accurately by adjusting parameters in the control system and implementing a predictive algorithm. A computer model was developed to emulate the tracking system. Using this model as a test bed, a self-tuning algorithm was developed to tune the system gains. The model yielded important findings concerning factors that affect the gains. The self-tuning algorithms will provide the concepts to write a program to automatically tune the gains in the real system. The section concerning orbiter position calibration provides a comparison to previous work that had been performed for plant growth. It provided the conceptualized routines required to visually determine the orbiter position and orientation. Furthermore, it identified the types of information which are required to flow between the robot controller and the vision system.
Dealing with the time-varying parameter problem of robot manipulators performing path tracking tasks
NASA Technical Reports Server (NTRS)
Song, Y. D.; Middleton, R. H.
1992-01-01
Many robotic applications involve time-varying payloads during the operation of the robot. It is therefore of interest to consider control schemes that deal with time-varying parameters. Using the properties of the element by element (or Hadarmad) product of matrices, we obtain the robot dynamics in parameter-isolated form, from which a new control scheme is developed. The controller proposed yields zero asymptotic tracking errors when applied to robotic systems with time-varying parameters by using a switching type control law. The results obtained are global in the initial state of the robot, and can be applied to rapidly varying systems.
Mark Tracking: Position/orientation measurements using 4-circle mark and its tracking experiments
NASA Technical Reports Server (NTRS)
Kanda, Shinji; Okabayashi, Keijyu; Maruyama, Tsugito; Uchiyama, Takashi
1994-01-01
Future space robots require position and orientation tracking with visual feedback control to track and capture floating objects and satellites. We developed a four-circle mark that is useful for this purpose. With this mark, four geometric center positions as feature points can be extracted from the mark by simple image processing. We also developed a position and orientation measurement method that uses the four feature points in our mark. The mark gave good enough image measurement accuracy to let space robots approach and contact objects. A visual feedback control system using this mark enabled a robot arm to track a target object accurately. The control system was able to tolerate a time delay of 2 seconds.
Cheng, Long; Hou, Zeng-Guang; Tan, Min; Zhang, W J
2012-10-01
The trajectory tracking problem of a closed-chain five-bar robot is studied in this paper. Based on an error transformation function and the backstepping technique, an approximation-based tracking algorithm is proposed, which can guarantee the control performance of the robotic system in both the stable and transient phases. In particular, the overshoot, settling time, and final tracking error of the robotic system can be all adjusted by properly setting the parameters in the error transformation function. The radial basis function neural network (RBFNN) is used to compensate the complicated nonlinear terms in the closed-loop dynamics of the robotic system. The approximation error of the RBFNN is only required to be bounded, which simplifies the initial "trail-and-error" configuration of the neural network. Illustrative examples are given to verify the theoretical analysis and illustrate the effectiveness of the proposed algorithm. Finally, it is also shown that the proposed approximation-based controller can be simplified by a smart mechanical design of the closed-chain robot, which demonstrates the promise of the integrated design and control philosophy.
Visual perception system and method for a humanoid robot
NASA Technical Reports Server (NTRS)
Chelian, Suhas E. (Inventor); Linn, Douglas Martin (Inventor); Wampler, II, Charles W. (Inventor); Bridgwater, Lyndon (Inventor); Wells, James W. (Inventor); Mc Kay, Neil David (Inventor)
2012-01-01
A robotic system includes a humanoid robot with robotic joints each moveable using an actuator(s), and a distributed controller for controlling the movement of each of the robotic joints. The controller includes a visual perception module (VPM) for visually identifying and tracking an object in the field of view of the robot under threshold lighting conditions. The VPM includes optical devices for collecting an image of the object, a positional extraction device, and a host machine having an algorithm for processing the image and positional information. The algorithm visually identifies and tracks the object, and automatically adapts an exposure time of the optical devices to prevent feature data loss of the image under the threshold lighting conditions. A method of identifying and tracking the object includes collecting the image, extracting positional information of the object, and automatically adapting the exposure time to thereby prevent feature data loss of the image.
Yoo, Sung Jin; Park, Bong Seok
2017-09-06
This paper addresses a distributed connectivity-preserving synchronized tracking problem of multiple uncertain nonholonomic mobile robots with limited communication ranges. The information of the time-varying leader robot is assumed to be accessible to only a small fraction of follower robots. The main contribution of this paper is to introduce a new distributed nonlinear error surface for dealing with both the synchronized tracking and the preservation of the initial connectivity patterns among nonholonomic robots. Based on this nonlinear error surface, the recursive design methodology is presented to construct the approximation-based local adaptive tracking scheme at the robot dynamic level. Furthermore, a technical lemma is established to analyze the stability and the connectivity preservation of the total closed-loop control system in the Lyapunov sense. An example is provided to illustrate the effectiveness of the proposed methodology.
Olfaction and Hearing Based Mobile Robot Navigation for Odor/Sound Source Search
Song, Kai; Liu, Qi; Wang, Qi
2011-01-01
Bionic technology provides a new elicitation for mobile robot navigation since it explores the way to imitate biological senses. In the present study, the challenging problem was how to fuse different biological senses and guide distributed robots to cooperate with each other for target searching. This paper integrates smell, hearing and touch to design an odor/sound tracking multi-robot system. The olfactory robot tracks the chemical odor plume step by step through information fusion from gas sensors and airflow sensors, while two hearing robots localize the sound source by time delay estimation (TDE) and the geometrical position of microphone array. Furthermore, this paper presents a heading direction based mobile robot navigation algorithm, by which the robot can automatically and stably adjust its velocity and direction according to the deviation between the current heading direction measured by magnetoresistive sensor and the expected heading direction acquired through the odor/sound localization strategies. Simultaneously, one robot can communicate with the other robots via a wireless sensor network (WSN). Experimental results show that the olfactory robot can pinpoint the odor source within the distance of 2 m, while two hearing robots can quickly localize and track the olfactory robot in 2 min. The devised multi-robot system can achieve target search with a considerable success ratio and high stability. PMID:22319401
Neural net target-tracking system using structured laser patterns
NASA Astrophysics Data System (ADS)
Cho, Jae-Wan; Lee, Yong-Bum; Lee, Nam-Ho; Park, Soon-Yong; Lee, Jongmin; Choi, Gapchu; Baek, Sunghyun; Park, Dong-Sun
1996-06-01
In this paper, we describe a robot endeffector tracking system using sensory information from recently-announced structured pattern laser diodes, which can generate images with several different types of structured pattern. The neural network approach is employed to recognize the robot endeffector covering the situation of three types of motion: translation, scaling and rotation. Features for the neural network to detect the position of the endeffector are extracted from the preprocessed images. Artificial neural networks are used to store models and to match with unknown input features recognizing the position of the robot endeffector. Since a minimal number of samples are used for different directions of the robot endeffector in the system, an artificial neural network with the generalization capability can be utilized for unknown input features. A feedforward neural network with the generalization capability can be utilized for unknown input features. A feedforward neural network trained with the back propagation learning is used to detect the position of the robot endeffector. Another feedforward neural network module is used to estimate the motion from a sequence of images and to control movements of the robot endeffector. COmbining the tow neural networks for recognizing the robot endeffector and estimating the motion with the preprocessing stage, the whole system keeps tracking of the robot endeffector effectively.
Inspection Robot Based Mobile Sensing and Power Line Tracking for Smart Grid
Byambasuren, Bat-erdene; Kim, Donghan; Oyun-Erdene, Mandakh; Bold, Chinguun; Yura, Jargalbaatar
2016-01-01
Smart sensing and power line tracking is very important in a smart grid system. Illegal electricity usage can be detected by remote current measurement on overhead power lines using an inspection robot. There is a need for accurate detection methods of illegal electricity usage. Stable and correct power line tracking is a very prominent issue. In order to correctly track and make accurate measurements, the swing path of a power line should be previously fitted and predicted by a mathematical function using an inspection robot. After this, the remote inspection robot can follow the power line and measure the current. This paper presents a new power line tracking method using parabolic and circle fitting algorithms for illegal electricity detection. We demonstrate the effectiveness of the proposed tracking method by simulation and experimental results. PMID:26907274
Inspection Robot Based Mobile Sensing and Power Line Tracking for Smart Grid.
Byambasuren, Bat-Erdene; Kim, Donghan; Oyun-Erdene, Mandakh; Bold, Chinguun; Yura, Jargalbaatar
2016-02-19
Smart sensing and power line tracking is very important in a smart grid system. Illegal electricity usage can be detected by remote current measurement on overhead power lines using an inspection robot. There is a need for accurate detection methods of illegal electricity usage. Stable and correct power line tracking is a very prominent issue. In order to correctly track and make accurate measurements, the swing path of a power line should be previously fitted and predicted by a mathematical function using an inspection robot. After this, the remote inspection robot can follow the power line and measure the current. This paper presents a new power line tracking method using parabolic and circle fitting algorithms for illegal electricity detection. We demonstrate the effectiveness of the proposed tracking method by simulation and experimental results.
Active MRI tracking for robotic assisted FUS
NASA Astrophysics Data System (ADS)
Xiao, Xu; Huang, Zhihong; Melzer, Andreas
2017-03-01
MR guided FUS is a noninvasive method producing thermal necrosis at the position of tumors with high accuracy and temperature control. Because the typical size of the ultrasound focus is smaller than the area of interested treatment tissues, focus repositioning become necessary to achieve multiple sonications to cover the whole targeted area. Using MR compatible mechanical actuators could help the ultrasound beam to reach a wider treatment range than using electrical beam steering technique and more flexibility in position the transducer. An active MR tracking technique was combined into the MRgFUS system to help locating the position of the mechanical actuator and the FUS transducer. For this study, a precise agar reference model was designed and fabricated to test the performance of the active tracking technique when it was used on the MR-compatible robotics InnoMotion™ (IBSMM, Engineering spol. s r.o. / Ltd, Czech Republic). The precision, tracking range and positioning speed of the combined robotic FUS system were evaluated in this study. Compared to the existing MR guided HIFU systems, the combined robotic system with active tracking techniques provides a potential that allows the FUS treatment to operate in a larger spatial range and with a faster speed, which is one of the main challenges for organ motion tracking.
[Haptic tracking control for minimally invasive robotic surgery].
Xu, Zhaohong; Song, Chengli; Wu, Wenwu
2012-06-01
Haptic feedback plays a significant role in minimally invasive robotic surgery (MIRS). A major deficiency of the current MIRS is the lack of haptic perception for the surgeon, including the commercially available robot da Vinci surgical system. In this paper, a dynamics model of a haptic robot is established based on Newton-Euler method. Because it took some period of time in exact dynamics solution, we used a digital PID arithmetic dependent on robot dynamics to ensure real-time bilateral control, and it could improve tracking precision and real-time control efficiency. To prove the proposed method, an experimental system in which two Novint Falcon haptic devices acting as master-slave system has been developed. Simulations and experiments showed proposed methods could give instrument force feedbacks to operator, and bilateral control strategy is an effective method to master-slave MIRS. The proposed methods could be used to tele-robotic system.
Understanding of and applications for robot vision guidance at KSC
NASA Technical Reports Server (NTRS)
Shawaga, Lawrence M.
1988-01-01
The primary thrust of robotics at KSC is for the servicing of Space Shuttle remote umbilical docking functions. In order for this to occur, robots performing servicing operations must be capable of tracking a swaying Orbiter in Six Degrees of Freedom (6-DOF). Currently, in NASA KSC's Robotic Applications Development Laboratory (RADL), an ASEA IRB-90 industrial robot is being equipped with a real-time computer vision (hardware and software) system to allow it to track a simulated Orbiter interface (target) in 6-DOF. The real-time computer vision system effectively becomes the eyes for the lab robot, guiding it through a closed loop visual feedback system to move with the simulated Orbiter interface. This paper will address an understanding of this vision guidance system and how it will be applied to remote umbilical servicing at KSC. In addition, other current and future applications will be addressed.
Multisensor-based human detection and tracking for mobile service robots.
Bellotto, Nicola; Hu, Huosheng
2009-02-01
One of fundamental issues for service robots is human-robot interaction. In order to perform such a task and provide the desired services, these robots need to detect and track people in the surroundings. In this paper, we propose a solution for human tracking with a mobile robot that implements multisensor data fusion techniques. The system utilizes a new algorithm for laser-based leg detection using the onboard laser range finder (LRF). The approach is based on the recognition of typical leg patterns extracted from laser scans, which are shown to also be very discriminative in cluttered environments. These patterns can be used to localize both static and walking persons, even when the robot moves. Furthermore, faces are detected using the robot's camera, and the information is fused to the legs' position using a sequential implementation of unscented Kalman filter. The proposed solution is feasible for service robots with a similar device configuration and has been successfully implemented on two different mobile platforms. Several experiments illustrate the effectiveness of our approach, showing that robust human tracking can be performed within complex indoor environments.
Adaptive tracking control of a wheeled mobile robot via an uncalibrated camera system.
Dixon, W E; Dawson, D M; Zergeroglu, E; Behal, A
2001-01-01
This paper considers the problem of position/orientation tracking control of wheeled mobile robots via visual servoing in the presence of parametric uncertainty associated with the mechanical dynamics and the camera system. Specifically, we design an adaptive controller that compensates for uncertain camera and mechanical parameters and ensures global asymptotic position/orientation tracking. Simulation and experimental results are included to illustrate the performance of the control law.
Robust multiperson detection and tracking for mobile service and social robots.
Li, Liyuan; Yan, Shuicheng; Yu, Xinguo; Tan, Yeow Kee; Li, Haizhou
2012-10-01
This paper proposes an efficient system which integrates multiple vision models for robust multiperson detection and tracking for mobile service and social robots in public environments. The core technique is a novel maximum likelihood (ML)-based algorithm which combines the multimodel detections in mean-shift tracking. First, a likelihood probability which integrates detections and similarity to local appearance is defined. Then, an expectation-maximization (EM)-like mean-shift algorithm is derived under the ML framework. In each iteration, the E-step estimates the associations to the detections, and the M-step locates the new position according to the ML criterion. To be robust to the complex crowded scenarios for multiperson tracking, an improved sequential strategy to perform the mean-shift tracking is proposed. Under this strategy, human objects are tracked sequentially according to their priority order. To balance the efficiency and robustness for real-time performance, at each stage, the first two objects from the list of the priority order are tested, and the one with the higher score is selected. The proposed method has been successfully implemented on real-world service and social robots. The vision system integrates stereo-based and histograms-of-oriented-gradients-based human detections, occlusion reasoning, and sequential mean-shift tracking. Various examples to show the advantages and robustness of the proposed system for multiperson tracking from mobile robots are presented. Quantitative evaluations on the performance of multiperson tracking are also performed. Experimental results indicate that significant improvements have been achieved by using the proposed method.
Towards free 3D end-point control for robotic-assisted human reaching using binocular eye tracking.
Maimon-Dror, Roni O; Fernandez-Quesada, Jorge; Zito, Giuseppe A; Konnaris, Charalambos; Dziemian, Sabine; Faisal, A Aldo
2017-07-01
Eye-movements are the only directly observable behavioural signals that are highly correlated with actions at the task level, and proactive of body movements and thus reflect action intentions. Moreover, eye movements are preserved in many movement disorders leading to paralysis (or amputees) from stroke, spinal cord injury, Parkinson's disease, multiple sclerosis, and muscular dystrophy among others. Despite this benefit, eye tracking is not widely used as control interface for robotic interfaces in movement impaired patients due to poor human-robot interfaces. We demonstrate here how combining 3D gaze tracking using our GT3D binocular eye tracker with custom designed 3D head tracking system and calibration method enables continuous 3D end-point control of a robotic arm support system. The users can move their own hand to any location of the workspace by simple looking at the target and winking once. This purely eye tracking based system enables the end-user to retain free head movement and yet achieves high spatial end point accuracy in the order of 6 cm RMSE error in each dimension and standard deviation of 4 cm. 3D calibration is achieved by moving the robot along a 3 dimensional space filling Peano curve while the user is tracking it with their eyes. This results in a fully automated calibration procedure that yields several thousand calibration points versus standard approaches using a dozen points, resulting in beyond state-of-the-art 3D accuracy and precision.
2011-01-01
Background Practicing arm and gait movements with robotic assistance after neurologic injury can help patients improve their movement ability, but patients sometimes reduce their effort during training in response to the assistance. Reduced effort has been hypothesized to diminish clinical outcomes of robotic training. To better understand patient slacking, we studied the role of visual distraction and auditory feedback in modulating patient effort during a common robot-assisted tracking task. Methods Fourteen participants with chronic left hemiparesis from stroke, five control participants with chronic right hemiparesis and fourteen non-impaired healthy control participants, tracked a visual target with their arms while receiving adaptive assistance from a robotic arm exoskeleton. We compared four practice conditions: the baseline tracking task alone; tracking while also performing a visual distracter task; tracking with the visual distracter and sound feedback; and tracking with sound feedback. For the distracter task, symbols were randomly displayed in the corners of the computer screen, and the participants were instructed to click a mouse button when a target symbol appeared. The sound feedback consisted of a repeating beep, with the frequency of repetition made to increase with increasing tracking error. Results Participants with stroke halved their effort and doubled their tracking error when performing the visual distracter task with their left hemiparetic arm. With sound feedback, however, these participants increased their effort and decreased their tracking error close to their baseline levels, while also performing the distracter task successfully. These effects were significantly smaller for the participants who used their non-paretic arm and for the participants without stroke. Conclusions Visual distraction decreased participants effort during a standard robot-assisted movement training task. This effect was greater for the hemiparetic arm, suggesting that the increased demands associated with controlling an affected arm make the motor system more prone to slack when distracted. Providing an alternate sensory channel for feedback, i.e., auditory feedback of tracking error, enabled the participants to simultaneously perform the tracking task and distracter task effectively. Thus, incorporating real-time auditory feedback of performance errors might improve clinical outcomes of robotic therapy systems. PMID:21513561
NASA Astrophysics Data System (ADS)
Ou, Meiying; Sun, Haibin; Gu, Shengwei; Zhang, Yangyi
2017-11-01
This paper investigates the distributed finite-time trajectory tracking control for a group of nonholonomic mobile robots with time-varying unknown parameters and external disturbances. At first, the tracking error system is derived for each mobile robot with the aid of a global invertible transformation, which consists of two subsystems, one is a first-order subsystem and another is a second-order subsystem. Then, the two subsystems are studied respectively, and finite-time disturbance observers are proposed for each robot to estimate the external disturbances. Meanwhile, distributed finite-time tracking controllers are developed for each mobile robot such that all states of each robot can reach the desired value in finite time, where the desired reference value is assumed to be the trajectory of a virtual leader whose information is available to only a subset of the followers, and the followers are assumed to have only local interaction. The effectiveness of the theoretical results is finally illustrated by numerical simulations.
Chang, Yeong-Chan
2005-12-01
This paper addresses the problem of designing adaptive fuzzy-based (or neural network-based) robust controls for a large class of uncertain nonlinear time-varying systems. This class of systems can be perturbed by plant uncertainties, unmodeled perturbations, and external disturbances. Nonlinear H(infinity) control technique incorporated with adaptive control technique and VSC technique is employed to construct the intelligent robust stabilization controller such that an H(infinity) control is achieved. The problem of the robust tracking control design for uncertain robotic systems is employed to demonstrate the effectiveness of the developed robust stabilization control scheme. Therefore, an intelligent robust tracking controller for uncertain robotic systems in the presence of high-degree uncertainties can easily be implemented. Its solution requires only to solve a linear algebraic matrix inequality and a satisfactorily transient and asymptotical tracking performance is guaranteed. A simulation example is made to confirm the performance of the developed control algorithms.
Ali, S M; Reisner, L A; King, B; Cao, A; Auner, G; Klein, M; Pandya, A K
2008-01-01
A redesigned motion control system for the medical robot Aesop allows automating and programming its movements. An IR eye tracking system has been integrated with this control interface to implement an intelligent, autonomous eye gaze-based laparoscopic positioning system. A laparoscopic camera held by Aesop can be moved based on the data from the eye tracking interface to keep the user's gaze point region at the center of a video feedback monitor. This system setup provides autonomous camera control that works around the surgeon, providing an optimal robotic camera platform.
NASA Astrophysics Data System (ADS)
Zou, Yanbiao; Chen, Tao
2018-06-01
To address the problem of low welding precision caused by the poor real-time tracking performance of common welding robots, a novel seam tracking system with excellent real-time tracking performance and high accuracy is designed based on the morphological image processing method and continuous convolution operator tracker (CCOT) object tracking algorithm. The system consists of a six-axis welding robot, a line laser sensor, and an industrial computer. This work also studies the measurement principle involved in the designed system. Through the CCOT algorithm, the weld feature points are determined in real time from the noise image during the welding process, and the 3D coordinate values of these points are obtained according to the measurement principle to control the movement of the robot and the torch in real time. Experimental results show that the sensor has a frequency of 50 Hz. The welding torch runs smoothly with a strong arc light and splash interference. Tracking error can reach ±0.2 mm, and the minimal distance between the laser stripe and the welding molten pool can reach 15 mm, which can significantly fulfill actual welding requirements.
Ando, Noriyasu; Emoto, Shuhei; Kanzaki, Ryohei
2016-12-19
Robotic odor source localization has been a challenging area and one to which biological knowledge has been expected to contribute, as finding odor sources is an essential task for organism survival. Insects are well-studied organisms with regard to odor tracking, and their behavioral strategies have been applied to mobile robots for evaluation. This "bottom-up" approach is a fundamental way to develop biomimetic robots; however, the biological analyses and the modeling of behavioral mechanisms are still ongoing. Therefore, it is still unknown how such a biological system actually works as the controller of a robotic platform. To answer this question, we have developed an insect-controlled robot in which a male adult silkmoth (Bombyx mori) drives a robot car in response to odor stimuli; this can be regarded as a prototype of a future insect-mimetic robot. In the cockpit of the robot, a tethered silkmoth walked on an air-supported ball and an optical sensor measured the ball rotations. These rotations were translated into the movement of the two-wheeled robot. The advantage of this "hybrid" approach is that experimenters can manipulate any parameter of the robot, which enables the evaluation of the odor-tracking capability of insects and provides useful suggestions for robotic odor-tracking. Furthermore, these manipulations are non-invasive ways to alter the sensory-motor relationship of a pilot insect and will be a useful technique for understanding adaptive behaviors.
Chiang, Mao-Hsiung; Lin, Hao-Ting; Hou, Chien-Lun
2011-01-01
In this paper, a stereo vision 3D position measurement system for a three-axial pneumatic parallel mechanism robot arm is presented. The stereo vision 3D position measurement system aims to measure the 3D trajectories of the end-effector of the robot arm. To track the end-effector of the robot arm, the circle detection algorithm is used to detect the desired target and the SAD algorithm is used to track the moving target and to search the corresponding target location along the conjugate epipolar line in the stereo pair. After camera calibration, both intrinsic and extrinsic parameters of the stereo rig can be obtained, so images can be rectified according to the camera parameters. Thus, through the epipolar rectification, the stereo matching process is reduced to a horizontal search along the conjugate epipolar line. Finally, 3D trajectories of the end-effector are computed by stereo triangulation. The experimental results show that the stereo vision 3D position measurement system proposed in this paper can successfully track and measure the fifth-order polynomial trajectory and sinusoidal trajectory of the end-effector of the three- axial pneumatic parallel mechanism robot arm. PMID:22319408
Thermal tracking in mobile robots for leak inspection activities.
Ibarguren, Aitor; Molina, Jorge; Susperregi, Loreto; Maurtua, Iñaki
2013-10-09
Maintenance tasks are crucial for all kind of industries, especially in extensive industrial plants, like solar thermal power plants. The incorporation of robots is a key issue for automating inspection activities, as it will allow a constant and regular control over the whole plant. This paper presents an autonomous robotic system to perform pipeline inspection for early detection and prevention of leakages in thermal power plants, based on the work developed within the MAINBOT (http://www.mainbot.eu) European project. Based on the information provided by a thermographic camera, the system is able to detect leakages in the collectors and pipelines. Beside the leakage detection algorithms, the system includes a particle filter-based tracking algorithm to keep the target in the field of view of the camera and to avoid the irregularities of the terrain while the robot patrols the plant. The information provided by the particle filter is further used to command a robot arm, which handles the camera and ensures that the target is always within the image. The obtained results show the suitability of the proposed approach, adding a tracking algorithm to improve the performance of the leakage detection system.
Thermal Tracking in Mobile Robots for Leak Inspection Activities
Ibarguren, Aitor; Molina, Jorge; Susperregi, Loreto; Maurtua, Iñaki
2013-01-01
Maintenance tasks are crucial for all kind of industries, especially in extensive industrial plants, like solar thermal power plants. The incorporation of robots is a key issue for automating inspection activities, as it will allow a constant and regular control over the whole plant. This paper presents an autonomous robotic system to perform pipeline inspection for early detection and prevention of leakages in thermal power plants, based on the work developed within the MAINBOT (http://www.mainbot.eu) European project. Based on the information provided by a thermographic camera, the system is able to detect leakages in the collectors and pipelines. Beside the leakage detection algorithms, the system includes a particle filter-based tracking algorithm to keep the target in the field of view of the camera and to avoid the irregularities of the terrain while the robot patrols the plant. The information provided by the particle filter is further used to command a robot arm, which handles the camera and ensures that the target is always within the image. The obtained results show the suitability of the proposed approach, adding a tracking algorithm to improve the performance of the leakage detection system. PMID:24113684
Robot Tracking of Human Subjects in Field Environments
NASA Technical Reports Server (NTRS)
Graham, Jeffrey; Shillcutt, Kimberly
2003-01-01
Future planetary exploration will involve both humans and robots. Understanding and improving their interaction is a main focus of research in the Intelligent Systems Branch at NASA's Johnson Space Center. By teaming intelligent robots with astronauts on surface extra-vehicular activities (EVAs), safety and productivity can be improved. The EVA Robotic Assistant (ERA) project was established to study the issues of human-robot teams, to develop a testbed robot to assist space-suited humans in exploration tasks, and to experimentally determine the effectiveness of an EVA assistant robot. A companion paper discusses the ERA project in general, its history starting with ASRO (Astronaut-Rover project), and the results of recent field tests in Arizona. This paper focuses on one aspect of the research, robot tracking, in greater detail: the software architecture and algorithms. The ERA robot is capable of moving towards and/or continuously following mobile or stationary targets or sequences of targets. The contributions made by this research include how the low-level pose data is assembled, normalized and communicated, how the tracking algorithm was generalized and implemented, and qualitative performance reports from recent field tests.
Computer hardware and software for robotic control
NASA Technical Reports Server (NTRS)
Davis, Virgil Leon
1987-01-01
The KSC has implemented an integrated system that coordinates state-of-the-art robotic subsystems. It is a sensor based real-time robotic control system performing operations beyond the capability of an off-the-shelf robot. The integrated system provides real-time closed loop adaptive path control of position and orientation of all six axes of a large robot; enables the implementation of a highly configurable, expandable testbed for sensor system development; and makes several smart distributed control subsystems (robot arm controller, process controller, graphics display, and vision tracking) appear as intelligent peripherals to a supervisory computer coordinating the overall systems.
3D Visual Tracking of an Articulated Robot in Precision Automated Tasks
Alzarok, Hamza; Fletcher, Simon; Longstaff, Andrew P.
2017-01-01
The most compelling requirements for visual tracking systems are a high detection accuracy and an adequate processing speed. However, the combination between the two requirements in real world applications is very challenging due to the fact that more accurate tracking tasks often require longer processing times, while quicker responses for the tracking system are more prone to errors, therefore a trade-off between accuracy and speed, and vice versa is required. This paper aims to achieve the two requirements together by implementing an accurate and time efficient tracking system. In this paper, an eye-to-hand visual system that has the ability to automatically track a moving target is introduced. An enhanced Circular Hough Transform (CHT) is employed for estimating the trajectory of a spherical target in three dimensions, the colour feature of the target was carefully selected by using a new colour selection process, the process relies on the use of a colour segmentation method (Delta E) with the CHT algorithm for finding the proper colour of the tracked target, the target was attached to the six degree of freedom (DOF) robot end-effector that performs a pick-and-place task. A cooperation of two Eye-to Hand cameras with their image Averaging filters are used for obtaining clear and steady images. This paper also examines a new technique for generating and controlling the observation search window in order to increase the computational speed of the tracking system, the techniques is named Controllable Region of interest based on Circular Hough Transform (CRCHT). Moreover, a new mathematical formula is introduced for updating the depth information of the vision system during the object tracking process. For more reliable and accurate tracking, a simplex optimization technique was employed for the calculation of the parameters for camera to robotic transformation matrix. The results obtained show the applicability of the proposed approach to track the moving robot with an overall tracking error of 0.25 mm. Also, the effectiveness of CRCHT technique in saving up to 60% of the overall time required for image processing. PMID:28067860
Algorithms and Sensors for Small Robot Path Following
NASA Technical Reports Server (NTRS)
Hogg, Robert W.; Rankin, Arturo L.; Roumeliotis, Stergios I.; McHenry, Michael C.; Helmick, Daniel M.; Bergh, Charles F.; Matthies, Larry
2002-01-01
Tracked mobile robots in the 20 kg size class are under development for applications in urban reconnaissance. For efficient deployment, it is desirable for teams of robots to be able to automatically execute path following behaviors, with one or more followers tracking the path taken by a leader. The key challenges to enabling such a capability are (l) to develop sensor packages for such small robots that can accurately determine the path of the leader and (2) to develop path following algorithms for the subsequent robots. To date, we have integrated gyros, accelerometers, compass/inclinometers, odometry, and differential GPS into an effective sensing package. This paper describes the sensor package, sensor processing algorithm, and path tracking algorithm we have developed for the leader/follower problem in small robots and shows the result of performance characterization of the system. We also document pragmatic lessons learned about design, construction, and electromagnetic interference issues particular to the performance of state sensors on small robots.
Finite-time tracking control for multiple non-holonomic mobile robots based on visual servoing
NASA Astrophysics Data System (ADS)
Ou, Meiying; Li, Shihua; Wang, Chaoli
2013-12-01
This paper investigates finite-time tracking control problem of multiple non-holonomic mobile robots via visual servoing. It is assumed that the pinhole camera is fixed to the ceiling, and camera parameters are unknown. The desired reference trajectory is represented by a virtual leader whose states are available to only a subset of the followers, and the followers have only interaction. First, the camera-objective visual kinematic model is introduced by utilising the pinhole camera model for each mobile robot. Second, a unified tracking error system between camera-objective visual servoing model and desired reference trajectory is introduced. Third, based on the neighbour rule and by using finite-time control method, continuous distributed cooperative finite-time tracking control laws are designed for each mobile robot with unknown camera parameters, where the communication topology among the multiple mobile robots is assumed to be a directed graph. Rigorous proof shows that the group of mobile robots converges to the desired reference trajectory in finite time. Simulation example illustrates the effectiveness of our method.
The Co-simulation of Humanoid Robot Based on Solidworks, ADAMS and Simulink
NASA Astrophysics Data System (ADS)
Song, Dalei; Zheng, Lidan; Wang, Li; Qi, Weiwei; Li, Yanli
A simulation method of adaptive controller is proposed for the humanoid robot system based on co-simulation of Solidworks, ADAMS and Simulink. A complex mathematical modeling process is avoided by this method, and the real time dynamic simulating function of Simulink would be exerted adequately. This method could be generalized to other complicated control system. This method is adopted to build and analyse the model of humanoid robot. The trajectory tracking and adaptive controller design also proceed based on it. The effect of trajectory tracking is evaluated by fitting-curve theory of least squares method. The anti-interference capability of the robot is improved a lot through comparative analysis.
NASA Astrophysics Data System (ADS)
Wray, J. D.
2003-05-01
The robotic observatory telescope must point precisely on the target object, and then track autonomously to a fraction of the FWHM of the system PSF for durations of ten to twenty minutes or more. It must retain this precision while continuing to function at rates approaching thousands of observations per night for all its years of useful life. These stringent requirements raise new challenges unique to robotic telescope systems design. Critical design considerations are driven by the applicability of the above requirements to all systems of the robotic observatory, including telescope and instrument systems, telescope-dome enclosure systems, combined electrical and electronics systems, environmental (e.g. seeing) control systems and integrated computer control software systems. Traditional telescope design considerations include the effects of differential thermal strain, elastic flexure, plastic flexure and slack or backlash with respect to focal stability, optical alignment and angular pointing and tracking precision. Robotic observatory design must holistically encapsulate these traditional considerations within the overall objective of maximized long-term sustainable precision performance. This overall objective is accomplished through combining appropriate mechanical and dynamical system characteristics with a full-time real-time telescope mount model feedback computer control system. Important design considerations include: identifying and reducing quasi-zero-backlash; increasing size to increase precision; directly encoding axis shaft rotation; pointing and tracking operation via real-time feedback between precision mount model and axis mounted encoders; use of monolithic construction whenever appropriate for sustainable mechanical integrity; accelerating dome motion to eliminate repetitive shock; ducting internal telescope air to outside dome; and the principal design criteria: maximizing elastic repeatability while minimizing slack, plastic deformation and hysteresis to facilitate long-term repeatably precise pointing and tracking performance.
Kim, Yonjae; Leonard, Simon; Shademan, Azad; Krieger, Axel; Kim, Peter C W
2014-06-01
Current surgical robots are controlled by a mechanical master located away from the patient, tracking surgeon's hands by wire and pulleys or mechanical linkage. Contactless hand tracking for surgical robot control is an attractive alternative, because it can be executed with minimal footprint at the patient's bedside without impairing sterility, while eliminating current disassociation between surgeon and patient. We compared technical and technologic feasibility of contactless hand tracking to the current clinical standard master controllers. A hand-tracking system (Kinect™-based 3Gear), a wire-based mechanical master (Mantis Duo), and a clinical mechanical linkage master (da Vinci) were evaluated for technical parameters with strong clinical relevance: system latency, static noise, robot slave tremor, and controller range. Five experienced surgeons performed a skill comparison study, evaluating the three different master controllers for efficiency and accuracy in peg transfer and pointing tasks. da Vinci had the lowest latency of 89 ms, followed by Mantis with 374 ms and 3Gear with 576 ms. Mantis and da Vinci produced zero static error. 3Gear produced average static error of 0.49 mm. The tremor of the robot used by the 3Gear and Mantis system had a radius of 1.7 mm compared with 0.5 mm for da Vinci. The three master controllers all had similar range. The surgeons took 1.98 times longer to complete the peg transfer task with the 3Gear system compared with Mantis, and 2.72 times longer with Mantis compared with da Vinci (p value 2.1e-9). For the pointer task, surgeons were most accurate with da Vinci with average error of 0.72 mm compared with Mantis's 1.61 mm and 3Gear's 2.41 mm (p value 0.00078). Contactless hand-tracking technology as a surgical master can execute simple surgical tasks. Whereas traditional master controllers outperformed, given that contactless hand-tracking is a first-generation technology, clinical potential is promising and could become a reality with some technical improvements.
Applied estimation for hybrid dynamical systems using perceptional information
NASA Astrophysics Data System (ADS)
Plotnik, Aaron M.
This dissertation uses the motivating example of robotic tracking of mobile deep ocean animals to present innovations in robotic perception and estimation for hybrid dynamical systems. An approach to estimation for hybrid systems is presented that utilizes uncertain perceptional information about the system's mode to improve tracking of its mode and continuous states. This results in significant improvements in situations where previously reported methods of estimation for hybrid systems perform poorly due to poor distinguishability of the modes. The specific application that motivates this research is an automatic underwater robotic observation system that follows and films individual deep ocean animals. A first version of such a system has been developed jointly by the Stanford Aerospace Robotics Laboratory and Monterey Bay Aquarium Research Institute (MBARI). This robotic observation system is successfully fielded on MBARI's ROVs, but agile specimens often evade the system. When a human ROV pilot performs this task, one advantage that he has over the robotic observation system in these situations is the ability to use visual perceptional information about the target, immediately recognizing any changes in the specimen's behavior mode. With the approach of the human pilot in mind, a new version of the robotic observation system is proposed which is extended to (a) derive perceptional information (visual cues) about the behavior mode of the tracked specimen, and (b) merge this dissimilar, discrete and uncertain information with more traditional continuous noisy sensor data by extending existing algorithms for hybrid estimation. These performance enhancements are enabled by integrating techniques in hybrid estimation, computer vision and machine learning. First, real-time computer vision and classification algorithms extract a visual observation of the target's behavior mode. Existing hybrid estimation algorithms are extended to admit this uncertain but discrete observation, complementing the information available from more traditional sensors. State tracking is achieved using a new form of Rao-Blackwellized particle filter called the mode-observed Gaussian Particle Filter. Performance is demonstrated using data from simulation and data collected on actual specimens in the ocean. The framework for estimation using both traditional and perceptional information is easily extensible to other stochastic hybrid systems with mode-related perceptional observations available.
Testbed for remote telepresence research
NASA Astrophysics Data System (ADS)
Adnan, Sarmad; Cheatham, John B., Jr.
1992-11-01
Teleoperated robots offer solutions to problems associated with operations in remote and unknown environments, such as space. Teleoperated robots can perform tasks related to inspection, maintenance, and retrieval. A video camera can be used to provide some assistance in teleoperations, but for fine manipulation and control, a telepresence system that gives the operator a sense of actually being at the remote location is more desirable. A telepresence system comprised of a head-tracking stereo camera system, a kinematically redundant arm, and an omnidirectional mobile robot has been developed at the mechanical engineering department at Rice University. This paper describes the design and implementation of this system, its control hardware, and software. The mobile omnidirectional robot has three independent degrees of freedom that permit independent control of translation and rotation, thereby simulating a free flying robot in a plane. The kinematically redundant robot arm has eight degrees of freedom that assist in obstacle and singularity avoidance. The on-board control computers permit control of the robot from the dual hand controllers via a radio modem system. A head-mounted display system provides the user with a stereo view from a pair of cameras attached to the mobile robotics system. The head tracking camera system moves stereo cameras mounted on a three degree of freedom platform to coordinate with the operator's head movements. This telepresence system provides a framework for research in remote telepresence, and teleoperations for space.
Coordinated Dynamic Behaviors for Multirobot Systems With Collision Avoidance.
Sabattini, Lorenzo; Secchi, Cristian; Fantuzzi, Cesare
2017-12-01
In this paper, we propose a novel methodology for achieving complex dynamic behaviors in multirobot systems. In particular, we consider a multirobot system partitioned into two subgroups: 1) dependent and 2) independent robots. Independent robots are utilized as a control input, and their motion is controlled in such a way that the dependent robots solve a tracking problem, that is following arbitrarily defined setpoint trajectories, in a coordinated manner. The control strategy proposed in this paper explicitly addresses the collision avoidance problem, utilizing a null space-based behavioral approach: this leads to combining, in a non conflicting manner, the tracking control law with a collision avoidance strategy. The combination of these control actions allows the robots to execute their task in a safe way. Avoidance of collisions is formally proven in this paper, and the proposed methodology is validated by means of simulations and experiments on real robots.
NASA Technical Reports Server (NTRS)
Davis, V. Leon; Nordeen, Ross
1988-01-01
A laboratory for developing robotics technology for hazardous and repetitive Shuttle and payload processing activities is discussed. An overview of the computer hardware and software responsible for integrating the laboratory systems is given. The center's anthropomorphic robot is placed on a track allowing it to be moved to different stations. Various aspects of the laboratory equipment are described, including industrial robot arm control, smart systems integration, the supervisory computer, programmable process controller, real-time tracking controller, image processing hardware, and control display graphics. Topics of research include: automated loading and unloading of hypergolics for space vehicles and payloads; the use of mobile robotics for security, fire fighting, and hazardous spill operations; nondestructive testing for SRB joint and seal verification; Shuttle Orbiter radiator damage inspection; and Orbiter contour measurements. The possibility of expanding the laboratory in the future is examined.
Human-tracking strategies for a six-legged rescue robot based on distance and view
NASA Astrophysics Data System (ADS)
Pan, Yang; Gao, Feng; Qi, Chenkun; Chai, Xun
2016-03-01
Human tracking is an important issue for intelligent robotic control and can be used in many scenarios, such as robotic services and human-robot cooperation. Most of current human-tracking methods are targeted for mobile/tracked robots, but few of them can be used for legged robots. Two novel human-tracking strategies, view priority strategy and distance priority strategy, are proposed specially for legged robots, which enable them to track humans in various complex terrains. View priority strategy focuses on keeping humans in its view angle arrange with priority, while its counterpart, distance priority strategy, focuses on keeping human at a reasonable distance with priority. To evaluate these strategies, two indexes(average and minimum tracking capability) are defined. With the help of these indexes, the view priority strategy shows advantages compared with distance priority strategy. The optimization is done in terms of these indexes, which let the robot has maximum tracking capability. The simulation results show that the robot can track humans with different curves like square, circular, sine and screw paths. Two novel control strategies are proposed which specially concerning legged robot characteristics to solve human tracking problems more efficiently in rescue circumstances.
Cleaning Robot for Solar Panels in Solar Power Station
NASA Astrophysics Data System (ADS)
Hang, Lu-Bin; Shen, Cheng-Wei; Bian, Huai-Qiang; Wang, Yan
2016-05-01
The dust particles on solar panel surface have been a serious problem for the photovoltaic industry, a new monorail-tracked robot used for automatic cleaning of solar panel is presented in this paper. To meet the requirement of comprehensive and stable cleaning of PV array, the monorail-tracked pattern of robot is introduced based on the monorail structure technique. The running and striding mechanism are designed for mobility of robot on the solar panels. According to the carrying capacity and water circulation mechanism, a type of self-cleaning device with filtering system is developed. Combined with the computer software and communications technology, the control system is built in this robot, which can realize the functions of autonomous operation, positioning and monitoring. The application of this developed cleaning robot can actualize the Industrialization of automatic cleaning for PV components and have wide market prospect.
Ultra-Wideband Tracking System Design for Relative Navigation
NASA Technical Reports Server (NTRS)
Ni, Jianjun David; Arndt, Dickey; Bgo, Phong; Dekome, Kent; Dusl, John
2011-01-01
This presentation briefly discusses a design effort for a prototype ultra-wideband (UWB) time-difference-of-arrival (TDOA) tracking system that is currently under development at NASA Johnson Space Center (JSC). The system is being designed for use in localization and navigation of a rover in a GPS deprived environment for surface missions. In one application enabled by the UWB tracking, a robotic vehicle carrying equipments can autonomously follow a crewed rover from work site to work site such that resources can be carried from one landing mission to the next thereby saving up-mass. The UWB Systems Group at JSC has developed a UWB TDOA High Resolution Proximity Tracking System which can achieve sub-inch tracking accuracy of a target within the radius of the tracking baseline [1]. By extending the tracking capability beyond the radius of the tracking baseline, a tracking system is being designed to enable relative navigation between two vehicles for surface missions. A prototype UWB TDOA tracking system has been designed, implemented, tested, and proven feasible for relative navigation of robotic vehicles. Future work includes testing the system with the application code to increase the tracking update rate and evaluating the linear tracking baseline to improve the flexibility of antenna mounting on the following vehicle.
A Segway RMP-based robotic transport system
NASA Astrophysics Data System (ADS)
Nguyen, Hoa G.; Kogut, Greg; Barua, Ripan; Burmeister, Aaron; Pezeshkian, Narek; Powell, Darren; Farrington, Nathan; Wimmer, Matt; Cicchetto, Brett; Heng, Chana; Ramirez, Velia
2004-12-01
In the area of logistics, there currently is a capability gap between the one-ton Army robotic Multifunction Utility/Logistics and Equipment (MULE) vehicle and a soldier"s backpack. The Unmanned Systems Branch at Space and Naval Warfare Systems Center (SPAWAR Systems Center, or SSC), San Diego, with the assistance of a group of interns from nearby High Tech High School, has demonstrated enabling technologies for a solution that fills this gap. A small robotic transport system has been developed based on the Segway Robotic Mobility Platform (RMP). We have demonstrated teleoperated control of this robotic transport system, and conducted two demonstrations of autonomous behaviors. Both demonstrations involved a robotic transporter following a human leader. In the first demonstration, the transporter used a vision system running a continuously adaptive mean-shift filter to track and follow a human. In the second demonstration, the separation between leader and follower was significantly increased using Global Positioning System (GPS) information. The track of the human leader, with a GPS unit in his backpack, was sent wirelessly to the transporter, also equipped with a GPS unit. The robotic transporter traced the path of the human leader by following these GPS breadcrumbs. We have additionally demonstrated a robotic medical patient transport capability by using the Segway RMP to power a mock-up of the Life Support for Trauma and Transport (LSTAT) patient care platform, on a standard NATO litter carrier. This paper describes the development of our demonstration robotic transport system and the various experiments conducted.
Mobility of lightweight robots over snow
NASA Astrophysics Data System (ADS)
Lever, James H.; Shoop, Sally A.
2006-05-01
Snowfields are challenging terrain for lightweight (<50 kg) unmanned ground vehicles. Deep sinkage, high snowcompaction resistance, traction loss while turning and ingestion of snow into the drive train can cause immobility within a few meters of travel. However, for suitably designed vehicles, deep snow offers a smooth, uniform surface that can obliterate obstacles. Key requirements for good over-snow mobility are low ground pressure, large clearance relative to vehicle size and a drive system that tolerates cohesive snow. A small robot will invariably encounter deep snow relative to its ground clearance. Because a single snowstorm can easily deposit 30 cm of fresh snow, robots with ground clearance less than about 10 cm must travel over the snow rather than gain support from the underlying ground. This can be accomplished using low-pressure tracks (< 1.5 kPa). Even still, snow-compaction resistance can exceed 20% of vehicle weight. Also, despite relatively high traction coefficients for low track pressures, differential or skid steering is difficult because the outboard track can easily break traction as the vehicle attempts to turn against the snow. Short track lengths (relative to track separation) or coupled articulated robots offer steering solutions for deep snow. This paper presents preliminary guidance to design lightweight robots for good mobility over snow based on mobility theory and tests of PackBot, Talon and SnoBot, a custom-designed research robot. Because many other considerations constrain robot designs, this guidance can help with development of winterization kits to improve the over-snow performance of existing robots.
Gas Source Localization via Behaviour Based Mobile Robot and Weighted Arithmetic Mean
NASA Astrophysics Data System (ADS)
Yeon, Ahmad Shakaff Ali; Kamarudin, Kamarulzaman; Visvanathan, Retnam; Mamduh Syed Zakaria, Syed Muhammad; Zakaria, Ammar; Munirah Kamarudin, Latifah
2018-03-01
This work is concerned with the localization of gas source in dynamic indoor environment using a single mobile robot system. Algorithms such as Braitenberg, Zig-Zag and the combination of the two were implemented on the mobile robot as gas plume searching and tracing behaviours. To calculate the gas source location, a weighted arithmetic mean strategy was used. All experiments were done on an experimental testbed consisting of a large gas sensor array (LGSA) to monitor real-time gas concentration within the testbed. Ethanol gas was released within the testbed and the source location was marked using a pattern that can be tracked by a pattern tracking system. A pattern template was also mounted on the mobile robot to track the trajectory of the mobile robot. Measurements taken by the mobile robot and the LGSA were then compared to verify the experiments. A combined total of 36.5 hours of real time experimental runs were done and the typical results from such experiments were presented in this paper. From the results, we obtained gas source localization errors between 0.4m to 1.2m from the real source location.
Dai, Yanyan; Kim, YoonGu; Wee, SungGil; Lee, DongHa; Lee, SukGyu
2015-05-01
This paper describes a switching formation strategy for multi-robots with velocity constraints to avoid and cross obstacles. In the strategy, a leader robot plans a safe path using the geometric obstacle avoidance control method (GOACM). By calculating new desired distances and bearing angles with the leader robot, the follower robots switch into a safe formation. With considering collision avoidance, a novel robot priority model, based on the desired distance and bearing angle between the leader and follower robots, is designed during the obstacle avoidance process. The adaptive tracking control algorithm guarantees that the trajectory and velocity tracking errors converge to zero. To demonstrate the validity of the proposed methods, simulation and experiment results present that multi-robots effectively form and switch formation avoiding obstacles without collisions. Copyright © 2014 ISA. Published by Elsevier Ltd. All rights reserved.
Robust tracking of dexterous continuum robots: Fusing FBG shape sensing and stereo vision.
Rumei Zhang; Hao Liu; Jianda Han
2017-07-01
Robust and efficient tracking of continuum robots is important for improving patient safety during space-confined minimally invasive surgery, however, it has been a particularly challenging task for researchers. In this paper, we present a novel tracking scheme by fusing fiber Bragg grating (FBG) shape sensing and stereo vision to estimate the position of continuum robots. Previous visual tracking easily suffers from the lack of robustness and leads to failure, while the FBG shape sensor can only reconstruct the local shape with integral cumulative error. The proposed fusion is anticipated to compensate for their shortcomings and improve the tracking accuracy. To verify its effectiveness, the robots' centerline is recognized by morphology operation and reconstructed by stereo matching algorithm. The shape obtained by FBG sensor is transformed into distal tip position with respect to the camera coordinate system through previously calibrated registration matrices. An experimental platform was set up and repeated tracking experiments were carried out. The accuracy estimated by averaging the absolute positioning errors between shape sensing and stereo vision is 0.67±0.65 mm, 0.41±0.25 mm, 0.72±0.43 mm for x, y and z, respectively. Results indicate that the proposed fusion is feasible and can be used for closed-loop control of continuum robots.
Chiang, Mao-Hsiung; Lin, Hao-Ting
2011-01-01
This study aimed to develop a novel 3D parallel mechanism robot driven by three vertical-axial pneumatic actuators with a stereo vision system for path tracking control. The mechanical system and the control system are the primary novel parts for developing a 3D parallel mechanism robot. In the mechanical system, a 3D parallel mechanism robot contains three serial chains, a fixed base, a movable platform and a pneumatic servo system. The parallel mechanism are designed and analyzed first for realizing a 3D motion in the X-Y-Z coordinate system of the robot's end-effector. The inverse kinematics and the forward kinematics of the parallel mechanism robot are investigated by using the Denavit-Hartenberg notation (D-H notation) coordinate system. The pneumatic actuators in the three vertical motion axes are modeled. In the control system, the Fourier series-based adaptive sliding-mode controller with H(∞) tracking performance is used to design the path tracking controllers of the three vertical servo pneumatic actuators for realizing 3D path tracking control of the end-effector. Three optical linear scales are used to measure the position of the three pneumatic actuators. The 3D position of the end-effector is then calculated from the measuring position of the three pneumatic actuators by means of the kinematics. However, the calculated 3D position of the end-effector cannot consider the manufacturing and assembly tolerance of the joints and the parallel mechanism so that errors between the actual position and the calculated 3D position of the end-effector exist. In order to improve this situation, sensor collaboration is developed in this paper. A stereo vision system is used to collaborate with the three position sensors of the pneumatic actuators. The stereo vision system combining two CCD serves to measure the actual 3D position of the end-effector and calibrate the error between the actual and the calculated 3D position of the end-effector. Furthermore, to verify the feasibility of the proposed parallel mechanism robot driven by three vertical pneumatic servo actuators, a full-scale test rig of the proposed parallel mechanism pneumatic robot is set up. Thus, simulations and experiments for different complex 3D motion profiles of the robot end-effector can be successfully achieved. The desired, the actual and the calculated 3D position of the end-effector can be compared in the complex 3D motion control.
Ultra-Wideband Time-Difference-of-Arrival High Resolution 3D Proximity Tracking System
NASA Technical Reports Server (NTRS)
Ni, Jianjun; Arndt, Dickey; Ngo, Phong; Phan, Chau; Dekome, Kent; Dusl, John
2010-01-01
This paper describes a research and development effort for a prototype ultra-wideband (UWB) tracking system that is currently under development at NASA Johnson Space Center (JSC). The system is being studied for use in tracking of lunar./Mars rovers and astronauts during early exploration missions when satellite navigation systems are not available. U IATB impulse radio (UWB-IR) technology is exploited in the design and implementation of the prototype location and tracking system. A three-dimensional (3D) proximity tracking prototype design using commercially available UWB products is proposed to implement the Time-Difference- Of-Arrival (TDOA) tracking methodology in this research effort. The TDOA tracking algorithm is utilized for location estimation in the prototype system, not only to exploit the precise time resolution possible with UWB signals, but also to eliminate the need for synchronization between the transmitter and the receiver. Simulations show that the TDOA algorithm can achieve the fine tracking resolution with low noise TDOA estimates for close-in tracking. Field tests demonstrated that this prototype UWB TDOA High Resolution 3D Proximity Tracking System is feasible for providing positioning-awareness information in a 3D space to a robotic control system. This 3D tracking system is developed for a robotic control system in a facility called "Moonyard" at Honeywell Defense & System in Arizona under a Space Act Agreement.
Technological advances in robotic-assisted laparoscopic surgery.
Tan, Gerald Y; Goel, Raj K; Kaouk, Jihad H; Tewari, Ashutosh K
2009-05-01
In this article, the authors describe the evolution of urologic robotic systems and the current state-of-the-art features and existing limitations of the da Vinci S HD System (Intuitive Surgical, Inc.). They then review promising innovations in scaling down the footprint of robotic platforms, the early experience with mobile miniaturized in vivo robots, advances in endoscopic navigation systems using augmented reality technologies and tracking devices, the emergence of technologies for robotic natural orifice transluminal endoscopic surgery and single-port surgery, advances in flexible robotics and haptics, the development of new virtual reality simulator training platforms compatible with the existing da Vinci system, and recent experiences with remote robotic surgery and telestration.
Niu, Jie; Yang, Qianqian; Wang, Xiaoyun; Song, Rong
2017-01-01
Robot-aided rehabilitation has become an important technology to restore and reinforce motor functions of patients with extremity impairment, whereas it can be extremely challenging to achieve satisfactory tracking performance due to uncertainties and disturbances during rehabilitation training. In this paper, a wire-driven rehabilitation robot that can work over a three-dimensional space is designed for upper-limb rehabilitation, and sliding mode control with nonlinear disturbance observer is designed for the robot to deal with the problem of unpredictable disturbances during robot-assisted training. Then, simulation and experiments of trajectory tracking are carried out to evaluate the performance of the system, the position errors, and the output forces of the designed control scheme are compared with those of the traditional sliding mode control (SMC) scheme. The results show that the designed control scheme can effectively reduce the tracking errors and chattering of the output forces as compared with the traditional SMC scheme, which indicates that the nonlinear disturbance observer can reduce the effect of unpredictable disturbances. The designed control scheme for the wire-driven rehabilitation robot has potential to assist patients with stroke in performing repetitive rehabilitation training.
Scaled position-force tracking for wireless teleoperation of miniaturized surgical robotic system.
Guo, Jing; Liu, Chao; Poignet, Philippe
2014-01-01
Miniaturized surgical robotic system presents promising trend for reducing invasiveness during operation. However, cables used for power and communication may affect its performance. In this paper we chose Zigbee wireless communication as a means to replace communication cables for miniaturized surgical robot. Nevertheless, time delay caused by wireless communication presents a new challenge to performance and stability of the teleoperation system. We proposed a bilateral wireless teleoperation architecture taking into consideration of the effect of position-force scaling between operator and slave. Optimal position-force tracking performance is obtained and the overall system is shown to be passive with a simple condition on the scaling factors satisfied. Simulation studies verify the efficiency of the proposed scaled wireless teleoperation scheme.
Singh, Tarkeshwar; Perry, Christopher M; Herter, Troy M
2016-01-26
Robotic and virtual-reality systems offer tremendous potential for improving assessment and rehabilitation of neurological disorders affecting the upper extremity. A key feature of these systems is that visual stimuli are often presented within the same workspace as the hands (i.e., peripersonal space). Integrating video-based remote eye tracking with robotic and virtual-reality systems can provide an additional tool for investigating how cognitive processes influence visuomotor learning and rehabilitation of the upper extremity. However, remote eye tracking systems typically compute ocular kinematics by assuming eye movements are made in a plane with constant depth (e.g. frontal plane). When visual stimuli are presented at variable depths (e.g. transverse plane), eye movements have a vergence component that may influence reliable detection of gaze events (fixations, smooth pursuits and saccades). To our knowledge, there are no available methods to classify gaze events in the transverse plane for monocular remote eye tracking systems. Here we present a geometrical method to compute ocular kinematics from a monocular remote eye tracking system when visual stimuli are presented in the transverse plane. We then use the obtained kinematics to compute velocity-based thresholds that allow us to accurately identify onsets and offsets of fixations, saccades and smooth pursuits. Finally, we validate our algorithm by comparing the gaze events computed by the algorithm with those obtained from the eye-tracking software and manual digitization. Within the transverse plane, our algorithm reliably differentiates saccades from fixations (static visual stimuli) and smooth pursuits from saccades and fixations when visual stimuli are dynamic. The proposed methods provide advancements for examining eye movements in robotic and virtual-reality systems. Our methods can also be used with other video-based or tablet-based systems in which eye movements are performed in a peripersonal plane with variable depth.
Li, Luyang; Liu, Yun-Hui; Jiang, Tianjiao; Wang, Kai; Fang, Mu
2018-02-01
Despite tremendous efforts made for years, trajectory tracking control (TC) of a nonholonomic mobile robot (NMR) without global positioning system remains an open problem. The major reason is the difficulty to localize the robot by using its onboard sensors only. In this paper, a newly designed adaptive trajectory TC method is proposed for the NMR without its position, orientation, and velocity measurements. The controller is designed on the basis of a novel algorithm to estimate position and velocity of the robot online from visual feedback of an omnidirectional camera. It is theoretically proved that the proposed algorithm yields the TC errors to asymptotically converge to zero. Real-world experiments are conducted on a wheeled NMR to validate the feasibility of the control system.
System design of a hand-held mobile robot for craniotomy.
Kane, Gavin; Eggers, Georg; Boesecke, Robert; Raczkowsky, Jörg; Wörn, Heinz; Marmulla, Rüdiger; Mühling, Joachim
2009-01-01
This contribution reports the development and initial testing of a Mobile Robot System for Surgical Craniotomy, the Craniostar. A kinematic system based on a unicycle robot is analysed to provide local positioning through two spiked wheels gripping directly onto a patients skull. A control system based on a shared control system between both the Surgeon and Robot is employed in a hand-held design that is tested initially on plastic phantom and swine skulls. Results indicate that the system has substantially lower risk than present robotically assisted craniotomies, and despite being a hand-held mobile robot, the Craniostar is still capable of sub-millimetre accuracy in tracking along a trajectory and thus achieving an accurate transfer of pre-surgical plan to the operating room procedure, without the large impact of current medical robots based on modified industrial robots.
NASA Astrophysics Data System (ADS)
Yoo, Sung Jin
2016-11-01
This paper presents a theoretical design approach for output-feedback formation tracking of multiple mobile robots under wheel perturbations. It is assumed that these perturbations are unknown and the linear and angular velocities of the robots are unmeasurable. First, adaptive state observers for estimating unmeasurable velocities of the robots are developed under the robots' kinematics and dynamics including wheel perturbation effects. Then, we derive a virtual-structure-based formation tracker scheme according to the observer dynamic surface design procedure. The main difficulty of the output-feedback control design is to manage the coupling problems between unmeasurable velocities and unknown wheel perturbation effects. These problems are avoided by using the adaptive technique and the function approximation property based on fuzzy logic systems. From the Lyapunov stability analysis, it is shown that point tracking errors of each robot and synchronisation errors for the desired formation converge to an adjustable neighbourhood of the origin, while all signals in the controlled closed-loop system are semiglobally uniformly ultimately bounded.
Zhao, Ximei; Ren, Chengyi; Liu, Hao; Li, Haogyi
2014-12-01
Robotic catheter minimally invasive operation requires that the driver control system has the advantages of quick response, strong anti-jamming and real-time tracking of target trajectory. Since the catheter parameters of itself and movement environment and other factors continuously change, when the driver is controlled using traditional proportional-integral-derivative (PID), the controller gain becomes fixed once the PID parameters are set. It can not change with the change of the parameters of the object and environmental disturbance so that its change affects the position tracking accuracy, and may bring a large overshoot endangering patients' vessel. Therefore, this paper adopts fuzzy PID control method to adjust PID gain parameters in the tracking process in order to improve the system anti-interference ability, dynamic performance and tracking accuracy. The simulation results showed that the fuzzy PID control method had a fast tracking performance and a strong robustness. Compared with those of traditional PID control, the feasibility and practicability of fuzzy PID control are verified in a robotic catheter minimally invasive operation.
Robotic fish tracking method based on suboptimal interval Kalman filter
NASA Astrophysics Data System (ADS)
Tong, Xiaohong; Tang, Chao
2017-11-01
Autonomous Underwater Vehicle (AUV) research focused on tracking and positioning, precise guidance and return to dock and other fields. The robotic fish of AUV has become a hot application in intelligent education, civil and military etc. In nonlinear tracking analysis of robotic fish, which was found that the interval Kalman filter algorithm contains all possible filter results, but the range is wide, relatively conservative, and the interval data vector is uncertain before implementation. This paper proposes a ptimization algorithm of suboptimal interval Kalman filter. Suboptimal interval Kalman filter scheme used the interval inverse matrix with its worst inverse instead, is more approximate nonlinear state equation and measurement equation than the standard interval Kalman filter, increases the accuracy of the nominal dynamic system model, improves the speed and precision of tracking system. Monte-Carlo simulation results show that the optimal trajectory of sub optimal interval Kalman filter algorithm is better than that of the interval Kalman filter method and the standard method of the filter.
Adaptive control of space based robot manipulators
NASA Technical Reports Server (NTRS)
Walker, Michael W.; Wee, Liang-Boon
1991-01-01
For space based robots in which the base is free to move, motion planning and control is complicated by uncertainties in the inertial properties of the manipulator and its load. A new adaptive control method is presented for space based robots which achieves globally stable trajectory tracking in the presence of uncertainties in the inertial parameters of the system. A partition is made of the fifteen degree of freedom system dynamics into two parts: a nine degree of freedom invertible portion and a six degree of freedom noninvertible portion. The controller is then designed to achieve trajectory tracking of the invertible portion of the system. This portion consist of the manipulator joint positions and the orientation of the base. The motion of the noninvertible portion is bounded, but unpredictable. This portion consist of the position of the robot's base and the position of the reaction wheel.
NASA Astrophysics Data System (ADS)
Amengonu, Yawo H.; Kakad, Yogendra P.
2014-07-01
Quasivelocity techniques were applied to derive the dynamics of a Differential Wheeled Mobile Robot (DWMR) in the companion paper. The present paper formulates a control system design for trajectory tracking of this class of robots. The method develops a feedback linearization technique for the nonlinear system using dynamic extension algorithm. The effectiveness of the nonlinear controller is illustrated with simulation example.
Chiang, Mao-Hsiung; Lin, Hao-Ting
2011-01-01
This study aimed to develop a novel 3D parallel mechanism robot driven by three vertical-axial pneumatic actuators with a stereo vision system for path tracking control. The mechanical system and the control system are the primary novel parts for developing a 3D parallel mechanism robot. In the mechanical system, a 3D parallel mechanism robot contains three serial chains, a fixed base, a movable platform and a pneumatic servo system. The parallel mechanism are designed and analyzed first for realizing a 3D motion in the X-Y-Z coordinate system of the robot’s end-effector. The inverse kinematics and the forward kinematics of the parallel mechanism robot are investigated by using the Denavit-Hartenberg notation (D-H notation) coordinate system. The pneumatic actuators in the three vertical motion axes are modeled. In the control system, the Fourier series-based adaptive sliding-mode controller with H∞ tracking performance is used to design the path tracking controllers of the three vertical servo pneumatic actuators for realizing 3D path tracking control of the end-effector. Three optical linear scales are used to measure the position of the three pneumatic actuators. The 3D position of the end-effector is then calculated from the measuring position of the three pneumatic actuators by means of the kinematics. However, the calculated 3D position of the end-effector cannot consider the manufacturing and assembly tolerance of the joints and the parallel mechanism so that errors between the actual position and the calculated 3D position of the end-effector exist. In order to improve this situation, sensor collaboration is developed in this paper. A stereo vision system is used to collaborate with the three position sensors of the pneumatic actuators. The stereo vision system combining two CCD serves to measure the actual 3D position of the end-effector and calibrate the error between the actual and the calculated 3D position of the end-effector. Furthermore, to verify the feasibility of the proposed parallel mechanism robot driven by three vertical pneumatic servo actuators, a full-scale test rig of the proposed parallel mechanism pneumatic robot is set up. Thus, simulations and experiments for different complex 3D motion profiles of the robot end-effector can be successfully achieved. The desired, the actual and the calculated 3D position of the end-effector can be compared in the complex 3D motion control. PMID:22247676
NASA Astrophysics Data System (ADS)
Li, Chunguang; Inoue, Yoshio; Liu, Tao; Shibata, Kyoko; Oka, Koichi
Master-slave control is becoming increasingly popular in the development of robotic systems which can provide rehabilitation training for hemiplegic patients with a unilaterally disabled limb. However, the system structures and control strategies of existent master-slave systems are always complex. An innovative master-slave system implementing force feedback and motion tracking for a rehabilitation robot is presented in this paper. The system consists of two identical motors with a wired connection, and the two motors are located at the master and slave manipulator sites respectively. The slave motor tracks the motion of the master motor directly driven by a patient. As well, the interaction force produced at the slave site is fed back to the patient. Therefore, the impaired limb driven by the slave motor can imitate the motion of the healthy limb controlling the master motor, and the patient can regulate the control force of the healthy limb properly according to the force sensation. The force sensing and motion tracking are achieved simultaneously with neither force sensors nor sophisticated control algorithms. The system is characterized by simple structure, bidirectional controllability, energy recycling, and force feedback without a force sensor. Test experiments on a prototype were conducted, and the results appraise the advantages of the system and demonstrate the feasibility of the proposed control scheme for a rehabilitation robot.
Integrated mobile robot control
NASA Technical Reports Server (NTRS)
Amidi, Omead; Thorpe, Charles
1991-01-01
This paper describes the structure, implementation, and operation of a real-time mobile robot controller which integrates capabilities such as: position estimation, path specification and tracking, human interfaces, fast communication, and multiple client support. The benefits of such high-level capabilities in a low-level controller was shown by its implementation for the Navlab autonomous vehicle. In addition, performance results from positioning and tracking systems are reported and analyzed.
Robust Neural Sliding Mode Control of Robot Manipulators
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nguyen Tran Hiep; Pham Thuong Cat
2009-03-05
This paper proposes a robust neural sliding mode control method for robot tracking problem to overcome the noises and large uncertainties in robot dynamics. The Lyapunov direct method has been used to prove the stability of the overall system. Simulation results are given to illustrate the applicability of the proposed method.
NASA Technical Reports Server (NTRS)
Sarkar, Nilanjan; Yun, Xiaoping; Kumar, Vijay
1994-01-01
There are many examples of mechanical systems that require rolling contacts between two or more rigid bodies. Rolling contacts engender nonholonomic constraints in an otherwise holonomic system. In this article, we develop a unified approach to the control of mechanical systems subject to both holonomic and nonholonomic constraints. We first present a state space realization of a constrained system. We then discuss the input-output linearization and zero dynamics of the system. This approach is applied to the dynamic control of mobile robots. Two types of control algorithms for mobile robots are investigated: trajectory tracking and path following. In each case, a smooth nonlinear feedback is obtained to achieve asymptotic input-output stability and Lagrange stability of the overall system. Simulation results are presented to demonstrate the effectiveness of the control algorithms and to compare the performane of trajectory-tracking and path-following algorithms.
Wu, Fang; Vibhute, Akash; Soh, Gim Song; Wood, Kristin L; Foong, Shaohui
2017-05-28
Due to their efficient locomotion and natural tolerance to hazardous environments, spherical robots have wide applications in security surveillance, exploration of unknown territory and emergency response. Numerous studies have been conducted on the driving mechanism, motion planning and trajectory tracking methods of spherical robots, yet very limited studies have been conducted regarding the obstacle avoidance capability of spherical robots. Most of the existing spherical robots rely on the "hit and run" technique, which has been argued to be a reasonable strategy because spherical robots have an inherent ability to recover from collisions. Without protruding components, they will not become stuck and can simply roll back after running into bstacles. However, for small scale spherical robots that contain sensitive surveillance sensors and cannot afford to utilize heavy protective shells, the absence of obstacle avoidance solutions would leave the robot at the mercy of potentially dangerous obstacles. In this paper, a compact magnetic field-based obstacle detection and avoidance system has been developed for miniature spherical robots. It utilizes a passive magnetic field so that the system is both compact and power efficient. The proposed system can detect not only the presence, but also the approaching direction of a ferromagnetic obstacle, therefore, an intelligent avoidance behavior can be generated by adapting the trajectory tracking method with the detection information. Design optimization is conducted to enhance the obstacle detection performance and detailed avoidance strategies are devised. Experimental results are also presented for validation purposes.
NASA Astrophysics Data System (ADS)
Kim, Jung Hoon; Hur, Sung-Moon; Oh, Yonghwan
2018-03-01
This paper is concerned with performance analysis of proportional-derivative/proportional-integral-derivative (PD/PID) controller for bounded persistent disturbances in a robotic manipulator. Even though the notion of input-to-state stability (ISS) has been widely used to deal with the effect of disturbances in control of a robotic manipulator, the corresponding studies cannot be directly applied to the treatment of persistent disturbances occurred in robotic manipulators. This is because the conventional studies relevant to ISS consider the H∞ performance for robotic systems, which is confined to the treatment of decaying disturbances, i.e. the disturbances those in the L2 space. To deal with the effect of persistent disturbances in robotic systems, we first provide a new treatment of ISS in the L∞ sense because bounded persistent disturbances should be intrinsically regarded as elements of the L∞ space. We next derive state-space representations of trajectory tracking control in the robotic systems which allow us to define the problem formulations more clearly. We then propose a novel control law that has a PD/PID control form, by which the trajectory tracking system satisfies the reformulated ISS. Furthermore, we can obtain a theoretical argument about the L∞ gain from the disturbance to the regulated output through the proposed control law. Finally, experimental studies for a typical 3-degrees of freedom robotic manipulator are given to demonstrate the effectiveness of the method introduced in this paper.
Robotic vehicle with multiple tracked mobility platforms
Salton, Jonathan R [Albuquerque, NM; Buttz, James H [Albuquerque, NM; Garretson, Justin [Albuquerque, NM; Hayward, David R [Wetmore, CO; Hobart, Clinton G [Albuquerque, NM; Deuel, Jr., Jamieson K.
2012-07-24
A robotic vehicle having two or more tracked mobility platforms that are mechanically linked together with a two-dimensional coupling, thereby forming a composite vehicle of increased mobility. The robotic vehicle is operative in hazardous environments and can be capable of semi-submersible operation. The robotic vehicle is capable of remote controlled operation via radio frequency and/or fiber optic communication link to a remote operator control unit. The tracks have a plurality of track-edge scallop cut-outs that allow the tracks to easily grab onto and roll across railroad tracks, especially when crossing the railroad tracks at an oblique angle.
NASA Technical Reports Server (NTRS)
Dywer, T. A. W., III; Lee, G. K. F.
1984-01-01
In connection with the current interest in agile spacecraft maneuvers, it has become necessary to consider the nonlinear coupling effects of multiaxial rotation in the treatment of command generation and tracking problems. Multiaxial maneuvers will be required in military missions involving a fast acquisition of moving targets in space. In addition, such maneuvers are also needed for the efficient operation of robot manipulators. Attention is given to details regarding the direct nonlinear command generation and tracking, an approach which has been successfully applied to the design of control systems for V/STOL aircraft, linearizing transformations for spacecraft controlled with external thrusters, the case of flexible spacecraft dynamics, examples from robot dynamics, and problems of implementation and testing.
NASA Astrophysics Data System (ADS)
Lu, Qun; Yu, Li; Zhang, Dan; Zhang, Xuebo
2018-01-01
This paper presentsa global adaptive controller that simultaneously solves tracking and regulation for wheeled mobile robots with unknown depth and uncalibrated camera-to-robot extrinsic parameters. The rotational angle and the scaled translation between the current camera frame and the reference camera frame, as well as the ones between the desired camera frame and the reference camera frame can be calculated in real time by using the pose estimation techniques. A transformed system is first obtained, for which an adaptive controller is then designed to accomplish both tracking and regulation tasks, and the controller synthesis is based on Lyapunov's direct method. Finally, the effectiveness of the proposed method is illustrated by a simulation study.
Robotic active positioning for magnetic resonance-guided high-intensity focused ultrasound
NASA Astrophysics Data System (ADS)
Xiao, Xu; Huang, Zhihong; Volovick, Alexander; Melzer, Andreas
2012-11-01
Magnetic resonance (MR) guided High-intensity focused ultrasound (HIFU) is a noninvasive method producing thermal necrosis and cavitation at the position of tumors with high accuracy. Because the typical size of the high-intensity focused ultrasound focus are much smaller than the targeted tumor or other tissues, multiple sonications and focus repositioning become necessary for HIFU treatment. In order to reach a much wider range, manual repositioning or using MR compatible mechanical actuators could be used. The repositioning technique is a time consuming procedure because it needs a series of MR imaging to detect the transducer and markers preplaced on the mechanical devices. We combined an active tracking technique into the MR guided HIFU system. In this work, the robotic system used is the MR-compatible robotics from InnoMotion{trade mark, serif} (IBSMM, Engineering spol. s r.o. / Ltd, Czech) which is originally designed for MR-guided needle biopsy. The precision and positioning speed of the combined robotic HIFU system are evaluated in this study. Compared to the existing MR guided HIFU systems, the combined robotic system with active tracking techniques provides a potential that allows the HIFU treatment to operate in a larger spatial range and with a faster speed.
Remote-controlled vision-guided mobile robot system
NASA Astrophysics Data System (ADS)
Ande, Raymond; Samu, Tayib; Hall, Ernest L.
1997-09-01
Automated guided vehicles (AGVs) have many potential applications in manufacturing, medicine, space and defense. The purpose of this paper is to describe exploratory research on the design of the remote controlled emergency stop and vision systems for an autonomous mobile robot. The remote control provides human supervision and emergency stop capabilities for the autonomous vehicle. The vision guidance provides automatic operation. A mobile robot test-bed has been constructed using a golf cart base. The mobile robot (Bearcat) was built for the Association for Unmanned Vehicle Systems (AUVS) 1997 competition. The mobile robot has full speed control with guidance provided by a vision system and an obstacle avoidance system using ultrasonic sensors systems. Vision guidance is accomplished using two CCD cameras with zoom lenses. The vision data is processed by a high speed tracking device, communicating with the computer the X, Y coordinates of blobs along the lane markers. The system also has three emergency stop switches and a remote controlled emergency stop switch that can disable the traction motor and set the brake. Testing of these systems has been done in the lab as well as on an outside test track with positive results that show that at five mph the vehicle can follow a line and at the same time avoid obstacles.
Adaptive control of space-based robot manipulators
NASA Technical Reports Server (NTRS)
Walker, Michael W.; Wee, Liang-Boon
1991-01-01
A control method is presented that achieves globally stable trajectory tracking in the presence of uncertainties in the inertial parameters of the system. The 15-DOF system dynamics are divided into two components: a 9-DOF invertible portion and 6-DOF noninvertible portion. A controller is then designed to achieve trajectory tracking of the invertible portion of the system, which consists of the manipulator-joint positions and the orientation of the base. The motion of the noninvertible portion is bounded but otherwise unspecified. This portion of the system consists of the position of the robot's base and the position of the reaction wheels. A simulation is presented to demonstrate the effectiveness of the controller. A quadratic polynomial is used to generate the desired trajectory to illustrate the trajectory-tracking capability of the controller.
NASA Astrophysics Data System (ADS)
Chen, ChuXin; Trivedi, Mohan M.
1992-03-01
This research is focused on enhancing the overall productivity of an integrated human-robot system. A simulation, animation, visualization, and interactive control (SAVIC) environment has been developed for the design and operation of an integrated robotic manipulator system. This unique system possesses the abilities for multisensor simulation, kinematics and locomotion animation, dynamic motion and manipulation animation, transformation between real and virtual modes within the same graphics system, ease in exchanging software modules and hardware devices between real and virtual world operations, and interfacing with a real robotic system. This paper describes a working system and illustrates the concepts by presenting the simulation, animation, and control methodologies for a unique mobile robot with articulated tracks, a manipulator, and sensory modules.
Model-independent position domain sliding mode control for contour tracking of robotic manipulator
NASA Astrophysics Data System (ADS)
Yue, W. H.; Pano, V.; Ouyang, P. R.; Hu, Y. Q.
2017-01-01
In this paper, a new position domain feedback type sliding mode control (PDC-SMC) law is proposed for contour tracking control of multi-DOF (degree of freedom) nonlinear robotic manipulators focusing on the improvement of contour tracking performances. One feature of the proposed control law is its model-independent control scheme that can avoid calculation of the feedforward part in a standard SMC. The new control law takes the advantages of the high contour tracking performance of PD type feedback position domain control (PDC) and the robustness of SMC. Stability analysis is performed using the Lyapunov stability theory, and simulation studies are conducted to verify the effectiveness of the developed PDC-SMC control system. In addition, the effects of control parameters of the SMC on system performances are studied.
Yoo, Sung Jin; Park, Jin Bae; Choi, Yoon Ho
2008-10-01
In this paper, we propose a new robust output feedback control approach for flexible-joint electrically driven (FJED) robots via the observer dynamic surface design technique. The proposed method only requires position measurements of the FJED robots. To estimate the link and actuator velocity information of the FJED robots with model uncertainties, we develop an adaptive observer using self-recurrent wavelet neural networks (SRWNNs). The SRWNNs are used to approximate model uncertainties in both robot (link) dynamics and actuator dynamics, and all their weights are trained online. Based on the designed observer, the link position tracking controller using the estimated states is induced from the dynamic surface design procedure. Therefore, the proposed controller can be designed more simply than the observer backstepping controller. From the Lyapunov stability analysis, it is shown that all signals in a closed-loop adaptive system are uniformly ultimately bounded. Finally, the simulation results on a three-link FJED robot are presented to validate the good position tracking performance and robustness of the proposed control system against payload uncertainties and external disturbances.
EVA Robotic Assistant Project: Platform Attitude Prediction
NASA Technical Reports Server (NTRS)
Nickels, Kevin M.
2003-01-01
The Robotic Systems Technology Branch is currently working on the development of an EVA Robotic Assistant under the sponsorship of the Surface Systems Thrust of the NASA Cross Enterprise Technology Development Program (CETDP). This will be a mobile robot that can follow a field geologist during planetary surface exploration, carry his tools and the samples that he collects, and provide video coverage of his activity. Prior experiments have shown that for such a robot to be useful it must be able to follow the geologist at walking speed over any terrain of interest. Geologically interesting terrain tends to be rough rather than smooth. The commercial mobile robot that was recently purchased as an initial testbed for the EVA Robotic Assistant Project, an ATRV Jr., is capable of faster than walking speed outside but it has no suspension. Its wheels with inflated rubber tires are attached to axles that are connected directly to the robot body. Any angular motion of the robot produced by driving over rough terrain will directly affect the pointing of the on-board stereo cameras. The resulting image motion is expected to make tracking of the geologist more difficult. This will either require the tracker to search a larger part of the image to find the target from frame to frame or to search mechanically in pan and tilt whenever the image motion is large enough to put the target outside the image in the next frame. This project consists of the design and implementation of a Kalman filter that combines the output of the angular rate sensors and linear accelerometers on the robot to estimate the motion of the robot base. The motion of the stereo camera pair mounted on the robot that results from this motion as the robot drives over rough terrain is then straightforward to compute. The estimates may then be used, for example, to command the robot s on-board pan-tilt unit to compensate for the camera motion induced by the base movement. This has been accomplished in two ways: first, a standalone head stabilizer has been implemented and second, the estimates have been used to influence the search algorithm of the stereo tracking algorithm. Studies of the image motion of a tracked object indicate that the image motion of objects is suppressed while the robot crossing rough terrain. This work expands the range of speed and surface roughness over which the robot should be able to track and follow a field geologist and accept arm gesture commands from the geologist.
Learning-based position control of a closed-kinematic chain robot end-effector
NASA Technical Reports Server (NTRS)
Nguyen, Charles C.; Zhou, Zhen-Lei
1990-01-01
A trajectory control scheme whose design is based on learning theory, for a six-degree-of-freedom (DOF) robot end-effector built to study robotic assembly of NASA hardwares in space is presented. The control scheme consists of two control systems: the feedback control system and the learning control system. The feedback control system is designed using the concept of linearization about a selected operating point, and the method of pole placement so that the closed-loop linearized system is stabilized. The learning control scheme consisting of PD-type learning controllers, provides additional inputs to improve the end-effector performance after each trial. Experimental studies performed on a 2 DOF end-effector built at CUA, for three tracking cases show that actual trajectories approach desired trajectories as the number of trials increases. The tracking errors are substantially reduced after only five trials.
Rong, Wei; Tong, Kai Yu; Hu, Xiao Ling; Ho, Sze Kit
2015-03-01
An electromyography-driven robot system integrated with neuromuscular electrical stimulation (NMES) was developed to investigate its effectiveness on post-stroke rehabilitation. The performance of this system in assisting finger flexion/extension with different assistance combinations was evaluated in five stroke subjects. Then, a pilot study with 20-sessions training was conducted to evaluate the training's effectiveness. The results showed that combined assistance from the NMES-robot could improve finger movement accuracy, encourage muscle activation of the finger muscles and suppress excessive muscular activities in the elbow joint. When assistances from both NMES and the robot were 50% of their maximum assistances, finger-tracking performance had the best results, with the lowest root mean square error, greater range of motion, higher voluntary muscle activations of the finger joints and lower muscle co-contraction in the finger and elbow joints. Upper limb function improved after the 20-session training, indicated by the increased clinical scores of Fugl-Meyer Assessment, Action Research Arm Test and Wolf Motor Function Test. Muscle co-contraction was reduced in the finger and elbow joints reflected by the Modified Ashworth Scale. The findings demonstrated that an electromyography-driven NMES-robot used for chronic stroke improved hand function and tracking performance. Further research is warranted to validate the method on a larger scale. Implications for Rehabilitation The hand robotics and neuromuscular electrical stimulation (NMES) techniques are still separate systems in current post-stroke hand rehabilitation. This is the first study to investigate the combined effects of the NMES and robot on hand rehabilitation. The finger tracking performance was improved with the combined assistance from the EMG-driven NMES-robot hand system. The assistance from the robot could improve the finger movement accuracy and the assistance from the NMES could reduce the muscle co-contraction on finger and elbow joints. The upper limb functions were improved on chronic stroke patients after the pilot study of 20-session hand training with the combined assistance from the EMG-driven NMES-robot. The muscle spasticity on finger and elbow joints was reduced after the training.
Combined Feature Based and Shape Based Visual Tracker for Robot Navigation
NASA Technical Reports Server (NTRS)
Deans, J.; Kunz, C.; Sargent, R.; Park, E.; Pedersen, L.
2005-01-01
We have developed a combined feature based and shape based visual tracking system designed to enable a planetary rover to visually track and servo to specific points chosen by a user with centimeter precision. The feature based tracker uses invariant feature detection and matching across a stereo pair, as well as matching pairs before and after robot movement in order to compute an incremental 6-DOF motion at each tracker update. This tracking method is subject to drift over time, which can be compensated by the shape based method. The shape based tracking method consists of 3D model registration, which recovers 6-DOF motion given sufficient shape and proper initialization. By integrating complementary algorithms, the combined tracker leverages the efficiency and robustness of feature based methods with the precision and accuracy of model registration. In this paper, we present the algorithms and their integration into a combined visual tracking system.
Patient motion tracking in the presence of measurement errors.
Haidegger, Tamás; Benyó, Zoltán; Kazanzides, Peter
2009-01-01
The primary aim of computer-integrated surgical systems is to provide physicians with superior surgical tools for better patient outcome. Robotic technology is capable of both minimally invasive surgery and microsurgery, offering remarkable advantages for the surgeon and the patient. Current systems allow for sub-millimeter intraoperative spatial positioning, however certain limitations still remain. Measurement noise and unintended changes in the operating room environment can result in major errors. Positioning errors are a significant danger to patients in procedures involving robots and other automated devices. We have developed a new robotic system at the Johns Hopkins University to support cranial drilling in neurosurgery procedures. The robot provides advanced visualization and safety features. The generic algorithm described in this paper allows for automated compensation of patient motion through optical tracking and Kalman filtering. When applied to the neurosurgery setup, preliminary results show that it is possible to identify patient motion within 700 ms, and apply the appropriate compensation with an average of 1.24 mm positioning error after 2 s of setup time.
Optimal Control Method of Robot End Position and Orientation Based on Dynamic Tracking Measurement
NASA Astrophysics Data System (ADS)
Liu, Dalong; Xu, Lijuan
2018-01-01
In order to improve the accuracy of robot pose positioning and control, this paper proposed a dynamic tracking measurement robot pose optimization control method based on the actual measurement of D-H parameters of the robot, the parameters is taken with feedback compensation of the robot, according to the geometrical parameters obtained by robot pose tracking measurement, improved multi sensor information fusion the extended Kalan filter method, with continuous self-optimal regression, using the geometric relationship between joint axes for kinematic parameters in the model, link model parameters obtained can timely feedback to the robot, the implementation of parameter correction and compensation, finally we can get the optimal attitude angle, realize the robot pose optimization control experiments were performed. 6R dynamic tracking control of robot joint robot with independent research and development is taken as experimental subject, the simulation results show that the control method improves robot positioning accuracy, and it has the advantages of versatility, simplicity, ease of operation and so on.
Stinton, S K; Siebold, R; Freedberg, H; Jacobs, C; Branch, T P
2016-03-01
The purpose of this study was to: (1) determine whether a robotic tibial rotation device and an electromagnetic tracking system could accurately reproduce the clinical dial test at 30° of knee flexion; (2) compare rotation data captured at the footplates of the robotic device to tibial rotation data measured using an electromagnetic sensor on the proximal tibia. Thirty-two unilateral ACL-reconstructed patients were examined using a robotic tibial rotation device that mimicked the dial test. The data reported in this study is only from the healthy legs of these patients. Torque was applied through footplates and was measured using servomotors. Lower leg motion was measured at the foot using the motors. Tibial motion was also measured through an electromagnetic tracking system and a sensor on the proximal tibia. Load-deformation curves representing rotational motion of the foot and tibia were compared using Pearson's correlation coefficients. Off-axis motions including medial-lateral translation and anterior-posterior translation were also measured using the electromagnetic system. The robotic device and electromagnetic system were able to provide axial rotation data and translational data for the tibia during the dial test. Motion measured at the foot was not correlated to motion of the tibial tubercle in internal rotation or in external rotation. The position of the tibial tubercle was 26.9° ± 11.6° more internally rotated than the foot at torque 0 Nm. Medial-lateral translation and anterior-posterior translation were combined to show the path of the tubercle in the coronal plane during tibial rotation. The information captured during a manual dial test includes both rotation of the tibia and proximal tibia translation. All of this information can be captured using a robotic tibial axial rotation device with an electromagnetic tracking system. The pathway of the tibial tubercle during tibial axial rotation can provide additional information about knee instability without relying on side-to-side comparison between knees. The translation of the proximal tibia is important information that must be considered in addition to axial rotation of the tibia when performing a dial test whether done manually or with a robotic device. Instrumented foot position cannot provide the same information. IV.
Robotic Attention Processing And Its Application To Visual Guidance
NASA Astrophysics Data System (ADS)
Barth, Matthew; Inoue, Hirochika
1988-03-01
This paper describes a method of real-time visual attention processing for robots performing visual guidance. This robot attention processing is based on a novel vision processor, the multi-window vision system that was developed at the University of Tokyo. The multi-window vision system is unique in that it only processes visual information inside local area windows. These local area windows are quite flexible in their ability to move anywhere on the visual screen, change their size and shape, and alter their pixel sampling rate. By using these windows for specific attention tasks, it is possible to perform high speed attention processing. The primary attention skills of detecting motion, tracking an object, and interpreting an image are all performed at high speed on the multi-window vision system. A basic robotic attention scheme using the attention skills was developed. The attention skills involved detection and tracking of salient visual features. The tracking and motion information thus obtained was utilized in producing the response to the visual stimulus. The response of the attention scheme was quick enough to be applicable to the real-time vision processing tasks of playing a video 'pong' game, and later using an automobile driving simulator. By detecting the motion of a 'ball' on a video screen and then tracking the movement, the attention scheme was able to control a 'paddle' in order to keep the ball in play. The response was faster than that of a human's, allowing the attention scheme to play the video game at higher speeds. Further, in the application to the driving simulator, the attention scheme was able to control both direction and velocity of a simulated vehicle following a lead car. These two applications show the potential of local visual processing in its use for robotic attention processing.
Mobile robots IV; Proceedings of the Meeting, Philadelphia, PA, Nov. 6, 7, 1989
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wolfe, W.J.; Chun, W.H.
1990-01-01
The present conference on mobile robot systems discusses high-speed machine perception based on passive sensing, wide-angle optical ranging, three-dimensional path planning for flying/crawling robots, navigation of autonomous mobile intelligence in an unstructured natural environment, mechanical models for the locomotion of a four-articulated-track robot, a rule-based command language for a semiautonomous Mars rover, and a computer model of the structured light vision system for a Mars rover. Also discussed are optical flow and three-dimensional information for navigation, feature-based reasoning trail detection, a symbolic neural-net production system for obstacle avoidance and navigation, intelligent path planning for robot navigation in an unknown environment,more » behaviors from a hierarchical control system, stereoscopic TV systems, the REACT language for autonomous robots, and a man-amplifying exoskeleton.« less
Near-Optimal Tracking Control of Mobile Robots Via Receding-Horizon Dual Heuristic Programming.
Lian, Chuanqiang; Xu, Xin; Chen, Hong; He, Haibo
2016-11-01
Trajectory tracking control of wheeled mobile robots (WMRs) has been an important research topic in control theory and robotics. Although various tracking control methods with stability have been developed for WMRs, it is still difficult to design optimal or near-optimal tracking controller under uncertainties and disturbances. In this paper, a near-optimal tracking control method is presented for WMRs based on receding-horizon dual heuristic programming (RHDHP). In the proposed method, a backstepping kinematic controller is designed to generate desired velocity profiles and the receding horizon strategy is used to decompose the infinite-horizon optimal control problem into a series of finite-horizon optimal control problems. In each horizon, a closed-loop tracking control policy is successively updated using a class of approximate dynamic programming algorithms called finite-horizon dual heuristic programming (DHP). The convergence property of the proposed method is analyzed and it is shown that the tracking control system based on RHDHP is asymptotically stable by using the Lyapunov approach. Simulation results on three tracking control problems demonstrate that the proposed method has improved control performance when compared with conventional model predictive control (MPC) and DHP. It is also illustrated that the proposed method has lower computational burden than conventional MPC, which is very beneficial for real-time tracking control.
NASA Astrophysics Data System (ADS)
Zhang, Haichong K.; Aalamifar, Fereshteh; Boctor, Emad M.
2016-04-01
Synthetic aperture for ultrasound is a technique utilizing a wide aperture in both transmit and receive to enhance the ultrasound image quality. The limitation of synthetic aperture is the maximum available aperture size limit determined by the physical size of ultrasound probe. We propose Synthetic-Tracked Aperture Ultrasound (STRATUS) imaging system to overcome the limitation by extending the beamforming aperture size through ultrasound probe tracking. With a setup involving a robotic arm, the ultrasound probe is moved using the robotic arm, while the positions on a scanning trajectory are tracked in real-time. Data from each pose are synthesized to construct a high resolution image. In previous studies, we have demonstrated the feasibility through phantom experiments. However, various additional factors such as real-time data collection or motion artifacts should be taken into account when the in vivo target becomes the subject. In this work, we build a robot-based STRATUS imaging system with continuous data collection capability considering the practical implementation. A curvilinear array is used instead of a linear array to benefit from its wider capture angle. We scanned human forearms under two scenarios: one submerged the arm in the water tank under 10 cm depth, and the other directly scanned the arm from the surface. The image contrast improved 5.51 dB, and 9.96 dB for the underwater scan and the direct scan, respectively. The result indicates the practical feasibility of STRATUS imaging system, and the technique can be potentially applied to the wide range of human body.
Novel graphical environment for virtual and real-world operations of tracked mobile manipulators
NASA Astrophysics Data System (ADS)
Chen, ChuXin; Trivedi, Mohan M.; Azam, Mir; Lassiter, Nils T.
1993-08-01
A simulation, animation, visualization and interactive control (SAVIC) environment has been developed for the design and operation of an integrated mobile manipulator system. This unique system possesses the abilities for (1) multi-sensor simulation, (2) kinematics and locomotion animation, (3) dynamic motion and manipulation animation, (4) transformation between real and virtual modes within the same graphics system, (5) ease in exchanging software modules and hardware devices between real and virtual world operations, and (6) interfacing with a real robotic system. This paper describes a working system and illustrates the concepts by presenting the simulation, animation and control methodologies for a unique mobile robot with articulated tracks, a manipulator, and sensory modules.
Software development to support sensor control of robot arc welding
NASA Technical Reports Server (NTRS)
Silas, F. R., Jr.
1986-01-01
The development of software for a Digital Equipment Corporation MINC-23 Laboratory Computer to provide functions of a workcell host computer for Space Shuttle Main Engine (SSME) robotic welding is documented. Routines were written to transfer robot programs between the MINC and an Advanced Robotic Cyro 750 welding robot. Other routines provide advanced program editing features while additional software allows communicatin with a remote computer aided design system. Access to special robot functions were provided to allow advanced control of weld seam tracking and process control for future development programs.
NASA Astrophysics Data System (ADS)
Yang, Xinxin; Ge, Shuzhi Sam; He, Wei
2018-04-01
In this paper, both the closed-form dynamics and adaptive robust tracking control of a space robot with two-link flexible manipulators under unknown disturbances are developed. The dynamic model of the system is described with assumed modes approach and Lagrangian method. The flexible manipulators are represented as Euler-Bernoulli beams. Based on singular perturbation technique, the displacements/joint angles and flexible modes are modelled as slow and fast variables, respectively. A sliding mode control is designed for trajectories tracking of the slow subsystem under unknown but bounded disturbances, and an adaptive sliding mode control is derived for slow subsystem under unknown slowly time-varying disturbances. An optimal linear quadratic regulator method is proposed for the fast subsystem to damp out the vibrations of the flexible manipulators. Theoretical analysis validates the stability of the proposed composite controller. Numerical simulation results demonstrate the performance of the closed-loop flexible space robot system.
A Computational Model of Spatial Development
NASA Astrophysics Data System (ADS)
Hiraki, Kazuo; Sashima, Akio; Phillips, Steven
Psychological experiments on children's development of spatial knowledge suggest experience at self-locomotion with visual tracking as important factors. Yet, the mechanism underlying development is unknown. We propose a robot that learns to mentally track a target object (i.e., maintaining a representation of an object's position when outside the field-of-view) as a model for spatial development. Mental tracking is considered as prediction of an object's position given the previous environmental state and motor commands, and the current environment state resulting from movement. Following Jordan & Rumelhart's (1992) forward modeling architecture the system consists of two components: an inverse model of sensory input to desired motor commands; and a forward model of motor commands to desired sensory input (goals). The robot was tested on the `three cups' paradigm (where children are required to select the cup containing the hidden object under various movement conditions). Consistent with child development, without the capacity for self-locomotion the robot's errors are self-center based. When given the ability of self-locomotion the robot responds allocentrically.
Unicompartmental knee arthroplasty: is robotic technology more accurate than conventional technique?
Citak, Mustafa; Suero, Eduardo M; Citak, Musa; Dunbar, Nicholas J; Branch, Sharon H; Conditt, Michael A; Banks, Scott A; Pearle, Andrew D
2013-08-01
Robotic-assisted unicompartmental knee arthroplasty (UKA) with rigid bone fixation "can significantly improve implant placement and leg alignment. The aim of this cadaveric study was to determine whether the use of robotic systems with dynamic bone tracking would provide more accurate UKA implant positioning compared to the conventional manual technique. Three-dimensional CT-based preoperative plans were created to determine the desired position and orientation for the tibial and femoral components. For each pair of cadaver knees, UKA was performed using traditional instrumentation on the left side and using a haptic robotic system on the right side. Postoperative CT scans were obtained and 3D-to-3D iterative closest point registration was performed. Implant position and orientation were compared to the preoperative plan. Surgical RMS errors for femoral component placement were within 1.9 mm and 3.7° in all directions of the planned implant position for the robotic group, while RMS errors for the manual group were within 5.4mm and 10.2°. Average RMS errors for tibial component placement were within 1.4mm and 5.0° in all directions for the robotic group; while, for the manual group, RMS errors were within 5.7 mm and 19.2°. UKA was more precise using a semiactive robotic system with dynamic bone tracking technology compared to the manual technique. Copyright © 2012 Elsevier B.V. All rights reserved.
Song, Shuang; Zhang, Changchun; Liu, Li; Meng, Max Q-H
2018-02-01
Flexible surgical robot can work in confined and complex environments, which makes it a good option for minimally invasive surgery. In order to utilize flexible manipulators in complicated and constrained surgical environments, it is of great significance to monitor the position and shape of the curvilinear manipulator in real time during the procedures. In this paper, we propose a magnetic tracking-based planar shape sensing and navigation system for flexible surgical robots in the transoral surgery. The system can provide the real-time tip position and shape information of the robot during the operation. We use wire-driven flexible robot to serve as the manipulator. It has three degrees of freedom. A permanent magnet is mounted at the distal end of the robot. Its magnetic field can be sensed with a magnetic sensor array. Therefore, position and orientation of the tip can be estimated utilizing a tracking method. A shape sensing algorithm is then carried out to estimate the real-time shape based on the tip pose. With the tip pose and shape display in the 3D reconstructed CT model, navigation can be achieved. Using the proposed system, we carried out planar navigation experiments on a skull phantom to touch three different target positions under the navigation of the skull display interface. During the experiments, the real-time shape has been well monitored and distance errors between the robot tip and the targets in the skull have been recorded. The mean navigation error is [Formula: see text] mm, while the maximum error is 3.2 mm. The proposed method provides the advantages that no sensors are needed to mount on the robot and no line-of-sight problem. Experimental results verified the feasibility of the proposed method.
Dynamic inverse models in human-cyber-physical systems
NASA Astrophysics Data System (ADS)
Robinson, Ryan M.; Scobee, Dexter R. R.; Burden, Samuel A.; Sastry, S. Shankar
2016-05-01
Human interaction with the physical world is increasingly mediated by automation. This interaction is characterized by dynamic coupling between robotic (i.e. cyber) and neuromechanical (i.e. human) decision-making agents. Guaranteeing performance of such human-cyber-physical systems will require predictive mathematical models of this dynamic coupling. Toward this end, we propose a rapprochement between robotics and neuromechanics premised on the existence of internal forward and inverse models in the human agent. We hypothesize that, in tele-robotic applications of interest, a human operator learns to invert automation dynamics, directly translating from desired task to required control input. By formulating the model inversion problem in the context of a tracking task for a nonlinear control system in control-a_ne form, we derive criteria for exponential tracking and show that the resulting dynamic inverse model generally renders a portion of the physical system state (i.e., the internal dynamics) unobservable from the human operator's perspective. Under stability conditions, we show that the human can achieve exponential tracking without formulating an estimate of the system's state so long as they possess an accurate model of the system's dynamics. These theoretical results are illustrated using a planar quadrotor example. We then demonstrate that the automation can intervene to improve performance of the tracking task by solving an optimal control problem. Performance is guaranteed to improve under the assumption that the human learns and inverts the dynamic model of the altered system. We conclude with a discussion of practical limitations that may hinder exact dynamic model inversion.
Object tracking with adaptive HOG detector and adaptive Rao-Blackwellised particle filter
NASA Astrophysics Data System (ADS)
Rosa, Stefano; Paleari, Marco; Ariano, Paolo; Bona, Basilio
2012-01-01
Scenarios for a manned mission to the Moon or Mars call for astronaut teams to be accompanied by semiautonomous robots. A prerequisite for human-robot interaction is the capability of successfully tracking humans and objects in the environment. In this paper we present a system for real-time visual object tracking in 2D images for mobile robotic systems. The proposed algorithm is able to specialize to individual objects and to adapt to substantial changes in illumination and object appearance during tracking. The algorithm is composed by two main blocks: a detector based on Histogram of Oriented Gradient (HOG) descriptors and linear Support Vector Machines (SVM), and a tracker which is implemented by an adaptive Rao-Blackwellised particle filter (RBPF). The SVM is re-trained online on new samples taken from previous predicted positions. We use the effective sample size to decide when the classifier needs to be re-trained. Position hypotheses for the tracked object are the result of a clustering procedure applied on the set of particles. The algorithm has been tested on challenging video sequences presenting strong changes in object appearance, illumination, and occlusion. Experimental tests show that the presented method is able to achieve near real-time performances with a precision of about 7 pixels on standard video sequences of dimensions 320 × 240.
Lower Limb Rehabilitation Using Patient Data
Saadat, Mozafar
2016-01-01
The aim of this study is to investigate the performance of a 6-DoF parallel robot in tracking the movement of the foot trajectory of a paretic leg during a single stride. The foot trajectories of nine patients with a paretic leg including both males and females have been measured and analysed by a Vicon system in a gait laboratory. Based on kinematic and dynamic analysis of a 6-DoF UPS parallel robot, an algorithm was developed in MATLAB to calculate the length of the actuators and their required forces during all trajectories. The workspace and singularity points of the robot were then investigated in nine different cases. A 6-DoF UPS parallel robot prototype with high repeatability was designed and built in order to simulate a single stride. Results showed that the robot was capable of tracking all of the trajectories with the maximum position error of 1.2 mm. PMID:27721648
Mobility Systems For Robotic Vehicles
NASA Astrophysics Data System (ADS)
Chun, Wendell
1987-02-01
The majority of existing robotic systems can be decomposed into five distinct subsystems: locomotion, control/man-machine interface (MMI), sensors, power source, and manipulator. When designing robotic vehicles, there are two main requirements: first, to design for the environment and second, for the task. The environment can be correlated with known missions. This can be seen by analyzing existing mobile robots. Ground mobile systems are generally wheeled, tracked, or legged. More recently, underwater vehicles have gained greater attention. For example, Jason Jr. made history by surveying the sunken luxury liner, the Titanic. The next big surge of robotic vehicles will be in space. This will evolve as a result of NASA's commitment to the Space Station. The foreseeable robots will interface with current systems as well as standalone, free-flying systems. A space robotic vehicle is similar to its underwater counterpart with very few differences. Their commonality includes missions and degrees-of-freedom. The issues of stability and communication are inherent in both systems and environment.
NASA Astrophysics Data System (ADS)
Heath Pastore, Tracy; Barnes, Mitchell; Hallman, Rory
2005-05-01
Robot technology is developing at a rapid rate for both commercial and Department of Defense (DOD) applications. As a result, the task of managing both technology and experience information is growing. In the not-to-distant past, tracking development efforts of robot platforms, subsystems and components was not too difficult, expensive, or time consuming. To do the same today is a significant undertaking. The Mobile Robot Knowledge Base (MRKB) provides the robotics community with a web-accessible, centralized resource for sharing information, experience, and technology to more efficiently and effectively meet the needs of the robot system user. The resource includes searchable information on robot components, subsystems, mission payloads, platforms, and DOD robotics programs. In addition, the MRKB website provides a forum for technology and information transfer within the DOD robotics community and an interface for the Robotic Systems Pool (RSP). The RSP manages a collection of small teleoperated and semi-autonomous robotic platforms, available for loan to DOD and other qualified entities. The objective is to put robots in the hands of users and use the test data and fielding experience to improve robot systems.
Azizian, Mahdi; Khoshnam, Mahta; Najmaei, Nima; Patel, Rajni V
2014-09-01
Intra-operative imaging is widely used to provide visual feedback to a clinician when he/she performs a procedure. In visual servoing, surgical instruments and parts of tissue/body are tracked by processing the acquired images. This information is then used within a control loop to manoeuvre a robotic manipulator during a procedure. A comprehensive search of electronic databases was completed for the period 2000-2013 to provide a survey of the visual servoing applications in medical robotics. The focus is on medical applications where image-based tracking is used for closed-loop control of a robotic system. Detailed classification and comparative study of various contributions in visual servoing using endoscopic or direct visual images are presented and summarized in tables and diagrams. The main challenges in using visual servoing for medical robotic applications are identified and potential future directions are suggested. 'Supervised automation of medical robotics' is found to be a major trend in this field. Copyright © 2013 John Wiley & Sons, Ltd.
Development of haptic system for surgical robot
NASA Astrophysics Data System (ADS)
Gang, Han Gyeol; Park, Jiong Min; Choi, Seung-Bok; Sohn, Jung Woo
2017-04-01
In this paper, a new type of haptic system for surgical robot application is proposed and its performances are evaluated experimentally. The proposed haptic system consists of an effective master device and a precision slave robot. The master device has 3-DOF rotational motion as same as human wrist motion. It has lightweight structure with a gyro sensor and three small-sized MR brakes for position measurement and repulsive torque generation, respectively. The slave robot has 3-DOF rotational motion using servomotors, five bar linkage and a torque sensor is used to measure resistive torque. It has been experimentally demonstrated that the proposed haptic system has good performances on tracking control of desired position and repulsive torque. It can be concluded that the proposed haptic system can be effectively applied to the surgical robot system in real field.
Trifocal Tensor-Based Adaptive Visual Trajectory Tracking Control of Mobile Robots.
Chen, Jian; Jia, Bingxi; Zhang, Kaixiang
2017-11-01
In this paper, a trifocal tensor-based approach is proposed for the visual trajectory tracking task of a nonholonomic mobile robot equipped with a roughly installed monocular camera. The desired trajectory is expressed by a set of prerecorded images, and the robot is regulated to track the desired trajectory using visual feedback. Trifocal tensor is exploited to obtain the orientation and scaled position information used in the control system, and it works for general scenes owing to the generality of trifocal tensor. In the previous works, the start, current, and final images are required to share enough visual information to estimate the trifocal tensor. However, this requirement can be easily violated for perspective cameras with limited field of view. In this paper, key frame strategy is proposed to loosen this requirement, extending the workspace of the visual servo system. Considering the unknown depth and extrinsic parameters (installing position of the camera), an adaptive controller is developed based on Lyapunov methods. The proposed control strategy works for almost all practical circumstances, including both trajectory tracking and pose regulation tasks. Simulations are made based on the virtual experimentation platform (V-REP) to evaluate the effectiveness of the proposed approach.
Design and Performance Evaluation of a UWB Communication and Tracking System for Mini-AERCam
NASA Technical Reports Server (NTRS)
Barton, Richard J.
2005-01-01
NASA Johnson Space Center (JSC) is developing a low-volume, low-mass, robotic free-flying camera known as Mini-AERCam (Autonomous Extra-vehicular Robotic Camera) to assist the International Space Station (ISS) operations. Mini-AERCam is designed to provide astronauts and ground control real-time video for camera views of ISS. The system will assist ISS crewmembers and ground personnel to monitor ongoing operations and perform visual inspections of exterior ISS components without requiring extravehicular activity (EAV). Mini-AERCam consists of a great number of subsystems. Many institutions and companies have been involved in the R&D for this project. A Mini-AERCam ground control system has been studied at Texas A&M University [3]. The path planning and control algorithms that direct the motions of Mini-AERCam have been developed through the joint effort of Carnegie Mellon University and the Texas Robotics and Automation Center [5]. NASA JSC has designed a layered control architecture that integrates all functions of Mini-AERCam [8]. The research described in this report is part of a larger effort focused on the communication and tracking subsystem that is designed to perform three major tasks: 1. To transmit commands from ISS to Mini-AERCam for control of robotic camera motions (downlink); 2. To transmit real-time video from Mini-AERCam to ISS for inspections (uplink); 3. To track the position of Mini-AERCam for precise motion control. The ISS propagation environment is unique due to the nature of the ISS structure and multiple RF interference sources [9]. The ISS is composed of various truss segments, solar panels, thermal radiator panels, and modules for laboratories and crew accommodations. A tracking system supplemental to GPS is desirable both to improve accuracy and to eliminate the structural blockage due to the close proximity of the ISS which could at times limit the number of GPS satellites accessible to the Mini-AERCam. Ideally, the tracking system will be a passive component of the communication system which will need to operate in a time-varying multipath environment created as the robot camera moves over the ISS structure. In addition, due to many interference sources located on the ISS, SSO, LEO satellites and ground-based transmitters, selecting a frequency for the ISS and Mini-AERCam link which will coexist with all interferers poses a major design challenge. To meet all of these challenges, ultrawideband (UWB) radio technology is being studied for use in the Mini-AERCam communication and tracking subsystem. The research described in this report is focused on design and evaluation of passive tracking system algorithms based on UWB radio transmissions from mini-AERCam.
Person detection, tracking and following using stereo camera
NASA Astrophysics Data System (ADS)
Wang, Xiaofeng; Zhang, Lilian; Wang, Duo; Hu, Xiaoping
2018-04-01
Person detection, tracking and following is a key enabling technology for mobile robots in many human-robot interaction applications. In this article, we present a system which is composed of visual human detection, video tracking and following. The detection is based on YOLO(You only look once), which applies a single convolution neural network(CNN) to the full image, thus can predict bounding boxes and class probabilities directly in one evaluation. Then the bounding box provides initial person position in image to initialize and train the KCF(Kernelized Correlation Filter), which is a video tracker based on discriminative classifier. At last, by using a stereo 3D sparse reconstruction algorithm, not only the position of the person in the scene is determined, but also it can elegantly solve the problem of scale ambiguity in the video tracker. Extensive experiments are conducted to demonstrate the effectiveness and robustness of our human detection and tracking system.
Multi-Target Tracking for Swarm vs. Swarm UAV Systems
2012-09-01
Uhlmann, “Using covariance intersection for SLAM,” Robotics and Autonomous Systems, vol. 55, pp. 3–20, Jan. 2007. [10] R. B. G. Wolfgang Niehsen... Krause , J. Leskovec, and C. Guestrin, “Data association for topic intensity track- ing,” Proceedings of the 23rd international conference on Machine
UWB Tracking System Design for Free-Flyers
NASA Technical Reports Server (NTRS)
Ni, Jianjun; Arndt, Dickey; Phan, Chan; Ngo, Phong; Gross, Julia; Dusl, John
2004-01-01
This paper discusses an ultra-wideband (UWB) tracking system design effort for Mini-AERCam (Autonomous Extra-vehicular Robotic Camera), a free-flying video camera system under development at NASA Johnson Space Center for aid in surveillance around the International Space Station (ISS). UWB technology is exploited to implement the tracking system due to its properties, such as high data rate, fine time resolution, and low power spectral density. A system design using commercially available UWB products is proposed. A tracking algorithm TDOA (Time Difference of Arrival) that operates cooperatively with the UWB system is developed in this research effort. Matlab simulations show that the tracking algorithm can achieve fine tracking resolution with low noise TDOA data. Lab experiments demonstrate the UWB tracking capability with fine resolution.
Stereo vision tracking of multiple objects in complex indoor environments.
Marrón-Romera, Marta; García, Juan C; Sotelo, Miguel A; Pizarro, Daniel; Mazo, Manuel; Cañas, José M; Losada, Cristina; Marcos, Alvaro
2010-01-01
This paper presents a novel system capable of solving the problem of tracking multiple targets in a crowded, complex and dynamic indoor environment, like those typical of mobile robot applications. The proposed solution is based on a stereo vision set in the acquisition step and a probabilistic algorithm in the obstacles position estimation process. The system obtains 3D position and speed information related to each object in the robot's environment; then it achieves a classification between building elements (ceiling, walls, columns and so on) and the rest of items in robot surroundings. All objects in robot surroundings, both dynamic and static, are considered to be obstacles but the structure of the environment itself. A combination of a Bayesian algorithm and a deterministic clustering process is used in order to obtain a multimodal representation of speed and position of detected obstacles. Performance of the final system has been tested against state of the art proposals; test results validate the authors' proposal. The designed algorithms and procedures provide a solution to those applications where similar multimodal data structures are found.
Influence of control parameters on the joint tracking performance of a coaxial weld vision system
NASA Technical Reports Server (NTRS)
Gangl, K. J.; Weeks, J. L.
1985-01-01
The first phase of a series of evaluations of a vision-based welding control sensor for the Space Shuttle Main Engine Robotic Welding System is described. The robotic welding system is presently under development at the Marshall Space Flight Center. This evaluation determines the standard control response parameters necessary for proper trajectory of the welding torch along the joint.
Spatial and rotational quality assurance of 6DOF patient tracking systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Belcher, Andrew H.; Liu, Xinmin; Grelewicz, Zachary
Purpose: External tracking systems used for patient positioning and motion monitoring during radiotherapy are now capable of detecting both translations and rotations. In this work, the authors develop a novel technique to evaluate the 6 degree of freedom 6(DOF) (translations and rotations) performance of external motion tracking systems. The authors apply this methodology to an infrared marker tracking system and two 3D optical surface mapping systems in a common tumor 6DOF workspace. Methods: An in-house designed and built 6DOF parallel kinematics robotic motion phantom was used to perform motions with sub-millimeter and subdegree accuracy in a 6DOF workspace. An infraredmore » marker tracking system was first used to validate a calibration algorithm which associates the motion phantom coordinate frame to the camera frame. The 6DOF positions of the mobile robotic system in this space were then tracked and recorded independently by an optical surface tracking system after a cranial phantom was rigidly fixed to the moveable platform of the robotic stage. The calibration methodology was first employed, followed by a comprehensive 6DOF trajectory evaluation, which spanned a full range of positions and orientations in a 20 × 20 × 16 mm and 5° × 5° × 5° workspace. The intended input motions were compared to the calibrated 6DOF measured points. Results: The technique found the accuracy of the infrared (IR) marker tracking system to have maximal root-mean square error (RMSE) values of 0.18, 0.25, 0.07 mm, 0.05°, 0.05°, and 0.09° in left–right (LR), superior–inferior (SI), anterior–posterior (AP), pitch, roll, and yaw, respectively, comparing the intended 6DOF position and the measured position by the IR camera. Similarly, the 6DOF RSME discrepancy for the HD optical surface tracker yielded maximal values of 0.46, 0.60, 0.54 mm, 0.06°, 0.11°, and 0.08° in LR, SI, AP, pitch, roll, and yaw, respectively, over the same 6DOF evaluative workspace. An earlier generation 3D optical surface tracking unit was observed to have worse tracking capabilities than both the IR camera unit and the newer 3D surface tracking system with maximal RMSE of 0.69, 0.74, 0.47 mm, 0.28°, 0.19°, and 0.18°, in LR, SI, AP, pitch, roll, and yaw, respectively, in the same 6DOF evaluation space. Conclusions: The proposed technique was found to be effective at evaluating the performance of 6DOF patient tracking systems. All observed optical tracking systems were found to exhibit tracking capabilities at the sub-millimeter and subdegree level within a 6DOF workspace.« less
A discrete-time adaptive control scheme for robot manipulators
NASA Technical Reports Server (NTRS)
Tarokh, M.
1990-01-01
A discrete-time model reference adaptive control scheme is developed for trajectory tracking of robot manipulators. The scheme utilizes feedback, feedforward, and auxiliary signals, obtained from joint angle measurement through simple expressions. Hyperstability theory is utilized to derive the adaptation laws for the controller gain matrices. It is shown that trajectory tracking is achieved despite gross robot parameter variation and uncertainties. The method offers considerable design flexibility and enables the designer to improve the performance of the control system by adjusting free design parameters. The discrete-time adaptation algorithm is extremely simple and is therefore suitable for real-time implementation. Simulations and experimental results are given to demonstrate the performance of the scheme.
Efficient physics-based tracking of heart surface motion for beating heart surgery robotic systems.
Bogatyrenko, Evgeniya; Pompey, Pascal; Hanebeck, Uwe D
2011-05-01
Tracking of beating heart motion in a robotic surgery system is required for complex cardiovascular interventions. A heart surface motion tracking method is developed, including a stochastic physics-based heart surface model and an efficient reconstruction algorithm. The algorithm uses the constraints provided by the model that exploits the physical characteristics of the heart. The main advantage of the model is that it is more realistic than most standard heart models. Additionally, no explicit matching between the measurements and the model is required. The application of meshless methods significantly reduces the complexity of physics-based tracking. Based on the stochastic physical model of the heart surface, this approach considers the motion of the intervention area and is robust to occlusions and reflections. The tracking algorithm is evaluated in simulations and experiments on an artificial heart. Providing higher accuracy than the standard model-based methods, it successfully copes with occlusions and provides high performance even when all measurements are not available. Combining the physical and stochastic description of the heart surface motion ensures physically correct and accurate prediction. Automatic initialization of the physics-based cardiac motion tracking enables system evaluation in a clinical environment.
Robotically assisted ureteroscopy for kidney exploration.
Talari, Hadi F; Monfaredi, Reza; Wilson, Emmanuel; Blum, Emily; Bayne, Christopher; Peters, Craig; Zhang, Anlin; Cleary, Kevin
2017-02-01
Ureteroscopy is a minimally invasive procedure for diagnosis and treatment of a wide range of urinary tract pathologies. It is most commonly performed in the diagnostic work-up of hematuria and the diagnosis and treatment of upper urinary tract malignancies and calculi. Ergonomic and visualization challenges as well as radiation exposure are limitations to conventional ureteroscopy. For example, for diagnostic tumor inspection, the urologist has to maneuver the ureteroscope through each of the 6 to 12 calyces in the kidney under fluoroscopy to ensure complete surveillance. Therefore, we have been developing a robotic system to "power drive" a flexible fiber-optic ureteroscope with 3D tip tracking and pre-operative image overlay. Our goal is to provide the urologist precise control of the ureteroscope tip with less radiation exposure. Our prototype system allows control of the three degrees of freedom of the ureteroscope via brushless motors and a joystick interface. The robot provides a steady platform for controlling the ureteroscope. Furthermore, the robot design facilitates a quick "snap-in" of the ureteroscope, thus allowing the ureteroscope to be mounted midway through the procedure. We have completed the mechanical system and the controlling software and begun evaluation using a kidney phantom. We put MRI-compatible fiducials on the phantom and obtained MR images. We registered these images with the robot using an electromagnetic tracking system and paired-point registration. The system is described and initial evaluation results are given in this paper.
Control of articulated snake robot under dynamic active constraints.
Kwok, Ka-Wai; Vitiello, Valentina; Yang, Guang-Zhong
2010-01-01
Flexible, ergonomically enhanced surgical robots have important applications to transluminal endoscopic surgery, for which path-following and dynamic shape conformance are essential. In this paper, kinematic control of a snake robot for motion stabilisation under dynamic active constraints is addressed. The main objective is to enable the robot to track the visual target accurately and steadily on deforming tissue whilst conforming to pre-defined anatomical constraints. The motion tracking can also be augmented with manual control. By taking into account the physical limits in terms of maximum frequency response of the system (manifested as a delay between the input of the manipulator and the movement of the end-effector), we show the importance of visual-motor synchronisation for performing accurate smooth pursuit movements. Detailed user experiments are performed to demonstrate the practical value of the proposed control mechanism.
Using Visual Odometry to Estimate Position and Attitude
NASA Technical Reports Server (NTRS)
Maimone, Mark; Cheng, Yang; Matthies, Larry; Schoppers, Marcel; Olson, Clark
2007-01-01
A computer program in the guidance system of a mobile robot generates estimates of the position and attitude of the robot, using features of the terrain on which the robot is moving, by processing digitized images acquired by a stereoscopic pair of electronic cameras mounted rigidly on the robot. Developed for use in localizing the Mars Exploration Rover (MER) vehicles on Martian terrain, the program can also be used for similar purposes on terrestrial robots moving in sufficiently visually textured environments: examples include low-flying robotic aircraft and wheeled robots moving on rocky terrain or inside buildings. In simplified terms, the program automatically detects visual features and tracks them across stereoscopic pairs of images acquired by the cameras. The 3D locations of the tracked features are then robustly processed into an estimate of overall vehicle motion. Testing has shown that by use of this software, the error in the estimate of the position of the robot can be limited to no more than 2 percent of the distance traveled, provided that the terrain is sufficiently rich in features. This software has proven extremely useful on the MER vehicles during driving on sandy and highly sloped terrains on Mars.
Non-orthogonal tool/flange and robot/world calibration.
Ernst, Floris; Richter, Lars; Matthäus, Lars; Martens, Volker; Bruder, Ralf; Schlaefer, Alexander; Schweikard, Achim
2012-12-01
For many robot-assisted medical applications, it is necessary to accurately compute the relation between the robot's coordinate system and the coordinate system of a localisation or tracking device. Today, this is typically carried out using hand-eye calibration methods like those proposed by Tsai/Lenz or Daniilidis. We present a new method for simultaneous tool/flange and robot/world calibration by estimating a solution to the matrix equation AX = YB. It is computed using a least-squares approach. Because real robots and localisation are all afflicted by errors, our approach allows for non-orthogonal matrices, partially compensating for imperfect calibration of the robot or localisation device. We also introduce a new method where full robot/world and partial tool/flange calibration is possible by using localisation devices providing less than six degrees of freedom (DOFs). The methods are evaluated on simulation data and on real-world measurements from optical and magnetical tracking devices, volumetric ultrasound providing 3-DOF data, and a surface laser scanning device. We compare our methods with two classical approaches: the method by Tsai/Lenz and the method by Daniilidis. In all experiments, the new algorithms outperform the classical methods in terms of translational accuracy by up to 80% and perform similarly in terms of rotational accuracy. Additionally, the methods are shown to be stable: the number of calibration stations used has far less influence on calibration quality than for the classical methods. Our work shows that the new method can be used for estimating the relationship between the robot's and the localisation device's coordinate systems. The new method can also be used for deficient systems providing only 3-DOF data, and it can be employed in real-time scenarios because of its speed. Copyright © 2012 John Wiley & Sons, Ltd.
Control Algorithms for a Shape-shifting Tracked Robotic Vehicle Climbing Obstacles
2008-12-01
robot be- havioural skills. The Swiss Federal Institute of Technology is developing the shape-shifting robotic platform Octopus [6] (Figure l(b...and traverse steep (a) Lurker (b) Octopus (c) NUGV (d) Chaos (e) STRV Figure 1: Shape-shifting robotic vehicles in different research labs. DRDC...situations. The system is assumed stuck when vv?; + v~ + v’i) < 0.01 mls or Vx < O. Only forward movements are considered in this work, for this reason
Smart Braid Feedback for the Closed-Loop Control of Soft Robotic Systems.
Felt, Wyatt; Chin, Khai Yi; Remy, C David
2017-09-01
This article experimentally investigates the potential of using flexible, inductance-based contraction sensors in the closed-loop motion control of soft robots. Accurate motion control remains a highly challenging task for soft robotic systems. Precise models of the actuation dynamics and environmental interactions are often unavailable. This renders open-loop control impossible, while closed-loop control suffers from a lack of suitable feedback. Conventional motion sensors, such as linear or rotary encoders, are difficult to adapt to robots that lack discrete mechanical joints. The rigid nature of these sensors runs contrary to the aspirational benefits of soft systems. As truly soft sensor solutions are still in their infancy, motion control of soft robots has so far relied on laboratory-based sensing systems such as motion capture, electromagnetic (EM) tracking, or Fiber Bragg Gratings. In this article, we used embedded flexible sensors known as Smart Braids to sense the contraction of McKibben muscles through changes in inductance. We evaluated closed-loop control on two systems: a revolute joint and a planar, one degree of freedom continuum manipulator. In the revolute joint, our proposed controller compensated for elasticity in the actuator connections. The Smart Braid feedback allowed motion control with a steady-state root-mean-square (RMS) error of [1.5]°. In the continuum manipulator, Smart Braid feedback enabled tracking of the desired tip angle with a steady-state RMS error of [1.25]°. This work demonstrates that Smart Braid sensors can provide accurate position feedback in closed-loop motion control suitable for field applications of soft robotic systems.
NASA Astrophysics Data System (ADS)
Bagheri, Zahra M.; Cazzolato, Benjamin S.; Grainger, Steven; O'Carroll, David C.; Wiederman, Steven D.
2017-08-01
Objective. Many computer vision and robotic applications require the implementation of robust and efficient target-tracking algorithms on a moving platform. However, deployment of a real-time system is challenging, even with the computational power of modern hardware. Lightweight and low-powered flying insects, such as dragonflies, track prey or conspecifics within cluttered natural environments, illustrating an efficient biological solution to the target-tracking problem. Approach. We used our recent recordings from ‘small target motion detector’ neurons in the dragonfly brain to inspire the development of a closed-loop target detection and tracking algorithm. This model exploits facilitation, a slow build-up of response to targets which move along long, continuous trajectories, as seen in our electrophysiological data. To test performance in real-world conditions, we implemented this model on a robotic platform that uses active pursuit strategies based on insect behaviour. Main results. Our robot performs robustly in closed-loop pursuit of targets, despite a range of challenging conditions used in our experiments; low contrast targets, heavily cluttered environments and the presence of distracters. We show that the facilitation stage boosts responses to targets moving along continuous trajectories, improving contrast sensitivity and detection of small moving targets against textured backgrounds. Moreover, the temporal properties of facilitation play a useful role in handling vibration of the robotic platform. We also show that the adoption of feed-forward models which predict the sensory consequences of self-movement can significantly improve target detection during saccadic movements. Significance. Our results provide insight into the neuronal mechanisms that underlie biological target detection and selection (from a moving platform), as well as highlight the effectiveness of our bio-inspired algorithm in an artificial visual system.
Calibration of the motor-assisted robotic stereotaxy system: MARS.
Heinig, Maximilian; Hofmann, Ulrich G; Schlaefer, Alexander
2012-11-01
The motor-assisted robotic stereotaxy system presents a compact and light-weight robotic system for stereotactic neurosurgery. Our system is designed to position probes in the human brain for various applications, for example, deep brain stimulation. It features five fully automated axes. High positioning accuracy is of utmost importance in robotic neurosurgery. First, the key parameters of the robot's kinematics are determined using an optical tracking system. Next, the positioning errors at the center of the arc--which is equivalent to the target position in stereotactic interventions--are investigated using a set of perpendicular cameras. A modeless robot calibration method is introduced and evaluated. To conclude, the application accuracy of the robot is studied in a phantom trial. We identified the bending of the arc under load as the robot's main error source. A calibration algorithm was implemented to compensate for the deflection of the robot's arc. The mean error after the calibration was 0.26 mm, the 68.27th percentile was 0.32 mm, and the 95.45th was 0.50 mm. The kinematic properties of the robot were measured, and based on the results an appropriate calibration method was derived. With mean errors smaller than currently used mechanical systems, our results show that the robot's accuracy is appropriate for stereotactic interventions.
NASA Astrophysics Data System (ADS)
Zhang, Yachu; Zhao, Yuejin; Liu, Ming; Dong, Liquan; Kong, Lingqin; Liu, Lingling
2017-09-01
In contrast to humans, who use only visual information for navigation, many mobile robots use laser scanners and ultrasonic sensors along with vision cameras to navigate. This work proposes a vision-based robot control algorithm based on deep convolutional neural networks. We create a large 15-layer convolutional neural network learning system and achieve the advanced recognition performance. Our system is trained from end to end to map raw input images to direction in supervised mode. The images of data sets are collected in a wide variety of weather conditions and lighting conditions. Besides, the data sets are augmented by adding Gaussian noise and Salt-and-pepper noise to avoid overfitting. The algorithm is verified by two experiments, which are line tracking and obstacle avoidance. The line tracking experiment is proceeded in order to track the desired path which is composed of straight and curved lines. The goal of obstacle avoidance experiment is to avoid the obstacles indoor. Finally, we get 3.29% error rate on the training set and 5.1% error rate on the test set in the line tracking experiment, 1.8% error rate on the training set and less than 5% error rate on the test set in the obstacle avoidance experiment. During the actual test, the robot can follow the runway centerline outdoor and avoid the obstacle in the room accurately. The result confirms the effectiveness of the algorithm and our improvement in the network structure and train parameters
NASA Astrophysics Data System (ADS)
Petrochenko, Andrey; Konyakhin, Igor
2017-06-01
In connection with the development of robotics have become increasingly popular variety of three-dimensional reconstruction of the system mapping and image-set received from the optical sensors. The main objective of technical and robot vision is the detection, tracking and classification of objects of the space in which these systems and robots operate [15,16,18]. Two-dimensional images sometimes don't contain sufficient information to address those or other problems: the construction of the map of the surrounding area for a route; object identification, tracking their relative position and movement; selection of objects and their attributes to complement the knowledge base. Three-dimensional reconstruction of the surrounding space allows you to obtain information on the relative positions of objects, their shape, surface texture. Systems, providing training on the basis of three-dimensional reconstruction of the results of the comparison can produce two-dimensional images of three-dimensional model that allows for the recognition of volume objects on flat images. The problem of the relative orientation of industrial robots with the ability to build threedimensional scenes of controlled surfaces is becoming actual nowadays.
The robot's eyes - Stereo vision system for automated scene analysis
NASA Technical Reports Server (NTRS)
Williams, D. S.
1977-01-01
Attention is given to the robot stereo vision system which maintains the image produced by solid-state detector television cameras in a dynamic random access memory called RAPID. The imaging hardware consists of sensors (two solid-state image arrays using a charge injection technique), a video-rate analog-to-digital converter, the RAPID memory, and various types of computer-controlled displays, and preprocessing equipment (for reflexive actions, processing aids, and object detection). The software is aimed at locating objects and transversibility. An object-tracking algorithm is discussed and it is noted that tracking speed is in the 50-75 pixels/s range.
2011-02-07
Sensor UGVs (SUGV) or Disruptor UGVs, depending on their payload. The SUGVs included vision, GPS/IMU, and LIDAR systems for identifying and tracking...employed by all the MAGICian research groups. Objects of interest were tracked using standard LIDAR and Computer Vision template-based feature...tracking approaches. Mapping was solved through Multi-Agent particle-filter based Simultaneous Locali- zation and Mapping ( SLAM ). Our system contains
Experientally guided robots. [for planet exploration
NASA Technical Reports Server (NTRS)
Merriam, E. W.; Becker, J. D.
1974-01-01
This paper argues that an experientally guided robot is necessary to successfully explore far-away planets. Such a robot is characterized as having sense organs which receive sensory information from its environment and motor systems which allow it to interact with that environment. The sensori-motor information which it receives is organized into an experiential knowledge structure and this knowledge in turn is used to guide the robot's future actions. A summary is presented of a problem solving system which is being used as a test bed for developing such a robot. The robot currently engages in the behaviors of visual tracking, focusing down, and looking around in a simulated Martian landscape. Finally, some unsolved problems are outlined whose solutions are necessary before an experientally guided robot can be produced. These problems center around organizing the motivational and memory structure of the robot and understanding its high-level control mechanisms.
Automatic tracking of laparoscopic instruments for autonomous control of a cameraman robot.
Khoiy, Keyvan Amini; Mirbagheri, Alireza; Farahmand, Farzam
2016-01-01
An automated instrument tracking procedure was designed and developed for autonomous control of a cameraman robot during laparoscopic surgery. The procedure was based on an innovative marker-free segmentation algorithm for detecting the tip of the surgical instruments in laparoscopic images. A compound measure of Saturation and Value components of HSV color space was incorporated that was enhanced further using the Hue component and some essential characteristics of the instrument segment, e.g., crossing the image boundaries. The procedure was then integrated into the controlling system of the RoboLens cameraman robot, within a triple-thread parallel processing scheme, such that the tip is always kept at the center of the image. Assessment of the performance of the system on prerecorded real surgery movies revealed an accuracy rate of 97% for high quality images and about 80% for those suffering from poor lighting and/or blood, water and smoke noises. A reasonably satisfying performance was also observed when employing the system for autonomous control of the robot in a laparoscopic surgery phantom, with a mean time delay of 200ms. It was concluded that with further developments, the proposed procedure can provide a practical solution for autonomous control of cameraman robots during laparoscopic surgery operations.
Drift-Free Humanoid State Estimation fusing Kinematic, Inertial and LIDAR Sensing
2014-08-01
registration to this map and other objects in the robot’s vicinity while also contributing to direct low-level control of a Boston Dynamics Atlas robot ...requirements. I. INTRODUCTION Dynamic locomotion of legged robotic systems remains an open and challenging research problem whose solution will enable...humanoids to perform tasks and reach places inaccessible to wheeled or tracked robots . Several research institutions are developing walking and running
System Design and Locomotion of Superball, an Untethered Tensegrity Robot
NASA Technical Reports Server (NTRS)
Sabelhaus, Andrew P.; Bruce, Jonathan; Caluwaerts, Ken; Manovi, Pavlo; Firoozi, Roya Fallah; Dobi, Sarah; Agogino, Alice M.; Sunspiral, Vytas
2015-01-01
The Spherical Underactuated Planetary Exploration Robot ball (SUPERball) is an ongoing project within NASA Ames Research Center's Intelligent Robotics Group and the Dynamic Tensegrity Robotics Lab (DTRL). The current SUPERball is the first full prototype of this tensegrity robot platform, eventually destined for space exploration missions. This work, building on prior published discussions of individual components, presents the fully-constructed robot. Various design improvements are discussed, as well as testing results of the sensors and actuators that illustrate system performance. Basic low-level motor position controls are implemented and validated against sensor data, which show SUPERball to be uniquely suited for highly dynamic state trajectory tracking. Finally, SUPERball is shown in a simple example of locomotion. This implementation of a basic motion primitive shows SUPERball in untethered control.
Control of wheeled mobile robot in restricted environment
NASA Astrophysics Data System (ADS)
Ali, Mohammed A. H.; En, Chang Yong
2018-03-01
This paper presents a simulation and practical control system for wheeled mobile robot in restricted environment. A wheeled mobile robot with 3 wheels is fabricated and controlled by proportional derivative active force control (PD-AFC) to move in a pre-planned restricted environment to maintain the tracking errors at zero level. A control system with two loops, outer by PD controller and inner loop by Active Force Control, are designed to control the wheeled mobile robot. Fuzzy logic controller is implemented in the Active force Control to estimate the inertia matrix that will be used to calculate the actual torque applied on the wheeled mobile robot. The mobile robot is tested in two different trajectories, namely are circular and straight path. The actual path and desired path are compared.
Training industrial robots with gesture recognition techniques
NASA Astrophysics Data System (ADS)
Piane, Jennifer; Raicu, Daniela; Furst, Jacob
2013-01-01
In this paper we propose to use gesture recognition approaches to track a human hand in 3D space and, without the use of special clothing or markers, be able to accurately generate code for training an industrial robot to perform the same motion. The proposed hand tracking component includes three methods: a color-thresholding model, naïve Bayes analysis and Support Vector Machine (SVM) to detect the human hand. Next, it performs stereo matching on the region where the hand was detected to find relative 3D coordinates. The list of coordinates returned is expectedly noisy due to the way the human hand can alter its apparent shape while moving, the inconsistencies in human motion and detection failures in the cluttered environment. Therefore, the system analyzes the list of coordinates to determine a path for the robot to move, by smoothing the data to reduce noise and looking for significant points used to determine the path the robot will ultimately take. The proposed system was applied to pairs of videos recording the motion of a human hand in a „real‟ environment to move the end-affector of a SCARA robot along the same path as the hand of the person in the video. The correctness of the robot motion was determined by observers indicating that motion of the robot appeared to match the motion of the video.
Trajectory tracking control for a nonholonomic mobile robot under ROS
NASA Astrophysics Data System (ADS)
Lakhdar Besseghieur, Khadir; Trębiński, Radosław; Kaczmarek, Wojciech; Panasiuk, Jarosław
2018-05-01
In this paper, the implementation of the trajectory tracking control strategy on a ROS-based mobile robot is considered. Our test-bench is the nonholonomic mobile robot ‘TURTLEBOT’. ROS facilitates considerably setting-up a suitable environment to test the designed controller. Our aim is to develop a framework using ROS concepts so that a trajectory tracking controller can be implemented on any ROS-enabled mobile robot. Practical experiments with ‘TURTLEBOT’ are conducted to assess the framework reliability.
Dickstein-Fischer, Laurie; Fischer, Gregory S
2014-01-01
It is estimated that Autism Spectrum Disorder (ASD) affects 1 in 68 children. Early identification of an ASD is exceedingly important to the introduction of an intervention. We are developing a robot-assisted approach that will serve as an improved diagnostic and early intervention tool for children with autism. The robot, named PABI® (Penguin for Autism Behavioral Interventions), is a compact humanoid robot taking on an expressive cartoon-like embodiment. The robot is affordable, durable, and portable so that it can be used in various settings including schools, clinics, and the home. Thus enabling significantly enhanced and more readily available diagnosis and continuation of care. Through facial expressions, body motion, verbal cues, stereo vision-based tracking, and a tablet computer, the robot is capable of interacting meaningfully with an autistic child. Initial implementations of the robot, as part of a comprehensive treatment model (CTM), include Applied Behavioral Analysis (ABA) therapy where the child interacts with a tablet computer wirelessly interfaced with the robot. At the same time, the robot makes meaningful expressions and utterances and uses stereo cameras in eyes to track the child, maintain eye contact, and collect data such as affect and gaze direction for charting of progress. In this paper we present the clinical justification, anticipated usage with corresponding requirements, prototype development of the robotic system, and demonstration of a sample application for robot-assisted ABA therapy.
Novel locomotion via biological inspiration
NASA Astrophysics Data System (ADS)
Quinn, Roger D.; Boxerbaum, Alexander; Palmer, Luther; Chiel, Hillel; Diller, Eric; Hunt, Alexander; Bachmann, Richard
2011-05-01
Animal behavioral, physiological and neurobiological studies are providing a wealth of inspirational data for robot design and control. Several very different biologically inspired mobile robots will be reviewed. A robot called DIGbot is being developed that moves independent of the direction of gravity using Distributed Inward Gripping (DIG) as a rapid and robust attachment mechanism observed in climbing animals. DIGbot is an 18 degree of freedom hexapod with onboard power and control systems. Passive compliance in its feet, which is inspired by the flexible tarsus of the cockroach, increases the robustness of the adhesion strategy and enables DIGbot to execute large steps and stationary turns while walking on mesh screens. A Whegs™ robot, inspired by insect locomotion principles, is being developed that can be rapidly reconfigured between tracks and wheel-legs and carry GeoSystems Zipper Mast. The mechanisms that cause it to passively change its gait on irregular terrain have been integrated into its hubs for a compact and modular design. The robot is designed to move smoothly on moderately rugged terrain using its tracks and run on irregular terrain and stairs using its wheel-legs. We are also developing soft bodied robots that use peristalsis, the same method of locomotion earthworms use. We present a technique of using a braided mesh exterior to produce fluid waves of motion along the body of the robot that increase the robot's speed relative to previous designs. The concept is highly scalable, for endoscopes to water, oil or gas line inspection.
Mobile Robot Designed with Autonomous Navigation System
NASA Astrophysics Data System (ADS)
An, Feng; Chen, Qiang; Zha, Yanfang; Tao, Wenyin
2017-10-01
With the rapid development of robot technology, robots appear more and more in all aspects of life and social production, people also ask more requirements for the robot, one is that robot capable of autonomous navigation, can recognize the road. Take the common household sweeping robot as an example, which could avoid obstacles, clean the ground and automatically find the charging place; Another example is AGV tracking car, which can following the route and reach the destination successfully. This paper introduces a new type of robot navigation scheme: SLAM, which can build the environment map in a totally strange environment, and at the same time, locate its own position, so as to achieve autonomous navigation function.
Improving the transparency of a rehabilitation robot by exploiting the cyclic behaviour of walking.
van Dijk, W; van der Kooij, H; Koopman, B; van Asseldonk, E H F; van der Kooij, H
2013-06-01
To promote active participation of neurological patients during robotic gait training, controllers, such as "assist as needed" or "cooperative control", are suggested. Apart from providing support, these controllers also require that the robot should be capable of resembling natural, unsupported, walking. This means that they should have a transparent mode, where the interaction forces between the human and the robot are minimal. Traditional feedback-control algorithms do not exploit the cyclic nature of walking to improve the transparency of the robot. The purpose of this study was to improve the transparent mode of robotic devices, by developing two controllers that use the rhythmic behavior of gait. Both controllers use adaptive frequency oscillators and kernel-based non-linear filters. Kernelbased non-linear filters can be used to estimate signals and their time derivatives, as a function of the gait phase. The first controller learns the motor angle, associated with a certain joint angle pattern, and acts as a feed-forward controller to improve the torque tracking (including the zero-torque mode). The second controller learns the state of the mechanical system and compensates for the dynamical effects (e.g. the acceleration of robot masses). Both controllers have been tested separately and in combination on a small subject population. Using the feedforward controller resulted in an improved torque tracking of at least 52 percent at the hip joint, and 61 percent at the knee joint. When both controllers were active simultaneously, the interaction power between the robot and the human leg was reduced by at least 40 percent at the thigh, and 43 percent at the shank. These results indicate that: if a robotic task is cyclic, the torque tracking and transparency can be improved by exploiting the predictions of adaptive frequency oscillator and kernel-based nonlinear filters.
NASA Astrophysics Data System (ADS)
Zhang, Haichong K.; Fang, Ting Yun; Finocchi, Rodolfo; Boctor, Emad M.
2017-03-01
Three dimensional (3D) ultrasound imaging is becoming a standard mode for medical ultrasound diagnoses. Conventional 3D ultrasound imaging is mostly scanned either by using a two dimensional matrix array or by motorizing a one dimensional array in the elevation direction. However, the former system is not widely assessable due to its cost, and the latter one has limited resolution and field-of-view in the elevation axis. Here, we propose a 3D ultrasound imaging system based on the synthetic tracked aperture approach, in which a robotic arm is used to provide accurate tracking and motion. While the ultrasound probe is moved by a robotic arm, each probe position is tracked and can be used to reconstruct a wider field-of-view as there are no physical barriers that restrict the elevational scanning. At the same time, synthetic aperture beamforming provides a better resolution in the elevation axis. To synthesize the elevational information, the single focal point is regarded as the virtual element, and forward and backward delay-andsum are applied to the radio-frequency (RF) data collected through the volume. The concept is experimentally validated using a general ultrasound phantom, and the elevational resolution improvement of 2.54 and 2.13 times was measured at the target depths of 20 mm and 110 mm, respectively.
Gaze-contingent control for minimally invasive robotic surgery.
Mylonas, George P; Darzi, Ara; Yang, Guang Zhong
2006-09-01
Recovering tissue depth and deformation during robotically assisted minimally invasive procedures is an important step towards motion compensation, stabilization and co-registration with preoperative data. This work demonstrates that eye gaze derived from binocular eye tracking can be effectively used to recover 3D motion and deformation of the soft tissue. A binocular eye-tracking device was integrated into the stereoscopic surgical console. After calibration, the 3D fixation point of the participating subjects could be accurately resolved in real time. A CT-scanned phantom heart model was used to demonstrate the accuracy of gaze-contingent depth extraction and motion stabilization of the soft tissue. The dynamic response of the oculomotor system was assessed with the proposed framework by using autoregressive modeling techniques. In vivo data were also used to perform gaze-contingent decoupling of cardiac and respiratory motion. Depth reconstruction, deformation tracking, and motion stabilization of the soft tissue were possible with binocular eye tracking. The dynamic response of the oculomotor system was able to cope with frequencies likely to occur under most routine minimally invasive surgical operations. The proposed framework presents a novel approach towards the tight integration of a human and a surgical robot where interaction in response to sensing is required to be under the control of the operating surgeon.
SLAMM: Visual monocular SLAM with continuous mapping using multiple maps
Md. Sabri, Aznul Qalid; Loo, Chu Kiong; Mansoor, Ali Mohammed
2018-01-01
This paper presents the concept of Simultaneous Localization and Multi-Mapping (SLAMM). It is a system that ensures continuous mapping and information preservation despite failures in tracking due to corrupted frames or sensor’s malfunction; making it suitable for real-world applications. It works with single or multiple robots. In a single robot scenario the algorithm generates a new map at the time of tracking failure, and later it merges maps at the event of loop closure. Similarly, maps generated from multiple robots are merged without prior knowledge of their relative poses; which makes this algorithm flexible. The system works in real time at frame-rate speed. The proposed approach was tested on the KITTI and TUM RGB-D public datasets and it showed superior results compared to the state-of-the-arts in calibrated visual monocular keyframe-based SLAM. The mean tracking time is around 22 milliseconds. The initialization is twice as fast as it is in ORB-SLAM, and the retrieved map can reach up to 90 percent more in terms of information preservation depending on tracking loss and loop closure events. For the benefit of the community, the source code along with a framework to be run with Bebop drone are made available at https://github.com/hdaoud/ORBSLAMM. PMID:29702697
Gaze-contingent soft tissue deformation tracking for minimally invasive robotic surgery.
Mylonas, George P; Stoyanov, Danail; Deligianni, Fani; Darzi, Ara; Yang, Guang-Zhong
2005-01-01
The introduction of surgical robots in Minimally Invasive Surgery (MIS) has allowed enhanced manual dexterity through the use of microprocessor controlled mechanical wrists. Although fully autonomous robots are attractive, both ethical and legal barriers can prohibit their practical use in surgery. The purpose of this paper is to demonstrate that it is possible to use real-time binocular eye tracking for empowering robots with human vision by using knowledge acquired in situ. By utilizing the close relationship between the horizontal disparity and the depth perception varying with the viewing distance, it is possible to use ocular vergence for recovering 3D motion and deformation of the soft tissue during MIS procedures. Both phantom and in vivo experiments were carried out to assess the potential frequency limit of the system and its intrinsic depth recovery accuracy. The potential applications of the technique include motion stabilization and intra-operative planning in the presence of large tissue deformation.
Teleoperation of Robonaut Using Finger Tracking
NASA Technical Reports Server (NTRS)
Champoux, Rachel G.; Luo, Victor
2012-01-01
With the advent of new finger tracking systems, the idea of a more expressive and intuitive user interface is being explored and implemented. One practical application for this new kind of interface is that of teleoperating a robot. For humanoid robots, a finger tracking interface is required due to the level of complexity in a human-like hand, where a joystick isn't accurate. Moreover, for some tasks, using one's own hands allows the user to communicate their intentions more effectively than other input. The purpose of this project was to develop a natural user interface for someone to teleoperate a robot that is elsewhere. Specifically, this was designed to control Robonaut on the international space station to do tasks too dangerous and/or too trivial for human astronauts. This interface was developed by integrating and modifying 3Gear's software, which includes a library of gestures and the ability to track hands. The end result is an interface in which the user can manipulate objects in real time in the user interface. then, the information is relayed to a simulator, the stand in for Robonaut, at a slight delay.
Human-Vehicle Interface for Semi-Autonomous Operation of Uninhabited Aero Vehicles
NASA Technical Reports Server (NTRS)
Jones, Henry L.; Frew, Eric W.; Woodley, Bruce R.; Rock, Stephen M.
2001-01-01
The robustness of autonomous robotic systems to unanticipated circumstances is typically insufficient for use in the field. The many skills of human user often fill this gap in robotic capability. To incorporate the human into the system, a useful interaction between man and machine must exist. This interaction should enable useful communication to be exchanged in a natural way between human and robot on a variety of levels. This report describes the current human-robot interaction for the Stanford HUMMINGBIRD autonomous helicopter. In particular, the report discusses the elements of the system that enable multiple levels of communication. An intelligent system agent manages the different inputs given to the helicopter. An advanced user interface gives the user and helicopter a method for exchanging useful information. Using this human-robot interaction, the HUMMINGBIRD has carried out various autonomous search, tracking, and retrieval missions.
CISUS: an integrated 3D ultrasound system for IGT using a modular tracking API
NASA Astrophysics Data System (ADS)
Boctor, Emad M.; Viswanathan, Anand; Pieper, Steve; Choti, Michael A.; Taylor, Russell H.; Kikinis, Ron; Fichtinger, Gabor
2004-05-01
Ultrasound has become popular in clinical/surgical applications, both as the primary image guidance modality and also in conjunction with other modalities like CT or MRI. Three dimensional ultrasound (3DUS) systems have also demonstrated usefulness in image-guided therapy (IGT). At the same time, however, current lack of open-source and open-architecture multi-modal medical visualization systems prevents 3DUS from fulfilling its potential. Several stand-alone 3DUS systems, like Stradx or In-Vivo exist today. Although these systems have been found to be useful in real clinical setting, it is difficult to augment their functionality and integrate them in versatile IGT systems. To address these limitations, a robotic/freehand 3DUS open environment (CISUS) is being integrated into the 3D Slicer, an open-source research tool developed for medical image analysis and surgical planning. In addition, the system capitalizes on generic application programming interfaces (APIs) for tracking devices and robotic control. The resulting platform-independent open-source system may serve as a valuable tool to the image guided surgery community. Other researchers could straightforwardly integrate the generic CISUS system along with other functionalities (i.e. dual view visualization, registration, real-time tracking, segmentation, etc) to rapidly create their medical/surgical applications. Our current driving clinical application is robotically assisted and freehand 3DUS-guided liver ablation, which is fully being integrated under the CISUS-3D Slicer. Initial functionality and pre-clinical feasibility are demonstrated on phantom and ex-vivo animal models.
Evaluation of a completely robotized neurosurgical operating microscope.
Kantelhardt, Sven R; Finke, Markus; Schweikard, Achim; Giese, Alf
2013-01-01
Operating microscopes are essential for most neurosurgical procedures. Modern robot-assisted controls offer new possibilities, combining the advantages of conventional and automated systems. We evaluated the prototype of a completely robotized operating microscope with an integrated optical coherence tomography module. A standard operating microscope was fitted with motors and control instruments, with the manual control mode and balance preserved. In the robot mode, the microscope was steered by a remote control that could be fixed to a surgical instrument. External encoders and accelerometers tracked microscope movements. The microscope was additionally fitted with an optical coherence tomography-scanning module. The robotized microscope was tested on model systems. It could be freely positioned, without forcing the surgeon to take the hands from the instruments or avert the eyes from the oculars. Positioning error was about 1 mm, and vibration faded in 1 second. Tracking of microscope movements, combined with an autofocus function, allowed determination of the focus position within the 3-dimensional space. This constituted a second loop of navigation independent from conventional infrared reflector-based techniques. In the robot mode, automated optical coherence tomography scanning of large surface areas was feasible. The prototype of a robotized optical coherence tomography-integrated operating microscope combines the advantages of a conventional manually controlled operating microscope with a remote-controlled positioning aid and a self-navigating microscope system that performs automated positioning tasks such as surface scans. This demonstrates that, in the future, operating microscopes may be used to acquire intraoperative spatial data, volume changes, and structural data of brain or brain tumor tissue.
Human motion behavior while interacting with an industrial robot.
Bortot, Dino; Ding, Hao; Antonopolous, Alexandros; Bengler, Klaus
2012-01-01
Human workers and industrial robots both have specific strengths within industrial production. Advantageously they complement each other perfectly, which leads to the development of human-robot interaction (HRI) applications. Bringing humans and robots together in the same workspace may lead to potential collisions. The avoidance of such is a central safety requirement. It can be realized with sundry sensor systems, all of them decelerating the robot when the distance to the human decreases alarmingly and applying the emergency stop, when the distance becomes too small. As a consequence, the efficiency of the overall systems suffers, because the robot has high idle times. Optimized path planning algorithms have to be developed to avoid that. The following study investigates human motion behavior in the proximity of an industrial robot. Three different kinds of encounters between the two entities under three robot speed levels are prompted. A motion tracking system is used to capture the motions. Results show, that humans keep an average distance of about 0,5m to the robot, when the encounter occurs. Approximation of the workbenches is influenced by the robot in ten of 15 cases. Furthermore, an increase of participants' walking velocity with higher robot velocities is observed.
Robotically assisted ureteroscopy for kidney exploration
Talari, Hadi F.; Monfaredi, Reza; Wilson, Emmanuel; Blum, Emily; Bayne, Christopher; Peters, Craig; Zhang, Anlin; Cleary, Kevin
2018-01-01
Ureteroscopy is a minimally invasive procedure for diagnosis and treatment of a wide range of urinary tract pathologies. It is most commonly performed in the diagnostic work-up of hematuria and the diagnosis and treatment of upper urinary tract malignancies and calculi. Ergonomic and visualization challenges as well as radiation exposure are limitations to conventional ureteroscopy. For example, for diagnostic tumor inspection, the urologist has to maneuver the ureteroscope through each of the 6 to 12 calyces in the kidney under fluoroscopy to ensure complete surveillance. Therefore, we have been developing a robotic system to “power drive” a flexible fiber-optic ureteroscope with 3D tip tracking and pre-operative image overlay. Our goal is to provide the urologist precise control of the ureteroscope tip with less radiation exposure. Our prototype system allows control of the three degrees of freedom of the ureteroscope via brushless motors and a joystick interface. The robot provides a steady platform for controlling the ureteroscope. Furthermore, the robot design facilitates a quick “snap-in” of the ureteroscope, thus allowing the ureteroscope to be mounted midway through the procedure. We have completed the mechanical system and the controlling software and begun evaluation using a kidney phantom. We put MRI-compatible fiducials on the phantom and obtained MR images. We registered these images with the robot using an electromagnetic tracking system and paired-point registration. The system is described and initial evaluation results are given in this paper. PMID:29731536
Beyl, Tim; Nicolai, Philip; Comparetti, Mirko D; Raczkowsky, Jörg; De Momi, Elena; Wörn, Heinz
2016-07-01
Scene supervision is a major tool to make medical robots safer and more intuitive. The paper shows an approach to efficiently use 3D cameras within the surgical operating room to enable for safe human robot interaction and action perception. Additionally the presented approach aims to make 3D camera-based scene supervision more reliable and accurate. A camera system composed of multiple Kinect and time-of-flight cameras has been designed, implemented and calibrated. Calibration and object detection as well as people tracking methods have been designed and evaluated. The camera system shows a good registration accuracy of 0.05 m. The tracking of humans is reliable and accurate and has been evaluated in an experimental setup using operating clothing. The robot detection shows an error of around 0.04 m. The robustness and accuracy of the approach allow for an integration into modern operating room. The data output can be used directly for situation and workflow detection as well as collision avoidance.
NASA Technical Reports Server (NTRS)
Otaguro, W. S.; Kesler, L. O.; Land, K. C.; Rhoades, D. E.
1987-01-01
An intelligent tracker capable of robotic applications requiring guidance and control of platforms, robotic arms, and end effectors has been developed. This packaged system capable of supervised autonomous robotic functions is partitioned into a multiple processor/parallel processing configuration. The system currently interfaces to cameras but has the capability to also use three-dimensional inputs from scanning laser rangers. The inputs are fed into an image processing and tracking section where the camera inputs are conditioned for the multiple tracker algorithms. An executive section monitors the image processing and tracker outputs and performs all the control and decision processes. The present architecture of the system is presented with discussion of its evolutionary growth for space applications. An autonomous rendezvous demonstration of this system was performed last year. More realistic demonstrations in planning are discussed.
Synthetic Fiber Capstan Drives for Highly Efficient, Torque Controlled, Robotic Applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mazumdar, Anirban; Spencer, Steven James; Hobart, Clinton
Here this paper describes the design and performance of a synthetic rope on sheave drive system. This system uses synthetic ropes instead of steel cables to achieve low weight and a compact form factor. We demonstrate how this system is capable of 28-Hz torque control bandwidth, 95% efficiency, and quiet operation, making it ideal for use on legged robots and other dynamic physically interactive systems. Component geometry and tailored maintenance procedures are used to achieve high endurance. Endurance tests based on walking data predict that the ropes will survive roughly 247,000 cycles when used on large (90 kg), fully actuatedmore » bipedal robot systems. The drive systems have been incorporated into two novel bipedal robots capable of three-dimensional unsupported walking. Robot data illustrate effective torque tracking and nearly silent operation. Finally, comparisons with alternative transmission designs illustrate the size, weight, and endurance advantages of using this type of synthetic rope drive system.« less
Synthetic Fiber Capstan Drives for Highly Efficient, Torque Controlled, Robotic Applications
Mazumdar, Anirban; Spencer, Steven James; Hobart, Clinton; ...
2017-01-05
Here this paper describes the design and performance of a synthetic rope on sheave drive system. This system uses synthetic ropes instead of steel cables to achieve low weight and a compact form factor. We demonstrate how this system is capable of 28-Hz torque control bandwidth, 95% efficiency, and quiet operation, making it ideal for use on legged robots and other dynamic physically interactive systems. Component geometry and tailored maintenance procedures are used to achieve high endurance. Endurance tests based on walking data predict that the ropes will survive roughly 247,000 cycles when used on large (90 kg), fully actuatedmore » bipedal robot systems. The drive systems have been incorporated into two novel bipedal robots capable of three-dimensional unsupported walking. Robot data illustrate effective torque tracking and nearly silent operation. Finally, comparisons with alternative transmission designs illustrate the size, weight, and endurance advantages of using this type of synthetic rope drive system.« less
A Biocompatible Near-Infrared 3D Tracking System*
Decker, Ryan S.; Shademan, Azad; Opfermann, Justin D.; Leonard, Simon; Kim, Peter C. W.; Krieger, Axel
2017-01-01
A fundamental challenge in soft-tissue surgery is that target tissue moves and deforms, becomes occluded by blood or other tissue, and is difficult to differentiate from surrounding tissue. We developed small biocompatible near-infrared fluorescent (NIRF) markers with a novel fused plenoptic and NIR camera tracking system, enabling 3D tracking of tools and target tissue while overcoming blood and tissue occlusion in the uncontrolled, rapidly changing surgical environment. In this work, we present the tracking system and marker design and compare tracking accuracies to standard optical tracking methods using robotic experiments. At speeds of 1 mm/s, we observe tracking accuracies of 1.61 mm, degrading only to 1.71 mm when the markers are covered in blood and tissue. PMID:28129145
Biocompatible Near-Infrared Three-Dimensional Tracking System.
Decker, Ryan S; Shademan, Azad; Opfermann, Justin D; Leonard, Simon; Kim, Peter C W; Krieger, Axel
2017-03-01
A fundamental challenge in soft-tissue surgery is that target tissue moves and deforms, becomes occluded by blood or other tissue, and is difficult to differentiate from surrounding tissue. We developed small biocompatible near-infrared fluorescent (NIRF) markers with a novel fused plenoptic and NIR camera tracking system, enabling three-dimensional tracking of tools and target tissue while overcoming blood and tissue occlusion in the uncontrolled, rapidly changing surgical environment. In this work, we present the tracking system and marker design and compare tracking accuracies to standard optical tracking methods using robotic experiments. At speeds of 1 mm/s, we observe tracking accuracies of 1.61 mm, degrading only to 1.71 mm when the markers are covered in blood and tissue.
Combined virtual and real robotic test-bed for single operator control of multiple robots
NASA Astrophysics Data System (ADS)
Lee, Sam Y.-S.; Hunt, Shawn; Cao, Alex; Pandya, Abhilash
2010-04-01
Teams of heterogeneous robots with different dynamics or capabilities could perform a variety of tasks such as multipoint surveillance, cooperative transport and explorations in hazardous environments. In this study, we work with heterogeneous robots of semi-autonomous ground and aerial robots for contaminant localization. We developed a human interface system which linked every real robot to its virtual counterpart. A novel virtual interface has been integrated with Augmented Reality that can monitor the position and sensory information from video feed of ground and aerial robots in the 3D virtual environment, and improve user situational awareness. An operator can efficiently control the real multi-robots using the Drag-to-Move method on the virtual multi-robots. This enables an operator to control groups of heterogeneous robots in a collaborative way for allowing more contaminant sources to be pursued simultaneously. The advanced feature of the virtual interface system is guarded teleoperation. This can be used to prevent operators from accidently driving multiple robots into walls and other objects. Moreover, the feature of the image guidance and tracking is able to reduce operator workload.
Plugin-docking system for autonomous charging using particle filter
NASA Astrophysics Data System (ADS)
Koyasu, Hiroshi; Wada, Masayoshi
2017-03-01
Autonomous charging of the robot battery is one of the key functions for the sake of expanding working areas of the robots. To realize it, most of existing systems use custom docking stations or artificial markers. By the other words, they can only charge on a few specific outlets. If the limit can be removed, working areas of the robots significantly expands. In this paper, we describe a plugin-docking system for the autonomous charging, which does not require any custom docking stations or artificial markers. A single camera is used for recognizing the 3D position of an outlet socket. A particle filter-based image tracking algorithm which is robust to the illumination change is applied. The algorithm is implemented on a robot with an omnidirectional moving system. The experimental results show the effectiveness of our system.
Tick, David; Satici, Aykut C; Shen, Jinglin; Gans, Nicholas
2013-08-01
This paper presents a novel navigation and control system for autonomous mobile robots that includes path planning, localization, and control. A unique vision-based pose and velocity estimation scheme utilizing both the continuous and discrete forms of the Euclidean homography matrix is fused with inertial and optical encoder measurements to estimate the pose, orientation, and velocity of the robot and ensure accurate localization and control signals. A depth estimation system is integrated in order to overcome the loss of scale inherent in vision-based estimation. A path following control system is introduced that is capable of guiding the robot along a designated curve. Stability analysis is provided for the control system and experimental results are presented that prove the combined localization and control system performs with high accuracy.
Chanel, Laure-Anais; Nageotte, Florent; Vappou, Jonathan; Luo, Jianwen; Cuvillon, Loic; de Mathelin, Michel
2015-01-01
High Intensity Focused Ultrasound (HIFU) therapy is a very promising method for ablation of solid tumors. However, intra-abdominal organ motion, principally due to breathing, is a substantial limitation that results in incorrect tumor targeting. The objective of this work is to develop an all-in-one robotized HIFU system that can compensate motion in real-time during HIFU treatment. To this end, an ultrasound visual servoing scheme working at 20 Hz was designed. It relies on the motion estimation by using a fast ultrasonic speckle tracking algorithm and on the use of an interleaved imaging/HIFU sonication sequence for avoiding ultrasonic wave interferences. The robotized HIFU system was tested on a sample of chicken breast undergoing a vertical sinusoidal motion at 0.25 Hz. Sonications with and without motion compensation were performed in order to assess the effect of motion compensation on thermal lesions induced by HIFU. Motion was reduced by more than 80% thanks to this ultrasonic visual servoing system.
Nonlinear robust controller design for multi-robot systems with unknown payloads
NASA Technical Reports Server (NTRS)
Song, Y. D.; Anderson, J. N.; Homaifar, A.; Lai, H. Y.
1992-01-01
This work is concerned with the control problem of a multi-robot system handling a payload with unknown mass properties. Force constraints at the grasp points are considered. Robust control schemes are proposed that cope with the model uncertainty and achieve asymptotic path tracking. To deal with the force constraints, a strategy for optimally sharing the task is suggested. This strategy basically consists of two steps. The first detects the robots that need help and the second arranges that help. It is shown that the overall system is not only robust to uncertain payload parameters, but also satisfies the force constraints.
Real-time tracking of objects for a KC-135 microgravity experiment
NASA Technical Reports Server (NTRS)
Littlefield, Mark L.
1994-01-01
The design of a visual tracking system for use on the Extra-Vehicular Activity Helper/Retriever (EVAHR) is discussed. EVAHR is an autonomous robot designed to perform numerous tasks in an orbital microgravity environment. Since the ability to grasp a freely translating and rotating object is vital to the robot's mission, the EVAHR must analyze range image generated by the primary sensor. This allows EVAHR to locate and focus its sensors so that an accurate set of object poses can be determined and a grasp strategy planned. To test the visual tracking system being developed, a mathematical simulation was used to model the space station environment and maintain dynamics on the EVAHR and any other free floating objects. A second phase of the investigation consists of a series of experiments carried out aboard a KC-135 aircraft flying a parabolic trajectory to simulate microgravity.
Adaptive Shape Kernel-Based Mean Shift Tracker in Robot Vision System
2016-01-01
This paper proposes an adaptive shape kernel-based mean shift tracker using a single static camera for the robot vision system. The question that we address in this paper is how to construct such a kernel shape that is adaptive to the object shape. We perform nonlinear manifold learning technique to obtain the low-dimensional shape space which is trained by training data with the same view as the tracking video. The proposed kernel searches the shape in the low-dimensional shape space obtained by nonlinear manifold learning technique and constructs the adaptive kernel shape in the high-dimensional shape space. It can improve mean shift tracker performance to track object position and object contour and avoid the background clutter. In the experimental part, we take the walking human as example to validate that our method is accurate and robust to track human position and describe human contour. PMID:27379165
Three-dimensional face pose detection and tracking using monocular videos: tool and application.
Dornaika, Fadi; Raducanu, Bogdan
2009-08-01
Recently, we have proposed a real-time tracker that simultaneously tracks the 3-D head pose and facial actions in monocular video sequences that can be provided by low quality cameras. This paper has two main contributions. First, we propose an automatic 3-D face pose initialization scheme for the real-time tracker by adopting a 2-D face detector and an eigenface system. Second, we use the proposed methods-the initialization and tracking-for enhancing the human-machine interaction functionality of an AIBO robot. More precisely, we show how the orientation of the robot's camera (or any active vision system) can be controlled through the estimation of the user's head pose. Applications based on head-pose imitation such as telepresence, virtual reality, and video games can directly exploit the proposed techniques. Experiments on real videos confirm the robustness and usefulness of the proposed methods.
Maurice, Matthew J; Kaouk, Jihad H
2017-12-01
To assess the feasibility of radical perineal cystoprostatectomy using the latest generation purpose-built single-port robotic surgical system. In two male cadavers the da Vinci ® SP1098 Surgical System (Intuitive Surgical, Sunnyvale, CA, USA) was used to perform radical perineal cystoprostatectomy and bilateral extended pelvic lymph node dissection (ePLND). New features in this model include enhanced high-definition three-dimensional optics, improved instrument manoeuvrability, and a real-time instrument tracking and guidance system. The surgery was accomplished through a 3-cm perineal incision via a novel robotic single-port system, which accommodates three double-jointed articulating robotic instruments, an articulating camera, and an accessory laparoscopic instrument. The primary outcomes were technical feasibility, intraoperative complications, and total robotic operative time. The cases were completed successfully without conversion. There were no accidental punctures or lacerations. The robotic operative times were 197 and 202 min. In this preclinical model, robotic radical perineal cystoprostatectomy and ePLND was feasible using the SP1098 robotic platform. Further investigation is needed to assess the feasibility of urinary diversion using this novel approach and new technology. © 2017 The Authors BJU International © 2017 BJU International Published by John Wiley & Sons Ltd.
User-centric design of a personal assistance robot (FRASIER) for active aging.
Padir, Taşkin; Skorinko, Jeanine; Dimitrov, Velin
2015-01-01
We present our preliminary results from the design process for developing the Worcester Polytechnic Institute's personal assistance robot, FRASIER, as an intelligent service robot for enabling active aging. The robot capabilities include vision-based object detection, tracking the user and help with carrying heavy items such as grocery bags or cafeteria trays. This work-in-progress report outlines our motivation and approach to developing the next generation of service robots for the elderly. Our main contribution in this paper is the development of a set of specifications based on the adopted user-centered design process, and realization of the prototype system designed to meet these specifications.
UWB Tracking System Design with TDOA Algorithm
NASA Technical Reports Server (NTRS)
Ni, Jianjun; Arndt, Dickey; Ngo, Phong; Phan, Chau; Gross, Julia; Dusl, John; Schwing, Alan
2006-01-01
This presentation discusses an ultra-wideband (UWB) tracking system design effort using a tracking algorithm TDOA (Time Difference of Arrival). UWB technology is exploited to implement the tracking system due to its properties, such as high data rate, fine time resolution, and low power spectral density. A system design using commercially available UWB products is proposed. A two-stage weighted least square method is chosen to solve the TDOA non-linear equations. Matlab simulations in both two-dimensional space and three-dimensional space show that the tracking algorithm can achieve fine tracking resolution with low noise TDOA data. The error analysis reveals various ways to improve the tracking resolution. Lab experiments demonstrate the UWBTDOA tracking capability with fine resolution. This research effort is motivated by a prototype development project Mini-AERCam (Autonomous Extra-vehicular Robotic Camera), a free-flying video camera system under development at NASA Johnson Space Center for aid in surveillance around the International Space Station (ISS).
WE-G-213CD-06: Implementation of Real-Time Tumor Tracking Using Robotic Couch.
Buzurovic, I; Yu, Y; Podder, T
2012-06-01
The purpose of this study was to present a novel method for real- time tumor tracking using a commercially available robotic treatment couch, and to evaluate tumor tracking accuracy. Commercially available robotic couches are capable of positioning patients with high level of accuracy; however, currently there is no provision for compensating tumor motion using these systems. Elekta's existing commercial couch (PreciseTM Table) was used without changing its design. To establish the real-time couch motion for tracking, a novel control system was developed and implemented. The tabletop could be moved in horizontal plane (laterally and longitudinally) using two Maxon-24V motors with gearbox combination. Vertical motion was obtained using robust 70V-Rockwell Automation motor. For vertical motor position sensing, we used Model 755A-Accu- Coder encoder. Two Baumer-ITD_01_4mm shaft encoders were used for the lateral and longitudinal motions of the couch. Motors were connected to the Advance Motion Controls (AMC) amplifiers: for the vertical motion, motor AMC-20A20-INV amplifier was used, and two AMC-Z6A8 amplifiers were applied for the lateral and longitudinal couch motions. The Galil DMC-4133 controller was connected to standard PC computer using USB port. The system had two independent power supplies: Galil PSR-12- 24-12A, 24vdc power supply with diodes for controller and 24vdc motors and amplifiers, and Galil-PS300W72 72vdc power supply for vertical motion. Control algorithms were developed for position and velocity adjustment. The system was tested for real-time tracking in the range of 50mm in all 3 directions (superior-inferior, lateral, anterior- posterior). Accuracies were 0.15, 0.20, and 0.18mm, respectively. Repeatability of the desired motion was within ± 0.2mm. Experimental results of couch tracking show feasibility of real-time tumor tracking with high level of accuracy (within sub-millimeter range). This tracking technique potentially offers a simple and effective method to minimize healthy tissues irradiation.Acknowledgement: Study supported by Elekta,Ltd. Study supported by Elekta, Ltd. © 2012 American Association of Physicists in Medicine.
NASA Technical Reports Server (NTRS)
Hebert, Paul; Ma, Jeremy; Borders, James; Aydemir, Alper; Bajracharya, Max; Hudson, Nicolas; Shankar, Krishna; Karumanchi, Sisir; Douillard, Bertrand; Burdick, Joel
2015-01-01
The use of the cognitive capabilties of humans to help guide the autonomy of robotics platforms in what is typically called "supervised-autonomy" is becoming more commonplace in robotics research. The work discussed in this paper presents an approach to a human-in-the-loop mode of robot operation that integrates high level human cognition and commanding with the intelligence and processing power of autonomous systems. Our framework for a "Supervised Remote Robot with Guided Autonomy and Teleoperation" (SURROGATE) is demonstrated on a robotic platform consisting of a pan-tilt perception head, two 7-DOF arms connected by a single 7-DOF torso, mounted on a tracked-wheel base. We present an architecture that allows high-level supervisory commands and intents to be specified by a user that are then interpreted by the robotic system to perform whole body manipulation tasks autonomously. We use a concept of "behaviors" to chain together sequences of "actions" for the robot to perform which is then executed real time.
Robust tuning of robot control systems
NASA Technical Reports Server (NTRS)
Minis, I.; Uebel, M.
1992-01-01
The computed torque control problem is examined for a robot arm with flexible, geared, joint drive systems which are typical in many industrial robots. The standard computed torque algorithm is not directly applicable to this class of manipulators because of the dynamics introduced by the joint drive system. The proposed approach to computed torque control combines a computed torque algorithm with torque controller at each joint. Three such control schemes are proposed. The first scheme uses the joint torque control system currently implemented on the robot arm and a novel form of the computed torque algorithm. The other two use the standard computed torque algorithm and a novel model following torque control system based on model following techniques. Standard tasks and performance indices are used to evaluate the performance of the controllers. Both numerical simulations and experiments are used in evaluation. The study shows that all three proposed systems lead to improved tracking performance over a conventional PD controller.
Rodríguez-Canosa, Gonzalo; Giner, Jaime del Cerro; Barrientos, Antonio
2014-01-01
The detection and tracking of mobile objects (DATMO) is progressively gaining importance for security and surveillance applications. This article proposes a set of new algorithms and procedures for detecting and tracking mobile objects by robots that work collaboratively as part of a multirobot system. These surveillance algorithms are conceived of to work with data provided by long distance range sensors and are intended for highly reliable object detection in wide outdoor environments. Contrary to most common approaches, in which detection and tracking are done by an integrated procedure, the approach proposed here relies on a modular structure, in which detection and tracking are carried out independently, and the latter might accept input data from different detection algorithms. Two movement detection algorithms have been developed for the detection of dynamic objects by using both static and/or mobile robots. The solution to the overall problem is based on the use of a Kalman filter to predict the next state of each tracked object. Additionally, new tracking algorithms capable of combining dynamic objects lists coming from either one or various sources complete the solution. The complementary performance of the separated modular structure for detection and identification is evaluated and, finally, a selection of test examples discussed. PMID:24526305
Virtual Passive Controller for Robot Systems Using Joint Torque Sensors
NASA Technical Reports Server (NTRS)
Aldridge, Hal A.; Juang, Jer-Nan
1997-01-01
This paper presents a control method based on virtual passive dynamic control that will stabilize a robot manipulator using joint torque sensors and a simple joint model. The method does not require joint position or velocity feedback for stabilization. The proposed control method is stable in the sense of Lyaponov. The control method was implemented on several joints of a laboratory robot. The controller showed good stability robustness to system parameter error and to the exclusion of nonlinear dynamic effects on the joints. The controller enhanced position tracking performance and, in the absence of position control, dissipated joint energy.
Yoo, Sung Jin; Park, Jin Bae; Choi, Yoon Ho
2006-12-01
A new method for the robust control of flexible-joint (FJ) robots with model uncertainties in both robot dynamics and actuator dynamics is proposed. The proposed control system is a combination of the adaptive dynamic surface control (DSC) technique and the self-recurrent wavelet neural network (SRWNN). The adaptive DSC technique provides the ability to overcome the "explosion of complexity" problem in backstepping controllers. The SRWNNs are used to observe the arbitrary model uncertainties of FJ robots, and all their weights are trained online. From the Lyapunov stability analysis, their adaptation laws are induced, and the uniformly ultimately boundedness of all signals in a closed-loop adaptive system is proved. Finally, simulation results for a three-link FJ robot are utilized to validate the good position tracking performance and robustness against payload uncertainties and external disturbances of the proposed control system.
Sliding Mode Control (SMC) of Robot Manipulator via Intelligent Controllers
NASA Astrophysics Data System (ADS)
Kapoor, Neha; Ohri, Jyoti
2017-02-01
Inspite of so much research, key technical problem, naming chattering of conventional, simple and robust SMC is still a challenge to the researchers and hence limits its practical application. However, newly developed soft computing based techniques can provide solution. In order to have advantages of conventional and heuristic soft computing based control techniques, in this paper various commonly used intelligent techniques, neural network, fuzzy logic and adaptive neuro fuzzy inference system (ANFIS) have been combined with sliding mode controller (SMC). For validation, proposed hybrid control schemes have been implemented for tracking a predefined trajectory by robotic manipulator, incorporating structured and unstructured uncertainties in the system. After reviewing numerous papers, all the commonly occurring uncertainties like continuous disturbance, uniform random white noise, static friction like coulomb friction and viscous friction, dynamic friction like Dhal friction and LuGre friction have been inserted in the system. Various performance indices like norm of tracking error, chattering in control input, norm of input torque, disturbance rejection, chattering rejection have been used. Comparative results show that with almost eliminated chattering the intelligent SMC controllers are found to be more efficient over simple SMC. It has also been observed from results that ANFIS based controller has the best tracking performance with the reduced burden on the system. No paper in the literature has found to have all these structured and unstructured uncertainties together for motion control of robotic manipulator.
Robot Teleoperation and Perception Assistance with a Virtual Holographic Display
NASA Technical Reports Server (NTRS)
Goddard, Charles O.
2012-01-01
Teleoperation of robots in space from Earth has historically been dfficult. Speed of light delays make direct joystick-type control infeasible, so it is desirable to command a robot in a very high-level fashion. However, in order to provide such an interface, knowledge of what objects are in the robot's environment and how they can be interacted with is required. In addition, many tasks that would be desirable to perform are highly spatial, requiring some form of six degree of freedom input. These two issues can be combined, allowing the user to assist the robot's perception by identifying the locations of objects in the scene. The zSpace system, a virtual holographic environment, provides a virtual three-dimensional space superimposed over real space and a stylus tracking position and rotation inside of it. Using this system, a possible interface for this sort of robot control is proposed.
Future of robotic surgery in urology.
Rassweiler, Jens J; Autorino, Riccardo; Klein, Jan; Mottrie, Alex; Goezen, Ali Serdar; Stolzenburg, Jens-Uwe; Rha, Koon H; Schurr, Marc; Kaouk, Jihad; Patel, Vipul; Dasgupta, Prokar; Liatsikos, Evangelos
2017-12-01
To provide a comprehensive overview of the current status of the field of robotic systems for urological surgery and discuss future perspectives. A non-systematic literature review was performed using PubMed/Medline search electronic engines. Existing patents for robotic devices were researched using the Google search engine. Findings were also critically analysed taking into account the personal experience of the authors. The relevant patents for the first generation of the da Vinci platform will expire in 2019. New robotic systems are coming onto the stage. These can be classified according to type of console, arrangement of robotic arms, handles and instruments, and other specific features (haptic feedback, eye-tracking). The Telelap ALF-X robot uses an open console with eye-tracking, laparoscopy-like handles with haptic feedback, and arms mounted on separate carts; first clinical trials with this system were reported in 2016. The Medtronic robot provides an open console using three-dimensional high-definition video technology and three arms. The Avatera robot features a closed console with microscope-like oculars, four arms arranged on one cart, and 5-mm instruments with six degrees of freedom. The REVO-I consists of an open console and a four-arm arrangement on one cart; the first experiments with this system were published in 2016. Medicaroid uses a semi-open console and three robot arms attached to the operating table. Clinical trials of the SP 1098-platform using the da Vinci Xi for console-based single-port surgery were reported in 2015. The SPORT robot has been tested in animal experiments for single-port surgery. The SurgiBot represents a bedside solution for single-port surgery providing flexible tube-guided instruments. The Avicenna Roboflex has been developed for robotic flexible ureteroscopy, with promising early clinical results. Several console-based robots for laparoscopic multi- and single-port surgery are expected to come to market within the next 5 years. Future developments in the field of robotic surgery are likely to focus on the specific features of robotic arms, instruments, console, and video technology. The high technical standards of four da Vinci generations have set a high bar for upcoming devices. Ultimately, the implementation of these upcoming systems will depend on their clinical applicability and costs. How these technical developments will facilitate surgery and whether their use will translate into better outcomes for our patients remains to be determined. © 2017 The Authors BJU International © 2017 BJU International Published by John Wiley & Sons Ltd.
Optimized Assistive Human-Robot Interaction Using Reinforcement Learning.
Modares, Hamidreza; Ranatunga, Isura; Lewis, Frank L; Popa, Dan O
2016-03-01
An intelligent human-robot interaction (HRI) system with adjustable robot behavior is presented. The proposed HRI system assists the human operator to perform a given task with minimum workload demands and optimizes the overall human-robot system performance. Motivated by human factor studies, the presented control structure consists of two control loops. First, a robot-specific neuro-adaptive controller is designed in the inner loop to make the unknown nonlinear robot behave like a prescribed robot impedance model as perceived by a human operator. In contrast to existing neural network and adaptive impedance-based control methods, no information of the task performance or the prescribed robot impedance model parameters is required in the inner loop. Then, a task-specific outer-loop controller is designed to find the optimal parameters of the prescribed robot impedance model to adjust the robot's dynamics to the operator skills and minimize the tracking error. The outer loop includes the human operator, the robot, and the task performance details. The problem of finding the optimal parameters of the prescribed robot impedance model is transformed into a linear quadratic regulator (LQR) problem which minimizes the human effort and optimizes the closed-loop behavior of the HRI system for a given task. To obviate the requirement of the knowledge of the human model, integral reinforcement learning is used to solve the given LQR problem. Simulation results on an x - y table and a robot arm, and experimental implementation results on a PR2 robot confirm the suitability of the proposed method.
Marescaux, Jacques; Solerc, Luc
2004-06-01
Medical image processing leads to an improvement in patient care by guiding the surgical gesture. Three-dimensional models of patients that are generated from computed tomographic scans or magnetic resonance imaging allow improved surgical planning and surgical simulation that offers the opportunity for a surgeon to train the surgical gesture before performing it for real. These two preoperative steps can be used intra-operatively because of the development of augmented reality, which consists of superimposing the preoperative three-dimensional model of the patient onto the real intraoperative view. Augmented reality provides the surgeon with a view of the patient in transparency and can also guide the surgeon, thanks to the real-time tracking of surgical tools during the procedure. When adapted to robotic surgery, this tool tracking enables visual serving with the ability to automatically position and control surgical robotic arms in three dimensions. It is also now possible to filter physiologic movements such as breathing or the heart beat. In the future, by combining augmented reality and robotics, these image-guided robotic systems will enable automation of the surgical procedure, which will be the next revolution in surgery.
TH-AB-202-11: Spatial and Rotational Quality Assurance of 6DOF Patient Tracking Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Belcher, AH; Liu, X; Grelewicz, Z
2016-06-15
Purpose: External tracking systems used for patient positioning and motion monitoring during radiotherapy are now capable of detecting both translations and rotations (6DOF). In this work, we develop a novel technique to evaluate the 6DOF performance of external motion tracking systems. We apply this methodology to an infrared (IR) marker tracking system and two 3D optical surface mapping systems in a common tumor 6DOF workspace. Methods: An in-house designed and built 6DOF parallel kinematics robotic motion phantom was used to follow input trajectories with sub-millimeter and sub-degree accuracy. The 6DOF positions of the robotic system were then tracked and recordedmore » independently by three optical camera systems. A calibration methodology which associates the motion phantom and camera coordinate frames was first employed, followed by a comprehensive 6DOF trajectory evaluation, which spanned a full range of positions and orientations in a 20×20×16 mm and 5×5×5 degree workspace. The intended input motions were compared to the calibrated 6DOF measured points. Results: The technique found the accuracy of the IR marker tracking system to have maximal root mean square error (RMSE) values of 0.25 mm translationally and 0.09 degrees rotationally, in any one axis, comparing intended 6DOF positions to positions measured by the IR camera. The 6DOF RSME discrepancy for the first 3D optical surface tracking unit yielded maximal values of 0.60 mm and 0.11 degrees over the same 6DOF volume. An earlier generation 3D optical surface tracker was observed to have worse tracking capabilities than both the IR camera unit and the newer 3D surface tracking system with maximal RMSE of 0.74 mm and 0.28 degrees within the same 6DOF evaluation space. Conclusion: The proposed technique was effective at evaluating the performance of 6DOF patient tracking systems. All systems examined exhibited tracking capabilities at the sub-millimeter and sub-degree level within a 6DOF workspace.« less
Adaptive Tracking Control for Robots With an Interneural Computing Scheme.
Tsai, Feng-Sheng; Hsu, Sheng-Yi; Shih, Mau-Hsiang
2018-04-01
Adaptive tracking control of mobile robots requires the ability to follow a trajectory generated by a moving target. The conventional analysis of adaptive tracking uses energy minimization to study the convergence and robustness of the tracking error when the mobile robot follows a desired trajectory. However, in the case that the moving target generates trajectories with uncertainties, a common Lyapunov-like function for energy minimization may be extremely difficult to determine. Here, to solve the adaptive tracking problem with uncertainties, we wish to implement an interneural computing scheme in the design of a mobile robot for behavior-based navigation. The behavior-based navigation adopts an adaptive plan of behavior patterns learning from the uncertainties of the environment. The characteristic feature of the interneural computing scheme is the use of neural path pruning with rewards and punishment interacting with the environment. On this basis, the mobile robot can be exploited to change its coupling weights in paths of neural connections systematically, which can then inhibit or enhance the effect of flow elimination in the dynamics of the evolutionary neural network. Such dynamical flow translation ultimately leads to robust sensory-to-motor transformations adapting to the uncertainties of the environment. A simulation result shows that the mobile robot with the interneural computing scheme can perform fault-tolerant behavior of tracking by maintaining suitable behavior patterns at high frequency levels.
SU-E-T-570: New Quality Assurance Method Using Motion Tracking for 6D Robotic Couches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cheon, W; Cho, J; Ahn, S
Purpose: To accommodate geometrically accurate patient positioning, a robotic couch that is capable of 6-degrees of freedom has been introduced. However, conventional couch QA methods are not sufficient to enable the necessary accuracy of tests. Therefore, we have developed a camera based motion detection and geometry calibration system for couch QA. Methods: Employing a Visual-Tracking System (VTS, BonitaB10, Vicon, UK) which tracks infrared reflective(IR) markers, camera calibration was conducted using a 5.7 × 5.7 × 5.7 cm{sup 3} cube attached with IR markers at each corner. After positioning a robotic-couch at the origin with the cube on the table top,more » 3D coordinates of the cube’s eight corners were acquired by VTS in the VTS coordinate system. Next, positions in reference coordinates (roomcoordinates) were assigned using the known relation between each point. Finally, camera calibration was completed by finding a transformation matrix between VTS and reference coordinate systems and by applying a pseudo inverse matrix method. After the calibration, the accuracy of linear and rotational motions as well as couch sagging could be measured by analyzing the continuously acquired data of the cube while the couch moves to a designated position. Accuracy of the developed software was verified through comparison with measurement data when using a Laser tracker (FARO, Lake Mary, USA) for a robotic-couch installed for proton therapy. Results: VTS system could track couch motion accurately and measured position in room-coordinates. The VTS measurements and Laser tracker data agreed within 1% of difference for linear and rotational motions. Also because the program analyzes motion in 3-Dimension, it can compute couch sagging. Conclusion: Developed QA system provides submillimeter/ degree accuracy which fulfills the high-end couch QA. This work was supported by the National Research Foundation of Korea funded by Ministry of Science, ICT & Future Planning. (2013M2A2A7043507 and 2012M3A9B6055201)« less
2012-01-01
Background Novel stroke rehabilitation techniques that employ electrical stimulation (ES) and robotic technologies are effective in reducing upper limb impairments. ES is most effective when it is applied to support the patients’ voluntary effort; however, current systems fail to fully exploit this connection. This study builds on previous work using advanced ES controllers, and aims to investigate the feasibility of Stimulation Assistance through Iterative Learning (SAIL), a novel upper limb stroke rehabilitation system which utilises robotic support, ES, and voluntary effort. Methods Five hemiparetic, chronic stroke participants with impaired upper limb function attended 18, 1 hour intervention sessions. Participants completed virtual reality tracking tasks whereby they moved their impaired arm to follow a slowly moving sphere along a specified trajectory. To do this, the participants’ arm was supported by a robot. ES, mediated by advanced iterative learning control (ILC) algorithms, was applied to the triceps and anterior deltoid muscles. Each movement was repeated 6 times and ILC adjusted the amount of stimulation applied on each trial to improve accuracy and maximise voluntary effort. Participants completed clinical assessments (Fugl-Meyer, Action Research Arm Test) at baseline and post-intervention, as well as unassisted tracking tasks at the beginning and end of each intervention session. Data were analysed using t-tests and linear regression. Results From baseline to post-intervention, Fugl-Meyer scores improved, assisted and unassisted tracking performance improved, and the amount of ES required to assist tracking reduced. Conclusions The concept of minimising support from ES using ILC algorithms was demonstrated. The positive results are promising with respect to reducing upper limb impairments following stroke, however, a larger study is required to confirm this. PMID:22676920
Meadmore, Katie L; Hughes, Ann-Marie; Freeman, Chris T; Cai, Zhonglun; Tong, Daisy; Burridge, Jane H; Rogers, Eric
2012-06-07
Novel stroke rehabilitation techniques that employ electrical stimulation (ES) and robotic technologies are effective in reducing upper limb impairments. ES is most effective when it is applied to support the patients' voluntary effort; however, current systems fail to fully exploit this connection. This study builds on previous work using advanced ES controllers, and aims to investigate the feasibility of Stimulation Assistance through Iterative Learning (SAIL), a novel upper limb stroke rehabilitation system which utilises robotic support, ES, and voluntary effort. Five hemiparetic, chronic stroke participants with impaired upper limb function attended 18, 1 hour intervention sessions. Participants completed virtual reality tracking tasks whereby they moved their impaired arm to follow a slowly moving sphere along a specified trajectory. To do this, the participants' arm was supported by a robot. ES, mediated by advanced iterative learning control (ILC) algorithms, was applied to the triceps and anterior deltoid muscles. Each movement was repeated 6 times and ILC adjusted the amount of stimulation applied on each trial to improve accuracy and maximise voluntary effort. Participants completed clinical assessments (Fugl-Meyer, Action Research Arm Test) at baseline and post-intervention, as well as unassisted tracking tasks at the beginning and end of each intervention session. Data were analysed using t-tests and linear regression. From baseline to post-intervention, Fugl-Meyer scores improved, assisted and unassisted tracking performance improved, and the amount of ES required to assist tracking reduced. The concept of minimising support from ES using ILC algorithms was demonstrated. The positive results are promising with respect to reducing upper limb impairments following stroke, however, a larger study is required to confirm this.
Klancar, Gregor; Kristan, Matej; Kovacic, Stanislav; Orqueda, Omar
2004-07-01
In this paper a global vision scheme for estimation of positions and orientations of mobile robots is presented. It is applied to robot soccer application which is a fast dynamic game and therefore needs an efficient and robust vision system implemented. General applicability of the vision system can be found in other robot applications such as mobile transport robots in production, warehouses, attendant robots, fast vision tracking of targets of interest and entertainment robotics. Basic operation of the vision system is divided into two steps. In the first, the incoming image is scanned and pixels are classified into a finite number of classes. At the same time, a segmentation algorithm is used to find corresponding regions belonging to one of the classes. In the second step, all the regions are examined. Selection of the ones that are a part of the observed object is made by means of simple logic procedures. The novelty is focused on optimization of the processing time needed to finish the estimation of possible object positions. Better results of the vision system are achieved by implementing camera calibration and shading correction algorithm. The former corrects camera lens distortion, while the latter increases robustness to irregular illumination conditions.
Watching elderly and disabled person's physical condition by remotely controlled monorail robot
NASA Astrophysics Data System (ADS)
Nagasaka, Yasunori; Matsumoto, Yoshinori; Fukaya, Yasutoshi; Takahashi, Tomoichi; Takeshita, Toru
2001-10-01
We are developing a nursing system using robots and cameras. The cameras are mounted on a remote controlled monorail robot which moves inside a room and watches the elderly. It is necessary to pay attention to the elderly at home or nursing homes all time. This requires staffs to pay attention to them at every time. The purpose of our system is to help those staffs. This study intends to improve such situation. A host computer controls a monorail robot to go in front of the elderly using the images taken by cameras on the ceiling. A CCD camera is mounted on the monorail robot to take pictures of their facial expression or movements. The robot sends the images to a host computer that checks them whether something unusual happens or not. We propose a simple calibration method for positioning the monorail robots to track the moves of the elderly for keeping their faces at center of camera view. We built a small experiment system, and evaluated our camera calibration method and image processing algorithm.
Robust adaptive uniform exact tracking control for uncertain Euler-Lagrange system
NASA Astrophysics Data System (ADS)
Yang, Yana; Hua, Changchun; Li, Junpeng; Guan, Xinping
2017-12-01
This paper offers a solution to the robust adaptive uniform exact tracking control for uncertain nonlinear Euler-Lagrange (EL) system. An adaptive finite-time tracking control algorithm is designed by proposing a novel nonsingular integral terminal sliding-mode surface. Moreover, a new adaptive parameter tuning law is also developed by making good use of the system tracking errors and the adaptive parameter estimation errors. Thus, both the trajectory tracking and the parameter estimation can be achieved in a guaranteed time adjusted arbitrarily based on practical demands, simultaneously. Additionally, the control result for the EL system proposed in this paper can be extended to high-order nonlinear systems easily. Finally, a test-bed 2-DOF robot arm is set-up to demonstrate the performance of the new control algorithm.
CT fluoroscopy-guided robotically-assisted lung biopsy
NASA Astrophysics Data System (ADS)
Xu, Sheng; Fichtinger, Gabor; Taylor, Russell H.; Banovac, Filip; Cleary, Kevin
2006-03-01
Lung biopsy is a common interventional radiology procedure. One of the difficulties in performing the lung biopsy is that lesions move with respiration. This paper presents a new robotically assisted lung biopsy system for CT fluoroscopy that can automatically compensate for the respiratory motion during the intervention. The system consists of a needle placement robot to hold the needle on the CT scan plane, a radiolucent Z-frame for registration of the CT and robot coordinate systems, and a frame grabber to obtain the CT fluoroscopy image in real-time. The CT fluoroscopy images are used to noninvasively track the motion of a pulmonary lesion in real-time. The position of the lesion in the images is automatically determined by the image processing software and the motion of the robot is controlled to compensate for the lesion motion. The system was validated under CT fluoroscopy using a respiratory motion simulator. A swine study was also done to show the feasibility of the technique in a respiring animal.
Chan, Joshua L; Mazilu, Dumitru; Miller, Justin G; Hunt, Timothy; Horvath, Keith A; Li, Ming
2016-10-01
Real-time magnetic resonance imaging (rtMRI) guidance provides significant advantages during transcatheter aortic valve replacement (TAVR) as it provides superior real-time visualization and accurate device delivery tracking. However, performing a TAVR within an MRI scanner remains difficult due to a constrained procedural environment. To address these concerns, a magnetic resonance (MR)-compatible robotic system to assist in TAVR deployments was developed. This study evaluates the technical design and interface considerations of an MR-compatible robotic-assisted TAVR system with the purpose of demonstrating that such a system can be developed and executed safely and precisely in a preclinical model. An MR-compatible robotic surgical assistant system was built for TAVR deployment. This system integrates a 5-degrees of freedom (DoF) robotic arm with a 3-DoF robotic valve delivery module. A user interface system was designed for procedural planning and real-time intraoperative manipulation of the robot. The robotic device was constructed of plastic materials, pneumatic actuators, and fiber-optical encoders. The mechanical profile and MR compatibility of the robotic system were evaluated. The system-level error based on a phantom model was 1.14 ± 0.33 mm. A self-expanding prosthesis was successfully deployed in eight Yorkshire swine under rtMRI guidance. Post-deployment imaging and necropsy confirmed placement of the stent within 3 mm of the aortic valve annulus. These phantom and in vivo studies demonstrate the feasibility and advantages of robotic-assisted TAVR under rtMRI guidance. This robotic system increases the precision of valve deployments, diminishes environmental constraints, and improves the overall success of TAVR.
Sampson, Patrica; Freeman, Chris; Coote, Susan; Demain, Sara; Feys, Peter; Meadmore, Katie; Hughes, Ann-Marie
2016-02-01
Few interventions address multiple sclerosis (MS) arm dysfunction but robotics and functional electrical stimulation (FES) appear promising. This paper investigates the feasibility of combining FES with passive robotic support during virtual reality (VR) training tasks to improve upper limb function in people with multiple sclerosis (pwMS). The system assists patients in following a specified trajectory path, employing an advanced model-based paradigm termed iterative learning control (ILC) to adjust the FES to improve accuracy and maximise voluntary effort. Reaching tasks were repeated six times with ILC learning the optimum control action from previous attempts. A convenience sample of five pwMS was recruited from local MS societies, and the intervention comprised 18 one-hour training sessions over 10 weeks. The accuracy of tracking performance without FES and the amount of FES delivered during training were analyzed using regression analysis. Clinical functioning of the arm was documented before and after treatment with standard tests. Statistically significant results following training included: improved accuracy of tracking performance both when assisted and unassisted by FES; reduction in maximum amount of FES needed to assist tracking; and less impairment in the proximal arm that was trained. The system was well tolerated by all participants with no increase in muscle fatigue reported. This study confirms the feasibility of FES combined with passive robot assistance as a potentially effective intervention to improve arm movement and control in pwMS and provides the basis for a follow-up study.
Yang, Dan; Xu, Bin; Rao, Kaiyou; Sheng, Weihua
2018-01-24
Indoor occupants' positions are significant for smart home service systems, which usually consist of robot service(s), appliance control and other intelligent applications. In this paper, an innovative localization method is proposed for tracking humans' position in indoor environments based on passive infrared (PIR) sensors using an accessibility map and an A-star algorithm, aiming at providing intelligent services. First the accessibility map reflecting the visiting habits of the occupants is established through the integral training with indoor environments and other prior knowledge. Then the PIR sensors, which placement depends on the training results in the accessibility map, get the rough location information. For more precise positioning, the A-start algorithm is used to refine the localization, fused with the accessibility map and the PIR sensor data. Experiments were conducted in a mock apartment testbed. The ground truth data was obtained from an Opti-track system. The results demonstrate that the proposed method is able to track persons in a smart home environment and provide a solution for home robot localization.
Yang, Dan; Xu, Bin; Rao, Kaiyou; Sheng, Weihua
2018-01-01
Indoor occupants’ positions are significant for smart home service systems, which usually consist of robot service(s), appliance control and other intelligent applications. In this paper, an innovative localization method is proposed for tracking humans’ position in indoor environments based on passive infrared (PIR) sensors using an accessibility map and an A-star algorithm, aiming at providing intelligent services. First the accessibility map reflecting the visiting habits of the occupants is established through the integral training with indoor environments and other prior knowledge. Then the PIR sensors, which placement depends on the training results in the accessibility map, get the rough location information. For more precise positioning, the A-start algorithm is used to refine the localization, fused with the accessibility map and the PIR sensor data. Experiments were conducted in a mock apartment testbed. The ground truth data was obtained from an Opti-track system. The results demonstrate that the proposed method is able to track persons in a smart home environment and provide a solution for home robot localization. PMID:29364188
Digital-Electronic/Optical Apparatus Would Recognize Targets
NASA Technical Reports Server (NTRS)
Scholl, Marija S.
1994-01-01
Proposed automatic target-recognition apparatus consists mostly of digital-electronic/optical cross-correlator that processes infrared images of targets. Infrared images of unknown targets correlated quickly with images of known targets. Apparatus incorporates some features of correlator described in "Prototype Optical Correlator for Robotic Vision System" (NPO-18451), and some of correlator described in "Compact Optical Correlator" (NPO-18473). Useful in robotic system; to recognize and track infrared-emitting, moving objects as variously shaped hot workpieces on conveyor belt.
Robotic ICSI (intracytoplasmic sperm injection).
Lu, Zhe; Zhang, Xuping; Leung, Clement; Esfandiari, Navid; Casper, Robert F; Sun, Yu
2011-07-01
This paper is the first report of robotic intracytoplasmic sperm injection (ICSI). ICSI is a clinical procedure performed worldwide in fertility clinics, requiring pick-up of a single sperm and insertion of it into an oocyte (i.e., egg cell). Since its invention 20 years ago, ICSI has been conducted manually by a handful of highly skilled embryologists; however, success rates vary significantly among clinics due to poor reproducibility and inconsistency across operators. We leverage our work in robotic cell injection to realize robotic ICSI and aim ultimately, to standardize how clinical ICSI is performed. This paper presents some of the technical aspects of our robotic ICSI system, including a cell holding device, motion control, and computer vision algorithms. The system performs visual tracking of single sperm, robotic immobilization of sperm, aspiration of sperm with picoliter volume, and insertion of sperm into an oocyte with a high degree of reproducibility. The system requires minimal human involvement (requiring only a few computer mouse clicks), and is human operator skill independent. Using the hamster oocyte-human sperm model in preliminary trials, the robotic system demonstrated a high success rate of 90.0% and survival rate of 90.7% (n=120). © 2011 IEEE
Resquín, Francisco; Gonzalez-Vargas, Jose; Ibáñez, Jaime; Brunetti, Fernando; Pons, José Luis
2016-01-01
Hybrid robotic systems represent a novel research field, where functional electrical stimulation (FES) is combined with a robotic device for rehabilitation of motor impairment. Under this approach, the design of robust FES controllers still remains an open challenge. In this work, we aimed at developing a learning FES controller to assist in the performance of reaching movements in a simple hybrid robotic system setting. We implemented a Feedback Error Learning (FEL) control strategy consisting of a feedback PID controller and a feedforward controller based on a neural network. A passive exoskeleton complemented the FES controller by compensating the effects of gravity. We carried out experiments with healthy subjects to validate the performance of the system. Results show that the FEL control strategy is able to adjust the FES intensity to track the desired trajectory accurately without the need of a previous mathematical model. PMID:27990245
Autonomous mobile robotic system for supporting counterterrorist and surveillance operations
NASA Astrophysics Data System (ADS)
Adamczyk, Marek; Bulandra, Kazimierz; Moczulski, Wojciech
2017-10-01
Contemporary research on mobile robots concerns applications to counterterrorist and surveillance operations. The goal is to develop systems that are capable of supporting the police and special forces by carrying out such operations. The paper deals with a dedicated robotic system for surveillance of large objects such as airports, factories, military bases, and many others. The goal is to trace unauthorised persons who try to enter to the guarded area, document the intrusion and report it to the surveillance centre, and then warn the intruder by sound messages and eventually subdue him/her by stunning through acoustic effect of great power. The system consists of several parts. An armoured four-wheeled robot assures required mobility of the system. The robot is equipped with a set of sensors including 3D mapping system, IR and video cameras, and microphones. It communicates with the central control station (CCS) by means of a wideband wireless encrypted system. A control system of the robot can operate autonomously, and under remote control. In the autonomous mode the robot follows the path planned by the CCS. Once an intruder has been detected, the robot can adopt its plan to allow tracking him/her. Furthermore, special procedures of treatment of the intruder are applied including warning about the breach of the border of the protected area, and incapacitation of an appropriately selected very loud sound until a patrol of guards arrives. Once getting stuck the robot can contact the operator who can remotely solve the problem the robot is faced with.
A Scalable Distributed Approach to Mobile Robot Vision
NASA Technical Reports Server (NTRS)
Kuipers, Benjamin; Browning, Robert L.; Gribble, William S.
1997-01-01
This paper documents our progress during the first year of work on our original proposal entitled 'A Scalable Distributed Approach to Mobile Robot Vision'. We are pursuing a strategy for real-time visual identification and tracking of complex objects which does not rely on specialized image-processing hardware. In this system perceptual schemas represent objects as a graph of primitive features. Distributed software agents identify and track these features, using variable-geometry image subwindows of limited size. Active control of imaging parameters and selective processing makes simultaneous real-time tracking of many primitive features tractable. Perceptual schemas operate independently from the tracking of primitive features, so that real-time tracking of a set of image features is not hurt by latency in recognition of the object that those features make up. The architecture allows semantically significant features to be tracked with limited expenditure of computational resources, and allows the visual computation to be distributed across a network of processors. Early experiments are described which demonstrate the usefulness of this formulation, followed by a brief overview of our more recent progress (after the first year).
An intelligent, free-flying robot
NASA Technical Reports Server (NTRS)
Reuter, G. J.; Hess, C. W.; Rhoades, D. E.; Mcfadin, L. W.; Healey, K. J.; Erickson, J. D.
1988-01-01
The ground-based demonstration of EVA Retriever, a voice-supervised, intelligent, free-flying robot, is designed to evaluate the capability to retrieve objects (astronauts, equipment, and tools) which have accidentally separated from the Space Station. The major objective of the EVA Retriever Project is to design, develop, and evaluate an integrated robotic hardware and on-board software system which autonomously: (1) performs system activation and check-out, (2) searches for and acquires the target, (3) plans and executes a rendezvous while continuously tracking the target, (4) avoids stationary and moving obstacles, (5) reaches for and grapples the target, (6) returns to transfer the object, and (7) returns to base.
An intelligent, free-flying robot
NASA Technical Reports Server (NTRS)
Reuter, G. J.; Hess, C. W.; Rhoades, D. E.; Mcfadin, L. W.; Healey, K. J.; Erickson, J. D.; Phinney, Dale E.
1989-01-01
The ground based demonstration of the extensive extravehicular activity (EVA) Retriever, a voice-supervised, intelligent, free flying robot, is designed to evaluate the capability to retrieve objects (astronauts, equipment, and tools) which have accidentally separated from the Space Station. The major objective of the EVA Retriever Project is to design, develop, and evaluate an integrated robotic hardware and on-board software system which autonomously: (1) performs system activation and check-out; (2) searches for and acquires the target; (3) plans and executes a rendezvous while continuously tracking the target; (4) avoids stationary and moving obstacles; (5) reaches for and grapples the target; (6) returns to transfer the object; and (7) returns to base.
A laser tracking dynamic robot metrology instrument
NASA Technical Reports Server (NTRS)
Parker, G. A.; Mayer, J. R. R.
1989-01-01
Research work over several years has resulted in the development of a laser tracking instrument capable of dynamic 3-D measurements of robot end-effector trajectories. The instrument characteristics and experiments to measure the static and dynamic performance of a robot in an industrial manufacturing environment are described. The use of this technology for space applications is examined.
An open-source framework for testing tracking devices using Lego Mindstorms
NASA Astrophysics Data System (ADS)
Jomier, Julien; Ibanez, Luis; Enquobahrie, Andinet; Pace, Danielle; Cleary, Kevin
2009-02-01
In this paper, we present an open-source framework for testing tracking devices in surgical navigation applications. At the core of image-guided intervention systems is the tracking interface that handles communication with the tracking device and gathers tracking information. Given that the correctness of tracking information is critical for protecting patient safety and for ensuring the successful execution of an intervention, the tracking software component needs to be thoroughly tested on a regular basis. Furthermore, with widespread use of extreme programming methodology that emphasizes continuous and incremental testing of application components, testing design becomes critical. While it is easy to automate most of the testing process, it is often more difficult to test components that require manual intervention such as tracking device. Our framework consists of a robotic arm built from a set of Lego Mindstorms and an open-source toolkit written in C++ to control the robot movements and assess the accuracy of the tracking devices. The application program interface (API) is cross-platform and runs on Windows, Linux and MacOS. We applied this framework for the continuous testing of the Image-Guided Surgery Toolkit (IGSTK), an open-source toolkit for image-guided surgery and shown that regression testing on tracking devices can be performed at low cost and improve significantly the quality of the software.
NASA Astrophysics Data System (ADS)
Kryuchkov, B. I.; Usov, V. M.; Chertopolokhov, V. A.; Ronzhin, A. L.; Karpov, A. A.
2017-05-01
Extravehicular activity (EVA) on the lunar surface, necessary for the future exploration of the Moon, involves extensive use of robots. One of the factors of safe EVA is a proper interaction between cosmonauts and robots in extreme environments. This requires a simple and natural man-machine interface, e.g. multimodal contactless interface based on recognition of gestures and cosmonaut's poses. When travelling in the "Follow Me" mode (master/slave), a robot uses onboard tools for tracking cosmonaut's position and movements, and on the basis of these data builds its itinerary. The interaction in the system "cosmonaut-robot" on the lunar surface is significantly different from that on the Earth surface. For example, a man, dressed in a space suit, has limited fine motor skills. In addition, EVA is quite tiring for the cosmonauts, and a tired human being less accurately performs movements and often makes mistakes. All this leads to new requirements for the convenient use of the man-machine interface designed for EVA. To improve the reliability and stability of human-robot communication it is necessary to provide options for duplicating commands at the task stages and gesture recognition. New tools and techniques for space missions must be examined at the first stage of works in laboratory conditions, and then in field tests (proof tests at the site of application). The article analyzes the methods of detection and tracking of movements and gesture recognition of the cosmonaut during EVA, which can be used for the design of human-machine interface. A scenario for testing these methods by constructing a virtual environment simulating EVA on the lunar surface is proposed. Simulation involves environment visualization and modeling of the use of the "vision" of the robot to track a moving cosmonaut dressed in a spacesuit.
Chai, Xun; Gao, Feng; Pan, Yang; Qi, Chenkun; Xu, Yilin
2015-04-22
Coordinate identification between vision systems and robots is quite a challenging issue in the field of intelligent robotic applications, involving steps such as perceiving the immediate environment, building the terrain map and planning the locomotion automatically. It is now well established that current identification methods have non-negligible limitations such as a difficult feature matching, the requirement of external tools and the intervention of multiple people. In this paper, we propose a novel methodology to identify the geometric parameters of 3D vision systems mounted on robots without involving other people or additional equipment. In particular, our method focuses on legged robots which have complex body structures and excellent locomotion ability compared to their wheeled/tracked counterparts. The parameters can be identified only by moving robots on a relatively flat ground. Concretely, an estimation approach is provided to calculate the ground plane. In addition, the relationship between the robot and the ground is modeled. The parameters are obtained by formulating the identification problem as an optimization problem. The methodology is integrated on a legged robot called "Octopus", which can traverse through rough terrains with high stability after obtaining the identification parameters of its mounted vision system using the proposed method. Diverse experiments in different environments demonstrate our novel method is accurate and robust.
Magician Simulator. A Realistic Simulator for Heterogenous Teams of Autonomous Robots
2011-01-18
IMU, and LIDAR systems for identifying and tracking mobile OOI at long range (>20m), providing early warnings and allowing neutralization from a... LIDAR and Computer Vision template-based feature tracking approaches. Mapping was solved through Multi-Agent particle-filter based Simultaneous...Locali- zation and Mapping ( SLAM ). Our system contains two maps, a physical map and an influence map (location of hostile OOI, explored and unexplored
Applications of artificial intelligence in safe human-robot interactions.
Najmaei, Nima; Kermani, Mehrdad R
2011-04-01
The integration of industrial robots into the human workspace presents a set of unique challenges. This paper introduces a new sensory system for modeling, tracking, and predicting human motions within a robot workspace. A reactive control scheme to modify a robot's operations for accommodating the presence of the human within the robot workspace is also presented. To this end, a special class of artificial neural networks, namely, self-organizing maps (SOMs), is employed for obtaining a superquadric-based model of the human. The SOM network receives information of the human's footprints from the sensory system and infers necessary data for rendering the human model. The model is then used in order to assess the danger of the robot operations based on the measured as well as predicted human motions. This is followed by the introduction of a new reactive control scheme that results in the least interferences between the human and robot operations. The approach enables the robot to foresee an upcoming danger and take preventive actions before the danger becomes imminent. Simulation and experimental results are presented in order to validate the effectiveness of the proposed method.
An MRI-Compatible Robotic System With Hybrid Tracking for MRI-Guided Prostate Intervention
Krieger, Axel; Iordachita, Iulian I.; Guion, Peter; Singh, Anurag K.; Kaushal, Aradhana; Ménard, Cynthia; Pinto, Peter A.; Camphausen, Kevin; Fichtinger, Gabor
2012-01-01
This paper reports the development, evaluation, and first clinical trials of the access to the prostate tissue (APT) II system—a scanner independent system for magnetic resonance imaging (MRI)-guided transrectal prostate interventions. The system utilizes novel manipulator mechanics employing a steerable needle channel and a novel six degree-of-freedom hybrid tracking method, comprising passive fiducial tracking for initial registration and subsequent incremental motion measurements. Targeting accuracy of the system in prostate phantom experiments and two clinical human-subject procedures is shown to compare favorably with existing systems using passive and active tracking methods. The portable design of the APT II system, using only standard MRI image sequences and minimal custom scanner interfacing, allows the system to be easily used on different MRI scanners. PMID:22009867
LABRADOR: a learning autonomous behavior-based robot for adaptive detection and object retrieval
NASA Astrophysics Data System (ADS)
Yamauchi, Brian; Moseley, Mark; Brookshire, Jonathan
2013-01-01
As part of the TARDEC-funded CANINE (Cooperative Autonomous Navigation in a Networked Environment) Program, iRobot developed LABRADOR (Learning Autonomous Behavior-based Robot for Adaptive Detection and Object Retrieval). LABRADOR was based on the rugged, man-portable, iRobot PackBot unmanned ground vehicle (UGV) equipped with an explosives ordnance disposal (EOD) manipulator arm and a custom gripper. For LABRADOR, we developed a vision-based object learning and recognition system that combined a TLD (track-learn-detect) filter based on object shape features with a color-histogram-based object detector. Our vision system was able to learn in real-time to recognize objects presented to the robot. We also implemented a waypoint navigation system based on fused GPS, IMU (inertial measurement unit), and odometry data. We used this navigation capability to implement autonomous behaviors capable of searching a specified area using a variety of robust coverage strategies - including outward spiral, random bounce, random waypoint, and perimeter following behaviors. While the full system was not integrated in time to compete in the CANINE competition event, we developed useful perception, navigation, and behavior capabilities that may be applied to future autonomous robot systems.
Automating High-Precision X-Ray and Neutron Imaging Applications with Robotics
Hashem, Joseph Anthony; Pryor, Mitch; Landsberger, Sheldon; ...
2017-03-28
Los Alamos National Laboratory and the University of Texas at Austin recently implemented a robotically controlled nondestructive testing (NDT) system for X-ray and neutron imaging. This system is intended to address the need for accurate measurements for a variety of parts and, be able to track measurement geometry at every imaging location, and is designed for high-throughput applications. This system was deployed in a beam port at a nuclear research reactor and in an operational inspection X-ray bay. The nuclear research reactor system consisted of a precision industrial seven-axis robot, 1.1-MW TRIGA research reactor, and a scintillator-mirror-camera-based imaging system. Themore » X-ray bay system incorporated the same robot, a 225-keV microfocus X-ray source, and a custom flat panel digital detector. The robotic positioning arm is programmable and allows imaging in multiple configurations, including planar, cylindrical, as well as other user defined geometries that provide enhanced engineering evaluation capability. The imaging acquisition device is coupled with the robot for automated image acquisition. The robot can achieve target positional repeatability within 17 μm in the 3-D space. Flexible automation with nondestructive imaging saves costs, reduces dosage, adds imaging techniques, and achieves better quality results in less time. Specifics regarding the robotic system and imaging acquisition and evaluation processes are presented. In conclusion, this paper reviews the comprehensive testing and system evaluation to affirm the feasibility of robotic NDT, presents the system configuration, and reviews results for both X-ray and neutron radiography imaging applications.« less
Automating High-Precision X-Ray and Neutron Imaging Applications with Robotics
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hashem, Joseph Anthony; Pryor, Mitch; Landsberger, Sheldon
Los Alamos National Laboratory and the University of Texas at Austin recently implemented a robotically controlled nondestructive testing (NDT) system for X-ray and neutron imaging. This system is intended to address the need for accurate measurements for a variety of parts and, be able to track measurement geometry at every imaging location, and is designed for high-throughput applications. This system was deployed in a beam port at a nuclear research reactor and in an operational inspection X-ray bay. The nuclear research reactor system consisted of a precision industrial seven-axis robot, 1.1-MW TRIGA research reactor, and a scintillator-mirror-camera-based imaging system. Themore » X-ray bay system incorporated the same robot, a 225-keV microfocus X-ray source, and a custom flat panel digital detector. The robotic positioning arm is programmable and allows imaging in multiple configurations, including planar, cylindrical, as well as other user defined geometries that provide enhanced engineering evaluation capability. The imaging acquisition device is coupled with the robot for automated image acquisition. The robot can achieve target positional repeatability within 17 μm in the 3-D space. Flexible automation with nondestructive imaging saves costs, reduces dosage, adds imaging techniques, and achieves better quality results in less time. Specifics regarding the robotic system and imaging acquisition and evaluation processes are presented. In conclusion, this paper reviews the comprehensive testing and system evaluation to affirm the feasibility of robotic NDT, presents the system configuration, and reviews results for both X-ray and neutron radiography imaging applications.« less
Moreno-Valenzuela, Javier; González-Hernández, Luis
2011-01-01
In this paper, a new control algorithm for operational space trajectory tracking control of robot arms is introduced. The new algorithm does not require velocity measurement and is based on (1) a primary controller which incorporates an algorithm to obtain synthesized velocity from joint position measurements and (2) a secondary controller which computes the desired joint acceleration and velocity required to achieve operational space motion control. The theory of singularly perturbed systems is crucial for the analysis of the closed-loop system trajectories. In addition, the practical viability of the proposed algorithm is explored through real-time experiments in a two degrees-of-freedom horizontal planar direct-drive arm. Copyright © 2010 ISA. Published by Elsevier Ltd. All rights reserved.
Sensing human hand motions for controlling dexterous robots
NASA Technical Reports Server (NTRS)
Marcus, Beth A.; Churchill, Philip J.; Little, Arthur D.
1988-01-01
The Dexterous Hand Master (DHM) system is designed to control dexterous robot hands such as the UTAH/MIT and Stanford/JPL hands. It is the first commercially available device which makes it possible to accurately and confortably track the complex motion of the human finger joints. The DHM is adaptable to a wide variety of human hand sizes and shapes, throughout their full range of motion.
Integrated mobile robot control
NASA Astrophysics Data System (ADS)
Amidi, Omead; Thorpe, Chuck E.
1991-03-01
This paper describes the strucwre implementation and operation of a real-time mobile robot controller which integrates capabilities such as: position estimation path specification and hacking human interfaces fast communication and multiple client support The benefits of such high-level capabilities in a low-level controller was shown by its implementation for the Naviab autonomous vehicle. In addition performance results from positioning and tracking systems are reported and analyzed.
Audio-Visual Perception System for a Humanoid Robotic Head
Viciana-Abad, Raquel; Marfil, Rebeca; Perez-Lorenzo, Jose M.; Bandera, Juan P.; Romero-Garces, Adrian; Reche-Lopez, Pedro
2014-01-01
One of the main issues within the field of social robotics is to endow robots with the ability to direct attention to people with whom they are interacting. Different approaches follow bio-inspired mechanisms, merging audio and visual cues to localize a person using multiple sensors. However, most of these fusion mechanisms have been used in fixed systems, such as those used in video-conference rooms, and thus, they may incur difficulties when constrained to the sensors with which a robot can be equipped. Besides, within the scope of interactive autonomous robots, there is a lack in terms of evaluating the benefits of audio-visual attention mechanisms, compared to only audio or visual approaches, in real scenarios. Most of the tests conducted have been within controlled environments, at short distances and/or with off-line performance measurements. With the goal of demonstrating the benefit of fusing sensory information with a Bayes inference for interactive robotics, this paper presents a system for localizing a person by processing visual and audio data. Moreover, the performance of this system is evaluated and compared via considering the technical limitations of unimodal systems. The experiments show the promise of the proposed approach for the proactive detection and tracking of speakers in a human-robot interactive framework. PMID:24878593
Sensor Control of Robot Arc Welding
NASA Technical Reports Server (NTRS)
Sias, F. R., Jr.
1983-01-01
The potential for using computer vision as sensory feedback for robot gas-tungsten arc welding is investigated. The basic parameters that must be controlled while directing the movement of an arc welding torch are defined. The actions of a human welder are examined to aid in determining the sensory information that would permit a robot to make reproducible high strength welds. Special constraints imposed by both robot hardware and software are considered. Several sensory modalities that would potentially improve weld quality are examined. Special emphasis is directed to the use of computer vision for controlling gas-tungsten arc welding. Vendors of available automated seam tracking arc welding systems and of computer vision systems are surveyed. An assessment is made of the state of the art and the problems that must be solved in order to apply computer vision to robot controlled arc welding on the Space Shuttle Main Engine.
Rong, Wei; Li, Waiming; Pang, Mankit; Hu, Junyan; Wei, Xijun; Yang, Bibo; Wai, Honwah; Zheng, Xiaoxiang; Hu, Xiaoling
2017-04-26
It is a challenge to reduce the muscular discoordination in the paretic upper limb after stroke in the traditional rehabilitation programs. In this study, a neuromuscular electrical stimulation (NMES) and robot hybrid system was developed for multi-joint coordinated upper limb physical training. The system could assist the elbow, wrist and fingers to conduct arm reaching out, hand opening/grasping and arm withdrawing by tracking an indicative moving cursor on the screen of a computer, with the support from the joint motors and electrical stimulations on target muscles, under the voluntary intention control by electromyography (EMG). Subjects with chronic stroke (n = 11) were recruited for the investigation on the assistive capability of the NMES-robot and the evaluation of the rehabilitation effectiveness through a 20-session device assisted upper limb training. In the evaluation, the movement accuracy measured by the root mean squared error (RMSE) during the tracking was significantly improved with the support from both the robot and NMES, in comparison with those without the assistance from the system (P < 0.05). The intra-joint and inter-joint muscular co-contractions measured by EMG were significantly released when the NMES was applied to the agonist muscles in the different phases of the limb motion (P < 0.05). After the physical training, significant improvements (P < 0.05) were captured by the clinical scores, i.e., Modified Ashworth Score (MAS, the elbow and the wrist), Fugl-Meyer Assessment (FMA), Action Research Arm Test (ARAT), and Wolf Motor Function Test (WMFT). The EMG-driven NMES-robotic system could improve the muscular coordination at the elbow, wrist and fingers. ClinicalTrials.gov. NCT02117089 ; date of registration: April 10, 2014.
Robotically assisted ureteroscopy for kidney exploration
NASA Astrophysics Data System (ADS)
Talari, Hadi F.; Monfaredi, Reza; Wilson, Emmanuel; Blum, Emily; Bayne, Christopher; Peters, Craig; Zhang, Anlin; Cleary, Kevin
2017-03-01
Ureteroscopy is a minimally invasive procedure for diagnosis and treatment of urinary tract pathology. Ergonomic and visualization challenges as well as radiation exposure are limitations to conventional ureteroscopy. Therefore, we have developed a robotic system to "power drive" a flexible ureteroscope with 3D tip tracking and pre-operative image overlay. The proposed system was evaluated using a kidney phantom registered to pre-operative MR images. Initial experiments show the potential of the device to provide additional assistance, precision, and guidance during urology procedures.
Multi-Robot FastSLAM for Large Domains
2007-03-01
Derr, D. Fox, A.B. Cremers , Integrating global position estimation and position tracking for mobile robots: The dynamic markov localization approach...Intelligence (AAAI), 2000. 53. Andrew J. Davison and David W. Murray. Simultaneous Localization and Map- Building Using Active Vision. IEEE...Wyeth, Michael Milford and David Prasser. A Modified Particle Filter for Simultaneous Robot Localization and Landmark Tracking in an Indoor
Model-based Robotic Dynamic Motion Control for the Robonaut 2 Humanoid Robot
NASA Technical Reports Server (NTRS)
Badger, Julia M.; Hulse, Aaron M.; Taylor, Ross C.; Curtis, Andrew W.; Gooding, Dustin R.; Thackston, Allison
2013-01-01
Robonaut 2 (R2), an upper-body dexterous humanoid robot, has been undergoing experimental trials on board the International Space Station (ISS) for more than a year. R2 will soon be upgraded with two climbing appendages, or legs, as well as a new integrated model-based control system. This control system satisfies two important requirements; first, that the robot can allow humans to enter its workspace during operation and second, that the robot can move its large inertia with enough precision to attach to handrails and seat track while climbing around the ISS. This is achieved by a novel control architecture that features an embedded impedance control law on the motor drivers called Multi-Loop control which is tightly interfaced with a kinematic and dynamic coordinated control system nicknamed RoboDyn that resides on centralized processors. This paper presents the integrated control algorithm as well as several test results that illustrate R2's safety features and performance.
Bekele, Esubalew T; Lahiri, Uttama; Swanson, Amy R.; Crittendon, Julie A.; Warren, Zachary E.; Sarkar, Nilanjan
2013-01-01
Emerging technology, especially robotic technology, has been shown to be appealing to children with autism spectrum disorders (ASD). Such interest may be leveraged to provide repeatable, accurate and individualized intervention services to young children with ASD based on quantitative metrics. However, existing robot-mediated systems tend to have limited adaptive capability that may impact individualization. Our current work seeks to bridge this gap by developing an adaptive and individualized robot-mediated technology for children with ASD. The system is composed of a humanoid robot with its vision augmented by a network of cameras for real-time head tracking using a distributed architecture. Based on the cues from the child’s head movement, the robot intelligently adapts itself in an individualized manner to generate prompts and reinforcements with potential to promote skills in the ASD core deficit area of early social orienting. The system was validated for feasibility, accuracy, and performance. Results from a pilot usability study involving six children with ASD and a control group of six typically developing (TD) children are presented. PMID:23221831
Concept and design philosophy of a person-accompanying robot
NASA Astrophysics Data System (ADS)
Mizoguchi, Hiroshi; Shigehara, Takaomi; Goto, Yoshiyasu; Hidai, Ken-ichi; Mishima, Taketoshi
1999-01-01
This paper proposes a person accompanying robot as a novel human collaborative robot. The person accompanying robot is such legged mobile robot that is possible to follow the person utilizing its vision. towards future aging society, human collaboration and human support are required as novel applications of robots. Such human collaborative robots share the same space with humans. But conventional robots are isolated from humans and lack the capability to observe humans. Study on human observing function of robot is crucial to realize novel robot such as service and pet robot. To collaborate and support humans properly human collaborative robot must have capability to observe and recognize humans. Study on human observing function of robot is crucial to realize novel robot such as service and pet robot. The authors are currently implementing a prototype of the proposed accompanying robot.As a base for the human observing function of the prototype robot, we have realized face tracking utilizing skin color extraction and correlation based tracking. We also develop a method for the robot to pick up human voice clearly and remotely by utilizing microphone arrays. Results of these preliminary study suggest feasibility of the proposed robot.
Percutaneous needle placement using laser guidance: a practical solution
NASA Astrophysics Data System (ADS)
Xu, Sheng; Kapoor, Ankur; Abi-Jaoudeh, Nadine; Imbesi, Kimberly; Hong, Cheng William; Mazilu, Dumitru; Sharma, Karun; Venkatesan, Aradhana M.; Levy, Elliot; Wood, Bradford J.
2013-03-01
In interventional radiology, various navigation technologies have emerged aiming to improve the accuracy of device deployment and potentially the clinical outcomes of minimally invasive procedures. While these technologies' performance has been explored extensively, their impact on daily clinical practice remains undetermined due to the additional cost and complexity, modification of standard devices (e.g. electromagnetic tracking), and different levels of experience among physicians. Taking these factors into consideration, a robotic laser guidance system for percutaneous needle placement is developed. The laser guidance system projects a laser guide line onto the skin entry point of the patient, helping the physician to align the needle with the planned path of the preoperative CT scan. To minimize changes to the standard workflow, the robot is integrated with the CT scanner via optical tracking. As a result, no registration between the robot and CT is needed. The robot can compensate for the motion of the equipment and keep the laser guide line aligned with the biopsy path in real-time. Phantom experiments showed that the guidance system can benefit physicians at different skill levels, while clinical studies showed improved accuracy over conventional freehand needle insertion. The technology is safe, easy to use, and does not involve additional disposable costs. It is our expectation that this technology can be accepted by interventional radiologists for CT guided needle placement procedures.
Telerobotic Excavator Designed to Compete in NASA's Lunabotics Mining Competition
NASA Technical Reports Server (NTRS)
Nash, Rodney; Santin, Cara; Yousef, Ahmed; Nguyen, Thien; Helferty, John; Pillapakkam, Shriram
2011-01-01
The second annual NASA Lunabotics Mining competition is to be held in May 23-28, 2011. The goal of the competition is for teams of university level students to design, build, test and compete with a fully integrated lunar excavator on a simulated lunar surface. Our team, named Lunar Solutions I, will be representing Temple University's College of Engineering in the competition. The team's main goal was to build a robot which is able to compete with other teams, and ultimately win the competition. The main challenge of the competition was to build a wireless robot that can excavate and collect a minimum of 10 kilograms of the regolith material within 15 minutes. The robot must also be designed to operate in conditions similar to those found on the lunar surface. The design of the lunar excavator is constrained by a set of requirements determined by NASA and detailed in the competition's rulebook. The excavator must have the ability to communicate with the "main base" wirelessly, and over a Wi-Fi network. Human operators are located at a remote site approximately 60 meters away from the simulated lunar surface upon which the robot must excavate the lunar regolith surface. During the competition, the robot will operate in a separate area from the control room in an area referred to as the "Lunarena." From the control room, the operators will have to control the robot using visual feedback from cameras placed both within the arena and on the robot. Using this visual feedback the human operators control the robots movement using both keyboard and joystick commands. In order to place in the competition, a minimum of 10 kg of regolith material has to be excavated, collected, and dumped into a specific location. For that reason, the robot must be provided with an effective and powerful excavation system. Our excavator uses tracks for the drive system. After performing extensive research and trade studies, we concluded that tracks would be the most effective method for transporting the excavator. When designing the excavation system, we analyzed several design options from the previous year's competition. We decided to use a front loader to collect the material, rather than a conveyer belt system or auger. Many of the designs from last year's competition used a conveyer belt mechanism to mine regolith and dump it into a temporary storage bin place on the robot. Using the front end loader approach allowed us to combine the scooping system and storage unit, which meant that the excavation system required less space.
Performance Evaluation and Benchmarking of Next Intelligent Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
del Pobil, Angel; Madhavan, Raj; Bonsignorio, Fabio
Performance Evaluation and Benchmarking of Intelligent Systems presents research dedicated to the subject of performance evaluation and benchmarking of intelligent systems by drawing from the experiences and insights of leading experts gained both through theoretical development and practical implementation of intelligent systems in a variety of diverse application domains. This contributed volume offers a detailed and coherent picture of state-of-the-art, recent developments, and further research areas in intelligent systems. The chapters cover a broad range of applications, such as assistive robotics, planetary surveying, urban search and rescue, and line tracking for automotive assembly. Subsystems or components described in this bookmore » include human-robot interaction, multi-robot coordination, communications, perception, and mapping. Chapters are also devoted to simulation support and open source software for cognitive platforms, providing examples of the type of enabling underlying technologies that can help intelligent systems to propagate and increase in capabilities. Performance Evaluation and Benchmarking of Intelligent Systems serves as a professional reference for researchers and practitioners in the field. This book is also applicable to advanced courses for graduate level students and robotics professionals in a wide range of engineering and related disciplines including computer science, automotive, healthcare, manufacturing, and service robotics.« less
Waspe, Adam C; McErlain, David D; Pitelka, Vasek; Holdsworth, David W; Lacefield, James C; Fenster, Aaron
2010-04-01
Preclinical research protocols often require insertion of needles to specific targets within small animal brains. To target biologically relevant locations in rodent brains more effectively, a robotic device has been developed that is capable of positioning a needle along oblique trajectories through a single burr hole in the skull under volumetric microcomputed tomography (micro-CT) guidance. An x-ray compatible stereotactic frame secures the head throughout the procedure using a bite bar, nose clamp, and ear bars. CT-to-robot registration enables structures identified in the image to be mapped to physical coordinates in the brain. Registration is accomplished by injecting a barium sulfate contrast agent as the robot withdraws the needle from predefined points in a phantom. Registration accuracy is affected by the robot-positioning error and is assessed by measuring the surface registration error for the fiducial and target needle tracks (FRE and TRE). This system was demonstrated in situ by injecting 200 microm tungsten beads into rat brains along oblique trajectories through a single burr hole on the top of the skull under micro-CT image guidance. Postintervention micro-CT images of each skull were registered with preintervention high-field magnetic resonance images of the brain to infer the anatomical locations of the beads. Registration using four fiducial needle tracks and one target track produced a FRE and a TRE of 96 and 210 microm, respectively. Evaluation with tissue-mimicking gelatin phantoms showed that locations could be targeted with a mean error of 154 +/- 113 microm. The integration of a robotic needle-positioning device with volumetric micro-CT image guidance should increase the accuracy and reduce the invasiveness of stereotactic needle interventions in small animals.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Waspe, Adam C.; McErlain, David D.; Pitelka, Vasek
Purpose: Preclinical research protocols often require insertion of needles to specific targets within small animal brains. To target biologically relevant locations in rodent brains more effectively, a robotic device has been developed that is capable of positioning a needle along oblique trajectories through a single burr hole in the skull under volumetric microcomputed tomography (micro-CT) guidance. Methods: An x-ray compatible stereotactic frame secures the head throughout the procedure using a bite bar, nose clamp, and ear bars. CT-to-robot registration enables structures identified in the image to be mapped to physical coordinates in the brain. Registration is accomplished by injecting amore » barium sulfate contrast agent as the robot withdraws the needle from predefined points in a phantom. Registration accuracy is affected by the robot-positioning error and is assessed by measuring the surface registration error for the fiducial and target needle tracks (FRE and TRE). This system was demonstrated in situ by injecting 200 {mu}m tungsten beads into rat brains along oblique trajectories through a single burr hole on the top of the skull under micro-CT image guidance. Postintervention micro-CT images of each skull were registered with preintervention high-field magnetic resonance images of the brain to infer the anatomical locations of the beads. Results: Registration using four fiducial needle tracks and one target track produced a FRE and a TRE of 96 and 210 {mu}m, respectively. Evaluation with tissue-mimicking gelatin phantoms showed that locations could be targeted with a mean error of 154{+-}113 {mu}m. Conclusions: The integration of a robotic needle-positioning device with volumetric micro-CT image guidance should increase the accuracy and reduce the invasiveness of stereotactic needle interventions in small animals.« less
Developments and Control of Biocompatible Conducting Polymer for Intracorporeal Continuum Robots.
Chikhaoui, Mohamed Taha; Benouhiba, Amine; Rougeot, Patrick; Rabenorosoa, Kanty; Ouisse, Morvan; Andreff, Nicolas
2018-04-30
Dexterity of robots is highly required when it comes to integration for medical applications. Major efforts have been conducted to increase the dexterity at the distal parts of medical robots. This paper reports on developments toward integrating biocompatible conducting polymers (CP) into inherently dexterous concentric tube robot paradigm. In the form of tri-layer thin structures, CP micro-actuators produce high strains while requiring less than 1 V for actuation. Fabrication, characterization, and first integrations of such micro-actuators are presented. The integration is validated in a preliminary telescopic soft robot prototype with qualitative and quantitative performance assessment of accurate position control for trajectory tracking scenarios. Further, CP micro-actuators are integrated to a laser steering system in a closed-loop control scheme with displacements up to 5 mm. Our first developments aim toward intracorporeal medical robotics, with miniaturized actuators to be embedded into continuum robots.
Iconic Gestures for Robot Avatars, Recognition and Integration with Speech.
Bremner, Paul; Leonards, Ute
2016-01-01
Co-verbal gestures are an important part of human communication, improving its efficiency and efficacy for information conveyance. One possible means by which such multi-modal communication might be realized remotely is through the use of a tele-operated humanoid robot avatar. Such avatars have been previously shown to enhance social presence and operator salience. We present a motion tracking based tele-operation system for the NAO robot platform that allows direct transmission of speech and gestures produced by the operator. To assess the capabilities of this system for transmitting multi-modal communication, we have conducted a user study that investigated if robot-produced iconic gestures are comprehensible, and are integrated with speech. Robot performed gesture outcomes were compared directly to those for gestures produced by a human actor, using a within participant experimental design. We show that iconic gestures produced by a tele-operated robot are understood by participants when presented alone, almost as well as when produced by a human. More importantly, we show that gestures are integrated with speech when presented as part of a multi-modal communication equally well for human and robot performances.
Decentralized adaptive control of robot manipulators with robust stabilization design
NASA Technical Reports Server (NTRS)
Yuan, Bau-San; Book, Wayne J.
1988-01-01
Due to geometric nonlinearities and complex dynamics, a decentralized technique for adaptive control for multilink robot arms is attractive. Lyapunov-function theory for stability analysis provides an approach to robust stabilization. Each joint of the arm is treated as a component subsystem. The adaptive controller is made locally stable with servo signals including proportional and integral gains. This results in the bound on the dynamical interactions with other subsystems. A nonlinear controller which stabilizes the system with uniform boundedness is used to improve the robustness properties of the overall system. As a result, the robot tracks the reference trajectories with convergence. This strategy makes computation simple and therefore facilitates real-time implementation.
NASA Astrophysics Data System (ADS)
Chen, Dechao; Zhang, Yunong
2017-10-01
Dual-arm redundant robot systems are usually required to handle primary tasks, repetitively and synchronously in practical applications. In this paper, a jerk-level synchronous repetitive motion scheme is proposed to remedy the joint-angle drift phenomenon and achieve the synchronous control of a dual-arm redundant robot system. The proposed scheme is novelly resolved at jerk level, which makes the joint variables, i.e. joint angles, joint velocities and joint accelerations, smooth and bounded. In addition, two types of dynamics algorithms, i.e. gradient-type (G-type) and zeroing-type (Z-type) dynamics algorithms, for the design of repetitive motion variable vectors, are presented in detail with the corresponding circuit schematics. Subsequently, the proposed scheme is reformulated as two dynamical quadratic programs (DQPs) and further integrated into a unified DQP (UDQP) for the synchronous control of a dual-arm robot system. The optimal solution of the UDQP is found by the piecewise-linear projection equation neural network. Moreover, simulations and comparisons based on a six-degrees-of-freedom planar dual-arm redundant robot system substantiate the operation effectiveness and tracking accuracy of the robot system with the proposed scheme for repetitive motion and synchronous control.
NASA Technical Reports Server (NTRS)
Tesar, Delbert; Tosunoglu, Sabri; Lin, Shyng-Her
1990-01-01
Research results on general serial robotic manipulators modeled with structural compliances are presented. Two compliant manipulator modeling approaches, distributed and lumped parameter models, are used in this study. System dynamic equations for both compliant models are derived by using the first and second order influence coefficients. Also, the properties of compliant manipulator system dynamics are investigated. One of the properties, which is defined as inaccessibility of vibratory modes, is shown to display a distinct character associated with compliant manipulators. This property indicates the impact of robot geometry on the control of structural oscillations. Example studies are provided to illustrate the physical interpretation of inaccessibility of vibratory modes. Two types of controllers are designed for compliant manipulators modeled by either lumped or distributed parameter techniques. In order to maintain the generality of the results, neither linearization is introduced. Example simulations are given to demonstrate the controller performance. The second type controller is also built for general serial robot arms and is adaptive in nature which can estimate uncertain payload parameters on-line and simultaneously maintain trajectory tracking properties. The relation between manipulator motion tracking capability and convergence of parameter estimation properties is discussed through example case studies. The effect of control input update delays on adaptive controller performance is also studied.
Event-Based Sensing and Control for Remote Robot Guidance: An Experimental Case
Santos, Carlos; Martínez-Rey, Miguel; Santiso, Enrique
2017-01-01
This paper describes the theoretical and practical foundations for remote control of a mobile robot for nonlinear trajectory tracking using an external localisation sensor. It constitutes a classical networked control system, whereby event-based techniques for both control and state estimation contribute to efficient use of communications and reduce sensor activity. Measurement requests are dictated by an event-based state estimator by setting an upper bound to the estimation error covariance matrix. The rest of the time, state prediction is carried out with the Unscented transformation. This prediction method makes it possible to select the appropriate instants at which to perform actuations on the robot so that guidance performance does not degrade below a certain threshold. Ultimately, we obtained a combined event-based control and estimation solution that drastically reduces communication accesses. The magnitude of this reduction is set according to the tracking error margin of a P3-DX robot following a nonlinear trajectory, remotely controlled with a mini PC and whose pose is detected by a camera sensor. PMID:28878144
Optimal Configuration of Human Motion Tracking Systems: A Systems Engineering Approach
NASA Technical Reports Server (NTRS)
Henderson, Steve
2005-01-01
Human motion tracking systems represent a crucial technology in the area of modeling and simulation. These systems, which allow engineers to capture human motion for study or replication in virtual environments, have broad applications in several research disciplines including human engineering, robotics, and psychology. These systems are based on several sensing paradigms, including electro-magnetic, infrared, and visual recognition. Each of these paradigms requires specialized environments and hardware configurations to optimize performance of the human motion tracking system. Ideally, these systems are used in a laboratory or other facility that was designed to accommodate the particular sensing technology. For example, electromagnetic systems are highly vulnerable to interference from metallic objects, and should be used in a specialized lab free of metal components.
Meadmore, Katie L; Cai, Zhonglun; Tong, Daisy; Hughes, Ann-Marie; Freeman, Chris T; Rogers, Eric; Burridge, Jane H
2011-01-01
A novel system has been developed which combines robotic therapy with electrical stimulation (ES) for upper limb stroke rehabilitation. This technology, termed SAIL: Stimulation Assistance through Iterative Learning, employs advanced model-based iterative learning control (ILC) algorithms to precisely assist participant's completion of 3D tracking tasks with their impaired arm. Data is reported from a preliminary study with unimpaired participants, and also from a single hemiparetic stroke participant with reduced upper limb function who has used the system in a clinical trial. All participants completed tasks which involved moving their (impaired) arm to follow an image of a slowing moving sphere along a trajectory. The participants' arm was supported by a robot and ES was applied to the triceps brachii and anterior deltoid muscles. During each task, the same tracking trajectory was repeated 6 times and ILC was used to compute the stimulation signals to be applied on the next iteration. Unimpaired participants took part in a single, one hour training session and the stroke participant undertook 18, 1 hour treatment sessions composed of tracking tasks varying in length, orientation and speed. The results reported describe changes in tracking ability and demonstrate feasibility of the SAIL system for upper limb rehabilitation. © 2011 IEEE
Model identification and vision-based H∞ position control of 6-DoF cable-driven parallel robots
NASA Astrophysics Data System (ADS)
Chellal, R.; Cuvillon, L.; Laroche, E.
2017-04-01
This paper presents methodologies for the identification and control of 6-degrees of freedom (6-DoF) cable-driven parallel robots (CDPRs). First a two-step identification methodology is proposed to accurately estimate the kinematic parameters independently and prior to the dynamic parameters of a physics-based model of CDPRs. Second, an original control scheme is developed, including a vision-based position controller tuned with the H∞ methodology and a cable tension distribution algorithm. The position is controlled in the operational space, making use of the end-effector pose measured by a motion-tracking system. A four-block H∞ design scheme with adjusted weighting filters ensures good trajectory tracking and disturbance rejection properties for the CDPR system, which is a nonlinear-coupled MIMO system with constrained states. The tension management algorithm generates control signals that maintain the cables under feasible tensions. The paper makes an extensive review of the available methods and presents an extension of one of them. The presented methodologies are evaluated by simulations and experimentally on a redundant 6-DoF INCA 6D CDPR with eight cables, equipped with a motion-tracking system.
Biologically inspired computation and learning in Sensorimotor Systems
NASA Astrophysics Data System (ADS)
Lee, Daniel D.; Seung, H. S.
2001-11-01
Networking systems presently lack the ability to intelligently process the rich multimedia content of the data traffic they carry. Endowing artificial systems with the ability to adapt to changing conditions requires algorithms that can rapidly learn from examples. We demonstrate the application of such learning algorithms on an inexpensive quadruped robot constructed to perform simple sensorimotor tasks. The robot learns to track a particular object by discovering the salient visual and auditory cues unique to that object. The system uses a convolutional neural network that automatically combines color, luminance, motion, and auditory information. The weights of the networks are adjusted using feedback from a teacher to reflect the reliability of the various input channels in the surrounding environment. Additionally, the robot is able to compensate for its own motion by adapting the parameters of a vestibular ocular reflex system.
Cost analysis of robotic versus laparoscopic general surgery procedures.
Higgins, Rana M; Frelich, Matthew J; Bosler, Matthew E; Gould, Jon C
2017-01-01
Robotic surgical systems have been used at a rapidly increasing rate in general surgery. Many of these procedures have been performed laparoscopically for years. In a surgical encounter, a significant portion of the total costs is associated with consumable supplies. Our hospital system has invested in a software program that can track the costs of consumable surgical supplies. We sought to determine the differences in cost of consumables with elective laparoscopic and robotic procedures for our health care organization. De-identified procedural cost and equipment utilization data were collected from the Surgical Profitability Compass Procedure Cost Manager System (The Advisory Board Company, Washington, DC) for our health care system for laparoscopic and robotic cholecystectomy, fundoplication, and inguinal hernia between the years 2013 and 2015. Outcomes were length of stay, case duration, and supply cost. Statistical analysis was performed using a t-test for continuous variables, and statistical significance was defined as p < 0.05. The total cost of consumable surgical supplies was significantly greater for all robotic procedures. Length of stay did not differ for fundoplication or cholecystectomy. Length of stay was greater for robotic inguinal hernia repair. Case duration was similar for cholecystectomy (84.3 robotic and 75.5 min laparoscopic, p = 0.08), but significantly longer for robotic fundoplication (197.2 robotic and 162.1 min laparoscopic, p = 0.01) and inguinal hernia repair (124.0 robotic and 84.4 min laparoscopic, p = ≪0.01). We found a significantly increased cost of general surgery procedures for our health care system when cases commonly performed laparoscopically are instead performed robotically. Our analysis is limited by the fact that we only included costs associated with consumable surgical supplies. The initial acquisition cost (over $1 million for robotic surgical system), depreciation, and service contract for the robotic and laparoscopic systems were not included in this analysis.
Chai, Xun; Gao, Feng; Pan, Yang; Qi, Chenkun; Xu, Yilin
2015-01-01
Coordinate identification between vision systems and robots is quite a challenging issue in the field of intelligent robotic applications, involving steps such as perceiving the immediate environment, building the terrain map and planning the locomotion automatically. It is now well established that current identification methods have non-negligible limitations such as a difficult feature matching, the requirement of external tools and the intervention of multiple people. In this paper, we propose a novel methodology to identify the geometric parameters of 3D vision systems mounted on robots without involving other people or additional equipment. In particular, our method focuses on legged robots which have complex body structures and excellent locomotion ability compared to their wheeled/tracked counterparts. The parameters can be identified only by moving robots on a relatively flat ground. Concretely, an estimation approach is provided to calculate the ground plane. In addition, the relationship between the robot and the ground is modeled. The parameters are obtained by formulating the identification problem as an optimization problem. The methodology is integrated on a legged robot called “Octopus”, which can traverse through rough terrains with high stability after obtaining the identification parameters of its mounted vision system using the proposed method. Diverse experiments in different environments demonstrate our novel method is accurate and robust. PMID:25912350
NASA Astrophysics Data System (ADS)
Roth, Eatai; Howell, Darrin; Beckwith, Cydney; Burden, Samuel A.
2017-05-01
Humans, interacting with cyber-physical systems (CPS), formulate beliefs about the system's dynamics. It is natural to expect that human operators, tasked with teleoperation, use these beliefs to control the remote robot. For tracking tasks in the resulting human-cyber-physical system (HCPS), theory suggests that human operators can achieve exponential tracking (in stable systems) without state estimation provided they possess an accurate model of the system's dynamics. This internalized inverse model, however, renders a portion of the system state unobservable to the human operator—the zero dynamics. Prior work shows humans can track through observable linear dynamics, thus we focus on nonlinear dynamics rendered unobservable through tracking control. We propose experiments to assess the human operator's ability to learn and invert such models, and distinguish this behavior from that achieved by pure feedback control.
Distributed multirobot sensing and tracking: a behavior-based approach
NASA Astrophysics Data System (ADS)
Parker, Lynne E.
1995-09-01
An important issue that arises in the automation of many large-scale surveillance and reconnaissance tasks is that of tracking the movements of (or maintaining passive contact with) objects navigating in a bounded area of interest. Oftentimes in these problems, the area to be monitored will move over time or will not permit fixed sensors, thus requiring a team of mobile sensors--or robots--to monitor the area collectively. In these situations, the robots must not only have mechanisms for determining how to track objects and how to fuse information from neighboring robots, but they must also have distributed control strategies for ensuring that the entire area of interest is continually covered to the greatest extent possible. This paper focuses on the distributed control issue by describing a proposed decentralized control mechanism that allows a team of robots to collectively track and monitor objects in an uncluttered area of interest. The approach is based upon an extension to the ALLIANCE behavior-based architecture that generalizes from the domain of loosely-coupled, independent applications to the domain of strongly cooperative applications, in which the action selection of a robot is dependent upon the actions selected by its teammates. We conclude the paper be describing our ongoing implementation of the proposed approach on a team of four mobile robots.
Localization and Tracking of Implantable Biomedical Sensors
Umay, Ilknur; Fidan, Barış; Barshan, Billur
2017-01-01
Implantable sensor systems are effective tools for biomedical diagnosis, visualization and treatment of various health conditions, attracting the interest of researchers, as well as healthcare practitioners. These systems efficiently and conveniently provide essential data of the body part being diagnosed, such as gastrointestinal (temperature, pH, pressure) parameter values, blood glucose and pressure levels and electrocardiogram data. Such data are first transmitted from the implantable sensor units to an external receiver node or network and then to a central monitoring and control (computer) unit for analysis, diagnosis and/or treatment. Implantable sensor units are typically in the form of mobile microrobotic capsules or implanted stationary (body-fixed) units. In particular, capsule-based systems have attracted significant research interest recently, with a variety of applications, including endoscopy, microsurgery, drug delivery and biopsy. In such implantable sensor systems, one of the most challenging problems is the accurate localization and tracking of the microrobotic sensor unit (e.g., robotic capsule) inside the human body. This article presents a literature review of the existing localization and tracking techniques for robotic implantable sensor systems with their merits and limitations and possible solutions of the proposed localization methods. The article also provides a brief discussion on the connection and cooperation of such techniques with wearable biomedical sensor systems. PMID:28335384
Soft brain-machine interfaces for assistive robotics: A novel control approach.
Schiatti, Lucia; Tessadori, Jacopo; Barresi, Giacinto; Mattos, Leonardo S; Ajoudani, Arash
2017-07-01
Robotic systems offer the possibility of improving the life quality of people with severe motor disabilities, enhancing the individual's degree of independence and interaction with the external environment. In this direction, the operator's residual functions must be exploited for the control of the robot movements and the underlying dynamic interaction through intuitive and effective human-robot interfaces. Towards this end, this work aims at exploring the potential of a novel Soft Brain-Machine Interface (BMI), suitable for dynamic execution of remote manipulation tasks for a wide range of patients. The interface is composed of an eye-tracking system, for an intuitive and reliable control of a robotic arm system's trajectories, and a Brain-Computer Interface (BCI) unit, for the control of the robot Cartesian stiffness, which determines the interaction forces between the robot and environment. The latter control is achieved by estimating in real-time a unidimensional index from user's electroencephalographic (EEG) signals, which provides the probability of a neutral or active state. This estimated state is then translated into a stiffness value for the robotic arm, allowing a reliable modulation of the robot's impedance. A preliminary evaluation of this hybrid interface concept provided evidence on the effective execution of tasks with dynamic uncertainties, demonstrating the great potential of this control method in BMI applications for self-service and clinical care.
Eizicovits, Danny; Edan, Yael; Tabak, Iris; Levy-Tzedek, Shelly
2018-01-01
Effective human-robot interactions in rehabilitation necessitates an understanding of how these should be tailored to the needs of the human. We report on a robotic system developed as a partner on a 3-D everyday task, using a gamified approach. To: (1) design and test a prototype system, to be ultimately used for upper-limb rehabilitation; (2) evaluate how age affects the response to such a robotic system; and (3) identify whether the robot's physical embodiment is an important aspect in motivating users to complete a set of repetitive tasks. 62 healthy participants, young (<30 yo) and old (>60 yo), played a 3D tic-tac-toe game against an embodied (a robotic arm) and a non-embodied (a computer-controlled lighting system) partner. To win, participants had to place three cups in sequence on a physical 3D grid. Cup picking-and-placing was chosen as a functional task that is often practiced in post-stroke rehabilitation. Movement of the participants was recorded using a Kinect camera. The timing of the participants' movement was primed by the response time of the system: participants moved slower when playing with the slower embodied system (p = 0.006). The majority of participants preferred the robot over the computer-controlled system. Slower response time of the robot compared to the computer-controlled one only affected the young group's motivation to continue playing. We demonstrated the feasibility of the system to encourage the performance of repetitive 3D functional movements, and track these movements. Young and old participants preferred to interact with the robot, compared with the non-embodied system. We contribute to the growing knowledge concerning personalized human-robot interactions by (1) demonstrating the priming of the human movement by the robotic movement - an important design feature, and (2) identifying response-speed as a design variable, the importance of which depends on the age of the user.
Dagnino, Giulio; Georgilas, Ioannis; Tarassoli, Payam; Atkins, Roger; Dogramadzi, Sanja
2016-03-01
Joint fracture surgery quality can be improved by robotic system with high-accuracy and high-repeatability fracture fragment manipulation. A new real-time vision-based system for fragment manipulation during robot-assisted fracture surgery was developed and tested. The control strategy was accomplished by merging fast open-loop control with vision-based control. This two-phase process is designed to eliminate the open-loop positioning errors by closing the control loop using visual feedback provided by an optical tracking system. Evaluation of the control system accuracy was performed using robot positioning trials, and fracture reduction accuracy was tested in trials on ex vivo porcine model. The system resulted in high fracture reduction reliability with a reduction accuracy of 0.09 mm (translations) and of [Formula: see text] (rotations), maximum observed errors in the order of 0.12 mm (translations) and of [Formula: see text] (rotations), and a reduction repeatability of 0.02 mm and [Formula: see text]. The proposed vision-based system was shown to be effective and suitable for real joint fracture surgical procedures, contributing a potential improvement of their quality.
Ao, Di; Song, Rong; Gao, JinWu
2017-08-01
Although the merits of electromyography (EMG)-based control of powered assistive systems have been certified, the factors that affect the performance of EMG-based human-robot cooperation, which are very important, have received little attention. This study investigates whether a more physiologically appropriate model could improve the performance of human-robot cooperation control for an ankle power-assist exoskeleton robot. To achieve the goal, an EMG-driven Hill-type neuromusculoskeletal model (HNM) and a linear proportional model (LPM) were developed and calibrated through maximum isometric voluntary dorsiflexion (MIVD). The two control models could estimate the real-time ankle joint torque, and HNM is more accurate and can account for the change of the joint angle and muscle dynamics. Then, eight healthy volunteers were recruited to wear the ankle exoskeleton robot and complete a series of sinusoidal tracking tasks in the vertical plane. With the various levels of assist based on the two calibrated models, the subjects were instructed to track the target displayed on the screen as accurately as possible by performing ankle dorsiflexion and plantarflexion. Two measurements, the root mean square error (RMSE) and root mean square jerk (RMSJ), were derived from the assistant torque and kinematic signals to characterize the movement performances, whereas the amplitudes of the recorded EMG signals from the tibialis anterior (TA) and the gastrocnemius (GAS) were obtained to reflect the muscular efforts. The results demonstrated that the muscular effort and smoothness of tracking movements decreased with an increase in the assistant ratio. Compared with LPM, subjects made lower physical efforts and generated smoother movements when using HNM, which implied that a more physiologically appropriate model could enable more natural and human-like human-robot cooperation and has potential value for improvement of human-exoskeleton interaction in future applications.
Survey of Motion Tracking Methods Based on Inertial Sensors: A Focus on Upper Limb Human Motion
Filippeschi, Alessandro; Schmitz, Norbert; Miezal, Markus; Bleser, Gabriele; Ruffaldi, Emanuele; Stricker, Didier
2017-01-01
Motion tracking based on commercial inertial measurements units (IMUs) has been widely studied in the latter years as it is a cost-effective enabling technology for those applications in which motion tracking based on optical technologies is unsuitable. This measurement method has a high impact in human performance assessment and human-robot interaction. IMU motion tracking systems are indeed self-contained and wearable, allowing for long-lasting tracking of the user motion in situated environments. After a survey on IMU-based human tracking, five techniques for motion reconstruction were selected and compared to reconstruct a human arm motion. IMU based estimation was matched against motion tracking based on the Vicon marker-based motion tracking system considered as ground truth. Results show that all but one of the selected models perform similarly (about 35 mm average position estimation error). PMID:28587178
Parallel Robot for Lower Limb Rehabilitation Exercises.
Rastegarpanah, Alireza; Saadat, Mozafar; Borboni, Alberto
2016-01-01
The aim of this study is to investigate the capability of a 6-DoF parallel robot to perform various rehabilitation exercises. The foot trajectories of twenty healthy participants have been measured by a Vicon system during the performing of four different exercises. Based on the kinematics and dynamics of a parallel robot, a MATLAB program was developed in order to calculate the length of the actuators, the actuators' forces, workspace, and singularity locus of the robot during the performing of the exercises. The calculated length of the actuators and the actuators' forces were used by motion analysis in SolidWorks in order to simulate different foot trajectories by the CAD model of the robot. A physical parallel robot prototype was built in order to simulate and execute the foot trajectories of the participants. Kinect camera was used to track the motion of the leg's model placed on the robot. The results demonstrate the robot's capability to perform a full range of various rehabilitation exercises.
Autonomous stair-climbing with miniature jumping robots.
Stoeter, Sascha A; Papanikolopoulos, Nikolaos
2005-04-01
The problem of vision-guided control of miniature mobile robots is investigated. Untethered mobile robots with small physical dimensions of around 10 cm or less do not permit powerful onboard computers because of size and power constraints. These challenges have, in the past, reduced the functionality of such devices to that of a complex remote control vehicle with fancy sensors. With the help of a computationally more powerful entity such as a larger companion robot, the control loop can be closed. Using the miniature robot's video transmission or that of an observer to localize it in the world, control commands can be computed and relayed to the inept robot. The result is a system that exhibits autonomous capabilities. The framework presented here solves the problem of climbing stairs with the miniature Scout robot. The robot's unique locomotion mode, the jump, is employed to hop one step at a time. Methods for externally tracking the Scout are developed. A large number of real-world experiments are conducted and the results discussed.
Parallel Robot for Lower Limb Rehabilitation Exercises
Saadat, Mozafar; Borboni, Alberto
2016-01-01
The aim of this study is to investigate the capability of a 6-DoF parallel robot to perform various rehabilitation exercises. The foot trajectories of twenty healthy participants have been measured by a Vicon system during the performing of four different exercises. Based on the kinematics and dynamics of a parallel robot, a MATLAB program was developed in order to calculate the length of the actuators, the actuators' forces, workspace, and singularity locus of the robot during the performing of the exercises. The calculated length of the actuators and the actuators' forces were used by motion analysis in SolidWorks in order to simulate different foot trajectories by the CAD model of the robot. A physical parallel robot prototype was built in order to simulate and execute the foot trajectories of the participants. Kinect camera was used to track the motion of the leg's model placed on the robot. The results demonstrate the robot's capability to perform a full range of various rehabilitation exercises. PMID:27799727
Hybrid tracking and control system for computer-aided retinal surgery
NASA Astrophysics Data System (ADS)
Ferguson, R. D.; Wright, Cameron H. G.; Rylander, Henry G., III; Welch, Ashley J.; Barrett, Steven F.
1996-05-01
We describe initial experimental results of a new hybrid digital and analog design for retinal tracking and laser beam control. Initial results demonstrate tracking rates which exceed the equivalent of 50 degrees per second in the eye, with automatic lesion pattern creation and robust loss of lock detection. Robotically assisted laser surgery to treat conditions such as diabetic retinopathy, macular degeneration, and retinal tears can now be realized under clinical conditions with requisite safety using standard video hardware and inexpensive optical components.
Robotic control and inspection verification
NASA Technical Reports Server (NTRS)
Davis, Virgil Leon
1991-01-01
Three areas of possible commercialization involving robots at the Kennedy Space Center (KSC) are discussed: a six degree-of-freedom target tracking system for remote umbilical operations; an intelligent torque sensing end effector for operating hand valves in hazardous locations; and an automatic radiator inspection device, a 13 by 65 foot robotic mechanism involving completely redundant motors, drives, and controls. Aspects concerning the first two innovations can be integrated to enable robots or teleoperators to perform tasks involving orientation and panal actuation operations that can be done with existing technology rather than waiting for telerobots to incorporate artificial intelligence (AI) to perform 'smart' autonomous operations. The third robot involves the application of complete control hardware redundancy to enable performance of work over and near expensive Space Shuttle hardware. The consumer marketplace may wish to explore commercialization of similiar component redundancy techniques for applications when a robot would not normally be used because of reliability concerns.
Li, Chong; Bi, Sheng; Zhang, Xuemin; Huo, Jianfei
2017-01-01
Numerous robots have been widely used to deliver rehabilitative training for hemiplegic patients to improve their functional ability. Because of the complexity and diversity of upper limb motion, customization of training patterns is one key factor during upper limb rehabilitation training. Most of the current rehabilitation robots cannot intelligently provide adaptive training parameters, and they have not been widely used in clinical rehabilitation. This article proposes a new end-effector upper limb rehabilitation robot, which is a two-link robotic arm with two active degrees of freedom. This work investigated the kinematics and dynamics of the robot system, the control system, and the realization of different rehabilitation therapies. We also explored the influence of constraint in rehabilitation therapies on interaction force and muscle activation. The deviation of the trajectory of the end effector and the required trajectory was less than 1 mm during the tasks, which demonstrated the movement accuracy of the robot. Besides, results also demonstrated the constraint exerted by the robot provided benefits for hemiplegic patients by changing muscle activation in the way similar to the movement pattern of the healthy subjects, which indicated that the robot can improve the patient's functional ability by training the normal movement pattern. PMID:29065614
Liu, Yali; Li, Chong; Ji, Linhong; Bi, Sheng; Zhang, Xuemin; Huo, Jianfei; Ji, Run
2017-01-01
Numerous robots have been widely used to deliver rehabilitative training for hemiplegic patients to improve their functional ability. Because of the complexity and diversity of upper limb motion, customization of training patterns is one key factor during upper limb rehabilitation training. Most of the current rehabilitation robots cannot intelligently provide adaptive training parameters, and they have not been widely used in clinical rehabilitation. This article proposes a new end-effector upper limb rehabilitation robot, which is a two-link robotic arm with two active degrees of freedom. This work investigated the kinematics and dynamics of the robot system, the control system, and the realization of different rehabilitation therapies. We also explored the influence of constraint in rehabilitation therapies on interaction force and muscle activation. The deviation of the trajectory of the end effector and the required trajectory was less than 1 mm during the tasks, which demonstrated the movement accuracy of the robot. Besides, results also demonstrated the constraint exerted by the robot provided benefits for hemiplegic patients by changing muscle activation in the way similar to the movement pattern of the healthy subjects, which indicated that the robot can improve the patient's functional ability by training the normal movement pattern.
Smart tissue anastomosis robot (STAR): a vision-guided robotics system for laparoscopic suturing.
Leonard, Simon; Wu, Kyle L; Kim, Yonjae; Krieger, Axel; Kim, Peter C W
2014-04-01
This paper introduces the smart tissue anastomosis robot (STAR). Currently, the STAR is a proof-of-concept for a vision-guided robotic system featuring an actuated laparoscopic suturing tool capable of executing running sutures from image-based commands. The STAR tool is designed around a commercially available laparoscopic suturing tool that is attached to a custom-made motor stage and the STAR supervisory control architecture that enables a surgeon to select and track incisions and the placement of stitches. The STAR supervisory-control interface provides two modes: A manual mode that enables a surgeon to specify the placement of each stitch and an automatic mode that automatically computes equally-spaced stitches based on an incision contour. Our experiments on planar phantoms demonstrate that the STAR in either mode is more accurate, up to four times more consistent and five times faster than surgeons using state-of-the-art robotic surgical system, four times faster than surgeons using manual Endo360(°)®, and nine times faster than surgeons using manual laparoscopic tools.
Technology transfer: Imaging tracker to robotic controller
NASA Technical Reports Server (NTRS)
Otaguro, M. S.; Kesler, L. O.; Land, Ken; Erwin, Harry; Rhoades, Don
1988-01-01
The transformation of an imaging tracker to a robotic controller is described. A multimode tracker was developed for fire and forget missile systems. The tracker locks on to target images within an acquisition window using multiple image tracking algorithms to provide guidance commands to missile control systems. This basic tracker technology is used with the addition of a ranging algorithm based on sizing a cooperative target to perform autonomous guidance and control of a platform for an Advanced Development Project on automation and robotics. A ranging tracker is required to provide the positioning necessary for robotic control. A simple functional demonstration of the feasibility of this approach was performed and described. More realistic demonstrations are under way at NASA-JSC. In particular, this modified tracker, or robotic controller, will be used to autonomously guide the Man Maneuvering Unit (MMU) to targets such as disabled astronauts or tools as part of the EVA Retriever efforts. It will also be used to control the orbiter's Remote Manipulator Systems (RMS) in autonomous approach and positioning demonstrations. These efforts will also be discussed.
Novel compliant actuator for wearable robotics applications.
Claros, M; Soto, R; Rodríguez, J J; Cantú, C; Contreras-Vidal, José L
2013-01-01
In the growing fields of wearable robotics, rehabilitation robotics, prosthetics, and walking robots, variable impedance and force actuators are being designed and implemented because of their ability to dynamically modulate the intrinsic viscoelastic properties such as stiffness and damping. This modulation is crucial to achieve an efficient and safe human-robot interaction that could lead to electronically generate useful emergent dynamical behaviors. In this work we propose a novel actuation system in which is implemented a control scheme based on equilibrium forces for an active joint capable to provide assistance/resistance as needed and also achieve minimal mechanical impedance when tracking the movement of the user limbs. The actuation system comprises a DC motor with a built in speed reducer, two force-sensing resistors (FSR), a mechanism which transmits to the FSRs the torque developed in the joint and a controller which regulate the amount of energy that is delivered to the DC motor. The proposed system showed more impedance reduction, by the effect of the controlled contact forces, compared with the ones in the reviewed literature.
Control of a free-flying robot manipulator system
NASA Technical Reports Server (NTRS)
Alexander, H.; Cannon, R. H., Jr.
1985-01-01
The goal of the research is to develop and test control strategies for a self-contained, free flying space robot. Such a robot would perform operations in space similar to those currently handled by astronauts during extravehicular activity (EVA). The focus of the work is to develop and carry out a program of research with a series of physical Satellite Robot Simulator Vehicles (SRSV's), two-dimensionally freely mobile laboratory models of autonomous free-flying space robots such as might perform extravehicular functions associated with operation of a space station or repair of orbiting satellites. The development of the SRSV and of some of the controller subsystems are discribed. The two-link arm was fitted to the SRSV base, and researchers explored the open-loop characteristics of the arm and thruster actuators. Work began on building the software foundation necessary for use of the on-board computer, as well as hardware and software for a local vision system for target identification and tracking.
NASA Astrophysics Data System (ADS)
Dima, M.; Francu, C.
2016-08-01
This paper presents a way to expand the field of use of the laser tracker and SmartTrack sensor localization device used in lately for the localisation of the end effector of the industrial robots to the localization of the mobile construction robots. The research paper presents the equipment along with its characteristics, determines the relationships for the localization coordinates by comparison to the forward kinematics of the industrial robot's spherical arm (positioning mechanism in spherical coordinates) and the orientation mechanism with three revolute axes. In the end of the paper the accuracy of the mobile robot's localization is analysed.
Rendezvous with connectivity preservation for multi-robot systems with an unknown leader
NASA Astrophysics Data System (ADS)
Dong, Yi
2018-02-01
This paper studies the leader-following rendezvous problem with connectivity preservation for multi-agent systems composed of uncertain multi-robot systems subject to external disturbances and an unknown leader, both of which are generated by a so-called exosystem with parametric uncertainty. By combining internal model design, potential function technique and adaptive control, two distributed control strategies are proposed to maintain the connectivity of the communication network, to achieve the asymptotic tracking of all the followers to the output of the unknown leader system, as well as to reject unknown external disturbances. It is also worth to mention that the uncertain parameters in the multi-robot systems and exosystem are further allowed to belong to unknown and unbounded sets when applying the second fully distributed control law containing a dynamic gain inspired by high-gain adaptive control or self-tuning regulator.
Remote ultrasound palpation for robotic interventions using absolute elastography.
Schneider, Caitlin; Baghani, Ali; Rohling, Robert; Salcudean, Septimiu
2012-01-01
Although robotic surgery has addressed many of the challenges presented by minimally invasive surgery, haptic feedback and the lack of knowledge of tissue stiffness is an unsolved problem. This paper presents a system for finding the absolute elastic properties of tissue using a freehand ultrasound scanning technique, which utilizes the da Vinci Surgical robot and a custom 2D ultrasound transducer for intraoperative use. An external exciter creates shear waves in the tissue, and a local frequency estimation method computes the shear modulus. Results are reported for both phantom and in vivo models. This system can be extended to any 6 degree-of-freedom tracking method and any 2D transducer to provide real-time absolute elastic properties of tissue.
Developing a multidisciplinary robotic surgery quality assessment program.
Gonsenhauser, Iahn; Abaza, Ronney; Mekhjian, Hagop; Moffatt-Bruce, Susan D
2012-01-01
The objective of this study was to test the feasibility of a novel quality-improvement (QI) program designed to incorporate multiple robotic surgical sub-specialties in one health care system. A robotic surgery quality assessment program was developed by The Ohio State University College of Medicine (OSUMC) in conjunction with The Ohio State University Medical Center Quality Improvement and Operations Department. A retrospective review of cases was performed using data interrogated from the OSUMC Information Warehouse from January 2007 through August 2009. Robotic surgery cases (n=2200) were assessed for operative times, length of stay (LOS), conversions, returns to surgery, readmissions and cancellations as potential quality indicators. An actionable and reproducible framework for the quality measurement and assessment of a multidisciplinary and interdepartmental robotic surgery program was successfully completed demonstrating areas for improvement opportunities. This report supports that standard quality indicators can be applied to multiple specialties within a health care system to develop a useful quality tracking and assessment tool in the highly specialized area of robotic surgery. © 2012 National Association for Healthcare Quality.
NASA Astrophysics Data System (ADS)
Guo, Dejun; Bourne, Joseph R.; Wang, Hesheng; Yim, Woosoon; Leang, Kam K.
2017-08-01
This paper presents the design and implementation of an adaptive-repetitive visual-servo control system for a moving high-flying vehicle (HFV) with an uncalibrated camera to monitor, track, and precisely control the movements of a low-flying vehicle (LFV) or mobile ground robot. Applications of this control strategy include the use of high-flying unmanned aerial vehicles (UAVs) with computer vision for monitoring, controlling, and coordinating the movements of lower altitude agents in areas, for example, where GPS signals may be unreliable or nonexistent. When deployed, a remote operator of the HFV defines the desired trajectory for the LFV in the HFV's camera frame. Due to the circular motion of the HFV, the resulting motion trajectory of the LFV in the image frame can be periodic in time, thus an adaptive-repetitive control system is exploited for regulation and/or trajectory tracking. The adaptive control law is able to handle uncertainties in the camera's intrinsic and extrinsic parameters. The design and stability analysis of the closed-loop control system is presented, where Lyapunov stability is shown. Simulation and experimental results are presented to demonstrate the effectiveness of the method for controlling the movement of a low-flying quadcopter, demonstrating the capabilities of the visual-servo control system for localization (i.e.,, motion capturing) and trajectory tracking control. In fact, results show that the LFV can be commanded to hover in place as well as track a user-defined flower-shaped closed trajectory, while the HFV and camera system circulates above with constant angular velocity. On average, the proposed adaptive-repetitive visual-servo control system reduces the average RMS tracking error by over 77% in the image plane and over 71% in the world frame compared to using just the adaptive visual-servo control law.
Real Time Target Tracking in a Phantom Using Ultrasonic Imaging
NASA Astrophysics Data System (ADS)
Xiao, X.; Corner, G.; Huang, Z.
In this paper we present a real-time ultrasound image guidance method suitable for tracking the motion of tumors. A 2D ultrasound based motion tracking system was evaluated. A robot was used to control the focused ultrasound and position it at the target that has been segmented from a real-time ultrasound video. Tracking accuracy and precision were investigated using a lesion mimicking phantom. Experiments have been conducted and results show sufficient efficiency of the image guidance algorithm. This work could be developed as the foundation for combining the real time ultrasound imaging tracking and MRI thermometry monitoring non-invasive surgery.
Coordinating teams of autonomous vehicles: an architectural perspective
NASA Astrophysics Data System (ADS)
Czichon, Cary; Peterson, Robert W.; Mettala, Erik G.; Vondrak, Ivo
2005-05-01
In defense-related robotics research, a mission level integration gap exists between mission tasks (tactical) performed by ground, sea, or air applications and elementary behaviors enacted by processing, communications, sensors, and weaponry resources (platform specific). The gap spans ensemble (heterogeneous team) behaviors, automatic MOE/MOP tracking, and tactical task modeling/simulation for virtual and mixed teams comprised of robotic and human combatants. This study surveys robotic system architectures, compares approaches for navigating problem/state spaces by autonomous systems, describes an architecture for an integrated, repository-based modeling, simulation, and execution environment, and outlines a multi-tiered scheme for robotic behavior components that is agent-based, platform-independent, and extendable via plug-ins. Tools for this integrated environment, along with a distributed agent framework for collaborative task performance are being developed by a U.S. Army funded SBIR project (RDECOM Contract N61339-04-C-0005).
Automatic localization of the da Vinci surgical instrument tips in 3-D transrectal ultrasound.
Mohareri, Omid; Ramezani, Mahdi; Adebar, Troy K; Abolmaesumi, Purang; Salcudean, Septimiu E
2013-09-01
Robot-assisted laparoscopic radical prostatectomy (RALRP) using the da Vinci surgical system is the current state-of-the-art treatment option for clinically confined prostate cancer. Given the limited field of view of the surgical site in RALRP, several groups have proposed the integration of transrectal ultrasound (TRUS) imaging in the surgical workflow to assist with accurate resection of the prostate and the sparing of the neurovascular bundles (NVBs). We previously introduced a robotic TRUS manipulator and a method for automatically tracking da Vinci surgical instruments with the TRUS imaging plane, in order to facilitate the integration of intraoperative TRUS in RALRP. Rapid and automatic registration of the kinematic frames of the da Vinci surgical system and the robotic TRUS probe manipulator is a critical component of the instrument tracking system. In this paper, we propose a fully automatic registration technique based on automatic 3-D TRUS localization of robot instrument tips pressed against the air-tissue boundary anterior to the prostate. The detection approach uses a multiscale filtering technique to identify and localize surgical instrument tips in the TRUS volume, and could also be used to detect other surface fiducials in 3-D ultrasound. Experiments have been performed using a tissue phantom and two ex vivo tissue samples to show the feasibility of the proposed methods. Also, an initial in vivo evaluation of the system has been carried out on a live anaesthetized dog with a da Vinci Si surgical system and a target registration error (defined as the root mean square distance of corresponding points after registration) of 2.68 mm has been achieved. Results show this method's accuracy and consistency for automatic registration of TRUS images to the da Vinci surgical system.
Position and orientation tracking system
Burks, Barry L.; DePiero, Fred W.; Armstrong, Gary A.; Jansen, John F.; Muller, Richard C.; Gee, Timothy F.
1998-01-01
A position and orientation tracking system presents a laser scanning appaus having two measurement pods, a control station, and a detector array. The measurement pods can be mounted in the dome of a radioactive waste storage silo. Each measurement pod includes dual orthogonal laser scanner subsystems. The first laser scanner subsystem is oriented to emit a first line laser in the pan direction. The second laser scanner is oriented to emit a second line laser in the tilt direction. Both emitted line lasers scan planes across the radioactive waste surface to encounter the detector array mounted on a target robotic vehicle. The angles of incidence of the planes with the detector array are recorded by the control station. Combining measurements describing each of the four planes provides data for a closed form solution of the algebraic transform describing the position and orientation of the target robotic vehicle.
Position and orientation tracking system
Burks, B.L.; DePiero, F.W.; Armstrong, G.A.; Jansen, J.F.; Muller, R.C.; Gee, T.F.
1998-05-05
A position and orientation tracking system presents a laser scanning apparatus having two measurement pods, a control station, and a detector array. The measurement pods can be mounted in the dome of a radioactive waste storage silo. Each measurement pod includes dual orthogonal laser scanner subsystems. The first laser scanner subsystem is oriented to emit a first line laser in the pan direction. The second laser scanner is oriented to emit a second line laser in the tilt direction. Both emitted line lasers scan planes across the radioactive waste surface to encounter the detector array mounted on a target robotic vehicle. The angles of incidence of the planes with the detector array are recorded by the control station. Combining measurements describing each of the four planes provides data for a closed form solution of the algebraic transform describing the position and orientation of the target robotic vehicle. 14 figs.
Integrating deliberative planning in a robot architecture
NASA Technical Reports Server (NTRS)
Elsaesser, Chris; Slack, Marc G.
1994-01-01
The role of planning and reactive control in an architecture for autonomous agents is discussed. The postulated architecture seperates the general robot intelligence problem into three interacting pieces: (1) robot reactive skills, i.e., grasping, object tracking, etc.; (2) a sequencing capability to differentially ativate the reactive skills; and (3) a delibrative planning capability to reason in depth about goals, preconditions, resources, and timing constraints. Within the sequencing module, caching techniques are used for handling routine activities. The planning system then builds on these cached solutions to routine tasks to build larger grain sized primitives. This eliminates large numbers of essentially linear planning problems. The architecture will be used in the future to incorporate in robots cognitive capabilites normally associated with intelligent behavior.
Experiments in teleoperator and autonomous control of space robotic vehicles
NASA Technical Reports Server (NTRS)
Alexander, Harold L.
1990-01-01
A research program and strategy are described which include fundamental teleoperation issues and autonomous-control issues of sensing and navigation for satellite robots. The program consists of developing interfaces for visual operation and studying the consequences of interface designs as well as developing navigation and control technologies based on visual interaction. A space-robot-vehicle simulator is under development for use in virtual-environment teleoperation experiments and neutral-buoyancy investigations. These technologies can be utilized in a study of visual interfaces to address tradeoffs between head-tracking and manual remote cameras, panel-mounted and helmet-mounted displays, and stereoscopic and monoscopic display systems. The present program can provide significant data for the development of control experiments for autonomously controlled satellite robots.
Iconic Gestures for Robot Avatars, Recognition and Integration with Speech
Bremner, Paul; Leonards, Ute
2016-01-01
Co-verbal gestures are an important part of human communication, improving its efficiency and efficacy for information conveyance. One possible means by which such multi-modal communication might be realized remotely is through the use of a tele-operated humanoid robot avatar. Such avatars have been previously shown to enhance social presence and operator salience. We present a motion tracking based tele-operation system for the NAO robot platform that allows direct transmission of speech and gestures produced by the operator. To assess the capabilities of this system for transmitting multi-modal communication, we have conducted a user study that investigated if robot-produced iconic gestures are comprehensible, and are integrated with speech. Robot performed gesture outcomes were compared directly to those for gestures produced by a human actor, using a within participant experimental design. We show that iconic gestures produced by a tele-operated robot are understood by participants when presented alone, almost as well as when produced by a human. More importantly, we show that gestures are integrated with speech when presented as part of a multi-modal communication equally well for human and robot performances. PMID:26925010
Jiang, Shan; Lou, Jinlong; Yang, Zhiyong; Dai, Jiansheng; Yu, Yan
2015-09-01
Biopsy and brachytherapy for small core breast cancer are always difficult medical problems in the field of cancer treatment. This research mainly develops a magnetic resonance imaging-guided high-precision robotic system for breast puncture treatment. First, a 5-degree-of-freedom tendon-based surgical robotic system is introduced in detail. What follows are the kinematic analysis and dynamical modeling of the robotic system, where a mathematic dynamic model is established using the Lagrange method and a lumped parameter tendon model is used to identify the nonlinear gain of the tendon-sheath transmission system. Based on the dynamical models, an adaptive proportional-integral-derivative controller with friction compensation is proposed for accurate position control. Through simulations using different sinusoidal input signals, we observe that the sinusoidal tracking error at 1/2π Hz is 0.41 mm. Finally, the experiments on tendon-sheath transmission and needle insertion performance are conducted, which show that the insertion precision is 0.68 mm in laboratory environment. © IMechE 2015.
An overview on real-time control schemes for wheeled mobile robot
NASA Astrophysics Data System (ADS)
Radzak, M. S. A.; Ali, M. A. H.; Sha’amri, S.; Azwan, A. R.
2018-04-01
The purpose of this paper is to review real-time control motion algorithms for wheeled mobile robot (WMR) when navigating in environment such as road. Its need a good controller to avoid collision with any disturbance and maintain a track error at zero level. The controllers are used with other aiding sensors to measure the WMR’s velocities, posture, and interference to estimate the required torque to be applied on the wheels of mobile robot. Four main categories for wheeled mobile robot control systems have been found in literature which are namely: Kinematic based controller, Dynamic based controllers, artificial intelligence based control system, and Active Force control. A MATLAB/Simulink software is the main software to simulate and implement the control system. The real-time toolbox in MATLAB/SIMULINK are used to receive/send data from sensors/to actuator with presence of disturbances, however others software such C, C++ and visual basic are rare to be used.
Direct adaptive robust tracking control for 6 DOF industrial robot with enhanced accuracy.
Yin, Xiuxing; Pan, Li
2018-01-01
A direct adaptive robust tracking control is proposed for trajectory tracking of 6 DOF industrial robot in the presence of parametric uncertainties, external disturbances and uncertain nonlinearities. The controller is designed based on the dynamic characteristics in the working space of the end-effector of the 6 DOF robot. The controller includes robust control term and model compensation term that is developed directly based on the input reference or desired motion trajectory. A projection-type parametric adaptation law is also designed to compensate for parametric estimation errors for the adaptive robust control. The feasibility and effectiveness of the proposed direct adaptive robust control law and the associated projection-type parametric adaptation law have been comparatively evaluated based on two 6 DOF industrial robots. The test results demonstrate that the proposed control can be employed to better maintain the desired trajectory tracking even in the presence of large parametric uncertainties and external disturbances as compared with PD controller and nonlinear controller. The parametric estimates also eventually converge to the real values along with the convergence of tracking errors, which further validate the effectiveness of the proposed parametric adaption law. Copyright © 2017 ISA. Published by Elsevier Ltd. All rights reserved.
Virtual remote center of motion control for needle placement robots.
Boctor, Emad M; Webster, Robert J; Mathieu, Herve; Okamura, Allison M; Fichtinger, Gabor
2004-01-01
We present an algorithm that enables percutaneous needle-placement procedures to be performed with unencoded, unregistered, minimally calibrated robots while removing the constraint of placing the needle tip on a mechanically enforced Remote Center of Motion (RCM). The algorithm requires only online tracking of the surgical tool and a five-degree-of-freedom (5-DOF) robot comprising three prismatic DOF and two rotational DOF. An incremental adaptive motion control cycle guides the needle to the insertion point and also orients it to align with the target-entry-point line. The robot executes RCM motion without having a physically constrained fulcrum point. The proof-of-concept prototype system achieved 0.78 mm translation accuracy and 1.4 degrees rotational accuracy (this is within the tracker accuracy) within 17 iterative steps (0.5-1 s). This research enables robotic assistant systems for image-guided percutaneous procedures to be prototyped/constructed more quickly and less expensively than has been previously possible. Since the clinical utility of such systems is clear and has been demonstrated in the literature, our work may help promote widespread clinical adoption of this technology by lowering system cost and complexity.
Fuzzy logic based robotic controller
NASA Technical Reports Server (NTRS)
Attia, F.; Upadhyaya, M.
1994-01-01
Existing Proportional-Integral-Derivative (PID) robotic controllers rely on an inverse kinematic model to convert user-specified cartesian trajectory coordinates to joint variables. These joints experience friction, stiction, and gear backlash effects. Due to lack of proper linearization of these effects, modern control theory based on state space methods cannot provide adequate control for robotic systems. In the presence of loads, the dynamic behavior of robotic systems is complex and nonlinear, especially where mathematical modeling is evaluated for real-time operators. Fuzzy Logic Control is a fast emerging alternative to conventional control systems in situations where it may not be feasible to formulate an analytical model of the complex system. Fuzzy logic techniques track a user-defined trajectory without having the host computer to explicitly solve the nonlinear inverse kinematic equations. The goal is to provide a rule-based approach, which is closer to human reasoning. The approach used expresses end-point error, location of manipulator joints, and proximity to obstacles as fuzzy variables. The resulting decisions are based upon linguistic and non-numerical information. This paper presents a solution to the conventional robot controller which is independent of computationally intensive kinematic equations. Computer simulation results of this approach as obtained from software implementation are also discussed.
2005-01-01
C. Hughes, Spacecraft Attitude Dynamics, New York, NY: Wiley, 1994. [8] H. K. Khalil, “Adaptive Output Feedback Control of Non- linear Systems...Closed-Loop Manipulator Control Using Quaternion Feedback ”, IEEE Trans. Robotics and Automation, Vol. 4, No. 4, pp. 434-440, (1988). [23] E...full-state feedback quaternion based controller de- veloped in [5] and focuses on the design of a general sub-task controller. This sub-task controller
Decentralized Adaptive Control For Robots
NASA Technical Reports Server (NTRS)
Seraji, Homayoun
1989-01-01
Precise knowledge of dynamics not required. Proposed scheme for control of multijointed robotic manipulator calls for independent control subsystem for each joint, consisting of proportional/integral/derivative feedback controller and position/velocity/acceleration feedforward controller, both with adjustable gains. Independent joint controller compensates for unpredictable effects, gravitation, and dynamic coupling between motions of joints, while forcing joints to track reference trajectories. Scheme amenable to parallel processing in distributed computing system wherein each joint controlled by relatively simple algorithm on dedicated microprocessor.
Symbiotic Navigation in Multi-Robot Systems with Remote Obstacle Knowledge Sharing
Ravankar, Abhijeet; Ravankar, Ankit A.; Kobayashi, Yukinori; Emaru, Takanori
2017-01-01
Large scale operational areas often require multiple service robots for coverage and task parallelism. In such scenarios, each robot keeps its individual map of the environment and serves specific areas of the map at different times. We propose a knowledge sharing mechanism for multiple robots in which one robot can inform other robots about the changes in map, like path blockage, or new static obstacles, encountered at specific areas of the map. This symbiotic information sharing allows the robots to update remote areas of the map without having to explicitly navigate those areas, and plan efficient paths. A node representation of paths is presented for seamless sharing of blocked path information. The transience of obstacles is modeled to track obstacles which might have been removed. A lazy information update scheme is presented in which only relevant information affecting the current task is updated for efficiency. The advantages of the proposed method for path planning are discussed against traditional method with experimental results in both simulation and real environments. PMID:28678193
A satellite orbital testbed for SATCOM using mobile robots
NASA Astrophysics Data System (ADS)
Shen, Dan; Lu, Wenjie; Wang, Zhonghai; Jia, Bin; Wang, Gang; Wang, Tao; Chen, Genshe; Blasch, Erik; Pham, Khanh
2016-05-01
This paper develops and evaluates a satellite orbital testbed (SOT) for satellite communications (SATCOM). SOT can emulate the 3D satellite orbit using the omni-wheeled robots and a robotic arm. The 3D motion of satellite is partitioned into the movements in the equatorial plane and the up-down motions in the vertical plane. The former actions are emulated by omni-wheeled robots while the up-down motions are performed by a stepped-motor-controlled-ball along a rod (robotic arm), which is attached to the robot. The emulated satellite positions will go to the measure model, whose results will be used to perform multiple space object tracking. Then the tracking results will go to the maneuver detection and collision alert. The satellite maneuver commands will be translated to robots commands and robotic arm commands. In SATCOM, the effects of jamming depend on the range and angles of the positions of satellite transponder relative to the jamming satellite. We extend the SOT to include USRP transceivers. In the extended SOT, the relative ranges and angles are implemented using omni-wheeled robots and robotic arms.
NASA Astrophysics Data System (ADS)
Massimiliano Capisani, Luca; Facchinetti, Tullio; Ferrara, Antonella
2010-08-01
This article presents the networked control of a robotic anthropomorphic manipulator based on a second-order sliding mode technique, where the control objective is to track a desired trajectory for the manipulator. The adopted control scheme allows an easy and effective distribution of the control algorithm over two networked machines. While the predictability of real-time tasks execution is achieved by the Soft Hard Real-Time Kernel (S.Ha.R.K.) real-time operating system, the communication is established via a standard Ethernet network. The performances of the control system are evaluated under different experimental system configurations using, to perform the experiments, a COMAU SMART3-S2 industrial robot, and the results are analysed to put into evidence the robustness of the proposed approach against possible network delays, packet losses and unmodelled effects.
Real-time tracking using stereo and motion: Visual perception for space robotics
NASA Technical Reports Server (NTRS)
Nishihara, H. Keith; Thomas, Hans; Huber, Eric; Reid, C. Ann
1994-01-01
The state-of-the-art in computing technology is rapidly attaining the performance necessary to implement many early vision algorithms at real-time rates. This new capability is helping to accelerate progress in vision research by improving our ability to evaluate the performance of algorithms in dynamic environments. In particular, we are becoming much more aware of the relative stability of various visual measurements in the presence of camera motion and system noise. This new processing speed is also allowing us to raise our sights toward accomplishing much higher-level processing tasks, such as figure-ground separation and active object tracking, in real-time. This paper describes a methodology for using early visual measurements to accomplish higher-level tasks; it then presents an overview of the high-speed accelerators developed at Teleos to support early visual measurements. The final section describes the successful deployment of a real-time vision system to provide visual perception for the Extravehicular Activity Helper/Retriever robotic system in tests aboard NASA's KC135 reduced gravity aircraft.
Robust dead reckoning system for mobile robots based on particle filter and raw range scan.
Duan, Zhuohua; Cai, Zixing; Min, Huaqing
2014-09-04
Robust dead reckoning is a complicated problem for wheeled mobile robots (WMRs), where the robots are faulty, such as the sticking of sensors or the slippage of wheels, for the discrete fault models and the continuous states have to be estimated simultaneously to reach a reliable fault diagnosis and accurate dead reckoning. Particle filters are one of the most promising approaches to handle hybrid system estimation problems, and they have also been widely used in many WMRs applications, such as pose tracking, SLAM, video tracking, fault identification, etc. In this paper, the readings of a laser range finder, which may be also interfered with by noises, are used to reach accurate dead reckoning. The main contribution is that a systematic method to implement fault diagnosis and dead reckoning in a particle filter framework concurrently is proposed. Firstly, the perception model of a laser range finder is given, where the raw scan may be faulty. Secondly, the kinematics of the normal model and different fault models for WMRs are given. Thirdly, the particle filter for fault diagnosis and dead reckoning is discussed. At last, experiments and analyses are reported to show the accuracy and efficiency of the presented method.
Robust Dead Reckoning System for Mobile Robots Based on Particle Filter and Raw Range Scan
Duan, Zhuohua; Cai, Zixing; Min, Huaqing
2014-01-01
Robust dead reckoning is a complicated problem for wheeled mobile robots (WMRs), where the robots are faulty, such as the sticking of sensors or the slippage of wheels, for the discrete fault models and the continuous states have to be estimated simultaneously to reach a reliable fault diagnosis and accurate dead reckoning. Particle filters are one of the most promising approaches to handle hybrid system estimation problems, and they have also been widely used in many WMRs applications, such as pose tracking, SLAM, video tracking, fault identification, etc. In this paper, the readings of a laser range finder, which may be also interfered with by noises, are used to reach accurate dead reckoning. The main contribution is that a systematic method to implement fault diagnosis and dead reckoning in a particle filter framework concurrently is proposed. Firstly, the perception model of a laser range finder is given, where the raw scan may be faulty. Secondly, the kinematics of the normal model and different fault models for WMRs are given. Thirdly, the particle filter for fault diagnosis and dead reckoning is discussed. At last, experiments and analyses are reported to show the accuracy and efficiency of the presented method. PMID:25192318
NASA Astrophysics Data System (ADS)
Ren, Wei
Cooperative control problems for multiple vehicle systems can be categorized as either formation control problems with applications to mobile robots, unmanned air vehicles, autonomous underwater vehicles, satellites, aircraft, spacecraft, and automated highway systems, or non-formation control problems such as task assignment, cooperative transport, cooperative role assignment, air traffic control, cooperative timing, and cooperative search. The cooperative control of multiple vehicle systems poses significant theoretical and practical challenges. For cooperative control strategies to be successful, numerous issues must be addressed. We consider three important and correlated issues: consensus seeking, formation keeping, and trajectory tracking. For consensus seeking, we investigate algorithms and protocols so that a team of vehicles can reach consensus on the values of the coordination data in the presence of imperfect sensors, communication dropout, sparse communication topologies, and noisy and unreliable communication links. The main contribution of this dissertation in this area is that we show necessary and/or sufficient conditions for consensus seeking with limited, unidirectional, and unreliable information exchange under fixed and switching interaction topologies (through either communication or sensing). For formation keeping, we apply a so-called "virtual structure" approach to spacecraft formation flying and multi-vehicle formation maneuvers. As a result, single vehicle path planning and trajectory generation techniques can be employed for the virtual structure while trajectory tracking strategies can be employed for each vehicle. The main contribution of this dissertation in this area is that we propose a decentralized architecture for multiple spacecraft formation flying in deep space with formation feedback introduced. This architecture ensures the necessary precision in the presence of actuator saturation, internal and external disturbances, and stringent inter-vehicle communication limitations. A constructive approach based on the satisficing control paradigm is also applied to multi-robot coordination in hardware. For trajectory tracking, we investigate nonlinear tracking controllers for fixed wing unmanned air vehicles and nonholonomic mobile robots with velocity and heading rate constraints. The main contribution of this dissertation in this area is that our proposed tracking controllers are shown to be robust to input uncertainties and measurement noise, and are computationally simple and can be implemented with low-cost, low-power microcontrollers. In addition, our approach allows piecewise continuous reference velocity and heading rate and can be extended to derive a variety of other trajectory tracking strategies.
Motion reconstruction of animal groups: From schooling fish to swarming mosquitoes
NASA Astrophysics Data System (ADS)
Butail, Sachit
The long-term goal of this research is to provide kinematic data for the design and validation of spatial models of collective behavior in animal groups. The specific research objective of this dissertation is to apply methods from nonlinear estimation and computer vision to construct multi-target tracking systems that process multi-view calibrated video to reconstruct the three-dimensional movement of animals in a group. We adapt the tracking systems for the study of two animal species: Danio aequipinnatus, a common species of schooling fish, and Anopheles gambiae, the most important vector of malaria in sub-Saharan Africa. Together these tracking systems span variability in target size on image, density, and movement. For tracking fish, we automatically initialize, predict, and reconstruct shape trajectories of multiple fish through occlusions. For mosquitoes, which appear as faded streaks on in-field footage, we provide methods to extract velocity information from the streaks, adaptively seek missing measurements, and resolve occlusions within a multi-hypothesis framework. In each case the research has yielded an unprecedented volume of trajectory data for subsequent analysis. We present kinematic data of fast-start response in fish schools and first-ever trajectories of wild mosquito swarming and mating events. The broader impact of this work is to advance the understanding of animal groups for the design of bio-inspired robotic systems, where, similar to the animal groups we study, the collective is able to perform tasks far beyond the capabilities of a single inexpensive robot.
Surrogate: A Body-Dexterous Mobile Manipulation Robot with a Tracked Base
NASA Technical Reports Server (NTRS)
Hebert, Paul (Inventor); Borders, James W. (Inventor); Hudson, Nicolas H. (Inventor); Kennedy, Brett A. (Inventor); Ma, Jeremy C. (Inventor); Bergh, Charles F. (Inventor)
2018-01-01
Robotics platforms in accordance with various embodiments of the invention can be utilized to implement highly dexterous robots capable of whole body motion. Robotics platforms in accordance with one embodiment of the invention include: a memory containing a whole body motion application; a spine, where the spine has seven degrees of freedom and comprises a spine actuator and three spine elbow joints that each include two spine joint actuators; at least one limb, where the at least one limb comprises a limb actuator and three limb elbow joints that each include two limb joint actuators; a tracked base; a connecting structure that connects the at least one limb to the spine; a second connecting structure that connects the spine to the tracked base; wherein the processor is configured by the whole body motion application to move the at least one limb and the spine to perform whole body motion.
Optimal critic learning for robot control in time-varying environments.
Wang, Chen; Li, Yanan; Ge, Shuzhi Sam; Lee, Tong Heng
2015-10-01
In this paper, optimal critic learning is developed for robot control in a time-varying environment. The unknown environment is described as a linear system with time-varying parameters, and impedance control is employed for the interaction control. Desired impedance parameters are obtained in the sense of an optimal realization of the composite of trajectory tracking and force regulation. Q -function-based critic learning is developed to determine the optimal impedance parameters without the knowledge of the system dynamics. The simulation results are presented and compared with existing methods, and the efficacy of the proposed method is verified.
Global Output-Feedback Control for Simultaneous Tracking and Stabilization of Wheeled Mobile Robots
NASA Astrophysics Data System (ADS)
Chang, J.; Zhang, L. J.; Xue, D.
A time-varying global output-feedback controller is presented that solves both tracking and stabilization for wheeled mobile robots simultaneously at the torque level. The controller synthesis is based on a coordinate transformation, Lyapunov direct method and backstepping technique. The performance of the proposed controller is demonstrated by simulation.
McMullen, David P.; Hotson, Guy; Katyal, Kapil D.; Wester, Brock A.; Fifer, Matthew S.; McGee, Timothy G.; Harris, Andrew; Johannes, Matthew S.; Vogelstein, R. Jacob; Ravitz, Alan D.; Anderson, William S.; Thakor, Nitish V.; Crone, Nathan E.
2014-01-01
To increase the ability of brain-machine interfaces (BMIs) to control advanced prostheses such as the modular prosthetic limb (MPL), we are developing a novel system: the Hybrid Augmented Reality Multimodal Operation Neural Integration Environment (HARMONIE). This system utilizes hybrid input, supervisory control, and intelligent robotics to allow users to identify an object (via eye tracking and computer vision) and initiate (via brain-control) a semi-autonomous reach-grasp-and-drop of the object by the MPL. Sequential iterations of HARMONIE were tested in two pilot subjects implanted with electrocorticographic (ECoG) and depth electrodes within motor areas. The subjects performed the complex task in 71.4% (20/28) and 67.7% (21/31) of trials after minimal training. Balanced accuracy for detecting movements was 91.1% and 92.9%, significantly greater than chance accuracies (p < 0.05). After BMI-based initiation, the MPL completed the entire task 100% (one object) and 70% (three objects) of the time. The MPL took approximately 12.2 seconds for task completion after system improvements implemented for the second subject. Our hybrid-BMI design prevented all but one baseline false positive from initiating the system. The novel approach demonstrated in this proof-of-principle study, using hybrid input, supervisory control, and intelligent robotics, addresses limitations of current BMIs. PMID:24760914
McMullen, David P; Hotson, Guy; Katyal, Kapil D; Wester, Brock A; Fifer, Matthew S; McGee, Timothy G; Harris, Andrew; Johannes, Matthew S; Vogelstein, R Jacob; Ravitz, Alan D; Anderson, William S; Thakor, Nitish V; Crone, Nathan E
2014-07-01
To increase the ability of brain-machine interfaces (BMIs) to control advanced prostheses such as the modular prosthetic limb (MPL), we are developing a novel system: the Hybrid Augmented Reality Multimodal Operation Neural Integration Environment (HARMONIE). This system utilizes hybrid input, supervisory control, and intelligent robotics to allow users to identify an object (via eye tracking and computer vision) and initiate (via brain-control) a semi-autonomous reach-grasp-and-drop of the object by the MPL. Sequential iterations of HARMONIE were tested in two pilot subjects implanted with electrocorticographic (ECoG) and depth electrodes within motor areas. The subjects performed the complex task in 71.4% (20/28) and 67.7% (21/31) of trials after minimal training. Balanced accuracy for detecting movements was 91.1% and 92.9%, significantly greater than chance accuracies (p < 0.05). After BMI-based initiation, the MPL completed the entire task 100% (one object) and 70% (three objects) of the time. The MPL took approximately 12.2 s for task completion after system improvements implemented for the second subject. Our hybrid-BMI design prevented all but one baseline false positive from initiating the system. The novel approach demonstrated in this proof-of-principle study, using hybrid input, supervisory control, and intelligent robotics, addresses limitations of current BMIs.
NASA Astrophysics Data System (ADS)
Mineo, Carmelo; MacLeod, Charles; Morozov, Maxim; Pierce, S. Gareth; Summan, Rahul; Rodden, Tony; Kahani, Danial; Powell, Jonathan; McCubbin, Paul; McCubbin, Coreen; Munro, Gavin; Paton, Scott; Watson, David
2017-02-01
Improvements in performance of modern robotic manipulators have in recent years allowed research aimed at development of fast automated non-destructive testing (NDT) of complex geometries. Contemporary robots are well adaptable to new tasks. Several robotic inspection prototype systems and a number of commercial products have been developed worldwide. This paper describes the latest progress in research focused at large composite aerospace components. A multi-robot flexible inspection cell is used to take the fundamental research and the feasibility studies to higher technology readiness levels, all set for the future industrial exploitation. The robot cell is equipped with high accuracy and high payload robots, mounted on 7 meter tracks, and an external rotary axis. A robotically delivered photogrammetry technique is first used to assess the position of the components placed within the robot working envelope and their deviation to CAD. Offline programming is used to generate a scan path for phased array ultrasonic testing (PAUT). PAUT is performed using a conformable wheel probe, with high data rate acquisition from PAUT controller. Real-time robot path-correction, based on force-torque control (FTC), is deployed to achieve the optimum ultrasonic coupling and repeatable data quality. New communication software is developed that enabled simultaneous control of the multiple robots performing different tasks and the acquisition of accurate positional data. All aspects of the system are controlled through a purposely developed graphic user interface that enables the flexible use of the unique set of hardware resources, the data acquisition, visualization and analysis.
Sensor control of robot arc welding
NASA Technical Reports Server (NTRS)
Sias, F. R., Jr.
1985-01-01
A basic problem in the application of robots for welding which is how to guide a torch along a weld seam using sensory information was studied. Improvement of the quality and consistency of certain Gas Tungsten Arc welds on the Space Shuttle Main Engine (SSME) that are too complex geometrically for conventional automation and therefore are done by hand was examined. The particular problems associated with space shuttle main egnine (SSME) manufacturing and weld-seam tracking with an emphasis on computer vision methods were analyzed. Special interface software for the MINC computr are developed which will allow it to be used both as a test system to check out the robot interface software and later as a development tool for further investigation of sensory systems to be incorporated in welding procedures.
Robust stability of second-order systems
NASA Technical Reports Server (NTRS)
Chuang, C.-H.
1993-01-01
A feedback linearization technique is used in conjunction with passivity concepts to design robust controllers for space robots. It is assumed that bounded modeling uncertainties exist in the inertia matrix and the vector representing the coriolis, centripetal, and friction forces. Under these assumptions, the controller guarantees asymptotic tracking of the joint variables. A Lagrangian approach is used to develop a dynamic model for space robots. Closed-loop simulation results are illustrated for a simple case of a single link planar manipulator with freely floating base.
On the reproducibility of expert-operated and robotic ultrasound acquisitions.
Kojcev, Risto; Khakzar, Ashkan; Fuerst, Bernhard; Zettinig, Oliver; Fahkry, Carole; DeJong, Robert; Richmon, Jeremy; Taylor, Russell; Sinibaldi, Edoardo; Navab, Nassir
2017-06-01
We present the evaluation of the reproducibility of measurements performed using robotic ultrasound imaging in comparison with expert-operated sonography. Robotic imaging for interventional procedures may be a valuable contribution, but requires reproducibility for its acceptance in clinical routine. We study this by comparing repeated measurements based on robotic and expert-operated ultrasound imaging. Robotic ultrasound acquisition is performed in three steps under user guidance: First, the patient is observed using a 3D camera on the robot end effector, and the user selects the region of interest. This allows for automatic planning of the robot trajectory. Next, the robot executes a sweeping motion following the planned trajectory, during which the ultrasound images and tracking data are recorded. As the robot is compliant, deviations from the path are possible, for instance due to patient motion. Finally, the ultrasound slices are compounded to create a volume. Repeated acquisitions can be performed automatically by comparing the previous and current patient surface. After repeated image acquisitions, the measurements based on acquisitions performed by the robotic system and expert are compared. Within our case series, the expert measured the anterior-posterior, longitudinal, transversal lengths of both of the left and right thyroid lobes on each of the 4 healthy volunteers 3 times, providing 72 measurements. Subsequently, the same procedure was performed using the robotic system resulting in a cumulative total of 144 clinically relevant measurements. Our results clearly indicated that robotic ultrasound enables more repeatable measurements. A robotic ultrasound platform leads to more reproducible data, which is of crucial importance for planning and executing interventions.
Real-time seam tracking control system based on line laser visions
NASA Astrophysics Data System (ADS)
Zou, Yanbiao; Wang, Yanbo; Zhou, Weilin; Chen, Xiangzhi
2018-07-01
A set of six-degree-of-freedom robotic welding automatic tracking platform was designed in this study to realize the real-time tracking of weld seams. Moreover, the feature point tracking method and the adaptive fuzzy control algorithm in the welding process were studied and analyzed. A laser vision sensor and its measuring principle were designed and studied, respectively. Before welding, the initial coordinate values of the feature points were obtained using morphological methods. After welding, the target tracking method based on Gaussian kernel was used to extract the real-time feature points of the weld. An adaptive fuzzy controller was designed to input the deviation value of the feature points and the change rate of the deviation into the controller. The quantization factors, scale factor, and weight function were adjusted in real time. The input and output domains, fuzzy rules, and membership functions were constantly updated to generate a series of smooth bias robot voltage. Three groups of experiments were conducted on different types of curve welds in a strong arc and splash noise environment using the welding current of 120 A short-circuit Metal Active Gas (MAG) Arc Welding. The tracking error was less than 0.32 mm and the sensor's metrical frequency can be up to 20 Hz. The end of the torch run smooth during welding. Weld trajectory can be tracked accurately, thereby satisfying the requirements of welding applications.
Autonomous Aerial Refueling Ground Test Demonstration—A Sensor-in-the-Loop, Non-Tracking Method
Chen, Chao-I; Koseluk, Robert; Buchanan, Chase; Duerner, Andrew; Jeppesen, Brian; Laux, Hunter
2015-01-01
An essential capability for an unmanned aerial vehicle (UAV) to extend its airborne duration without increasing the size of the aircraft is called the autonomous aerial refueling (AAR). This paper proposes a sensor-in-the-loop, non-tracking method for probe-and-drogue style autonomous aerial refueling tasks by combining sensitivity adjustments of a 3D Flash LIDAR camera with computer vision based image-processing techniques. The method overcomes the inherit ambiguity issues when reconstructing 3D information from traditional 2D images by taking advantage of ready to use 3D point cloud data from the camera, followed by well-established computer vision techniques. These techniques include curve fitting algorithms and outlier removal with the random sample consensus (RANSAC) algorithm to reliably estimate the drogue center in 3D space, as well as to establish the relative position between the probe and the drogue. To demonstrate the feasibility of the proposed method on a real system, a ground navigation robot was designed and fabricated. Results presented in the paper show that using images acquired from a 3D Flash LIDAR camera as real time visual feedback, the ground robot is able to track a moving simulated drogue and continuously narrow the gap between the robot and the target autonomously. PMID:25970254
Indirect iterative learning control for a discrete visual servo without a camera-robot model.
Jiang, Ping; Bamforth, Leon C A; Feng, Zuren; Baruch, John E F; Chen, YangQuan
2007-08-01
This paper presents a discrete learning controller for vision-guided robot trajectory imitation with no prior knowledge of the camera-robot model. A teacher demonstrates a desired movement in front of a camera, and then, the robot is tasked to replay it by repetitive tracking. The imitation procedure is considered as a discrete tracking control problem in the image plane, with an unknown and time-varying image Jacobian matrix. Instead of updating the control signal directly, as is usually done in iterative learning control (ILC), a series of neural networks are used to approximate the unknown Jacobian matrix around every sample point in the demonstrated trajectory, and the time-varying weights of local neural networks are identified through repetitive tracking, i.e., indirect ILC. This makes repetitive segmented training possible, and a segmented training strategy is presented to retain the training trajectories solely within the effective region for neural network approximation. However, a singularity problem may occur if an unmodified neural-network-based Jacobian estimation is used to calculate the robot end-effector velocity. A new weight modification algorithm is proposed which ensures invertibility of the estimation, thus circumventing the problem. Stability is further discussed, and the relationship between the approximation capability of the neural network and the tracking accuracy is obtained. Simulations and experiments are carried out to illustrate the validity of the proposed controller for trajectory imitation of robot manipulators with unknown time-varying Jacobian matrices.
Line following using a two camera guidance system for a mobile robot
NASA Astrophysics Data System (ADS)
Samu, Tayib; Kelkar, Nikhal; Perdue, David; Ruthemeyer, Michael A.; Matthews, Bradley O.; Hall, Ernest L.
1996-10-01
Automated unmanned guided vehicles have many potential applications in manufacturing, medicine, space and defense. A mobile robot has been designed for the 1996 Automated Unmanned Vehicle Society competition which was held in Orlando, Florida on July 15, 1996. The competition required the vehicle to follow solid and dashed lines around an approximately 800 ft. path while avoiding obstacles, overcoming terrain changes such as inclines and sand traps, and attempting to maximize speed. The purpose of this paper is to describe the algorithm developed for the line following. The line following algorithm images two windows and locates their centroid and with the knowledge that the points are on the ground plane, a mathematical and geometrical relationship between the image coordinates of the points and their corresponding ground coordinates are established. The angle of the line and minimum distance from the robot centroid are then calculated and used in the steering control. Two cameras are mounted on the robot with a camera on each side. One camera guides the robot and when it loses track of the line on its side, the robot control system automatically switches to the other camera. The test bed system has provided an educational experience for all involved and permits understanding and extending the state of the art in autonomous vehicle design.
Noncoherent Tracking of a Source of a Data-Modulated Signal
NASA Technical Reports Server (NTRS)
Arndt, G. Dickey; Ngo, Phong; Chen, Henry; Phan, Chau T.; Hill, Brent; Bourgeois, Brian; Dusl, John
2003-01-01
A proposed tracking receiver system containing three suitably positioned antenna elements and special signal-processing equipment would determine the direction of incidence of a microwave signal containing spread-spectrum digital data modulation. If the system were to contain two sets of antenna elements separated by a known baseline, it could determine the location of the transmitter as the intersection of the lines of incidence on the two antennas. Such systems could be used for diverse purposes in outer space and on Earth, including tracking astronauts and small robotic spacecraft working outside a spacecraft or space station, and locating cellular telephones from which distress calls have been made. The principle of operation does not require the transmission of a special identifying or distress signal by the cellular telephone or other transmitter to be tracked; instead, the system could utilize the data signal routinely sent by the transmitter, provided that the signal had the characteristics needed for processing.
Estimating Position of Mobile Robots From Omnidirectional Vision Using an Adaptive Algorithm.
Li, Luyang; Liu, Yun-Hui; Wang, Kai; Fang, Mu
2015-08-01
This paper presents a novel and simple adaptive algorithm for estimating the position of a mobile robot with high accuracy in an unknown and unstructured environment by fusing images of an omnidirectional vision system with measurements of odometry and inertial sensors. Based on a new derivation where the omnidirectional projection can be linearly parameterized by the positions of the robot and natural feature points, we propose a novel adaptive algorithm, which is similar to the Slotine-Li algorithm in model-based adaptive control, to estimate the robot's position by using the tracked feature points in image sequence, the robot's velocity, and orientation angles measured by odometry and inertial sensors. It is proved that the adaptive algorithm leads to global exponential convergence of the position estimation errors to zero. Simulations and real-world experiments are performed to demonstrate the performance of the proposed algorithm.
Evaluation of microsurgical tasks with OCT-guided and/or robot-assisted ophthalmic forceps
Yu, Haoran; Shen, Jin-Hui; Shah, Rohan J.; Simaan, Nabil; Joos, Karen M.
2015-01-01
Real-time intraocular optical coherence tomography (OCT) visualization of tissues with surgical feedback can enhance retinal surgery. An intraocular 23-gauge B-mode forward-imaging co-planar OCT-forceps, coupling connectors and algorithms were developed to form a unique ophthalmic surgical robotic system. Approach to the surface of a phantom or goat retina by a manual or robotic-controlled forceps, with and without real-time OCT guidance, was performed. Efficiency of lifting phantom membranes was examined. Placing the co-planar OCT imaging probe internal to the surgical tool reduced instrument shadowing and permitted constant tracking. Robotic assistance together with real-time OCT feedback improved depth perception accuracy. The first-generation integrated OCT-forceps was capable of peeling membrane phantoms despite smooth tips. PMID:25780736
Comparison of three different techniques for camera and motion control of a teleoperated robot.
Doisy, Guillaume; Ronen, Adi; Edan, Yael
2017-01-01
This research aims to evaluate new methods for robot motion control and camera orientation control through the operator's head orientation in robot teleoperation tasks. Specifically, the use of head-tracking in a non-invasive way, without immersive virtual reality devices was combined and compared with classical control modes for robot movements and camera control. Three control conditions were tested: 1) a condition with classical joystick control of both the movements of the robot and the robot camera, 2) a condition where the robot movements were controlled by a joystick and the robot camera was controlled by the user head orientation, and 3) a condition where the movements of the robot were controlled by hand gestures and the robot camera was controlled by the user head orientation. Performance, workload metrics and their evolution as the participants gained experience with the system were evaluated in a series of experiments: for each participant, the metrics were recorded during four successive similar trials. Results shows that the concept of robot camera control by user head orientation has the potential of improving the intuitiveness of robot teleoperation interfaces, specifically for novice users. However, more development is needed to reach a margin of progression comparable to a classical joystick interface. Copyright © 2016 Elsevier Ltd. All rights reserved.
Reference respiratory waveforms by minimum jerk model analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Anetai, Yusuke, E-mail: anetai@radonc.med.osaka-u.ac.jp; Sumida, Iori; Takahashi, Yutaka
Purpose: CyberKnife{sup ®} robotic surgery system has the ability to deliver radiation to a tumor subject to respiratory movements using Synchrony{sup ®} mode with less than 2 mm tracking accuracy. However, rapid and rough motion tracking causes mechanical tracking errors and puts mechanical stress on the robotic joint, leading to unexpected radiation delivery errors. During clinical treatment, patient respiratory motions are much more complicated, suggesting the need for patient-specific modeling of respiratory motion. The purpose of this study was to propose a novel method that provides a reference respiratory wave to enable smooth tracking for each patient. Methods: The minimummore » jerk model, which mathematically derives smoothness by means of jerk, or the third derivative of position and the derivative of acceleration with respect to time that is proportional to the time rate of force changed was introduced to model a patient-specific respiratory motion wave to provide smooth motion tracking using CyberKnife{sup ®}. To verify that patient-specific minimum jerk respiratory waves were being tracked smoothly by Synchrony{sup ®} mode, a tracking laser projection from CyberKnife{sup ®} was optically analyzed every 0.1 s using a webcam and a calibrated grid on a motion phantom whose motion was in accordance with three pattern waves (cosine, typical free-breathing, and minimum jerk theoretical wave models) for the clinically relevant superior–inferior directions from six volunteers assessed on the same node of the same isocentric plan. Results: Tracking discrepancy from the center of the grid to the beam projection was evaluated. The minimum jerk theoretical wave reduced the maximum-peak amplitude of radial tracking discrepancy compared with that of the waveforms modeled by cosine and typical free-breathing model by 22% and 35%, respectively, and provided smooth tracking for radial direction. Motion tracking constancy as indicated by radial tracking discrepancy affected by respiratory phase was improved in the minimum jerk theoretical model by 7.0% and 13% compared with that of the waveforms modeled by cosine and free-breathing model, respectively. Conclusions: The minimum jerk theoretical respiratory wave can achieve smooth tracking by CyberKnife{sup ®} and may provide patient-specific respiratory modeling, which may be useful for respiratory training and coaching, as well as quality assurance of the mechanical CyberKnife{sup ®} robotic trajectory.« less
Robot Position Sensor Fault Tolerance
NASA Technical Reports Server (NTRS)
Aldridge, Hal A.
1997-01-01
Robot systems in critical applications, such as those in space and nuclear environments, must be able to operate during component failure to complete important tasks. One failure mode that has received little attention is the failure of joint position sensors. Current fault tolerant designs require the addition of directly redundant position sensors which can affect joint design. A new method is proposed that utilizes analytical redundancy to allow for continued operation during joint position sensor failure. Joint torque sensors are used with a virtual passive torque controller to make the robot joint stable without position feedback and improve position tracking performance in the presence of unknown link dynamics and end-effector loading. Two Cartesian accelerometer based methods are proposed to determine the position of the joint. The joint specific position determination method utilizes two triaxial accelerometers attached to the link driven by the joint with the failed position sensor. The joint specific method is not computationally complex and the position error is bounded. The system wide position determination method utilizes accelerometers distributed on different robot links and the end-effector to determine the position of sets of multiple joints. The system wide method requires fewer accelerometers than the joint specific method to make all joint position sensors fault tolerant but is more computationally complex and has lower convergence properties. Experiments were conducted on a laboratory manipulator. Both position determination methods were shown to track the actual position satisfactorily. A controller using the position determination methods and the virtual passive torque controller was able to servo the joints to a desired position during position sensor failure.
Recent testing of a micro autonomous positioning system for multi-object instrumentation
NASA Astrophysics Data System (ADS)
Cochrane, W. A.; Atkinson, D. C.; Bailie, T. E. C.; Dickson, C.; Lim, T.; Luo, X.; Montgomery, D. M.; Schnetler, H.; Taylor, W. D.; Wilson, B.
2012-09-01
A multiple pick off mirror positioning sub-system has been developed as a solution for the deployment of mirrors within multi-object instrumentation such as the EAGLE instrument in the European Extremely Large Telescope (E-ELT). The positioning sub-system is a two wheeled differential steered friction drive robot with a footprint of approximately 20 x 20 mm. Controlled by RF communications there are two versions of the robot that exist. One is powered by a single cell lithium ion battery and the other utilises a power floor system. The robots use two brushless DC motors with 125:1 planetary gear heads for positioning in the coarse drive stages. A unique power floor allows the robots to be positioned at any location in any orientation on the focal plane. The design, linear repeatability tests, metrology and power continuity of the robot will be evaluated and presented in this paper. To gather photons from the objects of interest it is important to position POMs within a sphere of confusion of less than 10 μm, with an angular alignment better than 1 mrad. The robots potential of meeting these requirements will be described through the open-loop repeatability tests conducted with a Faro laser beam tracker. Tests have involved sending the robot step commands and automatically taking continuous measurements every three seconds. Currently the robot is capable of repeatedly travelling 233 mm within 0.307 mm at 5 mm/s. An analysis of the power floors reliability through the continuous monitoring of the voltage across the tracks with a Pico logger will also be presented.
Study of the Navigation Method for a Snake Robot Based on the Kinematics Model with MEMS IMU.
Zhao, Xu; Dou, Lihua; Su, Zhong; Liu, Ning
2018-03-16
A snake robot is a type of highly redundant mobile robot that significantly differs from a tracked robot, wheeled robot and legged robot. To address the issue of a snake robot performing self-localization in the application environment without assistant orientation, an autonomous navigation method is proposed based on the snake robot's motion characteristic constraints. The method realized the autonomous navigation of the snake robot with non-nodes and an external assistant using its own Micro-Electromechanical-Systems (MEMS) Inertial-Measurement-Unit (IMU). First, it studies the snake robot's motion characteristics, builds the kinematics model, and then analyses the motion constraint characteristics and motion error propagation properties. Second, it explores the snake robot's navigation layout, proposes a constraint criterion and the fixed relationship, and makes zero-state constraints based on the motion features and control modes of a snake robot. Finally, it realizes autonomous navigation positioning based on the Extended-Kalman-Filter (EKF) position estimation method under the constraints of its motion characteristics. With the self-developed snake robot, the test verifies the proposed method, and the position error is less than 5% of Total-Traveled-Distance (TDD). In a short-distance environment, this method is able to meet the requirements of a snake robot in order to perform autonomous navigation and positioning in traditional applications and can be extended to other familiar multi-link robots.
Regulation and Entrainment in Human-Robot Interaction
2000-01-01
applications for domestic, health care related, or entertainment based robots motivate the development of robots that can socially interact with, learn...picture shows WE-3RII, an expressive face robot developed at Waseda University. The middle right picture shows Robita, an upper-torso robot also... developed at Waseda University to track speaking turns. The far right picture shows our expressive robot, Kismet, developed at MIT. The two leftmost photos
Evaluation of a Home Biomonitoring Autonomous Mobile Robot.
Dorronzoro Zubiete, Enrique; Nakahata, Keigo; Imamoglu, Nevrez; Sekine, Masashi; Sun, Guanghao; Gomez, Isabel; Yu, Wenwei
2016-01-01
Increasing population age demands more services in healthcare domain. It has been shown that mobile robots could be a potential solution to home biomonitoring for the elderly. Through our previous studies, a mobile robot system that is able to track a subject and identify his daily living activities has been developed. However, the system has not been tested in any home living scenarios. In this study we did a series of experiments to investigate the accuracy of activity recognition of the mobile robot in a home living scenario. The daily activities tested in the evaluation experiment include watching TV and sleeping. A dataset recorded by a distributed distance-measuring sensor network was used as a reference to the activity recognition results. It was shown that the accuracy is not consistent for all the activities; that is, mobile robot could achieve a high success rate in some activities but a poor success rate in others. It was found that the observation position of the mobile robot and subject surroundings have high impact on the accuracy of the activity recognition, due to the variability of the home living daily activities and their transitional process. The possibility of improvement of recognition accuracy has been shown too.
Bergamasco, Massimo; Frisoli, Antonio; Fontana, Marco; Loconsole, Claudio; Leonardis, Daniele; Troncossi, Marco; Foumashi, Mohammad Mozaffari; Parenti-Castelli, Vincenzo
2011-01-01
This paper presents the preliminary results of the project BRAVO (Brain computer interfaces for Robotic enhanced Action in Visuo-motOr tasks). The objective of this project is to define a new approach to the development of assistive and rehabilitative robots for motor impaired users to perform complex visuomotor tasks that require a sequence of reaches, grasps and manipulations of objects. BRAVO aims at developing new robotic interfaces and HW/SW architectures for rehabilitation and regain/restoration of motor function in patients with upper limb sensorimotor impairment through extensive rehabilitation therapy and active assistance in the execution of Activities of Daily Living. The final system developed within this project will include a robotic arm exoskeleton and a hand orthosis that will be integrated together for providing force assistance. The main novelty that BRAVO introduces is the control of the robotic assistive device through the active prediction of intention/action. The system will actually integrate the information about the movement carried out by the user with a prediction of the performed action through an interpretation of current gaze of the user (measured through eye-tracking), brain activation (measured through BCI) and force sensor measurements. © 2011 IEEE
International Assessment of Unmanned Ground Vehicles
2008-02-01
research relevant to ground robotics include • Multi-sensor data fusion • Stereovision • Dedicated robots, including legged robots, tracked robots...Technology Laboratory has developed several mobile robots with leg - ged, wheeled, rolling, rowing, and hybrid locomotion. Areas of particular emphasis...117 UK Department of Trade and Industry ( DTI ) Global Watch Mission. November 2006. Mechatronics in Russia. 118 CRDI Web Site: http
DOE Office of Scientific and Technical Information (OSTI.GOV)
Neely, Jason C.; Sturgis, Beverly Rainwater; Byrne, Raymond Harry
This report contains the results of a research effort on advanced robot locomotion. The majority of this work focuses on walking robots. Walking robot applications include delivery of special payloads to unique locations that require human locomotion to exo-skeleton human assistance applications. A walking robot could step over obstacles and move through narrow openings that a wheeled or tracked vehicle could not overcome. It could pick up and manipulate objects in ways that a standard robot gripper could not. Most importantly, a walking robot would be able to rapidly perform these tasks through an intuitive user interface that mimics naturalmore » human motion. The largest obstacle arises in emulating stability and balance control naturally present in humans but needed for bipedal locomotion in a robot. A tracked robot is bulky and limited, but a wide wheel base assures passive stability. Human bipedal motion is so common that it is taken for granted, but bipedal motion requires active balance and stability control for which the analysis is non-trivial. This report contains an extensive literature study on the state-of-the-art of legged robotics, and it additionally provides the analysis, simulation, and hardware verification of two variants of a proto-type leg design.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Don, S; Cormack, R; Viswanathan, A
Purpose: To present a programmable robotic system for the accurate and fast deployment of an electromagnetic (EM) sensor for brachytherapy catheter localization. Methods: A robotic system for deployment of an EM sensor was designed and built. The system was programmed to increment the sensor position at specified time and space intervals. Sensor delivery accuracy was measured in a phantom using the localization of the EM sensor and tested in different environmental conditions. Accuracy was tested by measuring the distance between the physical locations reached by the sensor (measured by the EM tracker) and the intended programmed locations. Results: The systemmore » consisted of a stepper motor connected to drive wheels (that grip the cable to move the sensor) and a series of guides to connect to a brachytherapy transfer tube, all controlled by a programmable Arduino microprocessor. The total cost for parts was <$300. The positional accuracy of the sensor location was within 1 mm of the expected position provided by the motorized guide system. Acquisition speed to localize a brachytherapy catheter with 20 cm of active length was 10 seconds. The current design showed some cable slip and warping depending on environment temperature. Conclusion: The use of EM tracking for the localization of brachytherapy catheters has been previously demonstrated. Efficient data acquisition and artifact reduction requires fast and accurate deployment of an EM sensor in consistent, repeatable patterns, which cannot practically be achieved manually. The design of an inexpensive, programmable robot allowing for the precise deployment of stepping patterns was presented, and a prototype was built. Further engineering is necessary to ensure that the device provides efficient independent localization of brachytherapy catheters. This research was funded by the Kaye Family Award.« less
Teixidó, Mercè; Font, Davinia; Pallejà, Tomàs; Tresanchez, Marcel; Nogués, Miquel; Palacín, Jordi
2012-10-22
This work proposes the development of an embedded real-time fruit detection system for future automatic fruit harvesting. The proposed embedded system is based on an ARM Cortex-M4 (STM32F407VGT6) processor and an Omnivision OV7670 color camera. The future goal of this embedded vision system will be to control a robotized arm to automatically select and pick some fruit directly from the tree. The complete embedded system has been designed to be placed directly in the gripper tool of the future robotized harvesting arm. The embedded system will be able to perform real-time fruit detection and tracking by using a three-dimensional look-up-table (LUT) defined in the RGB color space and optimized for fruit picking. Additionally, two different methodologies for creating optimized 3D LUTs based on existing linear color models and fruit histograms were implemented in this work and compared for the case of red peaches. The resulting system is able to acquire general and zoomed orchard images and to update the relative tracking information of a red peach in the tree ten times per second.
Teixidó, Mercè; Font, Davinia; Pallejà, Tomàs; Tresanchez, Marcel; Nogués, Miquel; Palacín, Jordi
2012-01-01
This work proposes the development of an embedded real-time fruit detection system for future automatic fruit harvesting. The proposed embedded system is based on an ARM Cortex-M4 (STM32F407VGT6) processor and an Omnivision OV7670 color camera. The future goal of this embedded vision system will be to control a robotized arm to automatically select and pick some fruit directly from the tree. The complete embedded system has been designed to be placed directly in the gripper tool of the future robotized harvesting arm. The embedded system will be able to perform real-time fruit detection and tracking by using a three-dimensional look-up-table (LUT) defined in the RGB color space and optimized for fruit picking. Additionally, two different methodologies for creating optimized 3D LUTs based on existing linear color models and fruit histograms were implemented in this work and compared for the case of red peaches. The resulting system is able to acquire general and zoomed orchard images and to update the relative tracking information of a red peach in the tree ten times per second. PMID:23202040
Control of a HexaPOD treatment couch for robot-assisted radiotherapy.
Hermann, Christian; Ma, Lei; Wilbert, Jürgen; Baier, Kurt; Schilling, Klaus
2012-10-01
Moving tumors, for example in the vicinity of the lungs, pose a challenging problem in radiotherapy, as healthy tissue should not be irradiated. Apart from gating approaches, one standard method is to irradiate the complete volume within which a tumor moves plus a safety margin containing a considerable volume of healthy tissue. This work deals with a system for tumor motion compensation using the HexaPOD® robotic treatment couch (Medical Intelligence GmbH, Schwabmünchen, Germany). The HexaPOD, carrying the patient during treatment, is instructed to perform translational movements such that the tumor motion, from the beams-eye view of the linear accelerator, is eliminated. The dynamics of the HexaPOD are characterized by time delays, saturations, and other non-linearities that make the design of control a challenging task. The focus of this work lies on two control methods for the HexaPOD that can be used for reference tracking. The first method uses a model predictive controller based on a model gained through system identification methods, and the second method uses a position control scheme useful for reference tracking. We compared the tracking performance of both methods in various experiments with real hardware using ideal reference trajectories, prerecorded patient trajectories, and human volunteers whose breathing motion was compensated by the system.
Real-time intra-fraction-motion tracking using the treatment couch: a feasibility study
NASA Astrophysics Data System (ADS)
D'Souza, Warren D.; Naqvi, Shahid A.; Yu, Cedric X.
2005-09-01
Significant differences between planned and delivered treatments may occur due to respiration-induced tumour motion, leading to underdosing of parts of the tumour and overdosing of parts of the surrounding critical structures. Existing methods proposed to counter tumour motion include breath-holds, gating and MLC-based tracking. Breath-holds and gating techniques increase treatment time considerably, whereas MLC-based tracking is limited to two dimensions. We present an alternative solution in which a robotic couch moves in real time in response to organ motion. To demonstrate proof-of-principle, we constructed a miniature adaptive couch model consisting of two movable platforms that simulate tumour motion and couch motion, respectively. These platforms were connected via an electronic feedback loop so that the bottom platform responded to the motion of the top platform. We tested our model with a seven-field step-and-shoot delivery case in which we performed three film-based experiments: (1) static geometry, (2) phantom-only motion and (3) phantom motion with simulated couch motion. Our measurements demonstrate that the miniature couch was able to compensate for phantom motion to the extent that the dose distributions were practically indistinguishable from those in static geometry. Motivated by this initial success, we investigated a real-time couch compensation system consisting of a stereoscopic infra-red camera system interfaced to a robotic couch known as the Hexapod™, which responds in real time to any change in position detected by the cameras. Optical reflectors placed on a solid water phantom were used as surrogates for motion. We tested the effectiveness of couch-based motion compensation for fixed fields and a dynamic arc delivery cases. Due to hardware limitations, we performed film-based experiments (1), (2) and (3), with the robotic couch at a phantom motion period and dose rate of 16 s and 100 MU min-1, respectively. Analysis of film measurements showed near-equivalent dose distributions (<=2 mm agreement of corresponding isodose lines) for static geometry and motion-synchronized real-time robotic couch tracking-based radiation delivery.
Development of a compact continuum tubular robotic system for nasopharyngeal biopsy.
Wu, Liao; Song, Shuang; Wu, Keyu; Lim, Chwee Ming; Ren, Hongliang
2017-03-01
Traditional posterior nasopharyngeal biopsy using a flexible nasal endoscope has the risks of abrasion and injury to the nasal mucosa and thus causing trauma to the patient. Recently, a new class of robots known as continuum tubular robots (CTRs) provide a novel solution to the challenge with miniaturized size, curvilinear maneuverability, and capability of avoiding collision within the nasal environment. This paper presents a compact CTR which is 35 cm in total length, 10 cm in diameter, 2.15 kg in weight, and easy to be integrated with a robotic arm to perform more complicated operations. Structural design, end-effector design, and workspace analysis are described in detail. In addition, teleoperation of the CTR using a haptic input device is developed for position control in 3D space. Moreover, by integrating the robot with three electromagnetic tracking sensors, a navigation system together with a shape reconstruction algorithm is developed. Comprehensive experiments are conducted to test the functionality of the proposed prototype; experiment results show that under teleoperation, the system has an accuracy of 2.20 mm in following a linear path, an accuracy of 2.01 mm in following a circular path, and a latency time of 0.1 s. It is also found that the proposed shape reconstruction algorithm has a mean error of around 1 mm along the length of the tubes. Besides, the feasibility and effectiveness of the proposed robotic system being applied to posterior nasopharyngeal biopsy are demonstrated by a cadaver experiment. The proposed robotic system holds promise to enhance clinical operation in transnasal procedures.
Coordinated Control Of Mobile Robotic Manipulators
NASA Technical Reports Server (NTRS)
Seraji, Homayoun
1995-01-01
Computationally efficient scheme developed for on-line coordinated control of both manipulation and mobility of robots that include manipulator arms mounted on mobile bases. Applicable to variety of mobile robotic manipulators, including robots that move along tracks (typically, painting and welding robots), robots mounted on gantries and capable of moving in all three dimensions, wheeled robots, and compound robots (consisting of robots mounted on other robots). Theoretical basis discussed in several prior articles in NASA Tech Briefs, including "Increasing the Dexterity of Redundant Robots" (NPO-17801), "Redundant Robot Can Avoid Obstacles" (NPO-17852), "Configuration-Control Scheme Copes With Singularities" (NPO-18556), "More Uses for Configuration Control of Robots" (NPO-18607/NPO-18608).
Towards the development of a spring-based continuum robot for neurosurgery
NASA Astrophysics Data System (ADS)
Kim, Yeongjin; Cheng, Shing Shin; Desai, Jaydev P.
2015-03-01
Brain tumor is usually life threatening due to the uncontrolled growth of abnormal cells native to the brain or the spread of tumor cells from outside the central nervous system to the brain. The risks involved in carrying out surgery within such a complex organ can cause severe anxiety in cancer patients. However, neurosurgery, which remains one of the more effective ways of treating brain tumors focused in a confined volume, can have a tremendously increased success rate if the appropriate imaging modality is used for complete tumor removal. Magnetic resonance imaging (MRI) provides excellent soft-tissue contrast and is the imaging modality of choice for brain tumor imaging. MRI combined with continuum soft robotics has immense potential to be the revolutionary treatment technique in the field of brain cancer. It eliminates the concern of hand tremor and guarantees a more precise procedure. One of the prototypes of Minimally Invasive Neurosurgical Intracranial Robot (MINIR-II), which can be classified as a continuum soft robot, consists of a snake-like body made of three segments of rapid prototyped plastic springs. It provides improved dexterity with higher degrees of freedom and independent joint control. It is MRI-compatible, allowing surgeons to track and determine the real-time location of the robot relative to the brain tumor target. The robot was manufactured in a single piece using rapid prototyping technology at a low cost, allowing it to disposable after each use. MINIR-II has two DOFs at each segment with both joints controlled by two pairs of MRI-compatible SMA spring actuators. Preliminary motion tests have been carried out using vision-tracking method and the robot was able to move to different positions based on user commands.
IntelliTable: Inclusively-Designed Furniture with Robotic Capabilities.
Prescott, Tony J; Conran, Sebastian; Mitchinson, Ben; Cudd, Peter
2017-01-01
IntelliTable is a new proof-of-principle assistive technology system with robotic capabilities in the form of an elegant universal cantilever table able to move around by itself, or under user control. We describe the design and current capabilities of the table and the human-centered design methodology used in its development and initial evaluation. The IntelliTable study has delivered robotic platform programmed by a smartphone that can navigate around a typical home or care environment, avoiding obstacles, and positioning itself at the user's command. It can also be configured to navigate itself to pre-ordained places positions within an environment using ceiling tracking, responsive optical guidance and object-based sonar navigation.
Space station automation and robotics study. Operator-systems interface
NASA Technical Reports Server (NTRS)
1984-01-01
This is the final report of a Space Station Automation and Robotics Planning Study, which was a joint project of the Boeing Aerospace Company, Boeing Commercial Airplane Company, and Boeing Computer Services Company. The study is in support of the Advanced Technology Advisory Committee established by NASA in accordance with a mandate by the U.S. Congress. Boeing support complements that provided to the NASA Contractor study team by four aerospace contractors, the Stanford Research Institute (SRI), and the California Space Institute. This study identifies automation and robotics (A&R) technologies that can be advanced by requirements levied by the Space Station Program. The methodology used in the study is to establish functional requirements for the operator system interface (OSI), establish the technologies needed to meet these requirements, and to forecast the availability of these technologies. The OSI would perform path planning, tracking and control, object recognition, fault detection and correction, and plan modifications in connection with extravehicular (EV) robot operations.
Enhancing patient freedom in rehabilitation robotics using gaze-based intention detection.
Novak, Domen; Riener, Robert
2013-06-01
Several design strategies for rehabilitation robotics have aimed to improve patients' experiences using motivating and engaging virtual environments. This paper presents a new design strategy: enhancing patient freedom with a complex virtual environment that intelligently detects patients' intentions and supports the intended actions. A 'virtual kitchen' scenario has been developed in which many possible actions can be performed at any time, allowing patients to experiment and giving them more freedom. Remote eye tracking is used to detect the intended action and trigger appropriate support by a rehabilitation robot. This approach requires no additional equipment attached to the patient and has a calibration time of less than a minute. The system was tested on healthy subjects using the ARMin III arm rehabilitation robot. It was found to be technically feasible and usable by healthy subjects. However, the intention detection algorithm should be improved using better sensor fusion, and clinical tests with patients are needed to evaluate the system's usability and potential therapeutic benefits.
Dynamics and control of robot for capturing objects in space
NASA Astrophysics Data System (ADS)
Huang, Panfeng
Space robots are expected to perform intricate tasks in future space services, such as satellite maintenance, refueling, and replacing the orbital replacement unit (ORU). To realize these missions, the capturing operation may not be avoided. Such operations will encounter some challenges because space robots have some unique characteristics unfound on ground-based robots, such as, dynamic singularities, dynamic coupling between manipulator and space base, limited energy supply and working without a fixed base, and so on. In addition, since contacts and impacts may not be avoided during capturing operation. Therefore, dynamics and control problems of space robot for capturing objects are significant research topics if the robots are to be deployed for the space services. A typical servicing operation mainly includes three phases: capturing the object, berthing and docking the object, then repairing the target. Therefore, this thesis will focus on resolving some challenging problems during capturing the object, berthing and docking, and so on. In this thesis, I study and analyze the dynamics and control problems of space robot for capturing objects. This work has potential impact in space robotic applications. I first study the contact and impact dynamics of space robot and objects. I specifically focus on analyzing the impact dynamics and mapping the relationship of influence and speed. Then, I develop the fundamental theory for planning the minimum-collision based trajectory of space robot and designing the configuration of space robot at the moment of capture. To compensate for the attitude of the space base during the capturing approach operation, a new balance control concept which can effectively balance the attitude of the space base using the dynamic couplings is developed. The developed balance control concept helps to understand of the nature of space dynamic coupling, and can be readily applied to compensate or minimize the disturbance to the space base. After capturing the object, the space robot must complete the following two tasks: one is to berth the object, and the other is to re-orientate the attitude of the whole robot system for communication and power supply. Therefore, I propose a method to accomplish these two tasks simultaneously using manipulator motion only. The ultimate goal of space services is to realize the capture and manipulation autonomously. Therefore, I propose an affective approach based on learning human skill to track and capture the objects automatically in space. With human-teaching demonstration, the space robot is able to learn and abstract human tracking and capturing skill using an efficient neural-network learning architecture that combines flexible Cascade Neural Networks with Node Decoupled Extended Kalman Filtering (CNN-NDEKF). The simulation results attest that this approach is useful and feasible in tracking trajectory planning and capturing of space robot. Finally I propose a novel approach based on Genetic Algorithms (GAs) to optimize the approach trajectory of space robots in order to realize effective and stable operations. I complete the minimum-torque path planning in order to save the limited energy in space, and design the minimum jerk trajectory for the stabilization of the space manipulator and its space base. These optimal algorithms are very important and useful for the application of space robot.
Robot Tracer with Visual Camera
NASA Astrophysics Data System (ADS)
Jabbar Lubis, Abdul; Dwi Lestari, Yuyun; Dafitri, Haida; Azanuddin
2017-12-01
Robot is a versatile tool that can function replace human work function. The robot is a device that can be reprogrammed according to user needs. The use of wireless networks for remote monitoring needs can be utilized to build a robot that can be monitored movement and can be monitored using blueprints and he can track the path chosen robot. This process is sent using a wireless network. For visual robot using high resolution cameras to facilitate the operator to control the robot and see the surrounding circumstances.
Use of pharmacy delivery robots in intensive care units.
Summerfield, Marc R; Seagull, F Jacob; Vaidya, Neelesh; Xiao, Yan
2011-01-01
The use of pharmacy delivery robots in an institution's intensive care units was evaluated. In 2003, the University of Maryland Medical Center (UMMC) began a pilot program to determine the logistic capability and functional utility of robotic technology in the delivery of medications from satellite pharmacies to patient care units. Three satellite pharmacies currently used the robotic system. Five data sources (electronic robot activation records, logs, interviews, surveys, and observations) were used to assess five key aspects of robotic delivery: robot use, reliability, timeliness, cost minimization, and acceptance. A 19-item survey using a 7-point Likert-type scale was developed to determine if pharmacy delivery robots changed nurses' perception of pharmacy service. The components measured included general satisfaction, reliability, timeliness, stat orders, services, interaction with pharmacy, and status tracking. A total of 23 pre-implementation, 96 post-implementation, and 30 two-year follow-up surveys were completed. After implementation of the robotic delivery system, time from fax to label, order preparation time, and idle time for medications to be delivered decreased, while nurses' general satisfaction with the pharmacy and opinion of the reliability of pharmacy delivery significantly increased. Robotic delivery did not influence the perceived quality of delivery service or the timeliness of orders or stat orders. Robot reliability was a major issue for the technician but not for pharmacists, who did not have as much interaction with the devices. By considering the needs of UMMC and its patients and matching them with available technology, the institution was able to improve the medication-use process and timeliness of medication departure from the pharmacy.
NASA Astrophysics Data System (ADS)
Moore, John T.; Wiles, Andrew D.; Wedlake, Chris; Bainbridge, Daniel; Kiaii, Bob; Trejos, Ana Luisa; Patel, Rajni; Peters, Terry M.
2010-02-01
Trans-esophageal echocardiography (TEE) is a standard component of patient monitoring during most cardiac surgeries. In recent years magnetic tracking systems (MTS) have become sufficiently robust to function effectively in appropriately structured operating room environments. The ability to track a conventional multiplanar 2D TEE transducer in 3D space offers incredible potential by greatly expanding the cumulative field of view of cardiac anatomy beyond the limited field of view provided by 2D and 3D TEE technology. However, there is currently no TEE probe manufactured with MTS technology embedded in the transducer, which means sensors must be attached to the outer surface of the TEE. This leads to potential safety issues for patients, as well as potential damage to the sensor during procedures. This paper presents a standard 2D TEE probe fully integrated with MTS technology. The system is evaluated in an environment free of magnetic and electromagnetic disturbances, as well as a clinical operating room in the presence of a da Vinci robotic system. Our first integrated TEE device is currently being used in animal studies for virtual reality-enhanced ultrasound guidance of intracardiac surgeries, while the "second generation" TEE is in use in a clinical operating room as part of a project to measure perioperative heart shift and optimal port placement for robotic cardiac surgery. We demonstrate excellent system accuracy for both applications.
High-Performance 3D Articulated Robot Display
NASA Technical Reports Server (NTRS)
Powell, Mark W.; Torres, Recaredo J.; Mittman, David S.; Kurien, James A.; Abramyan, Lucy
2011-01-01
In the domain of telerobotic operations, the primary challenge facing the operator is to understand the state of the robotic platform. One key aspect of understanding the state is to visualize the physical location and configuration of the platform. As there is a wide variety of mobile robots, the requirements for visualizing their configurations vary diversely across different platforms. There can also be diversity in the mechanical mobility, such as wheeled, tracked, or legged mobility over surfaces. Adaptable 3D articulated robot visualization software can accommodate a wide variety of robotic platforms and environments. The visualization has been used for surface, aerial, space, and water robotic vehicle visualization during field testing. It has been used to enable operations of wheeled and legged surface vehicles, and can be readily adapted to facilitate other mechanical mobility solutions. The 3D visualization can render an articulated 3D model of a robotic platform for any environment. Given the model, the software receives real-time telemetry from the avionics system onboard the vehicle and animates the robot visualization to reflect the telemetered physical state. This is used to track the position and attitude in real time to monitor the progress of the vehicle as it traverses its environment. It is also used to monitor the state of any or all articulated elements of the vehicle, such as arms, legs, or control surfaces. The visualization can also render other sorts of telemetered states visually, such as stress or strains that are measured by the avionics. Such data can be used to color or annotate the virtual vehicle to indicate nominal or off-nominal states during operation. The visualization is also able to render the simulated environment where the vehicle is operating. For surface and aerial vehicles, it can render the terrain under the vehicle as the avionics sends it location information (GPS, odometry, or star tracking), and locate the vehicle over or on the terrain correctly. For long traverses over terrain, the visualization can stream in terrain piecewise in order to maintain the current area of interest for the operator without incurring unreasonable resource constraints on the computing platform. The visualization software is designed to run on laptops that can operate in field-testing environments without Internet access, which is a frequently encountered situation when testing in remote locations that simulate planetary environments such as Mars and other planetary bodies.
2007-01-01
Intelligent Robots and Systems, vol- ume 1, pp. 123–128, September 2002. [2] R. G. Brown and P. Y. Hwang . Introduction to Ran- dom Signals and Applied... Kalman Filter-based) method for calculat- ing a trajectory by tracking features at an unknown location on the Earth’s surface, provided the topography...Extended Kalman Filter (EKF) and an automatic target tracking algorithm. In the following section, the integration architecture is presented, which in
Robust perception algorithms for road and track autonomous following
NASA Astrophysics Data System (ADS)
Marion, Vincent; Lecointe, Olivier; Lewandowski, Cecile; Morillon, Joel G.; Aufrere, Romuald; Marcotegui, Beatrix; Chapuis, Roland; Beucher, Serge
2004-09-01
The French Military Robotic Study Program (introduced in Aerosense 2003), sponsored by the French Defense Procurement Agency and managed by Thales Airborne Systems as the prime contractor, focuses on about 15 robotic themes, which can provide an immediate "operational add-on value." The paper details the "road and track following" theme (named AUT2), which main purpose was to develop a vision based sub-system to automatically detect roadsides of an extended range of roads and tracks suitable to military missions. To achieve the goal, efforts focused on three main areas: (1) Improvement of images quality at algorithms inputs, thanks to the selection of adapted video cameras, and the development of a THALES patented algorithm: it removes in real time most of the disturbing shadows in images taken in natural environments, enhances contrast and lowers reflection effect due to films of water. (2) Selection and improvement of two complementary algorithms (one is segment oriented, the other region based) (3) Development of a fusion process between both algorithms, which feeds in real time a road model with the best available data. Each previous step has been developed so that the global perception process is reliable and safe: as an example, the process continuously evaluates itself and outputs confidence criteria qualifying roadside detection. The paper presents the processes in details, and the results got from passed military acceptance tests, which trigger the next step: autonomous track following (named AUT3).
Development of a 6DOF robotic motion phantom for radiation therapy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Belcher, Andrew H.; Liu, Xinmin; Grelewicz, Zachary
Purpose: The use of medical technology capable of tracking patient motion or positioning patients along 6 degree-of-freedom (6DOF) has steadily increased in the field of radiation therapy. However, due to the complex nature of tracking and performing 6DOF motion, it is critical that such technology is properly verified to be operating within specifications in order to ensure patient safety. In this study, a robotic motion phantom is presented that can be programmed to perform highly accurate motion along any X (left–right), Y (superior–inferior), Z (anterior–posterior), pitch (around X), roll (around Y), and yaw (around Z) axes. In addition, highly synchronizedmore » motion along all axes can be performed in order to simulate the dynamic motion of a tumor in 6D. The accuracy and reproducibility of this 6D motion were characterized. Methods: An in-house designed and built 6D robotic motion phantom was constructed following the Stewart–Gough parallel kinematics platform archetype. The device was controlled using an inverse kinematics formulation, and precise movements in all 6 degrees-of-freedom (X, Y, Z, pitch, roll, and yaw) were performed, both simultaneously and separately for each degree-of-freedom. Additionally, previously recorded 6D cranial and prostate motions were effectively executed. The robotic phantom movements were verified using a 15 fps 6D infrared marker tracking system and the measured trajectories were compared quantitatively to the intended input trajectories. The workspace, maximum 6D velocity, backlash, and weight load capabilities of the system were also established. Results: Evaluation of the 6D platform demonstrated translational root mean square error (RMSE) values of 0.14, 0.22, and 0.08 mm over 20 mm in X and Y and 10 mm in Z, respectively, and rotational RMSE values of 0.16°, 0.06°, and 0.08° over 10° of pitch, roll, and yaw, respectively. The robotic stage also effectively performed controlled 6D motions, as well as reproduced cranial trajectories over 15 min, with a maximal RMSE of 0.04 mm translationally and 0.04° rotationally, and a prostate trajectory over 2 min, with a maximal RMSE of 0.06 mm translationally and 0.04° rotationally. Conclusions: This 6D robotic phantom has proven to be accurate under clinical standards and capable of reproducing tumor motion in 6D. Such functionality makes the robotic phantom usable for either quality assurance or research purposes.« less
Butail, Sachit; Polverino, Giovanni; Phamduy, Paul; Del Sette, Fausto; Porfiri, Maurizio
2014-12-15
In animal studies, robots have been recently used as a valid tool for testing a wide spectrum of hypotheses. These robots often exploit visual or auditory cues to modulate animal behavior. The propensity of zebrafish, a model organism in biological studies, toward fish with similar color patterns and shape has been leveraged to design biologically inspired robots that successfully attract zebrafish in preference tests. With an aim of extending the application of such robots to field studies, here, we investigate the response of zebrafish to multiple robotic fish swimming at different speeds and in varying arrangements. A soft real-time multi-target tracking and control system remotely steers the robots in circular trajectories during the experimental trials. Our findings indicate a complex behavioral response of zebrafish to biologically inspired robots. More robots produce a significant change in salient measures of stress, with a fast robot swimming alone causing more freezing and erratic activity than two robots swimming slowly together. In addition, fish spend more time in the proximity of a robot when they swim far apart than when the robots swim close to each other. Increase in the number of robots also significantly alters the degree of alignment of fish motion with a robot. Results from this study are expected to advance our understanding of robot perception by live animals and aid in hypothesis-driven studies in unconstrained free-swimming environments. Copyright © 2014 Elsevier B.V. All rights reserved.
Telerobotic system concept for real-time soft-tissue imaging during radiotherapy beam delivery.
Schlosser, Jeffrey; Salisbury, Kenneth; Hristov, Dimitre
2010-12-01
The curative potential of external beam radiation therapy is critically dependent on having the ability to accurately aim radiation beams at intended targets while avoiding surrounding healthy tissues. However, existing technologies are incapable of real-time, volumetric, soft-tissue imaging during radiation beam delivery, when accurate target tracking is most critical. The authors address this challenge in the development and evaluation of a novel, minimally interfering, telerobotic ultrasound (U.S.) imaging system that can be integrated with existing medical linear accelerators (LINACs) for therapy guidance. A customized human-safe robotic manipulator was designed and built to control the pressure and pitch of an abdominal U.S. transducer while avoiding LINAC gantry collisions. A haptic device was integrated to remotely control the robotic manipulator motion and U.S. image acquisition outside the LINAC room. The ability of the system to continuously maintain high quality prostate images was evaluated in volunteers over extended time periods. Treatment feasibility was assessed by comparing a clinically deployed prostate treatment plan to an alternative plan in which beam directions were restricted to sectors that did not interfere with the transabdominal U.S. transducer. To demonstrate imaging capability concurrent with delivery, robot performance and U.S. target tracking in a phantom were tested with a 15 MV radiation beam active. Remote image acquisition and maintenance of image quality with the haptic interface was successfully demonstrated over 10 min periods in representative treatment setups of volunteers. Furthermore, the robot's ability to maintain a constant probe force and desired pitch angle was unaffected by the LINAC beam. For a representative prostate patient, the dose-volume histogram (DVH) for a plan with restricted sectors remained virtually identical to the DVH of a clinically deployed plan. With reduced margins, as would be enabled by real-time imaging, gross tumor volume coverage was identical while notable reductions of bladder and rectal volumes exposed to large doses were possible. The quality of U.S. images obtained during beam operation was not appreciably degraded by radiofrequency interference and 2D tracking of a phantom object in U.S. images obtained with the beam on/off yielded no significant differences. Remotely controlled robotic U.S. imaging is feasible in the radiotherapy environment and for the first time may offer real-time volumetric soft-tissue guidance concurrent with radiotherapy delivery.
Development of Robotics Applications in a Solid Propellant Mixing Laboratory
1988-06-01
implementation of robotic hardware and software into a laboratory environment requires a carefully structured series of phases which examines, in...strategy. The general methodology utilized in this project is discussed in Appendix A. The proposed laboratory robotics development program was structured ...Accessibility - Potential modifications - Safety precautions e) Robot Transport - Slider mechanisms - Linear tracks - Gantry configuration - Mobility f
Espinosa, Felipe; Santos, Carlos; Marrón-Romera, Marta; Pizarro, Daniel; Valdés, Fernando; Dongil, Javier
2011-01-01
This paper describes a relative localization system used to achieve the navigation of a convoy of robotic units in indoor environments. This positioning system is carried out fusing two sensorial sources: (a) an odometric system and (b) a laser scanner together with artificial landmarks located on top of the units. The laser source allows one to compensate the cumulative error inherent to dead-reckoning; whereas the odometry source provides less pose uncertainty in short trajectories. A discrete Extended Kalman Filter, customized for this application, is used in order to accomplish this aim under real time constraints. Different experimental results with a convoy of Pioneer P3-DX units tracking non-linear trajectories are shown. The paper shows that a simple setup based on low cost laser range systems and robot built-in odometry sensors is able to give a high degree of robustness and accuracy to the relative localization problem of convoy units for indoor applications. PMID:22164079
Espinosa, Felipe; Santos, Carlos; Marrón-Romera, Marta; Pizarro, Daniel; Valdés, Fernando; Dongil, Javier
2011-01-01
This paper describes a relative localization system used to achieve the navigation of a convoy of robotic units in indoor environments. This positioning system is carried out fusing two sensorial sources: (a) an odometric system and (b) a laser scanner together with artificial landmarks located on top of the units. The laser source allows one to compensate the cumulative error inherent to dead-reckoning; whereas the odometry source provides less pose uncertainty in short trajectories. A discrete Extended Kalman Filter, customized for this application, is used in order to accomplish this aim under real time constraints. Different experimental results with a convoy of Pioneer P3-DX units tracking non-linear trajectories are shown. The paper shows that a simple setup based on low cost laser range systems and robot built-in odometry sensors is able to give a high degree of robustness and accuracy to the relative localization problem of convoy units for indoor applications.
Chen, Gang; Song, Yongduan; Guan, Yanfeng
2018-03-01
This brief investigates the finite-time consensus tracking control problem for networked uncertain mechanical systems on digraphs. A new terminal sliding-mode-based cooperative control scheme is developed to guarantee that the tracking errors converge to an arbitrarily small bound around zero in finite time. All the networked systems can have different dynamics and all the dynamics are unknown. A neural network is used at each node to approximate the local unknown dynamics. The control schemes are implemented in a fully distributed manner. The proposed control method eliminates some limitations in the existing terminal sliding-mode-based consensus control methods and extends the existing analysis methods to the case of directed graphs. Simulation results on networked robot manipulators are provided to show the effectiveness of the proposed control algorithms.
Novel method for measuring a dense 3D strain map of robotic flapping wings
NASA Astrophysics Data System (ADS)
Li, Beiwen; Zhang, Song
2018-04-01
Measuring dense 3D strain maps of the inextensible membranous flapping wings of robots is of vital importance to the field of bio-inspired engineering. Conventional high-speed 3D videography methods typically reconstruct the wing geometries through measuring sparse points with fiducial markers, and thus cannot obtain the full-field mechanics of the wings in detail. In this research, we propose a novel system to measure a dense strain map of inextensible membranous flapping wings by developing a superfast 3D imaging system and a computational framework for strain analysis. Specifically, first we developed a 5000 Hz 3D imaging system based on the digital fringe projection technique using the defocused binary patterns to precisely measure the dynamic 3D geometries of rapidly flapping wings. Then, we developed a geometry-based algorithm to perform point tracking on the precisely measured 3D surface data. Finally, we developed a dense strain computational method using the Kirchhoff-Love shell theory. Experiments demonstrate that our method can effectively perform point tracking and measure a highly dense strain map of the wings without many fiducial markers.
NASA Astrophysics Data System (ADS)
Ji, Peng; Song, Aiguo; Song, Zimo; Liu, Yuqing; Jiang, Guohua; Zhao, Guopu
2017-02-01
In this paper, we describe a heading direction correction algorithm for a tracked mobile robot. To save hardware resources as far as possible, the mobile robot’s wrist camera is used as the only sensor, which is rotated to face stairs. An ensemble heading deviation detector is proposed to help the mobile robot correct its heading direction. To improve the generalization ability, a multi-scale Gabor filter is used to process the input image previously. Final deviation result is acquired by applying the majority vote strategy on all the classifiers’ results. The experimental results show that our detector is able to enable the mobile robot to correct its heading direction adaptively while it is climbing the stairs.
Ultrasound Imaging in Radiation Therapy: From Interfractional to Intrafractional Guidance
Western, Craig; Hristov, Dimitre
2015-01-01
External beam radiation therapy (EBRT) is included in the treatment regimen of the majority of cancer patients. With the proliferation of hypofractionated radiotherapy treatment regimens, such as stereotactic body radiation therapy (SBRT), interfractional and intrafractional imaging technologies are becoming increasingly critical to ensure safe and effective treatment delivery. Ultrasound (US)-based image guidance systems offer real-time, markerless, volumetric imaging with excellent soft tissue contrast, overcoming the limitations of traditional X-ray or computed tomography (CT)-based guidance for abdominal and pelvic cancer sites, such as the liver and prostate. Interfractional US guidance systems have been commercially adopted for patient positioning but suffer from systematic positioning errors induced by probe pressure. More recently, several research groups have introduced concepts for intrafractional US guidance systems leveraging robotic probe placement technology and real-time soft tissue tracking software. This paper reviews various commercial and research-level US guidance systems used in radiation therapy, with an emphasis on hardware and software technologies that enable the deployment of US imaging within the radiotherapy environment and workflow. Previously unpublished material on tissue tracking systems and robotic probe manipulators under development by our group is also included. PMID:26180704
A Unified Approach to Motion Control of Motion Robots
NASA Technical Reports Server (NTRS)
Seraji, H.
1994-01-01
This paper presents a simple on-line approach for motion control of mobile robots made up of a manipulator arm mounted on a mobile base. The proposed approach is equally applicable to nonholonomic mobile robots, such as rover-mounted manipulators and to holonomic mobile robots such as tracked robots or compound manipulators. The computational efficiency of the proposed control scheme makes it particularly suitable for real-time implementation.
Ai, Qingsong; Zhu, Chengxiang; Zuo, Jie; Liu, Quan; Xie, Sheng Q.; Yang, Ming
2017-01-01
A rehabilitation robot plays an important role in relieving the therapists’ burden and helping patients with ankle injuries to perform more accurate and effective rehabilitation training. However, a majority of current ankle rehabilitation robots are rigid and have drawbacks in terms of complex structure, poor flexibility and lack of safety. Taking advantages of pneumatic muscles’ good flexibility and light weight, we developed a novel two degrees of freedom (2-DOF) parallel compliant ankle rehabilitation robot actuated by pneumatic muscles (PMs). To solve the PM’s nonlinear characteristics during operation and to tackle the human-robot uncertainties in rehabilitation, an adaptive backstepping sliding mode control (ABS-SMC) method is proposed in this paper. The human-robot external disturbance can be estimated by an observer, who is then used to adjust the robot output to accommodate external changes. The system stability is guaranteed by the Lyapunov stability theorem. Experimental results on the compliant ankle rehabilitation robot show that the proposed ABS-SMC is able to estimate the external disturbance online and adjust the control output in real time during operation, resulting in a higher trajectory tracking accuracy and better response performance especially in dynamic conditions. PMID:29283406
Payne, Christopher J; Yang, Guang-Zhong
2014-08-01
Medical robots have evolved from autonomous systems to tele-operated platforms and mechanically-grounded, cooperatively-controlled robots. Whilst these approaches have seen both commercial and clinical success, uptake of these robots remains moderate because of their high cost, large physical footprint and long setup times. More recently, researchers have moved toward developing hand-held robots that are completely ungrounded and manipulated by surgeons in free space, in a similar manner to how conventional instruments are handled. These devices provide specific functions that assist the surgeon in accomplishing tasks that are otherwise challenging with manual manipulation. Hand-held robots have the advantages of being compact and easily integrated into the normal surgical workflow since there is typically little or no setup time. Hand-held devices can also have a significantly reduced cost to healthcare providers as they do not necessitate the complex, multi degree-of-freedom linkages that grounded robots require. However, the development of such devices is faced with many technical challenges, including miniaturization, cost and sterility, control stability, inertial and gravity compensation and robust instrument tracking. This review presents the emerging technical trends in hand-held medical robots and future development opportunities for promoting their wider clinical uptake.
Ai, Qingsong; Zhu, Chengxiang; Zuo, Jie; Meng, Wei; Liu, Quan; Xie, Sheng Q; Yang, Ming
2017-12-28
A rehabilitation robot plays an important role in relieving the therapists' burden and helping patients with ankle injuries to perform more accurate and effective rehabilitation training. However, a majority of current ankle rehabilitation robots are rigid and have drawbacks in terms of complex structure, poor flexibility and lack of safety. Taking advantages of pneumatic muscles' good flexibility and light weight, we developed a novel two degrees of freedom (2-DOF) parallel compliant ankle rehabilitation robot actuated by pneumatic muscles (PMs). To solve the PM's nonlinear characteristics during operation and to tackle the human-robot uncertainties in rehabilitation, an adaptive backstepping sliding mode control (ABS-SMC) method is proposed in this paper. The human-robot external disturbance can be estimated by an observer, who is then used to adjust the robot output to accommodate external changes. The system stability is guaranteed by the Lyapunov stability theorem. Experimental results on the compliant ankle rehabilitation robot show that the proposed ABS-SMC is able to estimate the external disturbance online and adjust the control output in real time during operation, resulting in a higher trajectory tracking accuracy and better response performance especially in dynamic conditions.
Biobotic insect swarm based sensor networks for search and rescue
NASA Astrophysics Data System (ADS)
Bozkurt, Alper; Lobaton, Edgar; Sichitiu, Mihail; Hedrick, Tyson; Latif, Tahmid; Dirafzoon, Alireza; Whitmire, Eric; Verderber, Alexander; Marin, Juan; Xiong, Hong
2014-06-01
The potential benefits of distributed robotics systems in applications requiring situational awareness, such as search-and-rescue in emergency situations, are indisputable. The efficiency of such systems requires robotic agents capable of coping with uncertain and dynamic environmental conditions. For example, after an earthquake, a tremendous effort is spent for days to reach to surviving victims where robotic swarms or other distributed robotic systems might play a great role in achieving this faster. However, current technology falls short of offering centimeter scale mobile agents that can function effectively under such conditions. Insects, the inspiration of many robotic swarms, exhibit an unmatched ability to navigate through such environments while successfully maintaining control and stability. We have benefitted from recent developments in neural engineering and neuromuscular stimulation research to fuse the locomotory advantages of insects with the latest developments in wireless networking technologies to enable biobotic insect agents to function as search-and-rescue agents. Our research efforts towards this goal include development of biobot electronic backpack technologies, establishment of biobot tracking testbeds to evaluate locomotion control efficiency, investigation of biobotic control strategies with Gromphadorhina portentosa cockroaches and Manduca sexta moths, establishment of a localization and communication infrastructure, modeling and controlling collective motion by learning deterministic and stochastic motion models, topological motion modeling based on these models, and the development of a swarm robotic platform to be used as a testbed for our algorithms.
Toward image guided robotic surgery: system validation.
Herrell, Stanley D; Kwartowitz, David Morgan; Milhoua, Paul M; Galloway, Robert L
2009-02-01
Navigation for current robotic assisted surgical techniques is primarily accomplished through a stereo pair of laparoscopic camera images. These images provide standard optical visualization of the surface but provide no subsurface information. Image guidance methods allow the visualization of subsurface information to determine the current position in relationship to that of tracked tools. A robotic image guided surgical system was designed and implemented based on our previous laboratory studies. A series of experiments using tissue mimicking phantoms with injected target lesions was performed. The surgeon was asked to resect "tumor" tissue with and without the augmentation of image guidance using the da Vinci robotic surgical system. Resections were performed and compared to an ideal resection based on the radius of the tumor measured from preoperative computerized tomography. A quantity called the resection ratio, that is the ratio of resected tissue compared to the ideal resection, was calculated for each of 13 trials and compared. The mean +/- SD resection ratio of procedures augmented with image guidance was smaller than that of procedures without image guidance (3.26 +/- 1.38 vs 9.01 +/- 1.81, p <0.01). Additionally, procedures using image guidance were shorter (average 8 vs 13 minutes). It was demonstrated that there is a benefit from the augmentation of laparoscopic video with updated preoperative images. Incorporating our image guided system into the da Vinci robotic system improved overall tissue resection, as measured by our metric. Adding image guidance to the da Vinci robotic surgery system may result in the potential for improvements such as the decreased removal of benign tissue while maintaining an appropriate surgical margin.
NASA Astrophysics Data System (ADS)
Yang, Juqing; Wang, Dayong; Fan, Baixing; Dong, Dengfeng; Zhou, Weihu
2017-03-01
In-situ intelligent manufacturing for large-volume equipment requires industrial robots with absolute high-accuracy positioning and orientation steering control. Conventional robots mainly employ an offline calibration technology to identify and compensate key robotic parameters. However, the dynamic and static parameters of a robot change nonlinearly. It is not possible to acquire a robot's actual parameters and control the absolute pose of the robot with a high accuracy within a large workspace by offline calibration in real-time. This study proposes a real-time online absolute pose steering control method for an industrial robot based on six degrees of freedom laser tracking measurement, which adopts comprehensive compensation and correction of differential movement variables. First, the pose steering control system and robot kinematics error model are constructed, and then the pose error compensation mechanism and algorithm are introduced in detail. By accurately achieving the position and orientation of the robot end-tool, mapping the computed Jacobian matrix of the joint variable and correcting the joint variable, the real-time online absolute pose compensation for an industrial robot is accurately implemented in simulations and experimental tests. The average positioning error is 0.048 mm and orientation accuracy is better than 0.01 deg. The results demonstrate that the proposed method is feasible, and the online absolute accuracy of a robot is sufficiently enhanced.
Indirect decentralized learning control
NASA Technical Reports Server (NTRS)
Longman, Richard W.; Lee, Soo C.; Phan, M.
1992-01-01
The new field of learning control develops controllers that learn to improve their performance at executing a given task, based on experience performing this specific task. In a previous work, the authors presented a theory of indirect learning control based on use of indirect adaptive control concepts employing simultaneous identification and control. This paper develops improved indirect learning control algorithms, and studies the use of such controllers in decentralized systems. The original motivation of the learning control field was learning in robots doing repetitive tasks such as on an assembly line. This paper starts with decentralized discrete time systems, and progresses to the robot application, modeling the robot as a time varying linear system in the neighborhood of the nominal trajectory, and using the usual robot controllers that are decentralized, treating each link as if it is independent of any coupling with other links. The basic result of the paper is to show that stability of the indirect learning controllers for all subsystems when the coupling between subsystems is turned off, assures convergence to zero tracking error of the decentralized indirect learning control of the coupled system, provided that the sample time in the digital learning controller is sufficiently short.
NASA Astrophysics Data System (ADS)
Dragone, Mauro; O'Donoghue, Ruadhan; Leonard, John J.; O'Hare, Gregory; Duffy, Brian; Patrikalakis, Andrew; Leederkerken, Jacques
2005-06-01
The paper describes an ongoing effort to enable autonomous mobile robots to play soccer in unstructured, everyday environments. Unlike conventional robot soccer competitions that are usually held on purpose-built robot soccer "fields", in our work we seek to develop the capability for robots to demonstrate aspects of soccer-playing in more diverse environments, such as schools, hospitals, or shopping malls, with static obstacles (furniture) and dynamic natural obstacles (people). This problem of "Soccer Anywhere" presents numerous research challenges including: (1) Simultaneous Localization and Mapping (SLAM) in dynamic, unstructured environments, (2) software control architectures for decentralized, distributed control of mobile agents, (3) integration of vision-based object tracking with dynamic control, and (4) social interaction with human participants. In addition to the intrinsic research merit of these topics, we believe that this capability would prove useful for outreach activities, in demonstrating robotics technology to primary and secondary school students, to motivate them to pursue careers in science and engineering.
Modeling and controlling a robotic convoy using guidance laws strategies.
Belkhouche, Fethi; Belkhouche, Boumediene
2005-08-01
This paper deals with the problem of modeling and controlling a robotic convoy. Guidance laws techniques are used to provide a mathematical formulation of the problem. The guidance laws used for this purpose are the velocity pursuit, the deviated pursuit, and the proportional navigation. The velocity pursuit equations model the robot's path under various sensors based control laws. A systematic study of the tracking problem based on this technique is undertaken. These guidance laws are applied to derive decentralized control laws for the angular and linear velocities. For the angular velocity, the control law is directly derived from the guidance laws after considering the relative kinematics equations between successive robots. The second control law maintains the distance between successive robots constant by controlling the linear velocity. This control law is derived by considering the kinematics equations between successive robots under the considered guidance law. Properties of the method are discussed and proven. Simulation results confirm the validity of our approach, as well as the validity of the properties of the method. Index Terms-Guidance laws, relative kinematics equations, robotic convoy, tracking.
NASA Astrophysics Data System (ADS)
Thienphrapa, Paul; Ramachandran, Bharat; Elhawary, Haytham; Taylor, Russell H.; Popovic, Aleksandra
2012-02-01
Free moving bodies in the heart pose a serious health risk as they may be released in the arteries causing blood flow disruption. These bodies may be the result of various medical conditions and trauma. The conventional approach to removing these objects involves open surgery with sternotomy, the use of cardiopulmonary bypass, and a wide resection of the heart muscle. We advocate a minimally invasive surgical approach using a flexible robotic end effector guided by 3D transesophageal echocardiography. In a phantom study, we track a moving body in a beating heart using a modified normalized cross-correlation method, with mean RMS errors of 2.3 mm. We previously found the foreign body motion to be fast and abrupt, rendering infeasible a retrieval method based on direct tracking. We proposed a strategy based on guiding a robot to the most spatially probable location of the fragment and securing it upon its reentry to said location. To improve efficacy in the context of a robotic retrieval system, we extend this approach by exploring multiple candidate capture locations. Salient locations are identified based on spatial probability, dwell time, and visit frequency; secondary locations are also examined. Aggregate results indicate that the location of highest spatial probability (50% occupancy) is distinct from the longest-dwelled location (0.84 seconds). Such metrics are vital in informing the design of a retrieval system and capture strategies, and they can be computed intraoperatively to select the best capture location based on constraints such as workspace, time, and device manipulability. Given the complex nature of fragment motion, the ability to analyze multiple capture locations is a desirable capability in an interventional system.
Sled Control and Safety System
NASA Technical Reports Server (NTRS)
Forrest, L. J.
1982-01-01
Computerized system for controlling motion of linear-track accelerator applied to other automated equipment, such as numerically-controlled machine tools and robot manipulators on assembly lines. System controls motions of sled with sine-wave signal created digitally by microprocessor. Dynamic parameters of sled motion are monitored so sled may be stopped safely if malfunction occurs. Sled is capable of sinusoidal accelerations up to 0.5 g with 125-kg load.
NASA Astrophysics Data System (ADS)
Wojtczyk, Martin; Panin, Giorgio; Röder, Thorsten; Lenz, Claus; Nair, Suraj; Heidemann, Rüdiger; Goudar, Chetan; Knoll, Alois
2010-01-01
After utilizing robots for more than 30 years for classic industrial automation applications, service robots form a constantly increasing market, although the big breakthrough is still awaited. Our approach to service robots was driven by the idea of supporting lab personnel in a biotechnology laboratory. After initial development in Germany, a mobile robot platform extended with an industrial manipulator and the necessary sensors for indoor localization and object manipulation, has been shipped to Bayer HealthCare in Berkeley, CA, USA, a global player in the sector of biopharmaceutical products, located in the San Francisco bay area. The determined goal of the mobile manipulator is to support the off-shift staff to carry out completely autonomous or guided, remote controlled lab walkthroughs, which we implement utilizing a recent development of our computer vision group: OpenTL - an integrated framework for model-based visual tracking.
Chen, Yuantao; Xu, Weihong; Kuang, Fangjun; Gao, Shangbing
2013-01-01
The efficient target tracking algorithm researches have become current research focus of intelligent robots. The main problems of target tracking process in mobile robot face environmental uncertainty. They are very difficult to estimate the target states, illumination change, target shape changes, complex backgrounds, and other factors and all affect the occlusion in tracking robustness. To further improve the target tracking's accuracy and reliability, we present a novel target tracking algorithm to use visual saliency and adaptive support vector machine (ASVM). Furthermore, the paper's algorithm has been based on the mixture saliency of image features. These features include color, brightness, and sport feature. The execution process used visual saliency features and those common characteristics have been expressed as the target's saliency. Numerous experiments demonstrate the effectiveness and timeliness of the proposed target tracking algorithm in video sequences where the target objects undergo large changes in pose, scale, and illumination.
High-performance object tracking and fixation with an online neural estimator.
Kumarawadu, Sisil; Watanabe, Keigo; Lee, Tsu-Tian
2007-02-01
Vision-based target tracking and fixation to keep objects that move in three dimensions in view is important for many tasks in several fields including intelligent transportation systems and robotics. Much of the visual control literature has focused on the kinematics of visual control and ignored a number of significant dynamic control issues that limit performance. In line with this, this paper presents a neural network (NN)-based binocular tracking scheme for high-performance target tracking and fixation with minimum sensory information. The procedure allows the designer to take into account the physical (Lagrangian dynamics) properties of the vision system in the control law. The design objective is to synthesize a binocular tracking controller that explicitly takes the systems dynamics into account, yet needs no knowledge of dynamic nonlinearities and joint velocity sensory information. The combined neurocontroller-observer scheme can guarantee the uniform ultimate bounds of the tracking, observer, and NN weight estimation errors under fairly general conditions on the controller-observer gains. The controller is tested and verified via simulation tests in the presence of severe target motion changes.
Trauma Pod: a semi-automated telerobotic surgical system.
Garcia, Pablo; Rosen, Jacob; Kapoor, Chetan; Noakes, Mark; Elbert, Greg; Treat, Michael; Ganous, Tim; Hanson, Matt; Manak, Joe; Hasser, Chris; Rohler, David; Satava, Richard
2009-06-01
The Trauma Pod (TP) vision is to develop a rapidly deployable robotic system to perform critical acute stabilization and/or surgical procedures, autonomously or in a teleoperative mode, on wounded soldiers in the battlefield who might otherwise die before treatment in a combat hospital could be provided. In the first phase of a project pursuing this vision, a robotic TP system was developed and its capability demonstrated by performing selected surgical procedures on a patient phantom. The system demonstrates the feasibility of performing acute stabilization procedures with the patient being the only human in the surgical cell. The teleoperated surgical robot is supported by autonomous robotic arms and subsystems that carry out scrub-nurse and circulating-nurse functions. Tool change and supply delivery are performed automatically and at least as fast as performed manually by nurses. Tracking and counting of the supplies is performed automatically. The TP system also includes a tomographic X-ray facility for patient diagnosis and two-dimensional (2D) fluoroscopic data to support interventions. The vast amount of clinical protocols generated in the TP system are recorded automatically. Automation and teleoperation capabilities form the basis for a more comprehensive acute diagnostic and management platform that will provide life-saving care in environments where surgical personnel are not present.
Cartesian control of redundant robots
NASA Technical Reports Server (NTRS)
Colbaugh, R.; Glass, K.
1989-01-01
A Cartesian-space position/force controller is presented for redundant robots. The proposed control structure partitions the control problem into a nonredundant position/force trajectory tracking problem and a redundant mapping problem between Cartesian control input F is a set member of the set R(sup m) and robot actuator torque T is a set member of the set R(sup n) (for redundant robots, m is less than n). The underdetermined nature of the F yields T map is exploited so that the robot redundancy is utilized to improve the dynamic response of the robot. This dynamically optimal F yields T map is implemented locally (in time) so that it is computationally efficient for on-line control; however, it is shown that the map possesses globally optimal characteristics. Additionally, it is demonstrated that the dynamically optimal F yields T map can be modified so that the robot redundancy is used to simultaneously improve the dynamic response and realize any specified kinematic performance objective (e.g., manipulability maximization or obstacle avoidance). Computer simulation results are given for a four degree of freedom planar redundant robot under Cartesian control, and demonstrate that position/force trajectory tracking and effective redundancy utilization can be achieved simultaneously with the proposed controller.
NASA Astrophysics Data System (ADS)
Sanjou, Michio; Nagasaka, Tsuyoshi
2017-11-01
The present study describes an automated system to measure the river flow velocity. A combination of the camera-tracking system and the Proportional/Integral/Derivative (PID) control could enable the boat-type robot to remain in position against the mainstream; this results in reasonable evaluation of the mean velocity by a duty ratio which corresponds to rotation speed of the screw propeller. A laser range finder module was installed to measure the local water depth. Reliable laboratory experiments with the prototype boat robot and electromagnetic velocimetry were conducted to obtain a calibration curve that connects the duty ratio and mean current velocity. The remaining accuracy in the target point was also examined quantitatively. The fluctuation in the spanwise direction is within half of the robot length. It was therefore found that the robot remains well within the target region. We used two-dimensional navigation tests to guarantee that the prototype moved smoothly to the target points and successfully measured the streamwise velocity profiles across the mainstream. Moreover, the present robot was found to move successfully not only in the laboratory flume but also in a small natural river. The robot could move smoothly from the starting point near the operator's site toward the target point where the velocity is measured, and it could evaluate the cross-sectional discharge.
An Intention-Driven Semi-autonomous Intelligent Robotic System for Drinking.
Zhang, Zhijun; Huang, Yongqian; Chen, Siyuan; Qu, Jun; Pan, Xin; Yu, Tianyou; Li, Yuanqing
2017-01-01
In this study, an intention-driven semi-autonomous intelligent robotic (ID-SIR) system is designed and developed to assist the severely disabled patients to live independently. The system mainly consists of a non-invasive brain-machine interface (BMI) subsystem, a robot manipulator and a visual detection and localization subsystem. Different from most of the existing systems remotely controlled by joystick, head- or eye tracking, the proposed ID-SIR system directly acquires the intention from users' brain. Compared with the state-of-art system only working for a specific object in a fixed place, the designed ID-SIR system can grasp any desired object in a random place chosen by a user and deliver it to his/her mouth automatically. As one of the main advantages of the ID-SIR system, the patient is only required to send one intention command for one drinking task and the autonomous robot would finish the rest of specific controlling tasks, which greatly eases the burden on patients. Eight healthy subjects attended our experiment, which contained 10 tasks for each subject. In each task, the proposed ID-SIR system delivered the desired beverage container to the mouth of the subject and then put it back to the original position. The mean accuracy of the eight subjects was 97.5%, which demonstrated the effectiveness of the ID-SIR system.
Adaptive control of an exoskeleton robot with uncertainties on kinematics and dynamics.
Brahmi, Brahim; Saad, Maarouf; Ochoa-Luna, Cristobal; Rahman, Mohammad H
2017-07-01
In this paper, we propose a new adaptive control technique based on nonlinear sliding mode control (JSTDE) taking into account kinematics and dynamics uncertainties. This approach is applied to an exoskeleton robot with uncertain kinematics and dynamics. The adaptation design is based on Time Delay Estimation (TDE). The proposed strategy does not necessitate the well-defined dynamic and kinematic models of the system robot. The updated laws are designed using Lyapunov-function to solve the adaptation problem systematically, proving the close loop stability and ensuring the convergence asymptotically of the outputs tracking errors. Experiments results show the effectiveness and feasibility of JSTDE technique to deal with the variation of the unknown nonlinear dynamics and kinematics of the exoskeleton model.
An adaptive actuator failure compensation scheme for two linked 2WD mobile robots
NASA Astrophysics Data System (ADS)
Ma, Yajie; Al-Dujaili, Ayad; Cocquempot, Vincent; El Badaoui El Najjar, Maan
2017-01-01
This paper develops a new adaptive compensation control scheme for two linked mobile robots with actuator failurs. A configuration with two linked two-wheel drive (2WD) mobile robots is proposed, and the modelling of its kinematics and dynamics are given. An adaptive failure compensation scheme is developed to compensate actuator failures, consisting of a kinematic controller and a multi-design integration based dynamic controller. The kinematic controller is a virtual one, and based on which, multiple adaptive dynamic control signals are designed which covers all possible failure cases. By combing these dynamic control signals, the dynamic controller is designed, which ensures system stability and asymptotic tracking properties. Simulation results verify the effectiveness of the proposed adaptive failure compensation scheme.
Aerial robot intelligent control method based on back-stepping
NASA Astrophysics Data System (ADS)
Zhou, Jian; Xue, Qian
2018-05-01
The aerial robot is characterized as strong nonlinearity, high coupling and parameter uncertainty, a self-adaptive back-stepping control method based on neural network is proposed in this paper. The uncertain part of the aerial robot model is compensated online by the neural network of Cerebellum Model Articulation Controller and robust control items are designed to overcome the uncertainty error of the system during online learning. At the same time, particle swarm algorithm is used to optimize and fix parameters so as to improve the dynamic performance, and control law is obtained by the recursion of back-stepping regression. Simulation results show that the designed control law has desired attitude tracking performance and good robustness in case of uncertainties and large errors in the model parameters.
Hadavand, Mostafa; Mirbagheri, Alireza; Behzadipour, Saeed; Farahmand, Farzam
2014-06-01
An effective master robot for haptic tele-surgery applications needs to provide a solution for the inversed movements of the surgical tool, in addition to sufficient workspace and manipulability, with minimal moving inertia. A novel 4 + 1-DOF mechanism was proposed, based on a triple parallelogram linkage, which provided a Remote Center of Motion (RCM) at the back of the user's hand. The kinematics of the robot was analyzed and a prototype was fabricated and evaluated by experimental tests. With a RCM at the back of the user's hand the actuators far from the end effector, the robot could produce the sensation of hand-inside surgery with minimal moving inertia. The target workspace was achieved with an acceptable manipulability. The trajectory tracking experiments revealed small errors, due to backlash at the joints. The proposed mechanism meets the basic requirements of an effective master robot for haptic tele-surgery applications. Copyright © 2013 John Wiley & Sons, Ltd.
Terrain interaction with the quarter scale beam walker
NASA Technical Reports Server (NTRS)
Chun, Wendell H.; Price, S.; Spiessbach, A.
1990-01-01
Frame walkers are a class of mobile robots that are robust and capable mobility platforms. Variations of the frame walker robot are in commercial use today. Komatsu Ltd. of Japan developed the Remotely Controlled Underwater Surveyor (ReCUS) and Normed Shipyards of France developed the Marine Robot (RM3). Both applications of the frame walker concept satisfied robotic mobility requirements that could not be met by a wheeled or tracked design. One vehicle design concept that falls within this class of mobile robots is the walking beam. A one-quarter scale prototype of the walking beam was built by Martin Marietta to evaluate the potential merits of utilizing the vehicle as a planetary rover. The initial phase of prototype rover testing was structured to evaluate the mobility performance aspects of the vehicle. Performance parameters such as vehicle power, speed, and attitude control were evaluated as a function of the environment in which the prototype vehicle was tested. Subsequent testing phases will address the integrated performance of the vehicle and a local navigation system.
Terrain Interaction With The Quarter Scale Beam Walker
NASA Astrophysics Data System (ADS)
Chun, Wendell H.; Price, R. S.; Spiessbach, Andrew J.
1990-03-01
Frame walkers are a class of mobile robots that are robust and capable mobility platforms. Variations of the frame walker robot are in commercial use today. Komatsu Ltd. of Japan developed the Remotely Controlled Underwater Surveyor (ReCUS) and Normed Shipyards of France developed the Marine Robot (RM3). Both applications of the frame walker concept satisfied robotic mobility requirements that could not be met by a wheeled or tracked design. One vehicle design concept that falls within this class of mobile robots is the walking beam. A one-quarter scale prototype of the walking beam was built by Martin Marietta to evaluate the potential merits of utilizing the vehicle as a planetary rover. The initial phase of prototype rover testing was structured to evaluate the mobility performance aspects of the vehicle. Performance parameters such as vehicle power, speed, and attitude control were evaluated as a function of the environment in which the prototype vehicle was tested. Subsequent testing phases will address the integrated performance of the vehicle and a local navigation system.
Plume-tracking robots: a new application of chemical sensors.
Ishid, H; Nakamoto, T; Moriizumi, T; Kikas, T; Janata, J
2001-04-01
Many animals have the ability to search for odor sources by tracking their plumes. Some of the key features of this search behavior have been successfully transferred to robot platforms, although the capabilities of animals are still beyond the current level of sensor technologies. The examples described in this paper are (1) incorporating into a wheeled robot the upwind surges and casting used by moths in tracking pheromone plumes, (2) extracting useful information from the response patterns of a chemical sensor array patterned after the spatially distributed chemoreceptors of some animals, and (3) mimicking the fanning behavior of silkworm moths to enhance the reception of chemical signals by drawing molecules from one direction. The achievements so far and current efforts are reviewed to illustrate the steps to be taken toward future development of this technology.
Impedance Control of the Rehabilitation Robot Based on Sliding Mode Control
NASA Astrophysics Data System (ADS)
Zhou, Jiawang; Zhou, Zude; Ai, Qingsong
As an auxiliary treatment, the 6-DOF parallel robot plays an important role in lower limb rehabilitation. In order to improve the efficiency and flexibility of the lower limb rehabilitation training, this paper studies the impedance controller based on the position control. A nonsingular terminal sliding mode control is developed to ensure the trajectory tracking precision and in contrast to traditional PID control strategy in the inner position loop, the system will be more stable. The stability of the system is proved by Lyapunov function to guarantee the convergence of the control errors. Simulation results validate the effectiveness of the target impedance model and show that the parallel robot can adjust gait trajectory online according to the human-machine interaction force to meet the gait request of patients, and changing the impedance parameters can meet the demands of different stages of rehabilitation training.
Robotic Variable Polarity Plasma Arc (VPPA) Welding
NASA Technical Reports Server (NTRS)
Jaffery, Waris S.
1993-01-01
The need for automated plasma welding was identified in the early stages of the Space Station Freedom Program (SSFP) because it requires approximately 1.3 miles of welding for assembly. As a result of the Variable Polarity Plasma Arc Welding (VPPAW) process's ability to make virtually defect-free welds in aluminum, it was chosen to fulfill the welding needs. Space Station Freedom will be constructed of 2219 aluminum utilizing the computer controlled VPPAW process. The 'Node Radial Docking Port', with it's saddle shaped weld path, has a constantly changing surface angle over 360 deg of the 282 inch weld. The automated robotic VPPAW process requires eight-axes of motion (six-axes of robot and two-axes of positioner movement). The robot control system is programmed to maintain Torch Center Point (TCP) orientation perpendicular to the part while the part positioner is tilted and rotated to maintain the vertical up orientation as required by the VPPAW process. The combined speed of the robot and the positioner are integrated to maintain a constant speed between the part and the torch. A laser-based vision sensor system has also been integrated to track the seam and map the surface of the profile during welding.
Robotic Variable Polarity Plasma Arc (VPPA) welding
NASA Astrophysics Data System (ADS)
Jaffery, Waris S.
1993-02-01
The need for automated plasma welding was identified in the early stages of the Space Station Freedom Program (SSFP) because it requires approximately 1.3 miles of welding for assembly. As a result of the Variable Polarity Plasma Arc Welding (VPPAW) process's ability to make virtually defect-free welds in aluminum, it was chosen to fulfill the welding needs. Space Station Freedom will be constructed of 2219 aluminum utilizing the computer controlled VPPAW process. The 'Node Radial Docking Port', with it's saddle shaped weld path, has a constantly changing surface angle over 360 deg of the 282 inch weld. The automated robotic VPPAW process requires eight-axes of motion (six-axes of robot and two-axes of positioner movement). The robot control system is programmed to maintain Torch Center Point (TCP) orientation perpendicular to the part while the part positioner is tilted and rotated to maintain the vertical up orientation as required by the VPPAW process. The combined speed of the robot and the positioner are integrated to maintain a constant speed between the part and the torch. A laser-based vision sensor system has also been integrated to track the seam and map the surface of the profile during welding.
Myoelectrically controlled wrist robot for stroke rehabilitation
2013-01-01
Background Robot-assisted rehabilitation is an advanced new technology in stroke rehabilitation to provide intensive training. Post-stroke motor recovery depends on active rehabilitation by voluntary participation of patient’s paretic motor system as early as possible in order to promote reorganization of brain. However, voluntary residual motor efforts to the affected limb have not been involved enough in most robot-assisted rehabilitation for patients after stroke. The objective of this study is to evaluate the feasibility of robot-assisted rehabilitation using myoelectric control on upper limb motor recovery. Methods In the present study, an exoskeleton-type rehabilitation robotic system was designed to provide voluntarily controlled assisted torque to the affected wrist. Voluntary intention was involved by using the residual surface electromyography (EMG) from flexor carpi radialis(FCR) and extensor carpi radialis (ECR)on the affected limb to control the mechanical assistance provided by the robotic system during wrist flexion and extension in a 20-session training. The system also applied constant resistant torque to the affected wrist during the training. Sixteen subjects after stroke had been recruited for evaluating the tracking performance and therapeutical effects of myoelectrically controlled robotic system. Results With the myoelectrically-controlled assistive torque, stroke survivors could reach a larger range of motion with a significant decrease in the EMG signal from the agonist muscles. The stroke survivors could be trained in the unreached range with their voluntary residual EMG on the paretic side. After 20-session rehabilitation training, there was a non-significant increase in the range of motion and a significant decrease in the root mean square error (RMSE) between the actual wrist angle and target angle. Significant improvements also could be found in muscle strength and clinical scales. Conclusions These results indicate that robot-aided therapy with voluntary participation of patient’s paretic motor system using myoelectric control might have positive effect on upper limb motor recovery. PMID:23758925
Intelligent surgical laser system configuration and software implementation
NASA Astrophysics Data System (ADS)
Hsueh, Chi-Fu T.; Bille, Josef F.
1992-06-01
An intelligent surgical laser system, which can help the ophthalmologist to achieve higher precision and control during their procedures, has been developed by ISL as model CLS 4001. In addition to the laser and laser delivery system, the system is also equipped with a vision system (IPU), robotics motion control (MCU), and a tracking closed loop system (ETS) that tracks the eye in three dimensions (X, Y and Z). The initial patient setup is computer controlled with guidance from the vision system. The tracking system is automatically engaged when the target is in position. A multi-level tracking system is developed by integrating our vision and tracking systems which have been able to maintain our laser beam precisely on target. The capabilities of the automatic eye setup and the tracking in three dimensions provides for improved accuracy and measurement repeatability. The system is operated through the Surgical Control Unit (SCU). The SCU communicates with the IPU and the MCU through both ethernet and RS232. Various scanning pattern (i.e., line, curve, circle, spiral, etc.) can be selected with given parameters. When a warning is activated, a voice message is played that will normally require a panel touch acknowledgement. The reliability of the system is ensured in three levels: (1) hardware, (2) software real time monitoring, and (3) user. The system is currently under clinical validation.
TENTACLE Multi-Camera Immersive Surveillance System Phase 2
2015-04-16
successful in solving the most challenging video analytics problems and taking the advanced research concepts into working systems for end- users in both...commercial, space and military applications. Notable successes include winning the DARPA Urban Challenge , software autonomy to guide the NASA robots (spirit... challenging urban environments. CMU is developing a scalable and extensible architecture, improving search/pursuit/tracking capabilities, and addressing
An assigned responsibility system for robotic teleoperation control.
Small, Nicolas; Lee, Kevin; Mann, Graham
2018-01-01
This paper proposes an architecture that explores a gap in the spectrum of existing strategies for robot control mode switching in adjustable autonomy. In situations where the environment is reasonably known and/or predictable, pre-planning these control changes could relieve robot operators of the additional task of deciding when and how to switch. Such a strategy provides a clear division of labour between the automation and the human operator(s) before the job even begins, allowing for individual responsibilities to be known ahead of time, limiting confusion and allowing rest breaks to be planned. Assigned Responsibility is a new form of adjustable autonomy-based teleoperation that allows the selective inclusion of automated control elements at key stages of a robot operation plan's execution. Progression through these stages is controlled by automatic goal accomplishment tracking. An implementation is evaluated through engineering tests and a usability study, demonstrating the viability of this approach and offering insight into its potential applications.
Direct model reference adaptive control of robotic arms
NASA Technical Reports Server (NTRS)
Kaufman, Howard; Swift, David C.; Cummings, Steven T.; Shankey, Jeffrey R.
1993-01-01
The results of controlling A PUMA 560 Robotic Manipulator and the NASA shuttle Remote Manipulator System (RMS) using a Command Generator Tracker (CGT) based Model Reference Adaptive Controller (DMRAC) are presented. Initially, the DMRAC algorithm was run in simulation using a detailed dynamic model of the PUMA 560. The algorithm was tuned on the simulation and then used to control the manipulator using minimum jerk trajectories as the desired reference inputs. The ability to track a trajectory in the presence of load changes was also investigated in the simulation. Satisfactory performance was achieved in both simulation and on the actual robot. The obtained responses showed that the algorithm was robust in the presence of sudden load changes. Because these results indicate that the DMRAC algorithm can indeed be successfully applied to the control of robotic manipulators, additional testing was performed to validate the applicability of DMRAC to simulated dynamics of the shuttle RMS.
Initial Experiments with the Leap Motion as a User Interface in Robotic Endonasal Surgery.
Travaglini, T A; Swaney, P J; Weaver, Kyle D; Webster, R J
The Leap Motion controller is a low-cost, optically-based hand tracking system that has recently been introduced on the consumer market. Prior studies have investigated its precision and accuracy, toward evaluating its usefulness as a surgical robot master interface. Yet due to the diversity of potential slave robots and surgical procedures, as well as the dynamic nature of surgery, it is challenging to make general conclusions from published accuracy and precision data. Thus, our goal in this paper is to explore the use of the Leap in the specific scenario of endonasal pituitary surgery. We use it to control a concentric tube continuum robot in a phantom study, and compare user performance using the Leap to previously published results using the Phantom Omni. We find that the users were able to achieve nearly identical average resection percentage and overall surgical duration with the Leap.
Initial Experiments with the Leap Motion as a User Interface in Robotic Endonasal Surgery
Travaglini, T. A.; Swaney, P. J.; Weaver, Kyle D.; Webster, R. J.
2016-01-01
The Leap Motion controller is a low-cost, optically-based hand tracking system that has recently been introduced on the consumer market. Prior studies have investigated its precision and accuracy, toward evaluating its usefulness as a surgical robot master interface. Yet due to the diversity of potential slave robots and surgical procedures, as well as the dynamic nature of surgery, it is challenging to make general conclusions from published accuracy and precision data. Thus, our goal in this paper is to explore the use of the Leap in the specific scenario of endonasal pituitary surgery. We use it to control a concentric tube continuum robot in a phantom study, and compare user performance using the Leap to previously published results using the Phantom Omni. We find that the users were able to achieve nearly identical average resection percentage and overall surgical duration with the Leap. PMID:26752501
An iterative learning control method with application for CNC machine tools
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kim, D.I.; Kim, S.
1996-01-01
A proportional, integral, and derivative (PID) type iterative learning controller is proposed for precise tracking control of industrial robots and computer numerical controller (CNC) machine tools performing repetitive tasks. The convergence of the output error by the proposed learning controller is guaranteed under a certain condition even when the system parameters are not known exactly and unknown external disturbances exist. As the proposed learning controller is repeatedly applied to the industrial robot or the CNC machine tool with the path-dependent repetitive task, the distance difference between the desired path and the actual tracked or machined path, which is one ofmore » the most significant factors in the evaluation of control performance, is progressively reduced. The experimental results demonstrate that the proposed learning controller can improve machining accuracy when the CNC machine tool performs repetitive machining tasks.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lee, Hau-Wei; Chen, Chieh-Li; Liu, Chien-Hung
Laser trackers are widely used in industry for tasks such as the assembly of airplanes and automobiles, contour measurement, and robot calibration. However, laser trackers are expensive, and the corresponding solution procedure is very complex. The influence of measurement uncertainties is also significant. This study proposes a three-dimensional space position measurement system which consists of two tracking modules, a zero tracking angle return subsystem, and a target quadrant photodiode (QPD). The target QPD is placed on the object being tracked. The origin locking method is used to keep the rays on the origin of the target QPD. The position ofmore » the target QPD is determined using triangulation since the two laser rays are projected onto one QPD. Modulation and demodulation are utilized to separate the coupled positional values. The experiment results show that measurement errors in the X, Y, and Z directions are less than {+-}0.05% when the measured object was moved by 300, 300, and 200 mm in the X, Y, and Z axes, respectively. The theoretical measurement error estimated from the measurement model is between {+-}0.02% and {+-}0.07% within the defined measurable range. The proposed system can be applied to the measurements of machine tools and robot arms.« less
Lee, Hau-Wei; Chen, Chieh-Li; Liu, Chien-Hung
2011-03-01
Laser trackers are widely used in industry for tasks such as the assembly of airplanes and automobiles, contour measurement, and robot calibration. However, laser trackers are expensive, and the corresponding solution procedure is very complex. The influence of measurement uncertainties is also significant. This study proposes a three-dimensional space position measurement system which consists of two tracking modules, a zero tracking angle return subsystem, and a target quadrant photodiode (QPD). The target QPD is placed on the object being tracked. The origin locking method is used to keep the rays on the origin of the target QPD. The position of the target QPD is determined using triangulation since the two laser rays are projected onto one QPD. Modulation and demodulation are utilized to separate the coupled positional values. The experiment results show that measurement errors in the X, Y, and Z directions are less than ±0.05% when the measured object was moved by 300, 300, and 200 mm in the X, Y, and Z axes, respectively. The theoretical measurement error estimated from the measurement model is between ±0.02% and ±0.07% within the defined measurable range. The proposed system can be applied to the measurements of machine tools and robot arms.
NASA Astrophysics Data System (ADS)
Lee, Hau-Wei; Chen, Chieh-Li; Liu, Chien-Hung
2011-03-01
Laser trackers are widely used in industry for tasks such as the assembly of airplanes and automobiles, contour measurement, and robot calibration. However, laser trackers are expensive, and the corresponding solution procedure is very complex. The influence of measurement uncertainties is also significant. This study proposes a three-dimensional space position measurement system which consists of two tracking modules, a zero tracking angle return subsystem, and a target quadrant photodiode (QPD). The target QPD is placed on the object being tracked. The origin locking method is used to keep the rays on the origin of the target QPD. The position of the target QPD is determined using triangulation since the two laser rays are projected onto one QPD. Modulation and demodulation are utilized to separate the coupled positional values. The experiment results show that measurement errors in the X, Y, and Z directions are less than ±0.05% when the measured object was moved by 300, 300, and 200 mm in the X, Y, and Z axes, respectively. The theoretical measurement error estimated from the measurement model is between ±0.02% and ±0.07% within the defined measurable range. The proposed system can be applied to the measurements of machine tools and robot arms.
Waspe, A C; Holdsworth, D W; Lacefield, J C; Fenster, A
2008-07-01
Preclinical research protocols often require the delivery of biological substances to specific targets in small animal disease models. To target biologically relevant locations in mice accurately, the needle positioning error needs to be < 200 μm. If targeting is inaccurate, experimental results can be inconclusive or misleading. We have developed a robotic manipulator that is capable of positioning a needle with a mean error < 100 μm. An apparatus and method were developed for integrating the needle-positioning robot with volumetric micro-computed tomography image guidance for interventions in small animals. Accurate image-to-robot registration is critical for integration as it enables targets identified in the image to be mapped to physical coordinates inside the animal. Registration is accomplished by injecting barium sulphate into needle tracks as the robot withdraws the needle from target points in a tissue-mimicking phantom. Registration accuracy is therefore affected by the positioning error of the robot and is assessed by measuring the point-to-line fiducial and target registration errors (FRE, TRE). Centroid points along cross-sectional slices of the track are determined using region growing segmentation followed by application of a center-of-mass algorithm. The centerline points are registered to needle trajectories in robot coordinates by applying an iterative closest point algorithm between points and lines. Implementing this procedure with four fiducial needle tracks produced a point-to-line FRE and TRE of 246 ± 58 μm and 194 ± 18 μm, respectively. The proposed registration technique produced a TRE < 200 μm, in the presence of robot positioning error, meeting design specification. © 2008 American Association of Physicists in Medicine.
Stochastic approach to error estimation for image-guided robotic systems.
Haidegger, Tamas; Gyõri, Sándor; Benyo, Balazs; Benyó, Zoltáán
2010-01-01
Image-guided surgical systems and surgical robots are primarily developed to provide patient safety through increased precision and minimal invasiveness. Even more, robotic devices should allow for refined treatments that are not possible by other means. It is crucial to determine the accuracy of a system, to define the expected overall task execution error. A major step toward this aim is to quantitatively analyze the effect of registration and tracking-series of multiplication of erroneous homogeneous transformations. First, the currently used models and algorithms are introduced along with their limitations, and a new, probability distribution based method is described. The new approach has several advantages, as it was demonstrated in our simulations. Primarily, it determines the full 6 degree of freedom accuracy of the point of interest, allowing for the more accurate use of advanced application-oriented concepts, such as Virtual Fixtures. On the other hand, it becomes feasible to consider different surgical scenarios with varying weighting factors.
Design and Development of a High Speed Sorting System Based on Machine Vision Guiding
NASA Astrophysics Data System (ADS)
Zhang, Wenchang; Mei, Jiangping; Ding, Yabin
In this paper, a vision-based control strategy to perform high speed pick-and-place tasks on automation product line is proposed, and relevant control software is develop. Using Delta robot to control a sucker to grasp disordered objects from one moving conveyer and then place them on the other in order. CCD camera gets one picture every time the conveyer moves a distance of ds. Objects position and shape are got after image processing. Target tracking method based on "Servo motor + synchronous conveyer" is used to fulfill the high speed porting operation real time. Experiments conducted on Delta robot sorting system demonstrate the efficiency and validity of the proposed vision-control strategy.
Ik Han, Seong; Lee, Jangmyung
2016-11-01
This paper presents finite-time sliding mode control (FSMC) with predefined constraints for the tracking error and sliding surface in order to obtain robust positioning of a robot manipulator with input nonlinearity due to an unknown deadzone and external disturbance. An assumed model feedforward FSMC was designed to avoid tedious identification procedures for the manipulator parameters and to obtain a fast response time. Two constraint switching control functions based on the tracking error and finite-time sliding surface were added to the FSMC to guarantee the predefined tracking performance despite the presence of an unknown deadzone and disturbance. The tracking error due to the deadzone and disturbance can be suppressed within the predefined error boundary simply by tuning the gain value of the constraint switching function and without the addition of an extra compensator. Therefore, the designed constraint controller has a simpler structure than conventional transformed error constraint methods and the sliding surface constraint scheme can also indirectly guarantee the tracking error constraint while being more stable than the tracking error constraint control. A simulation and experiment were performed on an articulated robot manipulator to validate the proposed control schemes. Copyright © 2016 ISA. Published by Elsevier Ltd. All rights reserved.
2018-01-01
Although the use of the surgical robot is rapidly expanding for various medical treatments, there still exist safety issues and concerns about robot-assisted surgeries due to limited vision through a laparoscope, which may cause compromised situation awareness and surgical errors requiring rapid emergency conversion to open surgery. To assist surgeon's situation awareness and preventive emergency response, this study proposes situation information guidance through a vision-based common algorithm architecture for automatic detection and tracking of intraoperative hemorrhage and surgical instruments. The proposed common architecture comprises the location of the object of interest using feature texture, morphological information, and the tracking of the object based on Kalman filter for robustness with reduced error. The average recall and precision of the instrument detection in four prostate surgery videos were 96% and 86%, and the accuracy of the hemorrhage detection in two prostate surgery videos was 98%. Results demonstrate the robustness of the automatic intraoperative object detection and tracking which can be used to enhance the surgeon's preventive state recognition during robot-assisted surgery. PMID:29854366
An orbital emulator for pursuit-evasion game theoretic sensor management
NASA Astrophysics Data System (ADS)
Shen, Dan; Wang, Tao; Wang, Gang; Jia, Bin; Wang, Zhonghai; Chen, Genshe; Blasch, Erik; Pham, Khanh
2017-05-01
This paper develops and evaluates an orbital emulator (OE) for space situational awareness (SSA). The OE can produce 3D satellite movements using capabilities generated from omni-wheeled robot and robotic arm motion methods. The 3D motion of a satellite is partitioned into the movements in the equatorial plane and the up-down motions in the vertical plane. The 3D actions are emulated by omni-wheeled robot models while the up-down motions are performed by a stepped-motor-controlled-ball along a rod (robotic arm), which is attached to the robot. For multiple satellites, a fast map-merging algorithm is integrated into the robot operating system (ROS) and simultaneous localization and mapping (SLAM) routines to locate the multiple robots in the scene. The OE is used to demonstrate a pursuit-evasion (PE) game theoretic sensor management algorithm, which models conflicts between a space-based-visible (SBV) satellite (as pursuer) and a geosynchronous (GEO) satellite (as evader). The cost function of the PE game is based on the informational entropy of the SBV-tracking-GEO scenario. GEO can maneuver using a continuous and low thruster. The hard-in-loop space emulator visually illustrates the SSA problem solution based PE game.
The use of interactive computer vision and robot hand controllers for enhancing manufacturing safety
NASA Technical Reports Server (NTRS)
Marzwell, Neville I.; Jacobus, Charles J.; Peurach, Thomas M.; Mitchell, Brian T.
1994-01-01
Current available robotic systems provide limited support for CAD-based model-driven visualization, sensing algorithm development and integration, and automated graphical planning systems. This paper describes ongoing work which provides the functionality necessary to apply advanced robotics to automated manufacturing and assembly operations. An interface has been built which incorporates 6-DOF tactile manipulation, displays for three dimensional graphical models, and automated tracking functions which depend on automated machine vision. A set of tools for single and multiple focal plane sensor image processing and understanding has been demonstrated which utilizes object recognition models. The resulting tool will enable sensing and planning from computationally simple graphical objects. A synergistic interplay between human and operator vision is created from programmable feedback received from the controller. This approach can be used as the basis for implementing enhanced safety in automated robotics manufacturing, assembly, repair and inspection tasks in both ground and space applications. Thus, an interactive capability has been developed to match the modeled environment to the real task environment for safe and predictable task execution.
The use of interactive computer vision and robot hand controllers for enhancing manufacturing safety
NASA Astrophysics Data System (ADS)
Marzwell, Neville I.; Jacobus, Charles J.; Peurach, Thomas M.; Mitchell, Brian T.
1994-02-01
Current available robotic systems provide limited support for CAD-based model-driven visualization, sensing algorithm development and integration, and automated graphical planning systems. This paper describes ongoing work which provides the functionality necessary to apply advanced robotics to automated manufacturing and assembly operations. An interface has been built which incorporates 6-DOF tactile manipulation, displays for three dimensional graphical models, and automated tracking functions which depend on automated machine vision. A set of tools for single and multiple focal plane sensor image processing and understanding has been demonstrated which utilizes object recognition models. The resulting tool will enable sensing and planning from computationally simple graphical objects. A synergistic interplay between human and operator vision is created from programmable feedback received from the controller. This approach can be used as the basis for implementing enhanced safety in automated robotics manufacturing, assembly, repair and inspection tasks in both ground and space applications. Thus, an interactive capability has been developed to match the modeled environment to the real task environment for safe and predictable task execution.
Modeling and control of tissue compression and temperature for automation in robot-assisted surgery.
Sinha, Utkarsh; Li, Baichun; Sankaranarayanan, Ganesh
2014-01-01
Robotic surgery is being used widely due to its various benefits that includes reduced patient trauma and increased dexterity and ergonomics for the operating surgeon. Making the whole or part of the surgical procedure autonomous increases patient safety and will enable the robotic surgery platform to be used in telesurgery. In this work, an Electrosurgery procedure that involves tissue compression and application of heat such as the coaptic vessel closure has been automated. A MIMO nonlinear model characterizing the tissue stiffness and conductance under compression was feedback linearized and tuned PID controllers were used to control the system to achieve both the displacement and temperature constraints. A reference input for both the constraints were chosen as a ramp and hold trajectory which reflect the real constraints that exist in an actual surgical procedure. Our simulations showed that the controllers successfully tracked the reference trajectories with minimal deviation and in finite time horizon. The MIMO system with controllers developed in this work can be used to drive a surgical robot autonomously and perform electrosurgical procedures such as coaptic vessel closures.
Real-time detection of moving objects from moving vehicles using dense stereo and optical flow
NASA Technical Reports Server (NTRS)
Talukder, Ashit; Matthies, Larry
2004-01-01
Dynamic scene perception is very important for autonomous vehicles operating around other moving vehicles and humans. Most work on real-time object tracking from moving platforms has used sparse features or assumed flat scene structures. We have recently extended a real-time, dense stereo system to include realtime, dense optical flow, enabling more comprehensive dynamic scene analysis. We describe algorithms to robustly estimate 6-DOF robot egomotion in the presence of moving objects using dense flow and dense stereo. We then use dense stereo and egomotion estimates to identify & other moving objects while the robot itself is moving. We present results showing accurate egomotion estimation and detection of moving people and vehicles under general 6-DOF motion of the robot and independently moving objects. The system runs at 18.3 Hz on a 1.4 GHz Pentium M laptop, computing 160x120 disparity maps and optical flow fields, egomotion, and moving object segmentation. We believe this is a significant step toward general unconstrained dynamic scene analysis for mobile robots, as well as for improved position estimation where GPS is unavailable.
Real-time detection of moving objects from moving vehicles using dense stereo and optical flow
NASA Technical Reports Server (NTRS)
Talukder, Ashit; Matthies, Larry
2004-01-01
Dynamic scene perception is very important for autonomous vehicles operating around other moving vehicles and humans. Most work on real-time object tracking from moving platforms has used sparse features or assumed flat scene structures. We have recently extended a real-time, dense stereo system to include real-time, dense optical flow, enabling more comprehensive dynamic scene analysis. We describe algorithms to robustly estimate 6-DOF robot egomotion in the presence of moving objects using dense flow and dense stereo. We then use dense stereo and egomotion estimates to identity other moving objects while the robot itself is moving. We present results showing accurate egomotion estimation and detection of moving people and vehicles under general 6-DOF motion of the robot and independently moving objects. The system runs at 18.3 Hz on a 1.4 GHz Pentium M laptop, computing 160x120 disparity maps and optical flow fields, egomotion, and moving object segmentation. We believe this is a significant step toward general unconstrained dynamic scene analysis for mobile robots, as well as for improved position estimation where GPS is unavailable.
Real-time Detection of Moving Objects from Moving Vehicles Using Dense Stereo and Optical Flow
NASA Technical Reports Server (NTRS)
Talukder, Ashit; Matthies, Larry
2004-01-01
Dynamic scene perception is very important for autonomous vehicles operating around other moving vehicles and humans. Most work on real-time object tracking from moving platforms has used sparse features or assumed flat scene structures. We have recently extended a real-time. dense stereo system to include realtime. dense optical flow, enabling more comprehensive dynamic scene analysis. We describe algorithms to robustly estimate 6-DOF robot egomotion in the presence of moving objects using dense flow and dense stereo. We then use dense stereo and egomotion estimates to identify other moving objects while the robot itself is moving. We present results showing accurate egomotion estimation and detection of moving people and vehicles under general 6DOF motion of the robot and independently moving objects. The system runs at 18.3 Hz on a 1.4 GHz Pentium M laptop. computing 160x120 disparity maps and optical flow fields, egomotion, and moving object segmentation. We believe this is a significant step toward general unconstrained dynamic scene analysis for mobile robots, as well as for improved position estimation where GPS is unavailable.
Development of quadruped walking locomotion gait generator using a hybrid method
NASA Astrophysics Data System (ADS)
Jasni, F.; Shafie, A. A.
2013-12-01
The earth, in many areas is hardly reachable by the wheeled or tracked locomotion system. Thus, walking locomotion system is becoming a favourite option for mobile robot these days. This is because of the ability of walking locomotion to move on the rugged and unlevel terrains. However, to develop a walking locomotion gait for a robot is not a simple task. Central Pattern Generator (CPGs) method is a biological inspired method that is introduced as a method to develop the gait for the walking robot recently to tackle the issue faced by the conventional method of pre-designed trajectory based method. However, research shows that even the CPG method do have some limitations. Thus, in this paper, a hybrid method that combines CPG and the pre-designed trajectory based method is introduced to develop a walking gait for quadruped walking robot. The 3-D foot trajectories and the joint angle trajectories developed using the proposed method are compared with the data obtained via the conventional method of pre-designed trajectory to confirm the performance.
PIMS sequencing extension: a laboratory information management system for DNA sequencing facilities.
Troshin, Peter V; Postis, Vincent Lg; Ashworth, Denise; Baldwin, Stephen A; McPherson, Michael J; Barton, Geoffrey J
2011-03-07
Facilities that provide a service for DNA sequencing typically support large numbers of users and experiment types. The cost of services is often reduced by the use of liquid handling robots but the efficiency of such facilities is hampered because the software for such robots does not usually integrate well with the systems that run the sequencing machines. Accordingly, there is a need for software systems capable of integrating different robotic systems and managing sample information for DNA sequencing services. In this paper, we describe an extension to the Protein Information Management System (PIMS) that is designed for DNA sequencing facilities. The new version of PIMS has a user-friendly web interface and integrates all aspects of the sequencing process, including sample submission, handling and tracking, together with capture and management of the data. The PIMS sequencing extension has been in production since July 2009 at the University of Leeds DNA Sequencing Facility. It has completely replaced manual data handling and simplified the tasks of data management and user communication. Samples from 45 groups have been processed with an average throughput of 10000 samples per month. The current version of the PIMS sequencing extension works with Applied Biosystems 3130XL 96-well plate sequencer and MWG 4204 or Aviso Theonyx liquid handling robots, but is readily adaptable for use with other combinations of robots. PIMS has been extended to provide a user-friendly and integrated data management solution for DNA sequencing facilities that is accessed through a normal web browser and allows simultaneous access by multiple users as well as facility managers. The system integrates sequencing and liquid handling robots, manages the data flow, and provides remote access to the sequencing results. The software is freely available, for academic users, from http://www.pims-lims.org/.
Study of the Navigation Method for a Snake Robot Based on the Kinematics Model with MEMS IMU
Dou, Lihua; Su, Zhong; Liu, Ning
2018-01-01
A snake robot is a type of highly redundant mobile robot that significantly differs from a tracked robot, wheeled robot and legged robot. To address the issue of a snake robot performing self-localization in the application environment without assistant orientation, an autonomous navigation method is proposed based on the snake robot’s motion characteristic constraints. The method realized the autonomous navigation of the snake robot with non-nodes and an external assistant using its own Micro-Electromechanical-Systems (MEMS) Inertial-Measurement-Unit (IMU). First, it studies the snake robot’s motion characteristics, builds the kinematics model, and then analyses the motion constraint characteristics and motion error propagation properties. Second, it explores the snake robot’s navigation layout, proposes a constraint criterion and the fixed relationship, and makes zero-state constraints based on the motion features and control modes of a snake robot. Finally, it realizes autonomous navigation positioning based on the Extended-Kalman-Filter (EKF) position estimation method under the constraints of its motion characteristics. With the self-developed snake robot, the test verifies the proposed method, and the position error is less than 5% of Total-Traveled-Distance (TDD). In a short-distance environment, this method is able to meet the requirements of a snake robot in order to perform autonomous navigation and positioning in traditional applications and can be extended to other familiar multi-link robots. PMID:29547515
Irradiation control parameters for computer-assisted laser photocoagulation of the retina
NASA Astrophysics Data System (ADS)
Naess, Espen; Molvik, Torstein; Barrett, Steven F.; Wright, Cameron H. G.; de Graaf, Peter W.
2001-06-01
A system for robotically assisted retinal surgery has been developed to rapidly and safely place lesions on the retina for photocoagulation therapy. This system provides real- time, motion stabilized lesion placement for typical irradiation times of 100 ms. The system consists of three main subsystems: a global, digital-based tracking subsystem; a fast, local analog tracking subsystem; and a confocal reflectance subsystem to control lesion parameters dynamically. We have reported on these subsystems in previous SPIE presentations. This paper concentrates on the development of the second hybrid system prototype. Considerable progress has been made toward reducing the footprint of the optical system, simplifying the user interface, fully characterizing the analog tracking system and using measurable lesion reflectance growth parameters to develop a noninvasive method to infer lesion depth. This method will allow dynamic control of laser dosimetry to provide similar lesions across the non-uniform retinal surface. These system improvements and progress toward a clinically significant system are covered in detail within this paper.
Tracked robot controllers for climbing obstacles autonomously
NASA Astrophysics Data System (ADS)
Vincent, Isabelle
2009-05-01
Research in mobile robot navigation has demonstrated some success in navigating flat indoor environments while avoiding obstacles. However, the challenge of analyzing complex environments to climb obstacles autonomously has had very little success due to the complexity of the task. Unmanned ground vehicles currently exhibit simple autonomous behaviours compared to the human ability to move in the world. This paper presents the control algorithms designed for a tracked mobile robot to autonomously climb obstacles by varying its tracks configuration. Two control algorithms are proposed to solve the autonomous locomotion problem for climbing obstacles. First, a reactive controller evaluates the appropriate geometric configuration based on terrain and vehicle geometric considerations. Then, a reinforcement learning algorithm finds alternative solutions when the reactive controller gets stuck while climbing an obstacle. The methodology combines reactivity to learning. The controllers have been demonstrated in box and stair climbing simulations. The experiments illustrate the effectiveness of the proposed approach for crossing obstacles.
Schwein, Adeline; Kramer, Benjamin; Chinnadurai, Ponraj; Virmani, Neha; Walker, Sean; O'Malley, Marcia; Lumsden, Alan B; Bismuth, Jean
2018-04-01
Combining three-dimensional (3D) catheter control with electromagnetic (EM) tracking-based navigation significantly reduced fluoroscopy time and improved robotic catheter movement quality in a previous in vitro pilot study. The aim of this study was to expound on previous results and to expand the value of EM tracking with a novel feature, assistednavigation, allowing automatic catheter orientation and semiautomatic vessel cannulation. Eighteen users navigated a robotic catheter in an aortic aneurysm phantom using an EM guidewire and a modified 9F robotic catheter with EM sensors at the tip of both leader and sheath. All users cannulated two targets, the left renal artery and posterior gate, using four visualization modes: (1) Standard fluoroscopy (control). (2) 2D biplane fluoroscopy showing real-time virtual catheter localization and orientation from EM tracking. (3) 2D biplane fluoroscopy with novel EM assisted navigation allowing the user to define the target vessel. The robotic catheter orients itself automatically toward the target; the user then only needs to advance the guidewire following this predefined optimized path to catheterize the vessel. Then, while advancing the catheter over the wire, the assisted navigation automatically modifies catheter bending and rotation in order to ensure smooth progression, avoiding loss of wire access. (4) Virtual 3D representation of the phantom showing real-time virtual catheter localization and orientation. Standard fluoroscopy was always available; cannulation and fluoroscopy times were noted for every mode and target cannulation. Quality of catheter movement was assessed by measuring the number of submovements of the catheter using the 3D coordinates of the EM sensors. A t-test was used to compare the standard fluoroscopy mode against EM tracking modes. EM tracking significantly reduced the mean fluoroscopy time (P < .001) and the number of submovements (P < .02) for both cannulation tasks. For the posterior gate, mean cannulation time was also significantly reduced when using EM tracking (P < .001). The use of novel EM assisted navigation feature (mode 3) showed further reduced cannulation time for the posterior gate (P = .002) and improved quality of catheter movement for the left renal artery cannulation (P = .021). These results confirmed the findings of a prior study that highlighted the value of combining 3D robotic catheter control and 3D navigation to improve safety and efficiency of endovascular procedures. The novel EM assisted navigation feature augments the robotic master/slave concept with automated catheter orientation toward the target and shows promising results in reducing procedure time and improving catheter motion quality. Copyright © 2017 Society for Vascular Surgery. Published by Elsevier Inc. All rights reserved.
Whitehurst, Sabrina V; Lockrow, Ernest G; Lendvay, Thomas S; Propst, Anthony M; Dunlow, Susan G; Rosemeyer, Christopher J; Gobern, Joseph M; White, Lee W; Skinner, Anna; Buller, Jerome L
2015-01-01
To compare the efficacy of simulation-based training between the Mimic dV- Trainer and traditional dry lab da Vinci robot training. A prospective randomized study analyzing the performance of 20 robotics-naive participants. Participants were enrolled in an online da Vinci Intuitive Surgical didactic training module, followed by training in use of the da Vinci standard surgical robot. Spatial ability tests were performed as well. Participants were randomly assigned to 1 of 2 training conditions: performance of 3 Fundamentals of Laparoscopic Surgery dry lab tasks using the da Vinci or performance of 4 dV-Trainer tasks. Participants in both groups performed all tasks to empirically establish proficiency criterion. Participants then performed the transfer task, a cystotomy closure using the daVinci robot on a live animal (swine) model. The performance of robotic tasks was blindly assessed by a panel of experienced surgeons using objective tracking data and using the validated Global Evaluative Assessment of Robotic Surgery (GEARS), a structured assessment tool. No statistically significant difference in surgeon performance was found between the 2 training conditions, dV-Trainer and da Vinci robot. Analysis of a 95% confidence interval for the difference in means (-0.803 to 0.543) indicated that the 2 methods are unlikely to differ to an extent that would be clinically meaningful. Based on the results of this study, a curriculum on the dV- Trainer was shown to be comparable to traditional da Vinci robot training. Therefore, we have identified that training on a virtual reality system may be an alternative to live animal training for future robotic surgeons. Published by Elsevier Inc.
Rusydi, Muhammad Ilhamdi; Sasaki, Minoru; Ito, Satoshi
2014-01-01
Biosignals will play an important role in building communication between machines and humans. One of the types of biosignals that is widely used in neuroscience are electrooculography (EOG) signals. An EOG has a linear relationship with eye movement displacement. Experiments were performed to construct a gaze motion tracking method indicated by robot manipulator movements. Three operators looked at 24 target points displayed on a monitor that was 40 cm in front of them. Two channels (Ch1 and Ch2) produced EOG signals for every single eye movement. These signals were converted to pixel units by using the linear relationship between EOG signals and gaze motion distances. The conversion outcomes were actual pixel locations. An affine transform method is proposed to determine the shift of actual pixels to target pixels. This method consisted of sequences of five geometry processes, which are translation-1, rotation, translation-2, shear and dilatation. The accuracy was approximately 0.86° ± 0.67° in the horizontal direction and 0.54° ± 0.34° in the vertical. This system successfully tracked the gaze motions not only in direction, but also in distance. Using this system, three operators could operate a robot manipulator to point at some targets. This result shows that the method is reliable in building communication between humans and machines using EOGs. PMID:24919013
Fusing human and machine skills for remote robotic operations
NASA Technical Reports Server (NTRS)
Schenker, Paul S.; Kim, Won S.; Venema, Steven C.; Bejczy, Antal K.
1991-01-01
The question of how computer assists can improve teleoperator trajectory tracking during both free and force-constrained motions is addressed. Computer graphics techniques which enable the human operator to both visualize and predict detailed 3D trajectories in real-time are reported. Man-machine interactive control procedures for better management of manipulator contact forces and positioning are also described. It is found that collectively, these novel advanced teleoperations techniques both enhance system performance and significantly reduce control problems long associated with teleoperations under time delay. Ongoing robotic simulations of the 1984 space shuttle Solar Maximum EVA Repair Mission are briefly described.
Results Of Automating A Photolithography Cell In A Clean Tunnel
NASA Astrophysics Data System (ADS)
June, David H.
1987-01-01
A prototype automated photobay was installed in an existing fab area utilizing flexible material handling techniques within a clean tunnel. The project objective was to prove design concepts of automated cassette-to-cassette handling within a clean tunnel that isolated operators from the wafers being processed. Material handling was by monorail track transport system to feed cassettes to pick and place robots. The robots loaded and unloaded cassettes of wafers to each of the various pieces of process equipment. The material handling algorithms, recipe downloading and statistical process control functions were all performed by custom software on the photobay cell controller.
Schlosser, Jeffrey; Gong, Ren Hui; Bruder, Ralf; Schweikard, Achim; Jang, Sungjune; Henrie, John; Kamaya, Aya; Koong, Albert; Chang, Daniel T; Hristov, Dimitre
2016-11-01
To present a system for robotic 4D ultrasound (US) imaging concurrent with radiotherapy beam delivery and estimate the proportion of liver stereotactic ablative body radiotherapy (SABR) cases in which robotic US image guidance can be deployed without interfering with clinically used VMAT beam configurations. The image guidance hardware comprises a 4D US machine, an optical tracking system for measuring US probe pose, and a custom-designed robot for acquiring hands-free US volumes. In software, a simulation environment incorporating the LINAC, couch, planning CT, and robotic US guidance hardware was developed. Placement of the robotic US hardware was guided by a target visibility map rendered on the CT surface by using the planning CT to simulate US propagation. The visibility map was validated in a prostate phantom and evaluated in patients by capturing live US from imaging positions suggested by the visibility map. In 20 liver SABR patients treated with VMAT, the simulation environment was used to virtually place the robotic hardware and US probe. Imaging targets were either planning target volumes (PTVs, range 5.9-679.5 ml) or gross tumor volumes (GTVs, range 0.9-343.4 ml). Presence or absence of mechanical interference with LINAC, couch, and patient body as well as interferences with treated beams was recorded. For PTV targets, robotic US guidance without mechanical interference was possible in 80% of the cases and guidance without beam interference was possible in 60% of the cases. For the smaller GTV targets, these proportions were 95% and 85%, respectively. GTV size (1/20), elongated shape (1/20), and depth (1/20) were the main factors limiting the availability of noninterfering imaging positions. The robotic US imaging system was deployed in two liver SABR patients during CT simulation with successful acquisition of 4D US sequences in different imaging positions. This study indicates that for VMAT liver SABR, robotic US imaging of a relevant internal target may be possible in 85% of the cases while using treatment plans currently deployed in the clinic. With beam replanning to account for the presence of robotic US guidance, intrafractional US may be an option for 95% of the liver SABR cases.
Cooperative Robots to Observe Moving Targets: Review.
Khan, Asif; Rinner, Bernhard; Cavallaro, Andrea
2018-01-01
The deployment of multiple robots for achieving a common goal helps to improve the performance, efficiency, and/or robustness in a variety of tasks. In particular, the observation of moving targets is an important multirobot application that still exhibits numerous open challenges, including the effective coordination of the robots. This paper reviews control techniques for cooperative mobile robots monitoring multiple targets. The simultaneous movement of robots and targets makes this problem particularly interesting, and our review systematically addresses this cooperative multirobot problem for the first time. We classify and critically discuss the control techniques: cooperative multirobot observation of multiple moving targets, cooperative search, acquisition, and track, cooperative tracking, and multirobot pursuit evasion. We also identify the five major elements that characterize this problem, namely, the coordination method, the environment, the target, the robot and its sensor(s). These elements are used to systematically analyze the control techniques. The majority of the studied work is based on simulation and laboratory studies, which may not accurately reflect real-world operational conditions. Importantly, while our systematic analysis is focused on multitarget observation, our proposed classification is useful also for related multirobot applications.
Remote Viewer for Maritime Robotics Software
NASA Technical Reports Server (NTRS)
Kuwata, Yoshiaki; Wolf, Michael; Huntsberger, Terrance L.; Howard, Andrew B.
2013-01-01
This software is a viewer program for maritime robotics software that provides a 3D visualization of the boat pose, its position history, ENC (Electrical Nautical Chart) information, camera images, map overlay, and detected tracks.
Manufacturing process applications team (MATeam)
NASA Technical Reports Server (NTRS)
Bangs, E. R.
1980-01-01
The objectives and activities of an aerospace technology transfer group are outlined and programs in various stages of progress are described including the orbital tube flaring device, infrared proximity sensor for robot positioning, laser stripping magnet wire, infrared imaging as welding process tracking system, carbide coating of cutting tools, nondestructive fracture toughness testing of titanium welds, portable solar system for agricultural applications, and an anerobic methane gas generator.
PRIMUS: autonomous navigation in open terrain with a tracked vehicle
NASA Astrophysics Data System (ADS)
Schaub, Guenter W.; Pfaendner, Alfred H.; Schaefer, Christoph
2004-09-01
The German experimental robotics program PRIMUS (PRogram for Intelligent Mobile Unmanned Systems) is focused on solutions for autonomous driving in unknown open terrain, over several project phases under specific realization aspects for more than 12 years. The main task of the program is to develop algorithms for a high degree of autonomous navigation skills with off-the-shelf available hardware/sensor technology and to integrate this into military vehicles. For obstacle detection a Dornier-3D-LADAR is integrated on a tracked vehicle "Digitized WIESEL 2". For road-following a digital video camera and a visual perception module from the Universitaet der Bundeswehr Munchen (UBM) has been integrated. This paper gives an overview of the PRIMUS program with a focus on the last program phase D (2001 - 2003). This includes the system architecture, the description of the modes of operation and the technology development with the focus on obstacle avoidance and obstacle classification using a 3-D LADAR. A collection of experimental results and a short look at the next steps in the German robotics program will conclude the paper.
Experiments in teleoperator and autonomous control of space robotic vehicles
NASA Technical Reports Server (NTRS)
Alexander, Harold L.
1991-01-01
A program of research embracing teleoperator and automatic navigational control of freely flying satellite robots is presented. Current research goals include: (1) developing visual operator interfaces for improved vehicle teleoperation; (2) determining the effects of different visual interface system designs on operator performance; and (3) achieving autonomous vision-based vehicle navigation and control. This research program combines virtual-environment teleoperation studies and neutral-buoyancy experiments using a space-robot simulator vehicle currently under development. Visual-interface design options under investigation include monoscopic versus stereoscopic displays and cameras, helmet-mounted versus panel-mounted display monitors, head-tracking versus fixed or manually steerable remote cameras, and the provision of vehicle-fixed visual cues, or markers, in the remote scene for improved sensing of vehicle position, orientation, and motion.
Autonomous intelligent assembly systems LDRD 105746 final report.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Anderson, Robert J.
2013-04-01
This report documents a three-year to develop technology that enables mobile robots to perform autonomous assembly tasks in unstructured outdoor environments. This is a multi-tier problem that requires an integration of a large number of different software technologies including: command and control, estimation and localization, distributed communications, object recognition, pose estimation, real-time scanning, and scene interpretation. Although ultimately unsuccessful in achieving a target brick stacking task autonomously, numerous important component technologies were nevertheless developed. Such technologies include: a patent-pending polygon snake algorithm for robust feature tracking, a color grid algorithm for uniquely identification and calibration, a command and control frameworkmore » for abstracting robot commands, a scanning capability that utilizes a compact robot portable scanner, and more. This report describes this project and these developed technologies.« less
Red ball ranging optimization based on dual camera ranging method
NASA Astrophysics Data System (ADS)
Kuang, Lei; Sun, Weijia; Liu, Jiaming; Tang, Matthew Wai-Chung
2018-05-01
In this paper, the process of positioning and moving to target red ball by NAO robot through its camera system is analyzed and improved using the dual camera ranging method. The single camera ranging method, which is adapted by NAO robot, was first studied and experimented. Since the existing error of current NAO Robot is not a single variable, the experiments were divided into two parts to obtain more accurate single camera ranging experiment data: forward ranging and backward ranging. Moreover, two USB cameras were used in our experiments that adapted Hough's circular method to identify a ball, while the HSV color space model was used to identify red color. Our results showed that the dual camera ranging method reduced the variance of error in ball tracking from 0.68 to 0.20.
The detrimental effect of friction on space microgravity robotics
NASA Technical Reports Server (NTRS)
Newman, Wyatt S.; Glosser, Gregory D.; Miller, Jeffrey H.; Rohn, Douglas
1992-01-01
The authors present an analysis of why control systems are ineffective in compensating for acceleration disturbances due to Coulomb friction. Linear arguments indicate that the effects of Coulomb friction on a body are most difficult to reject when the control actuator is separated from the body of compliance. The linear arguments were illustrated in a nonlinear simulation of optimal linear tracking control in the presence of nonlinear friction. The results of endpoint acceleration measurements for four robot designs are presented and are compared with simulation and to equivalent measurements on a human. It is concluded that Coulomb friction in common bearings and transmission induces unacceptable levels of endpoint acceleration, that these accelerations cannot be adequately attenuated by control, and that robots for microgravity work will require special design considerations for inherently low friction.
3D Perception Technologies for Surgical Operating Theatres.
Beyl, T; Schreiter, L; Nicolai, P; Raczkowsky, J; Wörn, H
2016-01-01
3D Perception technologies have been explored in various fields. This paper explores the application of such technologies for surgical operating theatres. Clinical applications can be found in workflow detection, tracking and analysis, collision avoidance with medical robots, perception of interaction between participants of the operation, training of the operation room crew, patient calibration and many more. In this paper a complete perception solution for the operating room is shown. The system is based on the ToF technology integrated to the Microsoft Kinect One implements a multi camera approach. Special emphasize is put on the tracking of the personnel and the evaluation of the system performance and accuracy.
Learning to Rapidly Re-Contact the Lost Plume in Chemical Plume Tracing
Cao, Meng-Li; Meng, Qing-Hao; Wang, Jia-Ying; Luo, Bing; Jing, Ya-Qi; Ma, Shu-Gen
2015-01-01
Maintaining contact between the robot and plume is significant in chemical plume tracing (CPT). In the time immediately following the loss of chemical detection during the process of CPT, Track-Out activities bias the robot heading relative to the upwind direction, expecting to rapidly re-contact the plume. To determine the bias angle used in the Track-Out activity, we propose an online instance-based reinforcement learning method, namely virtual trail following (VTF). In VTF, action-value is generalized from recently stored instances of successful Track-Out activities. We also propose a collaborative VTF (cVTF) method, in which multiple robots store their own instances, and learn from the stored instances, in the same database. The proposed VTF and cVTF methods are compared with biased upwind surge (BUS) method, in which all Track-Out activities utilize an offline optimized universal bias angle, in an indoor environment with three different airflow fields. With respect to our experimental conditions, VTF and cVTF show stronger adaptability to different airflow environments than BUS, and furthermore, cVTF yields higher success rates and time-efficiencies than VTF. PMID:25825974
Dual adaptive dynamic control of mobile robots using neural networks.
Bugeja, Marvin K; Fabri, Simon G; Camilleri, Liberato
2009-02-01
This paper proposes two novel dual adaptive neural control schemes for the dynamic control of nonholonomic mobile robots. The two schemes are developed in discrete time, and the robot's nonlinear dynamic functions are assumed to be unknown. Gaussian radial basis function and sigmoidal multilayer perceptron neural networks are used for function approximation. In each scheme, the unknown network parameters are estimated stochastically in real time, and no preliminary offline neural network training is used. In contrast to other adaptive techniques hitherto proposed in the literature on mobile robots, the dual control laws presented in this paper do not rely on the heuristic certainty equivalence property but account for the uncertainty in the estimates. This results in a major improvement in tracking performance, despite the plant uncertainty and unmodeled dynamics. Monte Carlo simulation and statistical hypothesis testing are used to illustrate the effectiveness of the two proposed stochastic controllers as applied to the trajectory-tracking problem of a differentially driven wheeled mobile robot.
NASA Astrophysics Data System (ADS)
Konacki, M.; Lejba, P.; Sybilski, P.; Pawłaszek, R.; Kozłowski, S.; Suchodolski, T.; Słonina, M.; Litwicki, M.; Sybilska, A.; Rogowska, B.; Kolb, U.; Burwitz, V.; Baader, J.; Groot, P.; Bloemen, S.; Ratajczak, M.; Hełminiak, K.; Borek, R.; Chodosiewicz, P.; Chimicz, A.
We present an update on the preparation of our assets that consists of a robotic network of eight optical telescopes and a laser ranging station for regular services in the SST domain. We report the development of new optical assets that include a double telescope system, Panoptes-1AB, and a new astrograph on our Solaris-3 telescope at the Siding Spring Observatory, Australia. Progress in the software development necessary for smooth SST operation includes a web based portal and an XML Azure Queue scheduling for the network giving easy access to our sensors. Astrometry24.net our new prototype cloud service for fast astrometry, streak detection and measurement with precision and performance results is also described. In the laser domain, for more than a year, Space Research Centre Borowiec laser station has regularly tracked space debris cooperative and uncooperative targets. The efforts of the stations’ staff have been focused on the tracking of typical rocket bodies from the LEO regime. Additionally, a second independent laser system fully dedicated to SST activities is under development. It will allow for an increased pace of operation of our consortium in the global SST laser domain.
Vicentini, Federico; Pedrocchi, Nicola; Malosio, Matteo; Molinari Tosatti, Lorenzo
2014-09-01
Robot-assisted neurorehabilitation often involves networked systems of sensors ("sensory rooms") and powerful devices in physical interaction with weak users. Safety is unquestionably a primary concern. Some lightweight robot platforms and devices designed on purpose include safety properties using redundant sensors or intrinsic safety design (e.g. compliance and backdrivability, limited exchange of energy). Nonetheless, the entire "sensory room" shall be required to be fail-safe and safely monitored as a system at large. Yet, sensor capabilities and control algorithms used in functional therapies require, in general, frequent updates or re-configurations, making a safety-grade release of such devices hardly sustainable in cost-effectiveness and development time. As such, promising integrated platforms for human-in-the-loop therapies could not find clinical application and manufacturing support because of lacking in the maintenance of global fail-safe properties. Under the general context of cross-machinery safety standards, the paper presents a methodology called SafeNet for helping in extending the safety rate of Human Robot Interaction (HRI) systems using unsafe components, including sensors and controllers. SafeNet considers, in fact, the robotic system as a device at large and applies the principles of functional safety (as in ISO 13489-1) through a set of architectural procedures and implementation rules. The enabled capability of monitoring a network of unsafe devices through redundant computational nodes, allows the usage of any custom sensors and algorithms, usually planned and assembled at therapy planning-time rather than at platform design-time. A case study is presented with an actual implementation of the proposed methodology. A specific architectural solution is applied to an example of robot-assisted upper-limb rehabilitation with online motion tracking. Copyright © 2014 Elsevier Ireland Ltd. All rights reserved.
Autonomous bone reposition around anatomical landmark for robot-assisted orthognathic surgery.
Woo, Sang-Yoon; Lee, Sang-Jeong; Yoo, Ji-Yong; Han, Jung-Joon; Hwang, Soon-Jung; Huh, Kyung-Hoe; Lee, Sam-Sun; Heo, Min-Suk; Choi, Soon-Chul; Yi, Won-Jin
2017-12-01
The purpose of this study was to develop a new method for enabling a robot to assist a surgeon in repositioning a bone segment to accurately transfer a preoperative virtual plan into the intraoperative phase in orthognathic surgery. We developed a robot system consisting of an arm with six degrees of freedom, a robot motion-controller, and a PC. An end-effector at the end of the robot arm transferred the movements of the robot arm to the patient's jawbone. The registration between the robot and CT image spaces was performed completely preoperatively, and the intraoperative registration could be finished using only position changes of the tracking tools at the robot end-effector and the patient's splint. The phantom's maxillomandibular complex (MMC) connected to the robot's end-effector was repositioned autonomously by the robot movements around an anatomical landmark of interest based on the tool center point (TCP) principle. The robot repositioned the MMC around the TCP of the incisor of the maxilla and the pogonion of the mandible following plans for real orthognathic patients. The accuracy of the robot's repositioning increased when an anatomical landmark for the TCP was close to the registration fiducials. In spite of this influence, we could increase the repositioning accuracy at the landmark by using the landmark itself as the TCP. With its ability to incorporate virtual planning using a CT image and autonomously execute the plan around an anatomical landmark of interest, the robot could help surgeons reposition bones more accurately and dexterously. Copyright © 2017 European Association for Cranio-Maxillo-Facial Surgery. Published by Elsevier Ltd. All rights reserved.
A Concept of the Differentially Driven Three Wheeled Robot
NASA Astrophysics Data System (ADS)
Kelemen, M.; Colville, D. J.; Kelemenová, T.; Virgala, I.; Miková, L.
2013-08-01
The paper deals with the concept of a differentially driven three wheeled robot. The main task for the robot is to follow the navigation black line on white ground. The robot also contains anti-collision sensors for avoiding obstacles on track. Students learn how to deal with signals from sensors and how to control DC motors. Students work with the controller and develop the locomotion algorithm and can attend a competition
AAPM and GEC-ESTRO guidelines for image-guided robotic brachytherapy: report of Task Group 192.
Podder, Tarun K; Beaulieu, Luc; Caldwell, Barrett; Cormack, Robert A; Crass, Jostin B; Dicker, Adam P; Fenster, Aaron; Fichtinger, Gabor; Meltsner, Michael A; Moerland, Marinus A; Nath, Ravinder; Rivard, Mark J; Salcudean, Tim; Song, Danny Y; Thomadsen, Bruce R; Yu, Yan
2014-10-01
In the last decade, there have been significant developments into integration of robots and automation tools with brachytherapy delivery systems. These systems aim to improve the current paradigm by executing higher precision and accuracy in seed placement, improving calculation of optimal seed locations, minimizing surgical trauma, and reducing radiation exposure to medical staff. Most of the applications of this technology have been in the implantation of seeds in patients with early-stage prostate cancer. Nevertheless, the techniques apply to any clinical site where interstitial brachytherapy is appropriate. In consideration of the rapid developments in this area, the American Association of Physicists in Medicine (AAPM) commissioned Task Group 192 to review the state-of-the-art in the field of robotic interstitial brachytherapy. This is a joint Task Group with the Groupe Européen de Curiethérapie-European Society for Radiotherapy & Oncology (GEC-ESTRO). All developed and reported robotic brachytherapy systems were reviewed. Commissioning and quality assurance procedures for the safe and consistent use of these systems are also provided. Manual seed placement techniques with a rigid template have an estimated in vivo accuracy of 3-6 mm. In addition to the placement accuracy, factors such as tissue deformation, needle deviation, and edema may result in a delivered dose distribution that differs from the preimplant or intraoperative plan. However, real-time needle tracking and seed identification for dynamic updating of dosimetry may improve the quality of seed implantation. The AAPM and GEC-ESTRO recommend that robotic systems should demonstrate a spatial accuracy of seed placement ≤1.0 mm in a phantom. This recommendation is based on the current performance of existing robotic brachytherapy systems and propagation of uncertainties. During clinical commissioning, tests should be conducted to ensure that this level of accuracy is achieved. These tests should mimic the real operating procedure as closely as possible. Additional recommendations on robotic brachytherapy systems include display of the operational state; capability of manual override; documented policies for independent check and data verification; intuitive interface displaying the implantation plan and visualization of needle positions and seed locations relative to the target anatomy; needle insertion in a sequential order; robot-clinician and robot-patient interactions robustness, reliability, and safety while delivering the correct dose at the correct site for the correct patient; avoidance of excessive force on radioactive sources; delivery confirmation of the required number or position of seeds; incorporation of a collision avoidance system; system cleaning, decontamination, and sterilization procedures. These recommendations are applicable to end users and manufacturers of robotic brachytherapy systems.
Effect of motor dynamics on nonlinear feedback robot arm control
NASA Technical Reports Server (NTRS)
Tarn, Tzyh-Jong; Li, Zuofeng; Bejczy, Antal K.; Yun, Xiaoping
1991-01-01
A nonlinear feedback robot controller that incorporates the robot manipulator dynamics and the robot joint motor dynamics is proposed. The manipulator dynamics and the motor dynamics are coupled to obtain a third-order-dynamic model, and differential geometric control theory is applied to produce a linearized and decoupled robot controller. The derived robot controller operates in the robot task space, thus eliminating the need for decomposition of motion commands into robot joint space commands. Computer simulations are performed to verify the feasibility of the proposed robot controller. The controller is further experimentally evaluated on the PUMA 560 robot arm. The experiments show that the proposed controller produces good trajectory tracking performances and is robust in the presence of model inaccuracies. Compared with a nonlinear feedback robot controller based on the manipulator dynamics only, the proposed robot controller yields conspicuously improved performance.
An approach to multivariable control of manipulators
NASA Technical Reports Server (NTRS)
Seraji, H.
1987-01-01
The paper presents simple schemes for multivariable control of multiple-joint robot manipulators in joint and Cartesian coordinates. The joint control scheme consists of two independent multivariable feedforward and feedback controllers. The feedforward controller is the minimal inverse of the linearized model of robot dynamics and contains only proportional-double-derivative (PD2) terms - implying feedforward from the desired position, velocity and acceleration. This controller ensures that the manipulator joint angles track any reference trajectories. The feedback controller is of proportional-integral-derivative (PID) type and is designed to achieve pole placement. This controller reduces any initial tracking error to zero as desired and also ensures that robust steady-state tracking of step-plus-exponential trajectories is achieved by the joint angles. Simple and explicit expressions of computation of the feedforward and feedback gains are obtained based on the linearized model of robot dynamics. This leads to computationally efficient schemes for either on-line gain computation or off-line gain scheduling to account for variations in the linearized robot model due to changes in the operating point. The joint control scheme is extended to direct control of the end-effector motion in Cartesian space. Simulation results are given for illustration.
Working and Learning with Knowledge in the Lobes of a Humanoid's Mind
NASA Technical Reports Server (NTRS)
Ambrose, Robert; Savely, Robert; Bluethmann, William; Kortenkamp, David
2003-01-01
Humanoid class robots must have sufficient dexterity to assist people and work in an environment designed for human comfort and productivity. This dexterity, in particular the ability to use tools, requires a cognitive understanding of self and the world that exceeds contemporary robotics. Our hypothesis is that the sense-think-act paradigm that has proven so successful for autonomous robots is missing one or more key elements that will be needed for humanoids to meet their full potential as autonomous human assistants. This key ingredient is knowledge. The presented work includes experiments conducted on the Robonaut system, a NASA and the Defense Advanced research Projects Agency (DARPA) joint project, and includes collaborative efforts with a DARPA Mobile Autonomous Robot Software technical program team of researchers at NASA, MIT, USC, NRL, UMass and Vanderbilt. The paper reports on results in the areas of human-robot interaction (human tracking, gesture recognition, natural language, supervised control), perception (stereo vision, object identification, object pose estimation), autonomous grasping (tactile sensing, grasp reflex, grasp stability) and learning (human instruction, task level sequences, and sensorimotor association).
A mathematical theory of learning control for linear discrete multivariable systems
NASA Technical Reports Server (NTRS)
Phan, Minh; Longman, Richard W.
1988-01-01
When tracking control systems are used in repetitive operations such as robots in various manufacturing processes, the controller will make the same errors repeatedly. Here consideration is given to learning controllers that look at the tracking errors in each repetition of the process and adjust the control to decrease these errors in the next repetition. A general formalism is developed for learning control of discrete-time (time-varying or time-invariant) linear multivariable systems. Methods of specifying a desired trajectory (such that the trajectory can actually be performed by the discrete system) are discussed, and learning controllers are developed. Stability criteria are obtained which are relatively easy to use to insure convergence of the learning process, and proper gain settings are discussed in light of measurement noise and system uncertainties.
Nonholonomic Ofject Tracking with Optical Sensors and Ofject Recognition Feedback
NASA Technical Reports Server (NTRS)
Goddard, R. E.; Hadaegh, F.
1994-01-01
Robotic controllers frequently operate under constraints. Often, the constraints are imperfectly or completely unknown. In this paper, the Lagrangian dynamics of a planar robot arm are expressed as a function of a globally unknown consraint.
Ultra-Wideband Angle-of-Arrival Tracking Systems
NASA Technical Reports Server (NTRS)
Arndt, G. Dickey; Ngo, Phong H.; Phan, Chau T.; Gross, Julia; Ni, Jianjun; Dusl, John
2010-01-01
Systems that measure the angles of arrival of ultra-wideband (UWB) radio signals and perform triangulation by use of those angles in order to locate the sources of those signals are undergoing development. These systems were originally intended for use in tracking UWB-transmitter-equipped astronauts and mobile robots on the surfaces of remote planets during early stages of exploration, before satellite-based navigation systems become operational. On Earth, these systems could be adapted to such uses as tracking UWB-transmitter-equipped firefighters inside buildings or in outdoor wildfire areas obscured by smoke. The same characteristics that have made UWB radio advantageous for fine resolution ranging, covert communication, and ground-penetrating radar applications in military and law-enforcement settings also contribute to its attractiveness for the present tracking applications. In particular, the waveform shape and the short duration of UWB pulses make it possible to attain the high temporal resolution (of the order of picoseconds) needed to measure angles of arrival with sufficient precision, and the low power spectral density of UWB pulses enables UWB radio communication systems to operate in proximity to other radio communication systems with little or no perceptible mutual interference.
Novel Door-opening Method for Six-legged Robots Based on Only Force Sensing
NASA Astrophysics Data System (ADS)
Chen, Zhi-Jun; Gao, Feng; Pan, Yang
2017-09-01
Current door-opening methods are mainly developed on tracked, wheeled and biped robots by applying multi-DOF manipulators and vision systems. However, door-opening methods for six-legged robots are seldom studied, especially using 0-DOF tools to operate and only force sensing to detect. A novel door-opening method for six-legged robots is developed and implemented to the six-parallel-legged robot. The kinematic model of the six-parallel-legged robot is established and the model of measuring the positional relationship between the robot and the door is proposed. The measurement model is completely based on only force sensing. The real-time trajectory planning method and the control strategy are designed. The trajectory planning method allows the maximum angle between the sagittal axis of the robot body and the normal line of the door plane to be 45º. A 0-DOF tool mounted to the robot body is applied to operate. By integrating with the body, the tool has 6 DOFs and enough workspace to operate. The loose grasp achieved by the tool helps release the inner force in the tool. Experiments are carried out to validate the method. The results show that the method is effective and robust in opening doors wider than 1 m. This paper proposes a novel door-opening method for six-legged robots, which notably uses a 0-DOF tool and only force sensing to detect and open the door.
Evolutionary multiobjective design of a flexible caudal fin for robotic fish.
Clark, Anthony J; Tan, Xiaobo; McKinley, Philip K
2015-11-25
Robotic fish accomplish swimming by deforming their bodies or other fin-like appendages. As an emerging class of embedded computing system, robotic fish are anticipated to play an important role in environmental monitoring, inspection of underwater structures, tracking of hazardous wastes and oil spills, and the study of live fish behaviors. While integration of flexible materials (into the fins and/or body) holds the promise of improved swimming performance (in terms of both speed and maneuverability) for these robots, such components also introduce significant design challenges due to the complex material mechanics and hydrodynamic interactions. The problem is further exacerbated by the need for the robots to meet multiple objectives (e.g., both speed and energy efficiency). In this paper, we propose an evolutionary multiobjective optimization approach to the design and control of a robotic fish with a flexible caudal fin. Specifically, we use the NSGA-II algorithm to investigate morphological and control parameter values that optimize swimming speed and power usage. Several evolved fin designs are validated experimentally with a small robotic fish, where fins of different stiffness values and sizes are printed with a multi-material 3D printer. Experimental results confirm the effectiveness of the proposed design approach in balancing the two competing objectives.
Addressing the Movement of a Freescale Robotic Car Using Neural Network
NASA Astrophysics Data System (ADS)
Horváth, Dušan; Cuninka, Peter
2016-12-01
This article deals with the management of a Freescale small robotic car along the predefined guide line. Controlling of the direction of movement of the robot is performed by neural networks, and scales (memory) of neurons are calculated by Hebbian learning from the truth tables as learning with a teacher. Reflexive infrared sensors serves as inputs. The results are experiments, which are used to compare two methods of mobile robot control - tracking lines.
Target motion tracking in MRI-guided transrectal robotic prostate biopsy.
Tadayyon, Hadi; Lasso, Andras; Kaushal, Aradhana; Guion, Peter; Fichtinger, Gabor
2011-11-01
MRI-guided prostate needle biopsy requires compensation for organ motion between target planning and needle placement. Two questions are studied and answered in this paper: 1) is rigid registration sufficient in tracking the targets with an error smaller than the clinically significant size of prostate cancer and 2) what is the effect of the number of intraoperative slices on registration accuracy and speed? we propose multislice-to-volume registration algorithms for tracking the biopsy targets within the prostate. Three orthogonal plus additional transverse intraoperative slices are acquired in the approximate center of the prostate and registered with a high-resolution target planning volume. Both rigid and deformable scenarios were implemented. Both simulated and clinical MRI-guided robotic prostate biopsy data were used to assess tracking accuracy. average registration errors in clinical patient data were 2.6 mm for the rigid algorithm and 2.1 mm for the deformable algorithm. rigid tracking appears to be promising. Three tracking slices yield significantly high registration speed with an affordable error.
SU-F-BRE-05: Development and Evaluation of a Real-Time Robotic 6D Quality Assurance Phantom
DOE Office of Scientific and Technical Information (OSTI.GOV)
Belcher, AH; Liu, X; Grelewicz, Z
Purpose: A 6 degree-of-freedom robotic phantom capable of reproducing dynamic tumor motion in 6D was designed to more effectively match solid tumor movements throughout pre-treatment scanning and radiation therapy. With the abundance of optical and x-ray 6D real-time tumor tracking methodologies clinically available, and the substantial dosimetric consequences of failing to consider tumor rotation as well as translation, this work presents the development and evaluation of a 6D instrument with the facility to improve quality assurance. Methods: An in-house designed and built 6D robotic motion phantom was constructed following the so-called Stewart-Gough parallel kinematics platform archetype. The device was thenmore » controlled using an inverse kinematics formulation, and precise movements in all six degrees of freedom (X, Y, Z, pitch, roll, and yaw) as well as previously obtained cranial motion, were effectively executed. The robotic phantom movements were verified using a 15 fps 6D infrared marker tracking system (Polaris, NDI), and quantitatively compared to the input trajectory. Thus, the accuracy and repeatability of 6D motion was investigated and the phantom performance was characterized. Results: Evaluation of the 6D platform demonstrated translational RMSE values of 0.196 mm, 0.260 mm, and 0.101 mm over 20 mm in X and Y and 10 mm in Z, respectively, and rotational RMSE values of 0.068 degrees, 0.0611 degrees, and 0.095 degrees over 10 degrees of pitch, roll, and yaw, respectively. The robotic stage also effectively performed controlled 6D motions, as well as reproduced cranial trajectories over 15 minutes, with a maximal RMSE of 0.044 mm translationally and 0.036 degrees rotationally. Conclusion: This 6D robotic phantom has proven to be accurate under clinical standards and capable of reproducing tumor motion in 6D. Consequently, such a robotics device has the potential to serve as a more effective system for IGRT QA that involves both translational and rotational dimensions. Research was partially funded by NIH Grant T32 EB002103-21 from NIBIB. Contents are solely the responsibility of the authors and do not necessarily represent the official views of the NIBIB or NIH.« less
Spatial Coverage Planning and Optimization for Planetary Exploration
NASA Technical Reports Server (NTRS)
Gaines, Daniel M.; Estlin, Tara; Chouinard, Caroline
2008-01-01
We are developing onboard planning and scheduling technology to enable in situ robotic explorers, such as rovers and aerobots, to more effectively assist scientists in planetary exploration. In our current work, we are focusing on situations in which the robot is exploring large geographical features such as craters, channels or regional boundaries. In to develop valid and high quality plans, the robot must take into account a range of scientific and engineering constraints and preferences. We have developed a system that incorporates multiobjective optimization and planning allowing the robot to generate high quality mission operations plans that respect resource limitations and mission constraints while attempting to maximize science and engineering objectives. An important scientific objective for the exploration of geological features is selecting observations that spatially cover an area of interest. We have developed a metric to enable an in situ explorer to reason about and track the spatial coverage quality of a plan. We describe this technique and show how it is combined in the overall multiobjective optimization and planning algorithm.
Recent results in visual servoing
NASA Astrophysics Data System (ADS)
Chaumette, François
2008-06-01
Visual servoing techniques consist in using the data provided by a vision sensor in order to control the motions of a dynamic system. Such systems are usually robot arms, mobile robots, aerial robots,… but can also be virtual robots for applications in computer animation, or even a virtual camera for applications in computer vision and augmented reality. A large variety of positioning tasks, or mobile target tracking, can be implemented by controlling from one to all the degrees of freedom of the system. Whatever the sensor configuration, which can vary from one on-board camera on the robot end-effector to several free-standing cameras, a set of visual features has to be selected at best from the image measurements available, allowing to control the degrees of freedom desired. A control law has also to be designed so that these visual features reach a desired value, defining a correct realization of the task. With a vision sensor providing 2D measurements, potential visual features are numerous, since as well 2D data (coordinates of feature points in the image, moments, …) as 3D data provided by a localization algorithm exploiting the extracted 2D measurements can be considered. It is also possible to combine 2D and 3D visual features to take the advantages of each approach while avoiding their respective drawbacks. From the selected visual features, the behavior of the system will have particular properties as for stability, robustness with respect to noise or to calibration errors, robot 3D trajectory, etc. The talk will present the main basic aspects of visual servoing, as well as technical advances obtained recently in the field inside the Lagadic group at INRIA/INRISA Rennes. Several application results will be also described.
Downey, John E; Weiss, Jeffrey M; Muelling, Katharina; Venkatraman, Arun; Valois, Jean-Sebastien; Hebert, Martial; Bagnell, J Andrew; Schwartz, Andrew B; Collinger, Jennifer L
2016-03-18
Recent studies have shown that brain-machine interfaces (BMIs) offer great potential for restoring upper limb function. However, grasping objects is a complicated task and the signals extracted from the brain may not always be capable of driving these movements reliably. Vision-guided robotic assistance is one possible way to improve BMI performance. We describe a method of shared control where the user controls a prosthetic arm using a BMI and receives assistance with positioning the hand when it approaches an object. Two human subjects with tetraplegia used a robotic arm to complete object transport tasks with and without shared control. The shared control system was designed to provide a balance between BMI-derived intention and computer assistance. An autonomous robotic grasping system identified and tracked objects and defined stable grasp positions for these objects. The system identified when the user intended to interact with an object based on the BMI-controlled movements of the robotic arm. Using shared control, BMI controlled movements and autonomous grasping commands were blended to ensure secure grasps. Both subjects were more successful on object transfer tasks when using shared control compared to BMI control alone. Movements made using shared control were more accurate, more efficient, and less difficult. One participant attempted a task with multiple objects and successfully lifted one of two closely spaced objects in 92 % of trials, demonstrating the potential for users to accurately execute their intention while using shared control. Integration of BMI control with vision-guided robotic assistance led to improved performance on object transfer tasks. Providing assistance while maintaining generalizability will make BMI systems more attractive to potential users. NCT01364480 and NCT01894802 .
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wen, N., E-mail: nwen1@hfhs.org; Snyder, K. C.; Qin, Y.
2016-05-15
Purpose: To evaluate the total systematic accuracy of a frameless, image guided stereotactic radiosurgery system. Methods: The localization accuracy and intermodality difference was determined by delivering radiation to an end-to-end prototype phantom, in which the targets were localized using optical surface monitoring system (OSMS), electromagnetic beacon-based tracking (Calypso®), cone-beam CT, “snap-shot” planar x-ray imaging, and a robotic couch. Six IMRT plans with jaw tracking and a flattening filter free beam were used to study the dosimetric accuracy for intracranial and spinal stereotactic radiosurgery treatment. Results: End-to-end localization accuracy of the system evaluated with the end-to-end phantom was 0.5 ± 0.2more » mm with a maximum deviation of 0.9 mm over 90 measurements (including jaw, MLC, and cone measurements for both auto and manual fusion) for single isocenter, single target treatment, 0.6 ± 0.4 mm for multitarget treatment with shared isocenter. Residual setup errors were within 0.1 mm for OSMS, and 0.3 mm for Calypso. Dosimetric evaluation based on absolute film dosimetry showed greater than 90% pass rate for all cases using a gamma criteria of 3%/1 mm. Conclusions: The authors’ experience demonstrates that the localization accuracy of the frameless image-guided system is comparable to robotic or invasive frame based radiosurgery systems.« less
Evaluation method on steering for the shape-shifting robot in different configurations
NASA Astrophysics Data System (ADS)
Chang, Jian; Li, Bin; Wang, Chong; Zheng, Huaibing; Li, Zhiqiang
2016-01-01
The evaluation method on steering is based on qualitative manner in existence, which causes the result inaccurate and fuzziness. It reduces the efficiency of process execution. So the method by quantitative manner for the shape-shifting robot in different configurations is proposed. Comparing to traditional evaluation method, the most important aspects which can influence the steering abilities of the robot in different configurations are researched in detail, including the energy, angular velocity, time and space. In order to improve the robustness of system, the ideal and slippage conditions are all considered by mathematical model. Comparing to the traditional weighting confirming method, the extent of robot steering method is proposed by the combination of subjective and objective weighting method. The subjective weighting method can show more preferences of the experts and is based on five-grade scale. The objective weighting method is based on information entropy to determine the factors. By the sensors fixed on the robot, the contract force between track grouser and ground, the intrinsic motion characteristics of robot are obtained and the experiment is done to prove the algorithm which is proposed as the robot in different common configurations. Through the method proposed in the article, fuzziness and inaccurate of the evaluation method has been solved, so the operators can choose the most suitable configuration of the robot to fulfil the different tasks more quickly and simply.
Robotic invasion of operation theatre and associated anaesthetic issues: A review.
Kakar, Prem N; Das, Jyotirmoy; Roy, Preeti Mittal; Pant, Vijaya
2011-01-01
A Robotic device is a powered, computer controlled manipulator with artificial sensing that can be reprogrammed to move and position tools to carry out a wide range of tasks. Robots and Telemanipulators were first developed by the National Aeronautics and Space Administration (NASA) for use in space exploration. Today's medical robotic systems were the brainchild of the United States Department of Defence's desire to decrease war casualties with the development of 'telerobotic surgery'. The 'master-slave' telemanipulator concept was developed for medical use in the early 1990s where the surgeon's (master) manual movements were transmitted to end-effector (slave) instruments at a remote site. Since then, the field of surgical robotics has undergone massive transformation and the future is even brighter. As expected, any new technique brings with it risks and the possibility of technical difficulties. The person who bears the brunt of complications or benefit from a new invention is the 'Patient'. Anaesthesiologists as always must do their part to be the patient's 'best man' in the perioperative period. We should be prepared for screening and selection of patients in a different perspective keeping in mind the steep learning curves of surgeons, long surgical hours, extreme patient positioning and other previously unknown anaesthetic challenges brought about by the surgical robot. In this article we have tried to track the development of surgical robots and consider the unique anaesthetic issues related to robot assisted surgeries.
Reliable vision-guided grasping
NASA Technical Reports Server (NTRS)
Nicewarner, Keith E.; Kelley, Robert B.
1992-01-01
Automated assembly of truss structures in space requires vision-guided servoing for grasping a strut when its position and orientation are uncertain. This paper presents a methodology for efficient and robust vision-guided robot grasping alignment. The vision-guided grasping problem is related to vision-guided 'docking' problems. It differs from other hand-in-eye visual servoing problems, such as tracking, in that the distance from the target is a relevant servo parameter. The methodology described in this paper is hierarchy of levels in which the vision/robot interface is decreasingly 'intelligent,' and increasingly fast. Speed is achieved primarily by information reduction. This reduction exploits the use of region-of-interest windows in the image plane and feature motion prediction. These reductions invariably require stringent assumptions about the image. Therefore, at a higher level, these assumptions are verified using slower, more reliable methods. This hierarchy provides for robust error recovery in that when a lower-level routine fails, the next-higher routine will be called and so on. A working system is described which visually aligns a robot to grasp a cylindrical strut. The system uses a single camera mounted on the end effector of a robot and requires only crude calibration parameters. The grasping procedure is fast and reliable, with a multi-level error recovery system.
Indirect decentralized repetitive control
NASA Technical Reports Server (NTRS)
Lee, Soo Cheol; Longman, Richard W.
1993-01-01
Learning control refers to controllers that learn to improve their performance at executing a given task, based on experience performing this specific task. In a previous work, the authors presented a theory of indirect decentralized learning control based on use of indirect adaptive control concepts employing simultaneous identification and control. This paper extends these results to apply to the indirect repetitive control problem in which a periodic (i.e., repetitive) command is given to a control system. Decentralized indirect repetitive control algorithms are presented that have guaranteed convergence to zero tracking error under very general conditions. The original motivation of the repetitive control and learning control fields was learning in robots doing repetitive tasks such as on an assembly line. This paper starts with decentralized discrete time systems, and progresses to the robot application, modeling the robot as a time varying linear system in the neighborhood of the desired trajectory. Decentralized repetitive control is natural for this application because the feedback control for link rotations is normally implemented in a decentralized manner, treating each link as if it is independent of the other links.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Wang, L; M Yang, Y; Nelson, B
Purpose: A novel end-to-end test system using a CCD camera and a scintillator based phantom (XRV-124, Logos Systems Int’l) capable of measuring the beam-by-beam delivery accuracy of Robotic Radiosurgery (CyberKnife) was developed and reported in our previous work. This work investigates its application in assessing the motion tracking (Synchrony) accuracy for CyberKnife. Methods: A QA plan with Anterior and Lateral beams (with 4 different collimator sizes) was created (Multiplan v5.3) for the XRV-124 phantom. The phantom was placed on a motion platform (superior and inferior movement), and the plans were delivered on the CyberKnife M6 system using four motion patterns:more » static, Sine- wave, Sine with 15° phase shift, and a patient breathing pattern composed of 2cm maximum motion with 4 second breathing cycle. Under integral recording mode, the time-averaged beam vectors (X, Y, Z) were measured by the phantom and compared with static delivery. In dynamic recording mode, the beam spots were recorded at a rate of 10 frames/second. The beam vector deviation from average position was evaluated against the various breathing patterns. Results: The average beam position of the six deliveries with no motion and three deliveries with Synchrony tracking on ideal motion (sinewave without phase shift) all agree within −0.03±0.00 mm, 0.10±0.04, and 0.04±0.03 in the X, Y, and X directions. Radiation beam width (FWHM) variations are within ±0.03 mm. Dynamic video record showed submillimeter tracking stability for both regular and irregular breathing pattern; however the tracking error up to 3.5 mm was observed when a 15 degree phase shift was introduced. Conclusion: The XRV-124 system is able to provide 3D and 4D targeting accuracy for CyberKnife delivery with Synchrony. The experimental results showed sub-millimeter delivery in phantom with excellent correlation in target to breathing motion. The accuracy was degraded when irregular motion and phase shift was introduced.« less
Operation of a Cartesian Robotic System in a Compact Microscope with Intelligent Controls
NASA Technical Reports Server (NTRS)
McDowell, Mark (Inventor)
2006-01-01
A Microscope Imaging System (CMIS) with intelligent controls is disclosed that provides techniques for scanning, identifying, detecting and tracking microscopic changes in selected characteristics or features of various surfaces including, but not limited to, cells, spheres, and manufactured products subject to difficult-to-see imperfections. The practice of the present invention provides applications that include colloidal hard spheres experiments, biological cell detection for patch clamping, cell movement and tracking, as well as defect identification in products, such as semiconductor devices, where surface damage can be significant, but difficult to detect. The CMIS system is a machine vision system, which combines intelligent image processing with remote control capabilities and provides the ability to autofocus on a microscope sample, automatically scan an image, and perform machine vision analysis on multiple samples simultaneously.
Navigation strategies for multiple autonomous mobile robots moving in formation
NASA Technical Reports Server (NTRS)
Wang, P. K. C.
1991-01-01
The problem of deriving navigation strategies for a fleet of autonomous mobile robots moving in formation is considered. Here, each robot is represented by a particle with a spherical effective spatial domain and a specified cone of visibility. The global motion of each robot in the world space is described by the equations of motion of the robot's center of mass. First, methods for formation generation are discussed. Then, simple navigation strategies for robots moving in formation are derived. A sufficient condition for the stability of a desired formation pattern for a fleet of robots each equipped with the navigation strategy based on nearest neighbor tracking is developed. The dynamic behavior of robot fleets consisting of three or more robots moving in formation in a plane is studied by means of computer simulation.
A multimodal imaging framework for enhanced robot-assisted partial nephrectomy guidance
NASA Astrophysics Data System (ADS)
Halter, Ryan J.; Wu, Xiaotian; Hartov, Alex; Seigne, John; Khan, Shadab
2015-03-01
Robot-assisted laparoscopic partial nephrectomies (RALPN) are performed to treat patients with locally confined renal carcinoma. There are well-documented benefits to performing partial (opposed to radical) kidney resections and to using robot-assisted laparoscopic (opposed to open) approaches. However, there are challenges in identifying tumor margins and critical benign structures including blood vessels and collecting systems during current RALPN procedures. The primary objective of this effort is to couple multiple image and data streams together to augment visual information currently provided to surgeons performing RALPN and ultimately ensure complete tumor resection and minimal damage to functional structures (i.e. renal vasculature and collecting systems). To meet this challenge we have developed a framework and performed initial feasibility experiments to couple pre-operative high-resolution anatomic images with intraoperative MRI, ultrasound (US) and optical-based surface mapping and kidney tracking. With these registered images and data streams, we aim to overlay the high-resolution contrast-enhanced anatomic (CT or MR) images onto the surgeon's view screen for enhanced guidance. To date we have integrated the following components of our framework: 1) a method for tracking an intraoperative US probe to extract the kidney surface and a set of embedded kidney markers, 2) a method for co-registering intraoperative US scans with pre-operative MR scans, and 3) a method for deforming pre-op scans to match intraoperative scans. These components have been evaluated through phantom studies to demonstrate protocol feasibility.
Johnson, Michelle J; Feng, Xin; Johnson, Laura M; Winters, Jack M
2007-03-01
There is a need to improve semi-autonomous stroke therapy in home environments often characterized by low supervision of clinical experts and low extrinsic motivation. Our distributed device approach to this problem consists of an integrated suite of low-cost robotic/computer-assistive technologies driven by a novel universal access software framework called UniTherapy. Our design strategy for personalizing the therapy, providing extrinsic motivation and outcome assessment is presented and evaluated. Three studies were conducted to evaluate the potential of the suite. A conventional force-reflecting joystick, a modified joystick therapy platform (TheraJoy), and a steering wheel platform (TheraDrive) were tested separately with the UniTherapy software. Stroke subjects with hemiparesis and able-bodied subjects completed tracking activities with the devices in different positions. We quantify motor performance across subject groups and across device platforms and muscle activation across devices at two positions in the arm workspace. Trends in the assessment metrics were consistent across devices with able-bodied and high functioning strokes subjects being significantly more accurate and quicker in their motor performance than low functioning subjects. Muscle activation patterns were different for shoulder and elbow across different devices and locations. The Robot/CAMR suite has potential for stroke rehabilitation. By manipulating hardware and software variables, we can create personalized therapy environments that engage patients, address their therapy need, and track their progress. A larger longitudinal study is still needed to evaluate these systems in under-supervised environments such as the home.
An image guidance system for positioning robotic cochlear implant insertion tools
NASA Astrophysics Data System (ADS)
Bruns, Trevor L.; Webster, Robert J.
2017-03-01
Cochlear implants must be inserted carefully to avoid damaging the delicate anatomical structures of the inner ear. This has motivated several approaches to improve the safety and efficacy of electrode array insertion by automating the process with specialized robotic or manual insertion tools. When such tools are used, they must be positioned at the entry point to the cochlea and aligned with the desired entry vector. This paper presents an image guidance system capable of accurately positioning a cochlear implant insertion tool. An optical tracking system localizes the insertion tool in physical space while a graphical user interface incorporates this with patient- specific anatomical data to provide error information to the surgeon in real-time. Guided by this interface, novice users successfully aligned the tool with an mean accuracy of 0.31 mm.
Weighted feature selection criteria for visual servoing of a telerobot
NASA Technical Reports Server (NTRS)
Feddema, John T.; Lee, C. S. G.; Mitchell, O. R.
1989-01-01
Because of the continually changing environment of a space station, visual feedback is a vital element of a telerobotic system. A real time visual servoing system would allow a telerobot to track and manipulate randomly moving objects. Methodologies for the automatic selection of image features to be used to visually control the relative position between an eye-in-hand telerobot and a known object are devised. A weighted criteria function with both image recognition and control components is used to select the combination of image features which provides the best control. Simulation and experimental results of a PUMA robot arm visually tracking a randomly moving carburetor gasket with a visual update time of 70 milliseconds are discussed.
NASA Astrophysics Data System (ADS)
Esteghamatian, Mehdi; Sarkar, Kripasindhu; Pautler, Stephen E.; Chen, Elvis C. S.; Peters, Terry M.
2012-02-01
Radical prostatectomy surgery (RP) is the gold standard for treatment of localized prostate cancer (PCa). Recently, emergence of minimally invasive techniques such as Laparoscopic Radical Prostatectomy (LRP) and Robot-Assisted Laparoscopic Radical Prostatectomy (RARP) has improved the outcomes for prostatectomy. However, it remains difficult for the surgeons to make informed decisions regarding resection margins and nerve sparing since the location of the tumor within the organ is not usually visible in a laparoscopic view. While MRI enables visualization of the salient structures and cancer foci, its efficacy in LRP is reduced unless it is fused into a stereoscopic view such that homologous structures overlap. Registration of the MRI image and peri-operative ultrasound image using a tracked probe can potentially be exploited to bring the pre-operative information into alignment with the patient coordinate system during the procedure. While doing so, prostate motion needs to be compensated in real-time to synchronize the stereoscopic view with the pre-operative MRI during the prostatectomy procedure. In this study, a point-based stereoscopic tracking technique is investigated to compensate for rigid prostate motion so that the same motion can be applied to the pre-operative images. This method benefits from stereoscopic tracking of the surface markers implanted over the surface of the prostate phantom. The average target registration error using this approach was 3.25+/-1.43mm.
Indirect learning control for nonlinear dynamical systems
NASA Technical Reports Server (NTRS)
Ryu, Yeong Soon; Longman, Richard W.
1993-01-01
In a previous paper, learning control algorithms were developed based on adaptive control ideas for linear time variant systems. The learning control methods were shown to have certain advantages over their adaptive control counterparts, such as the ability to produce zero tracking error in time varying systems, and the ability to eliminate repetitive disturbances. In recent years, certain adaptive control algorithms have been developed for multi-body dynamic systems such as robots, with global guaranteed convergence to zero tracking error for the nonlinear system euations. In this paper we study the relationship between such adaptive control methods designed for this specific class of nonlinear systems, and the learning control problem for such systems, seeking to converge to zero tracking error in following a specific command repeatedly, starting from the same initial conditions each time. The extension of these methods from the adaptive control problem to the learning control problem is seen to be trivial. The advantages and disadvantages of using learning control based on such adaptive control concepts for nonlinear systems, and the use of other currently available learning control algorithms are discussed.
Simultaneous Deployment and Tracking Multi-Robot Strategies with Connectivity Maintenance
Tardós, Javier; Aragues, Rosario; Sagüés, Carlos; Rubio, Carlos
2018-01-01
Multi-robot teams composed of ground and aerial vehicles have gained attention during the last few years. We present a scenario where both types of robots must monitor the same area from different view points. In this paper, we propose two Lloyd-based tracking strategies to allow the ground robots (agents) to follow the aerial ones (targets), keeping the connectivity between the agents. The first strategy establishes density functions on the environment so that the targets acquire more importance than other zones, while the second one iteratively modifies the virtual limits of the working area depending on the positions of the targets. We consider the connectivity maintenance due to the fact that coverage tasks tend to spread the agents as much as possible, which is addressed by restricting their motions so that they keep the links of a minimum spanning tree of the communication graph. We provide a thorough parametric study of the performance of the proposed strategies under several simulated scenarios. In addition, the methods are implemented and tested using realistic robotic simulation environments and real experiments. PMID:29558446
NASA Astrophysics Data System (ADS)
Petrişor, Silviu-Mihai; Bârsan, GhiÅ£Ä.
2013-12-01
The authors of this paper wish to highlight elements regarding the organology, functioning and simulation, in a real workspace, of a tracked mini robot structure destined for special applications in theatres of operation, a technological product which is subject to a national patent granted to our institution (patent no. RO a 2012 01051), the result of research activities undertaken under a contract won by national competition, a grant for young research teams, PN-RUTE- 2010 type. The issues outlined in this paper are aspects related to the original invention in comparison with other mini-robot structures, the inventors presenting succinctly the technological product description and its applicability both in the military and applicative area as well as in the educational one. Additionally, the advantages of using the technological product are shown in a real workspace, the constructive and functional solution before, finally, presenting, based on the modelling of the mechanical structure of the tilting module attached to the mini-robot, an application on the simulation and programming of the mini-robot under study.
Bilevel shared control for teleoperators
NASA Technical Reports Server (NTRS)
Hayati, Samad A. (Inventor); Venkataraman, Subramanian T. (Inventor)
1992-01-01
A shared system is disclosed for robot control including integration of the human and autonomous input modalities for an improved control. Autonomously planned motion trajectories are modified by a teleoperator to track unmodelled target motions, while nominal teleoperator motions are modified through compliance to accommodate geometric errors autonomously in the latter. A hierarchical shared system intelligently shares control over a remote robot between the autonomous and teleoperative portions of an overall control system. Architecture is hierarchical, and consists of two levels. The top level represents the task level, while the bottom, the execution level. In space applications, the performance of pure teleoperation systems depend significantly on the communication time delays between the local and the remote sites. Selection/mixing matrices are provided with entries which reflect how each input's signals modality is weighted. The shared control minimizes the detrimental effects caused by these time delays between earth and space.
Experimental validation of docking and capture using space robotics testbeds
NASA Technical Reports Server (NTRS)
Spofford, John; Schmitz, Eric; Hoff, William
1991-01-01
This presentation describes the application of robotic and computer vision systems to validate docking and capture operations for space cargo transfer vehicles. Three applications are discussed: (1) air bearing systems in two dimensions that yield high quality free-flying, flexible, and contact dynamics; (2) validation of docking mechanisms with misalignment and target dynamics; and (3) computer vision technology for target location and real-time tracking. All the testbeds are supported by a network of engineering workstations for dynamic and controls analyses. Dynamic simulation of multibody rigid and elastic systems are performed with the TREETOPS code. MATRIXx/System-Build and PRO-MATLAB/Simulab are the tools for control design and analysis using classical and modern techniques such as H-infinity and LQG/LTR. SANDY is a general design tool to optimize numerically a multivariable robust compensator with a user-defined structure. Mathematica and Macsyma are used to derive symbolically dynamic and kinematic equations.
NASA Astrophysics Data System (ADS)
Girach, Khalid; Bouazza-Marouf, K.; Kerr, David; Hewit, Jim
1994-11-01
The paper describes the investigations carried out to implement a line of sight control and communication link for a mobile robot vehicle for use in structured nuclear semi-hazardous environments. Line of sight free space optical laser communication links for remote teleoperation have important applications in hazardous environments. They have certain advantages over radio/microwave links and umbilical control such as greater protection against generation of and susceptance to electro-magnetic fields. The cable-less environment provides increased integrity and mechanical freedom to the mobile robot. However, to maintain the communication link, continuous point and tracking is required between the base station and the mobile vehicle. This paper presents a novel two ended optical tracking system utilizing the communication laser beams and photodetectors. The mobile robot is a six wheel drive vehicle with a manipulator arm which can operate in a variety of terrain. The operator obtains visual feedback information from cameras placed on the vehicle. From this information, the speed and direction of the vehicle can be controlled from a joystick panel. We describe the investigations carried out for the communication of analogue video and digital data signals over the laser link for speed and direction control.
Maintaining and Enhancing Diversity of Sampled Protein Conformations in Robotics-Inspired Methods.
Abella, Jayvee R; Moll, Mark; Kavraki, Lydia E
2018-01-01
The ability to efficiently sample structurally diverse protein conformations allows one to gain a high-level view of a protein's energy landscape. Algorithms from robot motion planning have been used for conformational sampling, and several of these algorithms promote diversity by keeping track of "coverage" in conformational space based on the local sampling density. However, large proteins present special challenges. In particular, larger systems require running many concurrent instances of these algorithms, but these algorithms can quickly become memory intensive because they typically keep previously sampled conformations in memory to maintain coverage estimates. In addition, robotics-inspired algorithms depend on defining useful perturbation strategies for exploring the conformational space, which is a difficult task for large proteins because such systems are typically more constrained and exhibit complex motions. In this article, we introduce two methodologies for maintaining and enhancing diversity in robotics-inspired conformational sampling. The first method addresses algorithms based on coverage estimates and leverages the use of a low-dimensional projection to define a global coverage grid that maintains coverage across concurrent runs of sampling. The second method is an automatic definition of a perturbation strategy through readily available flexibility information derived from B-factors, secondary structure, and rigidity analysis. Our results show a significant increase in the diversity of the conformations sampled for proteins consisting of up to 500 residues when applied to a specific robotics-inspired algorithm for conformational sampling. The methodologies presented in this article may be vital components for the scalability of robotics-inspired approaches.
NASA Astrophysics Data System (ADS)
Luy, N. T.
2018-04-01
The design of distributed cooperative H∞ optimal controllers for multi-agent systems is a major challenge when the agents' models are uncertain multi-input and multi-output nonlinear systems in strict-feedback form in the presence of external disturbances. In this paper, first, the distributed cooperative H∞ optimal tracking problem is transformed into controlling the cooperative tracking error dynamics in affine form. Second, control schemes and online algorithms are proposed via adaptive dynamic programming (ADP) and the theory of zero-sum differential graphical games. The schemes use only one neural network (NN) for each agent instead of three from ADP to reduce computational complexity as well as avoid choosing initial NN weights for stabilising controllers. It is shown that despite not using knowledge of cooperative internal dynamics, the proposed algorithms not only approximate values to Nash equilibrium but also guarantee all signals, such as the NN weight approximation errors and the cooperative tracking errors in the closed-loop system, to be uniformly ultimately bounded. Finally, the effectiveness of the proposed method is shown by simulation results of an application to wheeled mobile multi-robot systems.
Real-time multiple human perception with color-depth cameras on a mobile robot.
Zhang, Hao; Reardon, Christopher; Parker, Lynne E
2013-10-01
The ability to perceive humans is an essential requirement for safe and efficient human-robot interaction. In real-world applications, the need for a robot to interact in real time with multiple humans in a dynamic, 3-D environment presents a significant challenge. The recent availability of commercial color-depth cameras allow for the creation of a system that makes use of the depth dimension, thus enabling a robot to observe its environment and perceive in the 3-D space. Here we present a system for 3-D multiple human perception in real time from a moving robot equipped with a color-depth camera and a consumer-grade computer. Our approach reduces computation time to achieve real-time performance through a unique combination of new ideas and established techniques. We remove the ground and ceiling planes from the 3-D point cloud input to separate candidate point clusters. We introduce the novel information concept, depth of interest, which we use to identify candidates for detection, and that avoids the computationally expensive scanning-window methods of other approaches. We utilize a cascade of detectors to distinguish humans from objects, in which we make intelligent reuse of intermediary features in successive detectors to improve computation. Because of the high computational cost of some methods, we represent our candidate tracking algorithm with a decision directed acyclic graph, which allows us to use the most computationally intense techniques only where necessary. We detail the successful implementation of our novel approach on a mobile robot and examine its performance in scenarios with real-world challenges, including occlusion, robot motion, nonupright humans, humans leaving and reentering the field of view (i.e., the reidentification challenge), human-object and human-human interaction. We conclude with the observation that the incorporation of the depth information, together with the use of modern techniques in new ways, we are able to create an accurate system for real-time 3-D perception of humans by a mobile robot.
Duran, Cassidy; Estrada, Sean; O'Malley, Marcia; Lumsden, Alan B; Bismuth, Jean
2015-02-01
Endovascular robotics systems, now approved for clinical use in the United States and Europe, are seeing rapid growth in interest. Determining who has sufficient expertise for safe and effective clinical use remains elusive. Our aim was to analyze performance on a robotic platform to determine what defines an expert user. During three sessions, 21 subjects with a range of endovascular expertise and endovascular robotic experience (novices <2 hours to moderate-extensive experience with >20 hours) performed four tasks on a training model. All participants completed a 2-hour training session on the robot by a certified instructor. Completion times, global rating scores, and motion metrics were collected to assess performance. Electromagnetic tracking was used to capture and to analyze catheter tip motion. Motion analysis was based on derivations of speed and position including spectral arc length and total number of submovements (inversely proportional to proficiency of motion) and duration of submovements (directly proportional to proficiency). Ninety-eight percent of competent subjects successfully completed the tasks within the given time, whereas 91% of noncompetent subjects were successful. There was no significant difference in completion times between competent and noncompetent users except for the posterior branch (151 s:105 s; P = .01). The competent users had more efficient motion as evidenced by statistically significant differences in the metrics of motion analysis. Users with >20 hours of experience performed significantly better than those newer to the system, independent of prior endovascular experience. This study demonstrates that motion-based metrics can differentiate novice from trained users of flexible robotics systems for basic endovascular tasks. Efficiency of catheter movement, consistency of performance, and learning curves may help identify users who are sufficiently trained for safe clinical use of the system. This work will help identify the learning curve and specific movements that translate to expert robotic navigation. Copyright © 2015 Society for Vascular Surgery. Published by Elsevier Inc. All rights reserved.
PIMS sequencing extension: a laboratory information management system for DNA sequencing facilities
2011-01-01
Background Facilities that provide a service for DNA sequencing typically support large numbers of users and experiment types. The cost of services is often reduced by the use of liquid handling robots but the efficiency of such facilities is hampered because the software for such robots does not usually integrate well with the systems that run the sequencing machines. Accordingly, there is a need for software systems capable of integrating different robotic systems and managing sample information for DNA sequencing services. In this paper, we describe an extension to the Protein Information Management System (PIMS) that is designed for DNA sequencing facilities. The new version of PIMS has a user-friendly web interface and integrates all aspects of the sequencing process, including sample submission, handling and tracking, together with capture and management of the data. Results The PIMS sequencing extension has been in production since July 2009 at the University of Leeds DNA Sequencing Facility. It has completely replaced manual data handling and simplified the tasks of data management and user communication. Samples from 45 groups have been processed with an average throughput of 10000 samples per month. The current version of the PIMS sequencing extension works with Applied Biosystems 3130XL 96-well plate sequencer and MWG 4204 or Aviso Theonyx liquid handling robots, but is readily adaptable for use with other combinations of robots. Conclusions PIMS has been extended to provide a user-friendly and integrated data management solution for DNA sequencing facilities that is accessed through a normal web browser and allows simultaneous access by multiple users as well as facility managers. The system integrates sequencing and liquid handling robots, manages the data flow, and provides remote access to the sequencing results. The software is freely available, for academic users, from http://www.pims-lims.org/. PMID:21385349
Trajectory control of robot manipulators with closed-kinematic chain mechanism
NASA Technical Reports Server (NTRS)
Nguyen, Charles C.; Pooran, Farhad J.; Premack, Timothy
1987-01-01
The problem of Cartesian trajectory control of a closed-kinematic chain mechanism robot manipulator, recently built at CAIR to study the assembly of NASA hardware for the future Space Station, is considered. The study is performed by both computer simulation and experimentation for tracking of three different paths: a straight line, a sinusoid, and a circle. Linearization and pole placement methods are employed to design controller gains. Results show that the controllers are robust and there are good agreements between simulation and experimentation. The results also show excellent tracking quality and small overshoots.
A development of intelligent entertainment robot for home life
NASA Astrophysics Data System (ADS)
Kim, Cheoltaek; Lee, Ju-Jang
2005-12-01
The purpose of this paper was to present the study and design idea for entertainment robot with educational purpose (IRFEE). The robot has been designed for home life considering dependability and interaction. The developed robot has three objectives - 1. Develop autonomous robot, 2. Design robot considering mobility and robustness, 3. Develop robot interface and software considering entertainment and education functionalities. The autonomous navigation was implemented by active vision based SLAM and modified EPF algorithm. The two differential wheels, the pan-tilt were designed mobility and robustness and the exterior was designed considering esthetic element and minimizing interference. The speech and tracking algorithm provided the good interface with human. The image transfer and Internet site connection is needed for service of remote connection and educational purpose.
Robotics and telecommunication systems to provide better access to ultrasound expertise in the OR.
Angelini, L; Papaspyropoulos, V
2000-01-01
Surgery has begun to evolve as a result of the intense use of technological innovations. The result of this is better services for patients and enormous opportunities for the producers of biomedical instruments. The surgeon and the technologist are fast becoming allies in applying the latest developments of robotics, image treatment, simulation, sensors and telecommunications to surgery, in particular to the emerging field of minimally-invasive surgery. Ultrasonography is at present utilised both for diagnostic and therapeutic purposes in various fields. Intraoperative US examination can be of primary importance, especially when dealing with space-occupying lesions. The widening use of minimally-invasive surgery has furthered the development of US for use during this type of surgery. The success of a US examination requires not only a correct execution of the procedure, but also a correct interpretation of the images. We describe two projects that combine robotics and telecommunication systems to provide better access to US expertise in the operating room. The Midstep project has as its object the realisation of two robotic arms, one for the distant control of the US probe during laparoscopic surgery and the second to perform tele-interventional US. The second project, part of the Strategic CNR Project-'Robotics in Surgery', involves the realisation of a common platform for tracking and targeting surgical instruments in video-assisted surgery.
Mobile robotic sensors for perimeter detection and tracking.
Clark, Justin; Fierro, Rafael
2007-02-01
Mobile robot/sensor networks have emerged as tools for environmental monitoring, search and rescue, exploration and mapping, evaluation of civil infrastructure, and military operations. These networks consist of many sensors each equipped with embedded processors, wireless communication, and motion capabilities. This paper describes a cooperative mobile robot network capable of detecting and tracking a perimeter defined by a certain substance (e.g., a chemical spill) in the environment. Specifically, the contributions of this paper are twofold: (i) a library of simple reactive motion control algorithms and (ii) a coordination mechanism for effectively carrying out perimeter-sensing missions. The decentralized nature of the methodology implemented could potentially allow the network to scale to many sensors and to reconfigure when adding/deleting sensors. Extensive simulation results and experiments verify the validity of the proposed cooperative control scheme.
A unified perspective on robot control - The energy Lyapunov function approach
NASA Technical Reports Server (NTRS)
Wen, John T.
1990-01-01
A unified framework for the stability analysis of robot tracking control is presented. By using an energy-motivated Lyapunov function candidate, the closed-loop stability is shown for a large family of control laws sharing a common structure of proportional and derivative feedback and a model-based feedforward. The feedforward can be zero, partial or complete linearized dynamics, partial or complete nonlinear dynamics, or linearized or nonlinear dynamics with parameter adaptation. As result, the dichotomous approaches to the robot control problem based on the open-loop linearization and nonlinear Lyapunov analysis are both included in this treatment. Furthermore, quantitative estimates of the trade-offs between different schemes in terms of the tracking performance, steady state error, domain of convergence, realtime computation load and required a prior model information are derived.
A simple, inexpensive, and effective implementation of a vision-guided autonomous robot
NASA Astrophysics Data System (ADS)
Tippetts, Beau; Lillywhite, Kirt; Fowers, Spencer; Dennis, Aaron; Lee, Dah-Jye; Archibald, James
2006-10-01
This paper discusses a simple, inexpensive, and effective implementation of a vision-guided autonomous robot. This implementation is a second year entrance for Brigham Young University students to the Intelligent Ground Vehicle Competition. The objective of the robot was to navigate a course constructed of white boundary lines and orange obstacles for the autonomous competition. A used electric wheelchair was used as the robot base. The wheelchair was purchased from a local thrift store for $28. The base was modified to include Kegresse tracks using a friction drum system. This modification allowed the robot to perform better on a variety of terrains, resolving issues with last year's design. In order to control the wheelchair and retain the robust motor controls already on the wheelchair the wheelchair joystick was simply removed and replaced with a printed circuit board that emulated joystick operation and was capable of receiving commands through a serial port connection. Three different algorithms were implemented and compared: a purely reactive approach, a potential fields approach, and a machine learning approach. Each of the algorithms used color segmentation methods to interpret data from a digital camera in order to identify the features of the course. This paper will be useful to those interested in implementing an inexpensive vision-based autonomous robot.
Design of an autonomous teleoperated cargo transporting vehicle for lunar base operations
NASA Technical Reports Server (NTRS)
Holt, James; Lao, Tom; Monali, Nkoy
1989-01-01
At the turn of the century NASA plans to begin construction of a lunar base. The base will likely consist of developed areas (i.e., habitation, laboratory, landing and launching sites, power plant) separated from each other due to safety considerations. The Self-Repositioning Track Vehicle (SRTV) was designed to transport cargo between these base facilities. The SRTV operates by using two robotic arms to raise and position segments of track upon which the vehicle travels. The SRTV utilizes the semiautonomous mobility (SAM) method of teleoperation; actuator-controlled interlocking track sections; two robotic arms each with five degrees of freedom; and these materials: titanium for structural members and aluminum for shell members, with the possible use of light-weight, high-strength composites.
Measurement of Flexed Posture for Flexible Mono-Tread Mobile Track
NASA Astrophysics Data System (ADS)
Kinugasa, Tetsuya; Akagi, Tetsuya; Ishii, Kuniaki; Haji, Takafumi; Yoshida, Koji; Amano, Hisanori; Hayashi, Ryota; Tokuda, Kenichi; Iribe, Masatsugu; Osuka, Koichi
We have proposed Flexible Mono-tread mobile Track (FMT) as a mobile mechanism on rough terrain for rescue activity, environmental investigation and planetary explorer, etc. Generally speaking, one has to teleoperate robots under invisible condition. In order to operate the robots skillfully, it is necessary to detect not only condition around the robots and its position but also posture of the robots at any time. Since flexed posture of FMT decides turning radius and direction, it is important to know its posture. FMT has vertebral structure composed of vertebrae as rigid body and intervertebral disks made by flexible devices such as rubber cylinder and spring. Since the intervertebral disks flex in three dimension, traditional sensors such as potentiometers, rotary encoders and range finders can hardly use for measurement of its deformation. The purpose of the paper, therefore, is to measure flexed posture of FMT using a novel flexible displacement sensor. We prove that the flexed posture of FMT with five intervertebral disks can be detected through experiment.
Eye-in-Hand Manipulation for Remote Handling: Experimental Setup
NASA Astrophysics Data System (ADS)
Niu, Longchuan; Suominen, Olli; Aref, Mohammad M.; Mattila, Jouni; Ruiz, Emilio; Esque, Salvador
2018-03-01
A prototype for eye-in-hand manipulation in the context of remote handling in the International Thermonuclear Experimental Reactor (ITER)1 is presented in this paper. The setup consists of an industrial robot manipulator with a modified open control architecture and equipped with a pair of stereoscopic cameras, a force/torque sensor, and pneumatic tools. It is controlled through a haptic device in a mock-up environment. The industrial robot controller has been replaced by a single industrial PC running Xenomai that has a real-time connection to both the robot controller and another Linux PC running as the controller for the haptic device. The new remote handling control environment enables further development of advanced control schemes for autonomous and semi-autonomous manipulation tasks. This setup benefits from a stereovision system for accurate tracking of the target objects with irregular shapes. The overall environmental setup successfully demonstrates the required robustness and precision that remote handling tasks need.
On Using Exponential Parameter Estimators with an Adaptive Controller
NASA Technical Reports Server (NTRS)
Patre, Parag; Joshi, Suresh M.
2011-01-01
Typical adaptive controllers are restricted to using a specific update law to generate parameter estimates. This paper investigates the possibility of using any exponential parameter estimator with an adaptive controller such that the system tracks a desired trajectory. The goal is to provide flexibility in choosing any update law suitable for a given application. The development relies on a previously developed concept of controller/update law modularity in the adaptive control literature, and the use of a converse Lyapunov-like theorem. Stability analysis is presented to derive gain conditions under which this is possible, and inferences are made about the tracking error performance. The development is based on a class of Euler-Lagrange systems that are used to model various engineering systems including space robots and manipulators.
Exponentially Stabilizing Robot Control Laws
NASA Technical Reports Server (NTRS)
Wen, John T.; Bayard, David S.
1990-01-01
New class of exponentially stabilizing laws for joint-level control of robotic manipulators introduced. In case of set-point control, approach offers simplicity of proportion/derivative control architecture. In case of tracking control, approach provides several important alternatives to completed-torque method, as far as computational requirements and convergence. New control laws modified in simple fashion to obtain asymptotically stable adaptive control, when robot model and/or payload mass properties unknown.
Ibeas, Asier; de la Sen, Manuel
2006-10-01
The problem of controlling a tandem of robotic manipulators composing a teleoperation system with force reflection is addressed in this paper. The final objective of this paper is twofold: 1) to design a robust control law capable of ensuring closed-loop stability for robots with uncertainties and 2) to use the so-obtained control law to improve the tracking of each robot to its corresponding reference model in comparison with previously existing controllers when the slave is interacting with the obstacle. In this way, a multiestimation-based adaptive controller is proposed. Thus, the master robot is able to follow more accurately the constrained motion defined by the slave when interacting with an obstacle than when a single-estimation-based controller is used, improving the transparency property of the teleoperation scheme. The closed-loop stability is guaranteed if a minimum residence time, which might be updated online when unknown, between different controller parameterizations is respected. Furthermore, the analysis of the teleoperation and stability capabilities of the overall scheme is carried out. Finally, some simulation examples showing the working of the multiestimation scheme complete this paper.
Assessment of a visually guided autonomous exploration robot
NASA Astrophysics Data System (ADS)
Harris, C.; Evans, R.; Tidey, E.
2008-10-01
A system has been developed to enable a robot vehicle to autonomously explore and map an indoor environment using only visual sensors. The vehicle is equipped with a single camera, whose output is wirelessly transmitted to an off-board standard PC for processing. Visual features within the camera imagery are extracted and tracked, and their 3D positions are calculated using a Structure from Motion algorithm. As the vehicle travels, obstacles in its surroundings are identified and a map of the explored region is generated. This paper discusses suitable criteria for assessing the performance of the system by computer-based simulation and practical experiments with a real vehicle. Performance measures identified include the positional accuracy of the 3D map and the vehicle's location, the efficiency and completeness of the exploration and the system reliability. Selected results are presented and the effect of key system parameters and algorithms on performance is assessed. This work was funded by the Systems Engineering for Autonomous Systems (SEAS) Defence Technology Centre established by the UK Ministry of Defence.
NASA Astrophysics Data System (ADS)
Gwiazda, A.; Banas, W.; Sekala, A.; Foit, K.; Hryniewicz, P.; Kost, G.
2015-11-01
Process of workcell designing is limited by different constructional requirements. They are related to technological parameters of manufactured element, to specifications of purchased elements of a workcell and to technical characteristics of a workcell scene. This shows the complexity of the design-constructional process itself. The results of such approach are individually designed workcell suitable to the specific location and specific production cycle. Changing this parameters one must rebuild the whole configuration of a workcell. Taking into consideration this it is important to elaborate the base of typical elements of a robot kinematic chain that could be used as the tool for building Virtual modelling of kinematic chains of industrial robots requires several preparatory phase. Firstly, it is important to create a database element, which will be models of industrial robot arms. These models could be described as functional primitives that represent elements between components of the kinematic pairs and structural members of industrial robots. A database with following elements is created: the base kinematic pairs, the base robot structural elements, the base of the robot work scenes. The first of these databases includes kinematic pairs being the key component of the manipulator actuator modules. Accordingly, as mentioned previously, it includes the first stage rotary pair of fifth stage. This type of kinematic pairs was chosen due to the fact that it occurs most frequently in the structures of industrial robots. Second base consists of structural robot elements therefore it allows for the conversion of schematic structures of kinematic chains in the structural elements of the arm of industrial robots. It contains, inter alia, the structural elements such as base, stiff members - simple or angular units. They allow converting recorded schematic three-dimensional elements. Last database is a database of scenes. It includes elements of both simple and complex: simple models of technological equipment, conveyors models, models of the obstacles and like that. Using these elements it could be formed various production spaces (robotized workcells), in which it is possible to virtually track the operation of an industrial robot arm modelled in the system.
360-Degree Visual Detection and Target Tracking on an Autonomous Surface Vehicle
NASA Technical Reports Server (NTRS)
Wolf, Michael T; Assad, Christopher; Kuwata, Yoshiaki; Howard, Andrew; Aghazarian, Hrand; Zhu, David; Lu, Thomas; Trebi-Ollennu, Ashitey; Huntsberger, Terry
2010-01-01
This paper describes perception and planning systems of an autonomous sea surface vehicle (ASV) whose goal is to detect and track other vessels at medium to long ranges and execute responses to determine whether the vessel is adversarial. The Jet Propulsion Laboratory (JPL) has developed a tightly integrated system called CARACaS (Control Architecture for Robotic Agent Command and Sensing) that blends the sensing, planning, and behavior autonomy necessary for such missions. Two patrol scenarios are addressed here: one in which the ASV patrols a large harbor region and checks for vessels near a fixed asset on each pass and one in which the ASV circles a fixed asset and intercepts approaching vessels. This paper focuses on the ASV's central perception and situation awareness system, dubbed Surface Autonomous Visual Analysis and Tracking (SAVAnT), which receives images from an omnidirectional camera head, identifies objects of interest in these images, and probabilistically tracks the objects' presence over time, even as they may exist outside of the vehicle's sensor range. The integrated CARACaS/SAVAnT system has been implemented on U.S. Navy experimental ASVs and tested in on-water field demonstrations.
Performance capabilities of a JPL dual-arm advanced teleoperation system
NASA Technical Reports Server (NTRS)
Szakaly, Z. F.; Bejczy, A. K.
1991-01-01
The system comprises: (1) two PUMA 560 robot arms, each equipped with the latest JPL developed smart hands which contain 3-D force/moment and grasp force sensors; (2) two general purpose force reflecting hand controllers; (3) a NS32016 microprocessors based distributed computing system together with JPL developed universal motor controllers; (4) graphics display of sensor data; (5) capabilities for time delay experiments; and (6) automatic data recording capabilities. Several different types of control modes are implemented on this system using different feedback control techniques. Some of the control modes and the related feedback control techniques are described, and the achievable control performance for tracking position and force trajectories are reported. The interaction between position and force trajectory tracking is illustrated. The best performance is obtained by using a novel, task space error feedback technique.
Single-camera visual odometry to track a surgical X-ray C-arm base.
Esfandiari, Hooman; Lichti, Derek; Anglin, Carolyn
2017-12-01
This study provides a framework for a single-camera odometry system for localizing a surgical C-arm base. An application-specific monocular visual odometry system (a downward-looking consumer-grade camera rigidly attached to the C-arm base) is proposed in this research. The cumulative dead-reckoning estimation of the base is extracted based on frame-to-frame homography estimation. Optical-flow results are utilized to feed the odometry. Online positional and orientation parameters are then reported. Positional accuracy of better than 2% (of the total traveled distance) for most of the cases and 4% for all the cases studied and angular accuracy of better than 2% (of absolute cumulative changes in orientation) were achieved with this method. This study provides a robust and accurate tracking framework that not only can be integrated with the current C-arm joint-tracking system (i.e. TC-arm) but also is capable of being employed for similar applications in other fields (e.g. robotics).
Bilateral control of master-slave manipulators with constant time delay.
Forouzantabar, A; Talebi, H A; Sedigh, A K
2012-01-01
This paper presents a novel teleoperation controller for a nonlinear master-slave robotic system with constant time delay in communication channel. The proposed controller enables the teleoperation system to compensate human and environmental disturbances, while achieving master and slave position coordination in both free motion and contact situation. The current work basically extends the passivity based architecture upon the earlier work of Lee and Spong (2006) [14] to improve position tracking and consequently transparency in the face of disturbances and environmental contacts. The proposed controller employs a PID controller in each side to overcome some limitations of a PD controller and guarantee an improved performance. Moreover, by using Fourier transform and Parseval's identity in the frequency domain, we demonstrate that this new PID controller preserves the passivity of the system. Simulation and semi-experimental results show that the PID controller tracking performance is superior to that of the PD controller tracking performance in slave/environmental contacts. Copyright © 2011 ISA. Published by Elsevier Ltd. All rights reserved.
Robust Task Space Trajectory Tracking Control of Robotic Manipulators
NASA Astrophysics Data System (ADS)
Galicki, M.
2016-08-01
This work deals with the problem of the accurate task space trajectory tracking subject to finite-time convergence. Kinematic and dynamic equations of a redundant manipulator are assumed to be uncertain. Moreover, globally unbounded disturbances are allowed to act on the manipulator when tracking the trajectory by the end-effector. Furthermore, the movement is to be accomplished in such a way as to reduce both the manipulator torques and their oscillations thus eliminating the potential robot vibrations. Based on suitably defined task space non-singular terminal sliding vector variable and the Lyapunov stability theory, we propose a class of chattering-free robust controllers, based on the estimation of transpose Jacobian, which seem to be effective in counteracting both uncertain kinematics and dynamics, unbounded disturbances and (possible) kinematic and/or algorithmic singularities met on the robot trajectory. The numerical simulations carried out for a redundant manipulator of a SCARA type consisting of the three revolute kinematic pairs and operating in a two-dimensional task space, illustrate performance of the proposed controllers as well as comparisons with other well known control schemes.
Obstacle avoidance handling and mixed integer predictive control for space robots
NASA Astrophysics Data System (ADS)
Zong, Lijun; Luo, Jianjun; Wang, Mingming; Yuan, Jianping
2018-04-01
This paper presents a novel obstacle avoidance constraint and a mixed integer predictive control (MIPC) method for space robots avoiding obstacles and satisfying physical limits during performing tasks. Firstly, a novel kind of obstacle avoidance constraint of space robots, which needs the assumption that the manipulator links and the obstacles can be represented by convex bodies, is proposed by limiting the relative velocity between two closest points which are on the manipulator and the obstacle, respectively. Furthermore, the logical variables are introduced into the obstacle avoidance constraint, which have realized the constraint form is automatically changed to satisfy different obstacle avoidance requirements in different distance intervals between the space robot and the obstacle. Afterwards, the obstacle avoidance constraint and other system physical limits, such as joint angle ranges, the amplitude boundaries of joint velocities and joint torques, are described as inequality constraints of a quadratic programming (QP) problem by using the model predictive control (MPC) method. To guarantee the feasibility of the obtained multi-constraint QP problem, the constraints are treated as soft constraints and assigned levels of priority based on the propositional logic theory, which can realize that the constraints with lower priorities are always firstly violated to recover the feasibility of the QP problem. Since the logical variables have been introduced, the optimization problem including obstacle avoidance and system physical limits as prioritized inequality constraints is termed as MIPC method of space robots, and its computational complexity as well as possible strategies for reducing calculation amount are analyzed. Simulations of the space robot unfolding its manipulator and tracking the end-effector's desired trajectories with the existence of obstacles and physical limits are presented to demonstrate the effectiveness of the proposed obstacle avoidance strategy and MIPC control method of space robots.
Telerobot local-remote control architecture for space flight program applications
NASA Technical Reports Server (NTRS)
Zimmerman, Wayne; Backes, Paul; Steele, Robert; Long, Mark; Bon, Bruce; Beahan, John
1993-01-01
The JPL Supervisory Telerobotics (STELER) Laboratory has developed and demonstrated a unique local-remote robot control architecture which enables management of intermittent communication bus latencies and delays such as those expected for ground-remote operation of Space Station robotic systems via the Tracking and Data Relay Satellite System (TDRSS) communication platform. The current work at JPL in this area has focused on enhancing the technologies and transferring the control architecture to hardware and software environments which are more compatible with projected ground and space operational environments. At the local site, the operator updates the remote worksite model using stereo video and a model overlay/fitting algorithm which outputs the location and orientation of the object in free space. That information is relayed to the robot User Macro Interface (UMI) to enable programming of the robot control macros. This capability runs on a single Silicon Graphics Inc. machine. The operator can employ either manual teleoperation, shared control, or supervised autonomous control to manipulate the intended object. The remote site controller, called the Modular Telerobot Task Execution System (MOTES), runs in a multi-processor VME environment and performs the task sequencing, task execution, trajectory generation, closed loop force/torque control, task parameter monitoring, and reflex action. This paper describes the new STELER architecture implementation, and also documents the results of the recent autonomous docking task execution using the local site and MOTES.