Beik Mohammadi et al., 2019 - Google Patents
Mixed-reality deep reinforcement learning for a reach-to-grasp taskBeik Mohammadi et al., 2019
View PDF- Document ID
- 1472659692602384655
- Author
- Beik Mohammadi H
- Zamani M
- Kerzel M
- Wermter S
- Publication year
- Publication venue
- Artificial Neural Networks and Machine Learning–ICANN 2019: Theoretical Neural Computation: 28th International Conference on Artificial Neural Networks, Munich, Germany, September 17–19, 2019, Proceedings, Part I 28
External Links
Snippet
Abstract Deep Reinforcement Learning (DRL) has become successful across various robotic applications. However, DRL methods are not sample-efficient and require long learning times. We present an approach for online continuous deep reinforcement learning for a …
- 230000002787 reinforcement 0 title abstract description 33
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/39—Robotics, robotics to robotics hand
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B19/00—Programme-control systems
- G05B19/02—Programme-control systems electric
- G05B19/04—Programme control other than numerical control, i.e. in sequence controllers or logic controllers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B17/00—Systems involving the use of models or simulators of said systems
- G05B17/02—Systems involving the use of models or simulators of said systems electric
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1628—Programme controls characterised by the control loop
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Andrychowicz et al. | Learning dexterous in-hand manipulation | |
Shahid et al. | Continuous control actions learning and adaptation for robotic manipulation through reinforcement learning | |
JP7291185B2 (en) | Technologies for force and torque guided robot assembly | |
Dimeas et al. | Towards progressive automation of repetitive tasks through physical human-robot interaction | |
McDonald et al. | Guided imitation of task and motion planning | |
Beik Mohammadi et al. | Mixed-reality deep reinforcement learning for a reach-to-grasp task | |
JP2023526211A (en) | Distributed robot empirical learning | |
Ramamurthy et al. | Leveraging domain knowledge for reinforcement learning using MMC architectures | |
Fu et al. | Robot motor skill transfer with alternate learning in two spaces | |
Braun et al. | Incorporation of expert knowledge for learning robotic assembly tasks | |
Scheiderer et al. | Effects of domain randomization on simulation-to-reality transfer of reinforcement learning policies for industrial robots | |
Claassens | An RRT-based path planner for use in trajectory imitation | |
JP7531733B2 (en) | Transferring tasks between different domains | |
Nambiar et al. | Automation of unstructured production environment by applying reinforcement learning | |
Guo et al. | Robot path planning via deep reinforcement learning with improved reward function | |
Osa et al. | Goal-conditioned variational autoencoder trajectory primitives with continuous and discrete latent codes | |
Deng et al. | Sample-efficiency, stability and generalization analysis for deep reinforcement learning on robotic peg-in-hole assembly | |
Bai et al. | Bionic Hand Motion Control Method Based on Imitation of Human Hand Movements and Reinforcement Learning | |
Akbulut et al. | Bimanual rope manipulation skill synthesis through context dependent correction policy learning from human demonstration | |
Zhong et al. | Regressor-Based Model Adaptation for Shaping Deformable Linear Objects with Force Control | |
Kim et al. | Generalizing over uncertain dynamics for online trajectory generation | |
Yuan et al. | 6-DOF Industrial Manipulator Motion Planning Based on RRT-Connect Algorithm | |
Ruud | Reinforcement learning with the TIAGo research robot: manipulator arm control with actor-critic reinforcement learning | |
Liu et al. | A modified cartesian space DMPS model for robot motion generation | |
Wang et al. | Event-driven collision-free path planning for cooperative robots in dynamic environment |