Morales et al., 2021 - Google Patents
A survey on deep learning and deep reinforcement learning in robotics with a tutorial on deep reinforcement learningMorales et al., 2021
- Document ID
- 17229616862935592127
- Author
- Morales E
- Murrieta-Cid R
- Becerra I
- Esquivel-Basaldua M
- Publication year
- Publication venue
- Intelligent Service Robotics
External Links
Snippet
This article is about deep learning (DL) and deep reinforcement learning (DRL) works applied to robotics. Both tools have been shown to be successful in delivering data-driven solutions for robotics tasks, as well as providing a natural way to develop an end-to-end …
- 238000004805 robotic 0 title abstract description 80
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/004—Artificial life, i.e. computers simulating life
- G06N3/006—Artificial life, i.e. computers simulating life based on simulated virtual individual or collective life forms, e.g. single "avatar", social simulations, virtual worlds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/004—Artificial life, i.e. computers simulating life
- G06N3/008—Artificial life, i.e. computers simulating life based on physical entities controlled by simulated intelligence so as to replicate intelligent life forms, e.g. robots replicating pets or humans in their appearance or behavior
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
- G05B13/027—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/12—Computer systems based on biological models using genetic models
- G06N3/126—Genetic algorithms, i.e. information processing using digital simulations of the genetic system
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Morales et al. | A survey on deep learning and deep reinforcement learning in robotics with a tutorial on deep reinforcement learning | |
| Chen et al. | Stabilization approaches for reinforcement learning-based end-to-end autonomous driving | |
| Qureshi et al. | Motion planning networks | |
| Kahn et al. | Uncertainty-aware reinforcement learning for collision avoidance | |
| Tai et al. | A survey of deep network solutions for learning control in robotics: From reinforcement to imitation | |
| Chen et al. | Driving maneuvers prediction based autonomous driving control by deep Monte Carlo tree search | |
| Chernova et al. | Confidence-based policy learning from demonstration using gaussian mixture models | |
| Kilinc et al. | Reinforcement learning for robotic manipulation using simulated locomotion demonstrations | |
| Grigorescu et al. | Neurotrajectory: A neuroevolutionary approach to local state trajectory learning for autonomous vehicles | |
| Fan et al. | Learning resilient behaviors for navigation under uncertainty | |
| Levine | Motor skill learning with local trajectory methods | |
| Wen et al. | A deep residual reinforcement learning algorithm based on Soft Actor-Critic for autonomous navigation | |
| Dwivedi et al. | Continuous control of autonomous vehicles using plan-assisted deep reinforcement learning | |
| Abbatematteo et al. | Bootstrapping motor skill learning with motion planning | |
| Hirose et al. | Probabilistic visual navigation with bidirectional image prediction | |
| Zhang et al. | A Review on Robot Manipulation Methods in Human-Robot Interactions | |
| Mortensen et al. | Two-Stage Reinforcement Learning for Planetary Rover Navigation: Reducing the Reality Gap with Offline Noisy Data | |
| Palaio | DeepRL-based motion planning for indoor robot navigation | |
| Li et al. | Dynamic obstacle avoidance and grasping planning for mobile robotic arm in complex environment based on improved TD3: Y. Li et al. | |
| Chatzilygeroudis | Micro-data reinforcement learning for adaptive robots | |
| Hussein | Deep learning based approaches for imitation learning | |
| Moezzi | Towards Sample-Efficient Reinforcement Learning Methods for Robotic Manipulation Tasks | |
| László | Implementing a deep reinforcement learning model for autonomous driving | |
| Kazazis et al. | Development and evaluation of Reinforcement Learning models for the FOSSBot Open-Source educational robot | |
| Khaitan | Exploring Reinforcement Learning Approaches for Safety Critical Environments |