Bellegarda et al., 2020 - Google Patents
An online training method for augmenting mpc with deep reinforcement learningBellegarda et al., 2020
- Document ID
- 363685456475870432
- Author
- Bellegarda G
- Byl K
- Publication year
- Publication venue
- 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)
External Links
Snippet
Recent breakthroughs both in reinforcement learning and trajectory optimization have made significant advances towards real world robotic system deployment. Reinforcement learning (RL) can be applied to many problems without needing any modeling or intuition about the …
- 230000002787 reinforcement 0 title abstract description 17
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
- G05B13/027—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/12—Computer systems based on biological models using genetic models
- G06N3/126—Genetic algorithms, i.e. information processing using digital simulations of the genetic system
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B17/00—Systems involving the use of models or simulators of said systems
- G05B17/02—Systems involving the use of models or simulators of said systems electric
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Bellegarda et al. | An online training method for augmenting mpc with deep reinforcement learning | |
| Zhu et al. | A survey of sim-to-real transfer techniques applied to reinforcement learning for bioinspired robots | |
| Williams et al. | Information theoretic MPC for model-based reinforcement learning | |
| Samak et al. | Control strategies for autonomous vehicles | |
| Mordatch et al. | Interactive control of diverse complex characters with neural networks | |
| Park et al. | Inverse optimal control for humanoid locomotion | |
| Mehmood et al. | Application of deep reinforcement learning for tracking control of 3WD omnidirectional mobile robot | |
| Levine | Motor skill learning with local trajectory methods | |
| Lim et al. | Formation control of leader following unmanned ground vehicles using nonlinear model predictive control | |
| Alvarez-Padilla et al. | Real-time whole-body control of legged robots with model-predictive path integral control | |
| Melo et al. | Push recovery strategies through deep reinforcement learning | |
| Ganai et al. | Learning stabilization control from observations by learning lyapunov-like proxy models | |
| Liu et al. | Discrete-time hybrid automata learning: Legged locomotion meets skateboarding | |
| Zhang et al. | Trajectory-tracking control of robotic system via proximal policy optimization | |
| Bellegarda et al. | Combining benefits from trajectory optimization and deep reinforcement learning | |
| Xu et al. | Performance evaluation and optimization of human control strategy | |
| Oliveira et al. | Learning to race through coordinate descent bayesian optimisation | |
| Clark et al. | Evolving controllers for a transformable wheel mobile robot | |
| Tang et al. | Learning agile motor skills on quadrupedal robots using curriculum learning | |
| Demir et al. | Motion planning and control with randomized payloads using deep reinforcement learning | |
| Widulle et al. | Using reverse reinforcement learning for assembly tasks | |
| Carreras et al. | Application of SONQL for real-time learning of robot behaviors | |
| CN117474076A (en) | Adversarial inverse reinforcement learning landing method for the powered descent stage of the Mars probe | |
| Bai et al. | An improved DDPG algorithm based on evolution-guided transfer in reinforcement learning | |
| Ito et al. | Extended QDSEGA for controlling real robots-acquisition of locomotion patterns for snake-like robot |