Kolaric et al., 2020 - Google Patents
Local policy optimization for trajectory-centric reinforcement learningKolaric et al., 2020
View PDF- Document ID
- 12695073583234228621
- Author
- Kolaric P
- Jha D
- Raghunathan A
- Lewis F
- Benosman M
- Romeres D
- Nikovski D
- Publication year
- Publication venue
- 2020 IEEE International Conference on Robotics and Automation (ICRA)
External Links
Snippet
The goal of this paper is to present a method for simultaneous trajectory and local stabilizing policy optimization to generate local policies for trajectory-centric model-based reinforcement learning (MBRL). This is motivated by the fact that global policy optimization …
- 238000005457 optimization 0 title abstract description 52
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
- G05B13/042—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators in which a parameter or coefficient is automatically adjusted to optimise the performance
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
- G05B13/027—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/39—Robotics, robotics to robotics hand
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B17/00—Systems involving the use of models or simulators of said systems
- G05B17/02—Systems involving the use of models or simulators of said systems electric
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/02—Knowledge representation
- G06N5/022—Knowledge engineering, knowledge acquisition
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| EP3924884B1 (en) | System and method for robust optimization for trajectory-centric model-based reinforcement learning | |
| Choi et al. | Reinforcement learning for safety-critical control under model uncertainty, using control lyapunov functions and control barrier functions | |
| Dai et al. | Robust model predictive tracking control for robot manipulators with disturbances | |
| Castillo et al. | Intelligent adaptive model-based control of robotic dynamic systems with a hybrid fuzzy-neural approach | |
| Gribovskaya et al. | Learning non-linear multivariate dynamics of motion in robotic manipulators | |
| Rozo et al. | Learning optimal controllers in human-robot cooperative transportation tasks with position and force constraints | |
| Levine et al. | Learning neural network policies with guided policy search under unknown dynamics | |
| Krug et al. | Model predictive motion control based on generalized dynamical movement primitives | |
| Sacks et al. | Learning sampling distributions for model predictive control | |
| Kolaric et al. | Local policy optimization for trajectory-centric reinforcement learning | |
| Lee et al. | Gp-ilqg: Data-driven robust optimal control for uncertain nonlinear dynamical systems | |
| Li et al. | A unified perspective on multiple shooting in differential dynamic programming | |
| Fandel et al. | Development of reinforcement learning algorithm for 2-dof helicopter model | |
| Alberto et al. | Computed torque control with variable gains through Gaussian process regression | |
| Rezazadeh et al. | Learning contraction policies from offline data | |
| Haffemayer et al. | Model predictive control under hard collision avoidance constraints for a robotic arm | |
| Fan et al. | Robust identification of switching Markov ARX models using EM algorithm | |
| Boloka et al. | Knowledge transfer using model-based deep reinforcement learning | |
| Jha et al. | Local Policy Optimization for Trajectory-Centric Reinforcement Learning | |
| Kolaric et al. | Robust optimization for trajectory-centric model-based reinforcement learning | |
| Jha et al. | Robust optimization for trajectory-centric model-based reinforcement learning | |
| Baldauf et al. | Iterative learning-based model predictive control for mobile robots in space applications | |
| He et al. | A Barrier Pair Method for Safe Human-Robot Shared Autonomy | |
| Zometa et al. | Quantized deep path-following control on a microcontroller | |
| Shi et al. | A constrained framework based on IBLF for robot learning with human supervision |