Ganai et al., 2023 - Google Patents
Learning stabilization control from observations by learning lyapunov-like proxy modelsGanai et al., 2023
View PDF- Document ID
- 3208955406059212073
- Author
- Ganai M
- Hirayama C
- Chang Y
- Gao S
- Publication year
- Publication venue
- arXiv preprint arXiv:2303.02215
External Links
Snippet
The deployment of Reinforcement Learning to robotics applications faces the difficulty of reward engineering. Therefore, approaches have focused on creating reward functions by Learning from Observations (LfO) which is the task of learning policies from expert …
Classifications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
- G05B13/027—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
- G05B13/042—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators in which a parameter or coefficient is automatically adjusted to optimise the performance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B17/00—Systems involving the use of models or simulators of said systems
- G05B17/02—Systems involving the use of models or simulators of said systems electric
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Chang et al. | Stabilizing neural control using self-learned almost lyapunov critics | |
| Song et al. | Policy search for model predictive control with application to agile drone flight | |
| Pertsch et al. | Accelerating reinforcement learning with learned skill priors | |
| Han et al. | Actor-critic reinforcement learning for control with stability guarantee | |
| Fu et al. | One-shot learning of manipulation skills with online dynamics adaptation and neural network priors | |
| Koryakovskiy et al. | Model-plant mismatch compensation using reinforcement learning | |
| Chebotar et al. | Combining model-based and model-free updates for trajectory-centric reinforcement learning | |
| Li et al. | A policy search method for temporal logic specified reinforcement learning tasks | |
| Hanna et al. | Grounded action transformation for robot learning in simulation | |
| CN107748566B (en) | Underwater autonomous robot fixed depth control method based on reinforcement learning | |
| Ganai et al. | Learning stabilization control from observations by learning lyapunov-like proxy models | |
| Widmer et al. | Tuning legged locomotion controllers via safe bayesian optimization | |
| Leurent et al. | Approximate robust control of uncertain dynamical systems | |
| Bellegarda et al. | An online training method for augmenting mpc with deep reinforcement learning | |
| Levine | Motor skill learning with local trajectory methods | |
| Hafez et al. | Efficient intrinsically motivated robotic grasping with learning-adaptive imagination in latent space | |
| Torabi et al. | Sample-efficient adversarial imitation learning from observation | |
| Yang et al. | Mpr-rl: Multi-prior regularized reinforcement learning for knowledge transfer | |
| Ennen et al. | Learning robust manipulation skills with guided policy search via generative motor reflexes | |
| Valadas et al. | Learning low-dimensional strain models of soft robots by looking at the evolution of their shape with application to model-based control | |
| Carreno et al. | Adapting neural models with sequential monte carlo dropout | |
| Abbatematteo et al. | Bootstrapping motor skill learning with motion planning | |
| Bellegarda et al. | Combining benefits from trajectory optimization and deep reinforcement learning | |
| Pignat et al. | Generative adversarial training of product of policies for robust and adaptive movement primitives | |
| Chen et al. | CasIL: Cognizing and Imitating Skills via a Dual Cognition-Action Architecture |