Moezzi, 2024 - Google Patents
Towards Sample-Efficient Reinforcement Learning Methods for Robotic Manipulation TasksMoezzi, 2024
- Document ID
- 10941318954184863966
- Author
- Moezzi M
- Publication year
External Links
Snippet
Reinforcement learning (RL) provides a framework for solving complex tasks without pre- existing knowledge of the systems involved. These algorithms hold significant promise for control and robotics, areas that have historically depended on precise system dynamics and …
- 238000000034 method 0 title abstract description 201
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/04—Architectures, e.g. interconnection topology
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
- G06N99/005—Learning machines, i.e. computer in which a programme is changed according to experience gained by the machine itself during a complete run
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/0265—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion
- G05B13/027—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric the criterion being a learning criterion using neural networks only
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B13/00—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion
- G05B13/02—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric
- G05B13/04—Adaptive control systems, i.e. systems automatically adjusting themselves to have a performance which is optimum according to some preassigned criterion electric involving the use of models or simulators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computer systems utilising knowledge based models
- G06N5/04—Inference methods or devices
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6217—Design or setup of recognition systems and techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/50—Computer-aided design
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F19/00—Digital computing or data processing equipment or methods, specially adapted for specific applications
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| Lanillos et al. | Active inference in robotics and artificial agents: Survey and challenges | |
| Morales et al. | A survey on deep learning and deep reinforcement learning in robotics with a tutorial on deep reinforcement learning | |
| US11714996B2 (en) | Learning motor primitives and training a machine learning system using a linear-feedback-stabilized policy | |
| Li et al. | Prodmp: A unified perspective on dynamic and probabilistic movement primitives | |
| Shahid et al. | Continuous control actions learning and adaptation for robotic manipulation through reinforcement learning | |
| Bekey et al. | Neural networks in robotics | |
| Amarjyoti | Deep reinforcement learning for robotic manipulation-the state of the art | |
| Schaal et al. | Learning control in robotics | |
| Hanna et al. | Grounded action transformation for sim-to-real reinforcement learning | |
| Tavassoli et al. | Learning skills from demonstrations: A trend from motion primitives to experience abstraction | |
| Kargin et al. | A reinforcement learning approach for continuum robot control | |
| Levine | Motor skill learning with local trajectory methods | |
| Gawali et al. | Development of improved coyote optimization with deep neural network for intelligent skill knowledge transfer for human to robot interaction | |
| Rohrer | BECCA: Reintegrating AI for Natural World Interaction. | |
| Bonsignorio et al. | An imitation learning approach for the control of a low-cost low-accuracy robotic arm for unstructured environments | |
| Moezzi | Towards Sample-Efficient Reinforcement Learning Methods for Robotic Manipulation Tasks | |
| Learning | Tactile grasp refinement using deep reinforcement learning and analytic grasp stability metrics | |
| Zhang et al. | A Review on Robot Manipulation Methods in Human-Robot Interactions | |
| Grimes et al. | Learning nonparametric policies by imitation | |
| Navez | Contributions to the Concept of Embodied Intelligence in Soft Robotics through Control and Design Co-Optimization | |
| Chen et al. | Boosting Reinforcement Learning Algorithms in Continuous Robotic Reaching Tasks Using Adaptive Potential Functions | |
| Li | Reinforcement learning-based motion planning in partially observable environments for complex tasks | |
| Zheng | Data-Driven Robotic Manipulation of Deformable Objects Using Tactile Feedback: From Model-Free to Model-Based Approaches | |
| Weideman | Robot navigation in cluttered environments with deep reinforcement learning | |
| Gillen | Improving Reinforcement Learning for Robotics with Control and Dynamical Systems Theory |