Kumra et al., 2021 - Google Patents
Learning robotic manipulation tasks via task progress based Gaussian reward and loss adjusted explorationKumra et al., 2021
View PDF- Document ID
- 12787072088985677426
- Author
- Kumra S
- Joshi S
- Sahin F
- Publication year
- Publication venue
- IEEE Robotics and Automation Letters
External Links
Snippet
Multi-step manipulation tasks in unstructured environments are extremely challenging for a robot to learn. Such tasks interlace high-level reasoning that consists of the expected states that can be attained to achieve an overall task and low-level reasoning that decides what …
- 238000004805 robotic 0 title abstract description 35
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Seita et al. | Learning to rearrange deformable cables, fabrics, and bags with goal-conditioned transporter networks | |
Mu et al. | Maniskill: Generalizable manipulation skill benchmark with large-scale demonstrations | |
Shiarlis et al. | Taco: Learning task decomposition via temporal alignment for control | |
Kiatos et al. | Robust object grasping in clutter via singulation | |
Kumra et al. | Learning robotic manipulation tasks via task progress based Gaussian reward and loss adjusted exploration | |
Zhou et al. | 6dof grasp planning by optimizing a deep learning scoring function | |
Piergiovanni et al. | Learning real-world robot policies by dreaming | |
Tang et al. | Learning collaborative pushing and grasping policies in dense clutter | |
Toussaint et al. | Integrated motor control, planning, grasping and high-level reasoning in a blocks world using probabilistic inference | |
Shen et al. | Learning category-level generalizable object manipulation policy via generative adversarial self-imitation learning from demonstrations | |
Zeng | Learning visual affordances for robotic manipulation | |
Mohammed et al. | Pick and place objects in a cluttered scene using deep reinforcement learning | |
Yan et al. | Learning probabilistic multi-modal actor models for vision-based robotic grasping | |
Wang et al. | Policy learning in se (3) action spaces | |
Wang et al. | Learning of long-horizon sparse-reward robotic manipulator tasks with base controllers | |
Hwang et al. | Development of a mimic robot—Learning from demonstration incorporating object detection and multiaction recognition | |
Chen et al. | Dextransfer: Real world multi-fingered dexterous grasping with minimal human demonstrations | |
Chang et al. | Robot task learning from demonstration using petri nets | |
Khargonkar et al. | Neuralgrasps: Learning implicit representations for grasps of multiple robotic hands | |
Wang et al. | Bulletarm: An open-source robotic manipulation benchmark and learning framework | |
Zakaria et al. | Robotic control of the deformation of soft linear objects using deep reinforcement learning | |
Li et al. | Learning target-oriented push-grasping synergy in clutter with action space decoupling | |
Xu et al. | Efficient object manipulation to an arbitrary goal pose: Learning-based anytime prioritized planning | |
Ren et al. | Learning bifunctional push-grasping synergistic strategy for goal-agnostic and goal-oriented tasks | |
Ren et al. | Fast-learning grasping and pre-grasping via clutter quantization and Q-map masking |