Sepahvand et al., 2024 - Google Patents
Image-to-Joint Inverse Kinematic of a Supportive Continuum Arm Using Deep LearningSepahvand et al., 2024
View PDF- Document ID
- 6368843844440602056
- Author
- Sepahvand S
- Wang G
- Janabi-Sharifi F
- Publication year
- Publication venue
- arXiv preprint arXiv:2405.20248
External Links
Snippet
In this work, a deep learning-based technique is used to study the image-to-joint inverse kinematics of a tendondriven supportive continuum arm. An eye-off-hand configuration is considered by mounting a camera at a fixed pose with respect to the inertial frame attached …
- 230000003319 supportive effect 0 title abstract description 14
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1674—Programme controls characterised by safety, monitoring, diagnostic
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1602—Programme controls characterised by the control system, structure, architecture
- B25J9/1607—Calculation of inertia, jacobian matrixes and inverses
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
- G06N3/08—Learning methods
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/39—Robotics, robotics to robotics hand
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6217—Design or setup of recognition systems and techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
- G06K9/6232—Extracting features by transforming the feature space, e.g. multidimensional scaling; Mappings, e.g. subspace methods
- G06K9/6247—Extracting features by transforming the feature space, e.g. multidimensional scaling; Mappings, e.g. subspace methods based on an approximation criterion, e.g. principal component analysis
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Lee et al. | Making sense of vision and touch: Self-supervised learning of multimodal representations for contact-rich tasks | |
Hu et al. | 3-D deformable object manipulation using deep neural networks | |
Wang et al. | Dexgraspnet: A large-scale robotic dexterous grasp dataset for general objects based on simulation | |
Khalil et al. | Dexterous robotic manipulation of deformable objects with multi-sensory feedback-a review | |
Sutanto et al. | Learning latent space dynamics for tactile servoing | |
Mayer et al. | FFHNet: Generating multi-fingered robotic grasps for unknown objects in real-time | |
US20220402125A1 (en) | System and method for determining a grasping hand model | |
CN113119073A (en) | Mechanical arm system based on computer vision and machine learning and oriented to 3C assembly scene | |
Devgon et al. | Orienting novel 3D objects using self-supervised learning of rotation transforms | |
Zhu et al. | Robot obstacle avoidance system using deep reinforcement learning | |
Liu et al. | Robust robotic 3-D drawing using closed-loop planning and online picked pens | |
Sepahvand et al. | Image-to-Joint Inverse Kinematic of a Supportive Continuum Arm Using Deep Learning | |
Lu et al. | Visual-tactile robot grasping based on human skill learning from demonstrations using a wearable parallel hand exoskeleton | |
Funabashi et al. | Variable in-hand manipulations for tactile-driven robot hand via CNN-LSTM | |
Prasad et al. | Learning human-like hand reaching for human-robot handshaking | |
Chen et al. | Sequential motion primitives recognition of robotic arm task via human demonstration using hierarchical BiLSTM classifier | |
Liu et al. | DexRepNet: Learning dexterous robotic grasping network with geometric and spatial hand-object representations | |
Liu et al. | $\alpha $-MDF: An Attention-based Multimodal Differentiable Filter for Robot State Estimation | |
Palm et al. | Recognition of human grasps by time-clustering and fuzzy modeling | |
CN116423520A (en) | Mechanical arm track planning method based on vision and dynamic motion primitives | |
Añazco et al. | Human-like object grasping and relocation for an anthropomorphic robotic hand with natural hand pose priors in deep reinforcement learning | |
Yuan et al. | Tactile-Reactive Roller Grasper | |
Rolinat et al. | Human initiated grasp space exploration algorithm for an underactuated robot gripper using variational autoencoder | |
Rodriguez et al. | Learning postural synergies for categorical grasping through shape space registration | |
Hélénon et al. | Learning to Grasp: from Somewhere to Anywhere |