Schmidt et al., 2018 - Google Patents
Grasping of unknown objects using deep convolutional neural networks based on depth imagesSchmidt et al., 2018
View PDF- Document ID
- 6690525564310539205
- Author
- Schmidt P
- Vahrenkamp N
- Wächter M
- Asfour T
- Publication year
- Publication venue
- 2018 IEEE international conference on robotics and automation (ICRA)
External Links
Snippet
We present a data-driven, bottom-up, deep learning approach to robotic grasping of unknown objects using Deep Convolutional Neural Networks (DCNNs). The approach uses depth images of the scene as its sole input for synthesis of a single-grasp solution during …
- 230000001537 neural 0 title abstract description 7
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/36—Image preprocessing, i.e. processing the image information without deciding about the identity of the image
- G06K9/46—Extraction of features or characteristics of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6217—Design or setup of recognition systems and techniques; Extraction of features in feature space; Clustering techniques; Blind source separation
- G06K9/6232—Extracting features by transforming the feature space, e.g. multidimensional scaling; Mappings, e.g. subspace methods
- G06K9/6247—Extracting features by transforming the feature space, e.g. multidimensional scaling; Mappings, e.g. subspace methods based on an approximation criterion, e.g. principal component analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6267—Classification techniques
- G06K9/6268—Classification techniques relating to the classification paradigm, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/00221—Acquiring or recognising human faces, facial parts, facial sketches, facial expressions
- G06K9/00268—Feature extraction; Face representation
- G06K9/00281—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6201—Matching; Proximity measures
- G06K9/6202—Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
- G06K9/62—Methods or arrangements for recognition using electronic means
- G06K9/6288—Fusion techniques, i.e. combining data from various sources, e.g. sensor fusion
- G06K9/629—Fusion techniques, i.e. combining data from various sources, e.g. sensor fusion of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computer systems based on biological models
- G06N3/02—Computer systems based on biological models using neural network models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N99/00—Subject matter not provided for in other groups of this subclass
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1674—Programme controls characterised by safety, monitoring, diagnostic
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Schmidt et al. | Grasping of unknown objects using deep convolutional neural networks based on depth images | |
Newbury et al. | Deep learning approaches to grasp synthesis: A review | |
Nguyen et al. | Object-based affordances detection with convolutional neural networks and dense conditional random fields | |
Lundell et al. | Ddgc: Generative deep dexterous grasping in clutter | |
Bohg et al. | Data-driven grasp synthesis—a survey | |
US11741701B2 (en) | Autonomous task performance based on visual embeddings | |
Bohg et al. | Mind the gap-robotic grasping under incomplete observation | |
Hämäläinen et al. | Affordance learning for end-to-end visuomotor robot control | |
Tang et al. | Learning collaborative pushing and grasping policies in dense clutter | |
Ottenhaus et al. | Visuo-haptic grasping of unknown objects based on gaussian process implicit surfaces and deep learning | |
Suzuki et al. | Grasping of unknown objects on a planar surface using a single depth image | |
Faria et al. | Knowledge-based reasoning from human grasp demonstrations for robot grasp synthesis | |
Iqbal et al. | Toward sim-to-real directional semantic grasping | |
US20220402125A1 (en) | System and method for determining a grasping hand model | |
Bohg et al. | Task-based grasp adaptation on a humanoid robot | |
Huang et al. | Grasping novel objects with a dexterous robotic hand through neuroevolution | |
Ala et al. | A 3D-grasp synthesis algorithm to grasp unknown objects based on graspable boundary and convex segments | |
Chen et al. | A probabilistic framework for uncertainty-aware high-accuracy precision grasping of unknown objects | |
Hasegawa et al. | Graspfusion: Realizing complex motion by learning and fusing grasp modalities with instance segmentation | |
Aarno et al. | Early reactive grasping with second order 3D feature relations | |
Kushwaha et al. | Generating quality grasp rectangle using Pix2Pix GAN for intelligent robot grasping | |
Yang et al. | Predict robot grasp outcomes based on multi-modal information | |
Mohammed et al. | Color matching based approach for robotic grasping | |
Romero et al. | Human-to-robot mapping of grasps | |
Geng et al. | A Novel Real-time Grasping Method Cobimbed with YOLO and GDFCN |