Staretu et al., 2016 - Google Patents
Leap motion device used to control a real anthropomorphic gripperStaretu et al., 2016
View HTML- Document ID
- 15506465164959675301
- Author
- Staretu I
- Moldovan C
- Publication year
- Publication venue
- International Journal of Advanced Robotic Systems
External Links
Snippet
This paper presents for the first time the use of the Leap Motion device to control an anthropomorphic gripper with five fingers. First, a description of the Leap Motion device is presented, highlighting its main functional characteristics, followed by testing of its use for …
- 238000000034 method 0 abstract description 16
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03545—Pens or stylus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F1/00—Details of data-processing equipment not covered by groups G06F3/00 - G06F13/00, e.g. cooling, packaging or power supply specially adapted for computer application
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06K—RECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
- G06K9/00—Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F19/00—Digital computing or data processing equipment or methods, specially adapted for specific applications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06F—ELECTRICAL DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING; COUNTING
- G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Du et al. | Markerless kinect-based hand tracking for robot teleoperation | |
Lin et al. | Robot grasp planning based on demonstrated grasp strategies | |
Bachmann et al. | Review of three-dimensional human-computer interaction with focus on the leap motion controller | |
Staretu et al. | Leap motion device used to control a real anthropomorphic gripper | |
Moon et al. | Multiple kinect sensor fusion for human skeleton tracking using Kalman filtering | |
Ciocarlie et al. | Hand posture subspaces for dexterous robotic grasping | |
Xue et al. | Multimodal human hand motion sensing and analysis—A review | |
Krishnan et al. | SWIRL: A sequential windowed inverse reinforcement learning algorithm for robot tasks with delayed rewards | |
Eppner et al. | Exploitation of environmental constraints in human and robotic grasping | |
Bicchi | On the closure properties of robotic grasping | |
Lin et al. | Grasp planning to maximize task coverage | |
Lambrecht et al. | Spatial programming for industrial robots through task demonstration | |
Cicirelli et al. | A kinect-based gesture recognition approach for a natural human robot interface | |
Zapata-Impata et al. | Fast geometry-based computation of grasping points on three-dimensional point clouds | |
Lee et al. | Mimetic communication model with compliant physical contact in human—humanoid interaction | |
Fang et al. | Robotic teleoperation systems using a wearable multimodal fusion device | |
Cordella et al. | Human hand motion analysis and synthesis of optimal power grasps for a robotic hand | |
Dang et al. | Semantic grasping: planning task-specific stable robotic grasps | |
LaViola Jr | Context aware 3D gesture recognition for games and virtual reality | |
LaViola Jr | An introduction to 3D gestural interfaces | |
Chen et al. | Real-time multi-modal human–robot collaboration using gestures and speech | |
Zubrycki et al. | Using integrated vision systems: three gears and leap motion, to control a 3-finger dexterous gripper | |
Pozzi et al. | Hand closure model for planning top grasps with soft robotic hands | |
Chen et al. | Hand gesture modeling and recognition for human and robot interactive assembly using hidden markov models | |
Alonso et al. | Hand gesture recognition in real world scenarios using approximate string matching |