Nothing Special   »   [go: up one dir, main page]

US20190220098A1 - Gesture Operated Wrist Mounted Camera System - Google Patents

Gesture Operated Wrist Mounted Camera System Download PDF

Info

Publication number
US20190220098A1
US20190220098A1 US16/363,261 US201916363261A US2019220098A1 US 20190220098 A1 US20190220098 A1 US 20190220098A1 US 201916363261 A US201916363261 A US 201916363261A US 2019220098 A1 US2019220098 A1 US 2019220098A1
Authority
US
United States
Prior art keywords
user
camera
wristband
wristband device
sensor
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/363,261
Inventor
Vikas Gupta
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US16/363,261 priority Critical patent/US20190220098A1/en
Publication of US20190220098A1 publication Critical patent/US20190220098A1/en
Priority to US17/690,754 priority patent/US11861069B2/en
Priority to US18/503,762 priority patent/US20240152215A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/014Hand-worn input/output arrangements, e.g. data gloves
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/033Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
    • G06F3/0346Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/50Constructional details
    • H04N23/51Housings
    • H04N5/2252

Definitions

  • the present invention relates to wearable devices, and more particularly, to a wearable camera system.
  • Timing and framing are key factors in photography and videography that capture moments. People increasingly want to reduce the time it takes to capture moments.
  • Conventional camera solutions require too much preparation or pre-preparation to take a shot or setup a shot.
  • conventional solutions rely on mounts, optical and traditional viewfinders, and shutter buttons.
  • Conventional camera solutions include traditional digital pocket cameras or digital single-lens reflex cameras (DSLRs), mounted cameras, smartphones, and wearable computers.
  • DSLRs digital single-lens reflex cameras
  • the system comprises a wristband device that includes at least one sensor and a camera coupled to the wristband device.
  • the camera is controlled by at least one gesture determined using the at least one sensor.
  • the method comprises providing a wristband device that includes at least one sensor, coupling a camera to the wristband device, determining at least one gesture using the at least one sensor, and controlling the camera by using the at least one gesture.
  • FIG. 1 illustrates a system for capturing media in accordance with an embodiment.
  • FIG. 2 illustrates a system for sensor data classification by a wearable camera device in accordance with an embodiment.
  • FIG. 3 illustrates a method for capturing media in accordance with an embodiment.
  • FIG. 4 illustrates a method for capturing media in accordance with another embodiment.
  • FIG. 5 illustrates a user point of view of a wearable camera device in accordance with an embodiment.
  • FIG. 6 illustrates a side view of the wearable camera device in accordance with an embodiment.
  • FIG. 7 illustrates a subject point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 8 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 9 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 10 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 11 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 12 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 13 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 14 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 15 illustrates a user point of view of a multi-wearable camera device system in accordance with an embodiment.
  • FIG. 16 illustrates a user interface display of a wearable camera device in accordance with an embodiment.
  • FIG. 17 illustrates a method for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment.
  • FIG. 18 illustrates a method for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment.
  • FIG. 19 illustrates a subject point of view of a wearable camera device in accordance with an embodiment.
  • the present invention relates to wearable devices, and more particularly, to a wearable camera system.
  • the following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements.
  • Various modifications to the preferred embodiment and the generic principles and features described herein will be readily apparent to those skilled in the art.
  • the present invention is not intended to be limited to the embodiments shown but is to be accorded the widest scope consistent with the principles and features described herein.
  • Traditional camera solutions include digital pocket cameras (e.g., DSLRs), mounted cameras, smartphones, and camera eyeglasses.
  • Traditional camera solutions typically utilize one of three methods for framing a shot.
  • a first method utilizes optical viewfinders which suffer from the drawbacks including but not limited to requiring a user's eye to be in close proximity to an optical window which takes time and is inconvenient.
  • a second method utilizes digital viewfinders which suffer from the drawbacks including but not limited to requiring a digital screen, having a screen size and weight that limits how portable any solution can be, requiring software to display and update the screen which can be slow, requiring power and extra hardware for the screen, and providing an unnatural interface that requires the user to look at a digital representation of what they can see with their own eyes in front of them.
  • a third method utilizes no viewfinder so the user cannot frame a shot before taking it.
  • a user needs to retrieve a camera from storage, look through a viewfinder or struggle to frame a shot, and find and know how to operate the shutter button/mechanism.
  • the camera is typically sized and weighted to be held in the user's hand.
  • a system and method in accordance with the present invention addresses the aforementioned issues by providing a wearable camera system (wearable camera device) that is gesture operated and mounted/worn on the wrist of the user.
  • the gesture operated wearable camera device eliminates size, storage, retrieval, and operating knowledge constraints associated with traditional digital cameras.
  • the user of the wearable camera device does not need to retrieve any device or place it into storage (because the device is already worn around the user's wrist), does not need to find and know how to operate the shutter button, and does not need to setup any device by holding or using a mount.
  • the gesture operated wearable camera device eliminates constraints associated with traditional mounted cameras.
  • the user of the wearable camera device does not need to use a mount that requires additional time and effort and a receptacle and has a limited range of movement, does not need to predetermine the scope of the framing or use pre-mounting techniques, and does not need to rely on digital or optical viewfinders for framing.
  • the gesture operated wearable camera device eliminates constraints and the limited range of motion associated with traditional smartphone cameras.
  • the user of the gesture operated camera system does not need to access or unlock (e.g., using a pattern or passcode) the device before capturing an image, does not need to utilize a specific photo application to capture the image, and does not need to use a device that that is not optimized for photo taking.
  • the gesture operated wearable camera device eliminates constraints associated with traditional camera eyeglasses.
  • the user of the gesture operated wearable camera device does not need to utilize a lens fixed to the user's eye or any type of special eyewear, does not obscure the field of view with various actions, and does not include unintuitive camera triggers such as voice commands and winks.
  • the system and method in accordance with the present invention provides a wearable camera device that utilizes gesture recognition to allow for the user to take high quality photos/video/media quickly and efficiently.
  • the photos and/or video are taken by easily framing the photo at different angles by simply using the user's hand, without the need for a viewfinder, or any contact with hardware in the palm or fingers of the user's hand, or the use of a second/free hand to trigger a shutter button.
  • the user of the wearable camera device raises an empty hand, gestures with the same hand as if taking a photo with a traditional camera, and captures a photo (or video) within a known framed area.
  • the wearable camera device has the advantage of reducing the amount of time required to prepare, frame, and capture a media (photo/video).
  • the wearable camera device comprises a wristband device (band component) and a camera that is either coupled to the wristband device (so that the user could potential upgrade or interchange the type of the camera) or integrated within the wristband device.
  • the wristband device includes a plurality of embedded sensors including but not limited to microelectromechanical systems (MEMS) devices, gyroscopes, accelerometers, pressure sensors, optical sensors, biometric, electromagnetic sensors, and motion sensors that detect user movements and gestures and the wristband also includes a processor device/unit to analyze and classify the detected gestures.
  • MEMS microelectromechanical systems
  • the wearable camera device comprises a glove-like device that is in contact with the user's fingers and a mounted camera coupled to the glove-like device.
  • the wearable camera device does not include a wristband device portion and the sensor and processor and computing components are instead housed in the glove-like device that is worn by the user like a typical glove.
  • the camera is small in size and lightweight.
  • the camera includes a camera lens.
  • the camera lens is any of thumbnail sized, nano sized, the size of the tip of a pen, the size of a traditional camera lens located on a smartphone, and other small sizes.
  • the camera lens is smaller than thumbnail sized.
  • the camera is interchangeable with various types of lens so that the user can upgrade the wearable camera device.
  • the wristband device is interchanged with different colors and patterns and jewelry pieces to enable customizability.
  • the camera with camera lens is coupled to the wristband device and positioned near the wrist in a manner that allows the camera lens to fluidly, automatically, and continuously follow the movement and/or rotation of the user's wrist.
  • the camera lens is flexibly fixed to the wristband device near the outer edge of the user's hand. This allows the camera lens to move in line with the wrist when rotated or when the hand is moved.
  • the camera lens is rigidly fixed to a portion of the wristband device and is actuated electromechanically by electrical signals from the wristband device that are generated in response to wrist or hand movements that are detected by MEMS devices and sensors embedded within the wristband device.
  • the wristband device recognizes arm, wrist, hand, and finger gestures based on a variety of techniques by monitoring, filtering and detecting, classifying, and processing muscle movements, tendon movements, bone movements, wrist shape changes, hand shape shapes, finger shape changes, and/or bioelectromagnetics states, other user movements and changes to provide for sensor signal detection, feature extraction, and gesture recognition functionalities.
  • the wristband device utilizes wrist contour biometrics and a contour mapping mechanism as a primary input for gesture recognition.
  • additional sensors that detect additional inputs and sensor data are utilized by the wristband device to determine the gestures and perform the gesture recognition.
  • firmware running on an embedded system in the wristband device can monitor, filter, feature extract, classify, and interpret recognized gestures, and then transmit the recognized gestures as camera firmware commands to the camera to control the camera firmware, hardware components, and electromechanical functions of the camera.
  • the user's hand that is wearing the wristband device, the user's fingers, and the user's wrist are all oriented in particular orientations.
  • the muscles/tendons contract and/or relax, and the bones and skin of the user also move or change/shift positions.
  • the wristband device or band includes sensor arrays that are spaced apart from each other at predetermined distances and with known pitches so that the sensor arrays completely circumvent the user's wrist.
  • the wristband device includes sensor arrays that do not completely circumvent the user's wrist and instead are focused on a certain location or region. The sensor arrays either measure distance to the surface of the wrist or measure the pressure that the wrist exerts at each sensor position. Analog-to-digital converters (ADCs) then convert sensor signals detected by the sensor arrays to an array of values.
  • ADCs Analog-to-digital converters
  • the array of values that are collected from the user's wrist by the wristband device are a representation of the state of the user's wrist, hand, and fingers at the time point N.
  • the array of values are then filtered by the wristband device using a plurality of filters including but not limited to median filters and infinite impulse (IIR) filters to reduce/eliminate noise and motion artifacts, and extract/retain features that meet certain thresholds within a known tolerance criteria (i.e., distance or pressure measurement+/ ⁇ some acceptable error at each sensor position).
  • a known tolerance criteria i.e., distance or pressure measurement+/ ⁇ some acceptable error at each sensor position.
  • the wristband device includes a sensor data classification unit that includes pre-defined or pre-determined classifiers that contain features sets per each recognized gesture (e.g., particular arrangement of the user's wrist, hand, and/or fingers).
  • pre-defined classifiers and feature sets are pre-loaded in the wristband device's firmware from reduced training sets that are previously collected from sufficiently large and diverse training population data.
  • the pre-defined classifiers and feature sets are collected in real-time from a cloud-based database that stores data related to all of the users that are using the wearable camera device. In another embodiment, the pre-defined classifiers and feature sets are trained based on the current wearer/user of the wearable camera device over a predetermined time period. Training includes but is not limited to machine learning techniques, device feedback, and user feedback.
  • firmware of the wristband device issues a command/callback to either the firmware of the wristband device to perform an action by the wristband device (e.g., dismissing an alert or notification, uploading data to cloud storage, etc.) or to the firmware of the camera to perform a camera-related action by the camera (e.g., taking a photo, controlling camera shutter, changing camera modes, actuating the camera lens, etc.).
  • an action by the wristband device e.g., dismissing an alert or notification, uploading data to cloud storage, etc.
  • firmware of the camera to perform a camera-related action by the camera (e.g., taking a photo, controlling camera shutter, changing camera modes, actuating the camera lens, etc.).
  • the wristband device detects an array of sensor data values associated with user movements at a frequency of K. Based on the detected array of values and subsequent feature extraction and gesture recognition, a sequence of commands at time point N are extracted from the recognized gestures and are classified at events N+1/K, N+1/K*2, N+1/K*3, etc. These events are processed continuously in a loop to provide constant command control to the wearable camera device.
  • Other embodiments may use additional sensors including but not limited to Biometric electroencephalography (EEG) magnetoencephalography (MEG) sensors, and electromyography (EMG) sensors in combination with pressure and optical sensors to reduce noise, false positive features, or misclassification of gestures and commands.
  • EEG Biometric electroencephalography
  • MEG magnetoencephalography
  • EMG electromyography
  • the wristband device monitors movements from the user's arm, wrist, hand, fingers, and thumb.
  • the user's hand with the wearable camera device is raised and positioned to frame the photo/video.
  • a photo/video can be captured by a plurality of finger movements and gestures recognized by the wearable camera device.
  • the index finger and thumb are extended, approximately 90 degrees from each other, effectively creating one corner of a camera frame.
  • varying angles and positions between the user's fingers or the index finger and thumb are possible to create the camera frame while increasing usability and minimizing potential user fatigue.
  • the user can hold up all four fingers and extend the thumb approximately 90 degrees from the four fingers to frame the camera or the user can merely hold up one finger to frame the photo/video.
  • a plurality of user movements and gestures can be associated with a plurality of wearable camera device functions and that would be within the spirit and the scope of the present invention.
  • the position of the shutter plane is always substantially parallel to the plane created by the index finger and the thumb—even when the user's wrist is rotated.
  • the wristband device can detect the extended index finger and thumb by monitoring, processing, classifying, and recognizing muscle, tendon, and wrist movements, wrist contours, hand shapes, and movements using feature extraction and gesture recognition to control the camera.
  • the wristband device transmits an instruction/command/call to the camera to open the shutter and wait in a ready state to take/capture a photo or video.
  • the wristband device recognizes these gestures as a camera trigger gesture by monitoring, classifying, and extracting features from the finger and wrist movements using the plurality of embedded sensors that collects/detects muscle, tendon, bioelectromagnetics, and anatomical contour changes and the processor device that utilizes algorithms (e.g., machine learning classifiers) to analyze and classify the detected movements.
  • algorithms e.g., machine learning classifiers
  • the wristband device triggers the camera to capture a photo or video in response to the detected movements that are classified as a camera trigger gesture.
  • the wristband device detects other gestures for camera-related functions such as capturing photos/videos, changing/selecting operation modes, zooming in/out, etc.
  • the operation modes include but are not limited to adjusting the lens position by rotating the wristband device around the wrist.
  • the wristband device has a light emitting user interface (UI) that is sized to be displayed on the user's wrist and that can be manipulated either by the detected user movements/gestures or by direct user touch gestures on a display screen of the UI.
  • the UI can be either smaller than traditional video displays, similarly sized, or larger.
  • the UI can be an LCD display, LED display, or another type of display unit.
  • the wearable camera device or the wristband device itself includes a communication device (e.g., WiFi or Bluetooth receiver and transmitter) that includes any of WiFi and Bluetooth communication capabilities so that the sensor movement data that is detected and analyzed as well as the captured photos and videos can be wirelessly communicated to another device (e.g., smartphone, laptop, etc.) or to a cloud-computing storage system by the communication device.
  • the captured photos and videos can be automatically transmitted (or transmitted according to a predetermined schedule) to the device or cloud-computing storage system so that the user's information is seamlessly backed up.
  • the wristband device can be used in conjunction with a second wristband device on the user's opposing hand.
  • the first and the second wristband devices can communicate different modes, operations, and data to each other and also work in conjunction as a multi-camera system.
  • the multi-camera system can include additional gesture recognitions and features to enable advanced photo and video taking including but not limited to panorama photos/videos and 3D photos/videos.
  • the system and method in accordance with the present invention provide a gestured operated wrist-mounted camera system (wearable camera device) that is an unobtrusive accessory-type wristband camera that can be worn by the user for 24 hours a day so that the user can capture moments at any time or place. Therefore, the wearable camera device is convenient and quickly produces high quality and framed camera shots.
  • the need for a physical viewfinder (either optical or digital), a shutter button, and command and control buttons is eliminated. Readying, framing, sizing, and shutter operations do not require physical buttons.
  • the user can also personalize the wearable camera device by selecting a plurality of different styles, colors, designs, and patterns associated with the wristband.
  • the form factor of the wearable camera device provides camera hardware that is integrated into the wristband device so there is no need for costly hardware (physical or digital screen) or a viewfinder which reduces the footprint (size, power, complexity).
  • the wristband device is form-fitting and adjustable and can include a small adjustable positioned camera and lens.
  • the wristband device includes a display unit that can serve as a viewfinder for added functionality.
  • the camera is coupled to the wristband device at the junction of the pisiform and ulna. In another embodiment, the camera is coupled to the wristband device at or near the edge of an abductor digit minima. These positions allow for the plane of the field of view created naturally by the index finger and thumb (or other user's hand/finger orientations) to be parallel with the camera lens even when the wrist is maneuvered and rotated at any angle (eliminating any potential obstructions). In another embodiment, the position of the camera lens is electromechanically moveable to align with the movement of the hand and wrist.
  • the wearable camera device is able to monitor, recognize and interpret arm, wrist, hand, and finger gestures, movements, and changes by monitoring, detecting, processing, and classifying muscle movements, tendon movements, hand shapes, and wrist contours using feature extractions and gesture recognition.
  • the gestures that the wearable camera device can detect to initialize, ready the camera, and take photos and/or videos include but are not limited to a semicircle shape, circle shape, OK action, hang loose action, swipe left across index finger, swipe right across index finger, tap or double tap between fingers, tap or double tap across index finger, 90 degree framing, corner box framing, double view framing, framing, switching modes of operation (camera to video, etc.), traversing modes and options, selecting options, triggering a shutter, starting capture, motion capture, encoding, and zooming in/out.
  • the wristband device can include features outside of camera functionality, processing, and wireless communication including but not limited to a clock, timer, and Internet/email capabilities.
  • the wristband device includes an adjustable or rotatable (around the wrist) camera lens position and multiple lenses on the wristband (e.g., front, rear, facing, etc).
  • the wristband device includes perforations along the outside of the band that can emit light (e.g., light-emitting diodes or LEOs) to notify the user with different patterns, shapes, and designs that act as an interface display to provide feedback information to the user (e.g., photo taken, photo uploaded to cloud storage, error in taking photo, etc.).
  • the LED display and patterns can be programmable by the user or predetermined.
  • the wristband device is touch enabled to allow for touch gestures to be recognized directly on the wristband device (or the display unit of the wristband device) in addition to muscle, tendon, and bioelectromagnetics recognition.
  • the wristband device includes the ability to tap, drag, and flick (away) the emitted light based notifications and objects around the outer portion of the wristband. This allows the user to manipulate and adjust the position of the emitted display, to control the notifications once they are no longer relevant to the user, and to respond to certain notifications with various input, touch, and gesture responses.
  • the wristband device includes the ability to manipulate modes of operation, options, and the interface display by performing a gesture with only one hand (palm free).
  • the wristband device includes at least one accelerometer to aid in framing the shots and with orientation.
  • the wristband device includes a vibration sensor that can vibrate as part of the interface display that provides feedback to the user (in addition to the LED notifications).
  • the wristband device includes a speaker with audio output as part of the interface display so that additional feedback (in audio format) can be provided.
  • the wristband device includes a plurality of sensors that are embedded within and connected via circuitry to detect various data from the user and the user's surrounding environment.
  • the plurality of sensors can include but are not limited to any of or any combination of MEMS devices, gyroscopes, accelerometers, torque sensors, weight sensors, pressure sensors, magnetometers, temperature sensors, light sensors, cameras, microphones, GPS, wireless detection sensors, altitude sensors, blood pressure sensors, heart rate sensors, biometric sensors, radio frequently identification (RFID), near field communication (NFC), mobile communication, Wi-Fi, strain gauges, fingerprint sensors, smell sensors, gas sensors, chemical sensors, color sensors, sound sensors, acoustic sensors, ultraviolet sensors, electric field sensors, magnetic field sensors, gravity sensors, wind speed sensors, wind direction sensors, compass sensors, geo-locator sensors, polarized light sensors, infrared emitter sensors, and photo-reflective sensors.
  • MEMS devices gyroscopes
  • accelerometers torque sensors
  • weight sensors pressure sensors
  • magnetometers temperature sensors
  • the wristband device includes a processor device that analyzes the detected sensor data (from the plurality of sensors) using a sensor data classification unit that utilizes a plurality of algorithmic processes.
  • the plurality of algorithmic processes can include but is not limited to any of or any combination of back propagation, bayes networks, machine learning, deep learning, neural networks, fuzzy mean max neural networks, hidden Markov chains, hierarchical temporal memory, k nearest neighbor (KNN), adaboot, and histogram analysis.
  • the user utilizes a plurality of wristband devices worn on both hands.
  • wireless communication and synchronization between the plurality of wristband devices can provide multi-camera functionality including but not limited to any of new framing options, larger framing areas, 30 capability, and 360 degree capability.
  • the wearable camera device includes a capability of connecting to various networks (public or private).
  • FIG. 1 illustrates a system 100 for capturing media in accordance with an embodiment.
  • the media can include photos, video, and/or other types of media.
  • the system 100 is a wearable camera device that includes a wristband device 150 and a camera 134 coupled to the wristband device 150 .
  • the camera 134 comprises a lens, a barrel, and an actuator.
  • the wristband device 150 includes a plurality of components including but not limited to any of a power supply 102 , a random access memory (RAM) 104 , microcontroller (MCU) 106 , read-only memory (ROM) 108 , a clock 110 , a storage/memory device 112 (including but not limited to Flash memory), a radio 114 (including but not limited to Bluetooth and WiFi), an antenna 116 , a haptic/vibrator sensor 118 , an accelerometer 120 , a gyroscope 122 , an analog-to-digital converter (ADC) 124 , pre-amplifier/filter/amplifier/bias device 126 , a plurality of external sensors 128 (e.g., EKG, EEG, MEG, EMG, pressure, optical, etc.), an in-system programming component (ISP) 130 , and an internal sensor 132 .
  • the embedded accelerometer 120 and gyroscope 122 are included in
  • the plurality of external sensors 128 detect user movements (e.g., muscle or tendon movements) and the plurality of components of the wristband device 150 determine a plurality of gestures by monitoring, classifying, and extracting features from the detected user movements.
  • the plurality of gestures control various actions that are executed by either the wristband device 150 itself or the camera 134 including but not limited to taking pictures or videos, scrolling through various options, and scrolling through various modes.
  • FIG. 2 illustrates a system 200 for sensor data classification by a wearable camera device in accordance with an embodiment.
  • the system 200 includes an application component 202 , an operating system (OS) component 204 , and a plurality of drivers 206 .
  • the application component 202 includes inputs/outputs (I/O), a digitizer/classifier, a gesture detection component, a sensor data classification unit, and a commands component.
  • the OS 204 includes a scheduling/memory management component and a message passing component.
  • the each driver of a plurality of drivers 206 may include communication drivers, sensor drivers, general purpose I/O (GPIO) drivers, and file system (FS) drivers.
  • the system 200 detects user movements using the sensors of the wristband device and processes the detected user movements using the application component 202 and the sensor data classification unit to determine the gestures and associated commands that control the camera.
  • GPIO general purpose I/O
  • FS file system
  • FIG. 3 illustrates a method 300 for capturing media in accordance with an embodiment.
  • the media can include photos, video, and/or other types of media.
  • the method 300 comprises providing a wristband device that includes at least one sensor, via step 302 , coupling a camera to the wristband device, via step 304 , determining at least one gesture using the at least one sensor, via step 306 , and controlling the camera by using the at least one gesture, via step 308 .
  • the method further includes detecting, by the at least one sensor, user movements by using any of muscle, tendon, bioelectromagnetics, and anatomical contour changes (using the contour mapping mechanism) of a user.
  • a processor of the wristband device then analyzes the user movements using the sensor data classification unit to extract features from the user movements and to determine at least one gesture from the extracted features using various classifiers. For example, if the detected user movement is determined to be the at least one gesture of a user extending an index finger and thumb approximately 90 degrees from each other, the wristband device will instruct/control the camera to open a shutter and await the photo/video capture.
  • a communication device is coupled to both the wristband device and the camera, wherein the communication device transmits data from both the wristband device (e.g., gesture classifications) and the camera (e.g., photos and videos) to another device that comprises any of a smartphone, a laptop, a desktop, and a cloud-based server system.
  • FIG. 4 illustrates a method 400 for capturing media in accordance with another embodiment.
  • the media can include photos, video, and/or other types of media.
  • the method 400 comprises movements of a user using the wearable camera device occurring, via step 402 , voltage sensors detecting the user movements, via step 404 , the voltage sensors recording the detected user movements as a signal, via step 406 , the signal passing through a pre-amplifier, via step 408 , the pre-amplified signal passing through a plurality of filters to remove noise and additional motion artifacts, via step 410 , the filtered signal passing through an amplifier, via step 412 , and the amplified signal passing through an analog-to-digital converter (ADC), via step 414 .
  • ADC analog-to-digital converter
  • the filtered, amplified, and converted signal has features extracted from it, via step 416 , and these feature extractions result in gesture determinations and associated instructions, callbacks, and notifications, via step 418 .
  • the callbacks and notifications are registered by the wearable camera device, via step 420 , and the instructions or callbacks are transmitted from the wristband device to the camera for execution, via step 422 , which sets up certain camera controls and functions (e.g., taking a photo/video), via step 424 .
  • the controls and functions are encoded, via step 426 , which results in the triggering of the associated camera function (e.g., data capture), via step 428 , and the wearable camera device then transmits the captured media, via step 430 .
  • FIG. 5 illustrates a user point of view of a wearable camera device 500 in accordance with an embodiment.
  • the wearable camera device 500 includes a wristband device 502 that is coupled to a camera (or the camera is embedded within the wristband device 502 ).
  • the wristband device 502 also includes other hardware components and a muscle, tendon, finger gesture recognizer or sensor detection device.
  • the wearable camera device 500 is shown from the user's point of view when framing a photo/video and includes a frame of view 504 that serves as the user's viewfinder (in place of traditional digital or optical viewfinders).
  • the wearable camera device 500 does not include a traditional optical or digital viewfinder which allows the user greater flexibility in taking nature photos and videos by using the frame of view 504 .
  • a viewfinder is displayed on a display unit/screen (e.g., LCD/LED screen) of the user interface (UI) of the wristband device.
  • UI user interface
  • the user can verify the correct frame has been captured by checking the display unit of the wristband device that displays the signal from the camera lens.
  • the user's index finger and thumb are extended approximately 90 degrees from each other to frame the subject.
  • varying angles and positions between the user's fingers or the index finger and thumb are possible to create the camera frame while increasing usability and minimizing potential user fatigue.
  • the wristband's internal sensors detect the user movements and shapes (orientation of the index finger relative to the thumb) as a muscle movement.
  • the detected muscle movement is determined to be a certain gesture that readies the camera focus and shutter.
  • the wearable camera device 500 then awaits another detected user movement that is determined to be a certain gesture that will trigger another camera action (such as taking the photo/video).
  • the user wearing the wearable camera device faces his/her palm outwards and towards the subject with all four fingers raised (and possibly extending the thumb approximately 90 degrees from the four fingers or resting the thumb up against the index finger).
  • the wristband device would detect user movements, determine a gesture from the detected user movements, and then transmit a command/instruction to the camera based upon the determined gesture to carry out a camera action including but not limited to triggering a camera shutter action.
  • the user wears the wearable camera device (or specifically the wristband device portion) at a position that is rotated 180 degrees or the opposite of the normal wearing position.
  • the camera lens is facing the user thereby allowing the user to take “selfie” style captures using similar user movements and gestures (e.g., lowering one of the fingers that are raised, etc.).
  • FIG. 6 illustrates a side view of the wearable camera device 600 in accordance with an embodiment.
  • the wearable camera device 600 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 602 and a camera coupled to the wristband device 602 .
  • the camera includes a camera lens 604 focused on an object 608 .
  • the frame of view 606 is focused on the object 608 .
  • the user's index finger and thumb are in the same position (extended approximately 90 degrees from each other) and so the camera is once again in a ready position for when another user movement and associated gestured is determined.
  • FIG. 7 illustrates a subject point of view of the wearable camera device 700 in accordance with an embodiment.
  • the wearable camera device 700 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 702 and a camera coupled to the wristband device 702 .
  • the camera includes a camera lens 704 .
  • the user's index finger and thumb are in the same position (extended approximately 90 degrees from each other) and so the camera is once again in a ready position for when another user movement and associated gestured is determined.
  • FIG. 8 illustrates a user point of view of the wearable camera device 800 in accordance with an embodiment.
  • the wearable camera device 800 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 802 and a camera coupled to or integrated within the wristband device 802 .
  • the user's index finger and thumb have moved from the position of being extended approximately 90 degrees from each other to a button pressing simulation movement 806 .
  • the button pressing simulation movement 806 is when the user's index finger is slightly lowered.
  • a different type of user movement can be associated with the button pressing simulation movement 806 .
  • the wristband device 502 's sensors detect the user movement and the wristband device 502 's internal components and sensor data classification unit determine that a specific gesture associated with the detected user movement has occurred.
  • the determined gesture prompts the camera to take a picture/photo of an object 808 in the distance that is framed by a frame of view 804 .
  • the wristband device of the wearable camera device detects a plurality of sensor signals and determines a plurality of gestures using the plurality of sensor signals.
  • the plurality of sensor signals can represent either a single gesture (e.g., “clicking” motion of one finger, etc.) or a set of gestures in a specific sequence (e.g., sliding a finger to the left or right and tapping, etc.) that the firmware of the wristband device constitutes as a single action.
  • the firmware of the wristband device sends out an associated command to control the camera.
  • FIGS. 9-11 represent additional sets of gestures that the wristband device can determine as a single action.
  • FIG. 9 illustrates a user point of view of the wearable camera device 900 in accordance with an embodiment.
  • the wearable camera device 900 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 902 and a camera coupled to or integrated within the wristband device 902 .
  • the wristband device 902 has determined a gesture 904 that comprises the user tapping or double tapping the thumb.
  • the gesture 904 triggers the wearable camera device 900 to display the target/subject of the photo/video on the wristband device 902 .
  • the wristband device 902 includes a user interface display that can display the subject that is being photographed by the camera to ensure that the user has focused the camera correctly upon the target/subject that the user wants to photograph/video.
  • FIG. 10 illustrates a user point of view of the wearable camera device 1000 in accordance with an embodiment.
  • the wearable camera device 1000 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1002 and a camera coupled to or integrated within the wristband device 1002 .
  • the wristband device 1002 has determined a gesture 1004 that comprises the user swiping left with the thumb.
  • the gesture 1004 triggers the wearable camera device 1000 to change various modes of the camera and the mode change is displayed on the user interface of the wristband device 1002 .
  • the wristband device 1002 includes a user interface display that can display the varying modes.
  • the modes are varied using alternating LED patterns.
  • the modes are varied and the text (e.g., “Photo Taking”, “Video Taking”, “Night-Time Photo”, etc.) is displayed on the user interface display.
  • FIG. 11 illustrates a user point of view of the wearable camera device 1100 in accordance with an embodiment.
  • the wearable camera device 1100 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1102 and a camera coupled to or integrated within the wristband device 1102 .
  • the wristband device 1102 has determined a gesture 1104 that comprises the user swiping right with the thumb.
  • the gesture 1104 triggers the wearable camera device to also change various modes of the camera (in the opposite direction as swiping left so essentially the user can scroll left and right through various options) and the mode change is displayed on the user interface of the wristband device 1102 .
  • the wristband device 1102 includes a user interface display that can display the varying modes.
  • the modes are varied using alternating LED patterns.
  • the modes are varied and the text (e.g., “Photo Taking”, “Video Taking”, “Night-Time Photo”, etc.) is displayed on the user interface display.
  • FIG. 12 illustrates a user point of view of the wearable camera device 1200 in accordance with an embodiment.
  • the wearable camera device 1200 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1202 and a camera coupled to or integrated within the wristband device 1202 .
  • the wristband device 1202 has determined a gesture 1204 that comprises the user holding his/her fingers in a “C” shape.
  • the gesture 1204 triggers the wearable camera device 1200 to select the mode that the user has scrolled to (by swiping left or right).
  • the user can input various other hand gestures and shapes that are associated with mode selection.
  • FIG. 13 illustrates a user point of view of the wearable camera device 1300 in accordance with an embodiment.
  • the wearable camera device 1300 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1302 and a camera coupled to or integrated within the wristband device 1302 .
  • the wristband device 1302 has determined a gesture 1304 that comprises the user's fingers snapping.
  • the gesture 1304 triggers an “OK” action to the wearable camera device 1300 .
  • the “OK” action represents the user setting or acknowledging a particular setting or command that may be displayed to the user as feedback on the device interface output (e.g., LEOs, user interface display, etc.).
  • the device interface output e.g., LEOs, user interface display, etc.
  • FIG. 14 illustrates a user point of view of the wearable camera device 1400 in accordance with an embodiment.
  • the wearable camera device 1400 has components similar to the wearable camera device of FIG. 5 including a wristband device 1402 and a camera coupled to or integrated within the wristband device 1402 .
  • the wristband device 1402 has determined a gesture 1404 that comprises the user tapping or double tapping the index finger and the middle finger both to the thumb.
  • the gesture 1404 triggers an “OK” action to the wearable camera device 1400 .
  • the “OK” action represents the user setting or acknowledging a particular setting or command that may be displayed to the user as feedback on the device interface output (e.g., LEOs, user interface display, etc.).
  • the device interface output e.g., LEOs, user interface display, etc.
  • FIG. 15 illustrates a user point of view of a multi-wearable camera device system 1500 in accordance with an embodiment.
  • the multi-wearable camera device system 1500 includes a first wearable camera device 1502 coupled to a second wearable camera device 1504 .
  • the first and the second wearable camera devices 1502 - 1504 enhance the user's ability to frame a target/subject 1506 .
  • the first and the second wearable camera devices 1502 - 1504 communicate with each other and enable the user to select from various dual wristband modes.
  • the multi-wearable camera device system 1500 syncs and stitches together image and video capture thereby creating panorama or wide angle capture.
  • the multi-wearable camera device system 1500 provides syncing and capturing of 30 image or video captures, 30 stereoscope vision, and 30 zooming.
  • FIG. 16 illustrates a user interface display of a wearable camera device 1600 in accordance with an embodiment.
  • the wearable camera device 1600 includes a user interface display 1602 .
  • the user interface display 1602 comprises LEOs that display various patterns such as the pattern 1604 based upon camera settings and notifications.
  • the LEOs are integrated via laser cut holes cut into the user interface display 1602 .
  • the user can use hand gestures (waiving across the user interface display 1602 ) or touch gestures (pressing on the user interface display 1602 ) to respond to the various notifications or delete them.
  • hand gestures waiving across the user interface display 1602
  • touch gestures pressing on the user interface display 1602
  • the wearable camera device can include varying types of display units and user interfaces and that would be within the spirit and scope of the present invention.
  • FIG. 17 illustrates a method 1700 for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment.
  • the media can include photos, video, and/or other types of media.
  • the method 1700 represents two time points (N and N+1).
  • the user of the wearable camera device starts in position A (with the index finger and thumb approximately 90 degrees from each other) at time point N and ends in position B (with the index finger slightly lowered in a “clicking” motion) at time point N+1.
  • the wearable camera device detects the user's hand positioning and movements using a plurality of embedded sensors and determines various gestures from the detected positioning/movements, via step 1702 .
  • the wearable camera device utilizes a contour mapping mechanism that provides a contour map of the anatomical contours of the user's wrist that is wearing the wristband device of the wearable camera device, via step 1704 .
  • the wristband device (band and embedded sensors) position around the user's wrist is denoted by a solid line and the contour of the user's wrist is denoted by a dotted line.
  • the contour map of the user's wrist is in wrist contour shape one (1) and when the user's hand shifts in position from position A to position B at time point N+1, the contour map of the user's wrist is in wrist contour shape two (2).
  • the wristband device classifies the contour changes using feature extraction and associated classifiers from contour shapes, via step 1706 .
  • the key features of the classifier are plotted on a graph that depicts a distance in millimeters (mm) on the y-axis and that depicts the sensor position around the circumference of the wrist and based upon the contour map on the x-axis.
  • the plotted graph displays the wrist contour shape 1 in a first sensor position and when the user's hand is in position B at time point N+1, the plotted graph displays the wrist contour shape 2 in a second sensor position.
  • the wearable camera device After classifying the sensor positions to determine the gesture, the wearable camera device the determined gesture is associated with a certain command that is then transmitted to the camera, via step 1708 .
  • the wearable camera device determines the first sensor position (that is associated with a first gesture)
  • the camera receives a “camera ready” command (because that command is associated with the first gesture)
  • the wearable camera device determines the second sensor position after the user has changed orientation of his/her hand which is associated with a second gesture
  • the camera receives a “shutter press” command (because that command is associated with the second gesture) and the photo/video is captured by the wearable camera device.
  • a variety of contour shapes and sensor positions can be associated with a variety of gestures and subsequently with a variety of commands and that would be within the spirit and scope of the present invention.
  • FIG. 18 illustrates a method 1800 for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment.
  • the media can include photos, video, and/or other types of media.
  • the wearable camera device monitors the hand of a user wearing the wearable camera device (that includes a wristband device and a camera coupled to or integrated within the wristband device), via step 1802 .
  • the user movements cause muscle, skin, tendons, and bone to move as well allowing the wearable camera device to detect the user movements as sensor values, via step 1804 , by using embedded sensors within the wristband device of the wearable camera device at a time point N, via step 1806 .
  • the detected sensor values are converted into a signal using an ADC and stored as an array of values (AV), via step 1808 .
  • filters e.g., median filter
  • Steps 1802 - 1810 represent the “monitoring” phase of the method 1800 utilized by the wearable camera device.
  • the monitoring phase is repeated by the wearable camera device to create enough training data for certain gestures (G, G1, etc.), via step 1812 .
  • Once enough training data is created (or pre-downloaded into the wearable camera device) common features are identified for each gesture (G, G1, etc.) that are each associated with certain gesture classifiers (C, C1, etc.) that are each associated with certain actions (A, A1, etc.), via step 1814 .
  • the wearable camera device receives the determined gestures and creates a gesture classifier for each, via step 1816 .
  • Steps 1812 - 1816 represent the “classification” phase of the method 1800 utilized by the wearable camera device.
  • the wearable camera device After the classification phase has been completed by the wearable camera device and enough training data has been created, via step 1816 , the wearable camera device compares the filtered AV values to the gesture classifiers (created via step 1816 ) using matching methodologies including but not limited to machine learning and k-nearest neighbor algorithms (KNN), via step 1818 . If the features from the filtered AV values match a certain gesture classifier, the wearable camera device calls the action (transmits the command to the camera) associated with that certain gesture classifier, via step 1820 .
  • KNN machine learning and k-nearest neighbor algorithms
  • Steps 1818 - 1820 represent the “recognition” phase of the method 1800 utilized by the wearable camera device.
  • the action command or call (e.g., camera shutter) is received by the camera, via step 1822 , and the camera firmware initiates camera encoding of the camera controls and functions, via step 1824 .
  • the action command or call is encoded, via step 1826 , and the action is executed (e.g., the camera captures the image), via step 1828 .
  • the wearable camera device transmits the media to a local or remote storage, via step 1830 , and a user can view the media, via step 1832 .
  • the wearable camera device can associated a plurality of sensor values with gestures, classifiers, and actions and that would be within the spirit and scope of the present invention.
  • FIG. 19 illustrates a subject point of view of a wearable camera device 1900 in accordance with an embodiment.
  • the wearable camera device 1900 includes a wristband device 1902 (or band) that wraps around the wrist of the user and a camera 1904 that is coupled to or integrated within the wristband device 1902 .
  • the camera 1904 includes a camera lens 1906 and a camera sub assembly.
  • the camera lens 1906 is mechanically and flexibly affixed to the wristband device 1902 near the bottom of the user's palm.
  • the flex fixing camera lens 1906 mechanically moves with the movements/gestures of the user's wrist and palm.
  • the camera lens 1906 is rigidly fixed to a portion of the wristband device 1902 that is controllable by electromechanical actuation based on movements/gestures.
  • the subject point of view represents the view from a subject that is looking straight at the camera lens 1906 .
  • the subject point of view shows the palm and various fingers (middle, ring, pinky) of the user.
  • the wristband device 1902 includes a plurality of components including but not limited to sensors, display unit, hardware platform, battery, storage, and radio. In one embodiment, the wristband device 1902 rotates around the user's wrist enabling the camera lens 1906 to either face outward towards the subject or face inward towards the user for “selfie” picture capturing capability.
  • a system and method in accordance with the present invention discloses a wearable camera system (wearable camera device) for capturing media (photo/video).
  • the wearable camera device comprises a wristband device that includes at least one sensor.
  • the wearable camera device also comprises a camera coupled to the wristband device.
  • the camera is integrated within the wristband device as one overall device. The camera is controlled by at least one gesture determined from user movements and sensor data detected by the at least one sensor.
  • the wristband device includes additional hardware components including but not limited to a processor that analyzes the user movements detected by the at least one sensor and a memory device that stores the various data (detected user movements, determined gestures, etc.).
  • the at least one sensor detects the user movements by detecting any of muscle, tendon, bioelectromagnetics, and anatomical contour changes of the user.
  • the processor analyzes the detected user movements using a sensor data classification unit that utilizes filters and algorithms that extract features from the user movements to determine the at least one gesture.
  • the sensor data classification unit utilizes any of back propagation, Bayes networks, neural networks, and machine learning to determine the user gestures.
  • the at least one sensor is any of a gyroscope, an accelerometer, a pressure sensor, a temperature sensor, and a light sensor.
  • the camera is controlled without using any of an optical viewfinder or a digital viewfinder and instead uses a natural viewfinder created by the user's natural fingers.
  • the camera includes a lens that is small (e.g., thumbnail sized) and the lens is positioned to follow the rotational movement of the user. This ensures that the lens is never obstructed and is positioned on the user's wrist in a way that enables clear targeting of the photo/video/media subjects that the user wants to capture.
  • the wearable camera device can determine a plurality of gestures that trigger various camera actions.
  • the gesture includes a user extending an index finger and thumb approximately 90 degrees from each other to create one corner of a camera frame.
  • the camera shutter's plan is substantially parallel to a plane created by the extending of the index finger and thumb.
  • the user can determine various hand gestures and actions and correlate each of these hand gestures and actions to specific camera actions.
  • the wearable camera device determines additional gestures and associated camera actions.
  • the wristband device of the wearable camera device instructs/controls/transmits a message to the camera to open the shutter and await a photo/video/media capture gesture from the user.
  • the user provides a photo/video/media capture gesture by simulating a button pressure.
  • the user provides a photo/video/media capture gesture by another gesture that is either predetermined or inputted by the user such as snapping of the fingers.
  • the at least one gesture or plurality of gestures that the wearable camera device can determine includes any of initializing camera function, semicircle shapes, circle shapes, OK action, hang loose action, swipe left across index finger, swipe right across index finger, tap between fingers, double tap between fingers, witching modes, selecting options, zooming, triggering a shutter to start capture, and triggering a shutter to start motion capture.
  • the wearable camera device includes a communication device (e.g., transmitter/receiver device) coupled to both the wristband device and the camera (or just one of either the wristband device and the camera).
  • the communication device transmits data from either or both the wristband device (e.g., user movements, gestures, etc.) and the camera (e.g., photos, videos, media, etc.) to another device that comprises any of a smartphone and a cloud-based server system.
  • the wristband device includes a user interface display that comprises a plurality of light emitting diodes (LEDs) that produces various patterns associated with camera actions and notifications and alerts for the user.
  • LEDs light emitting diodes
  • a system and method in accordance with the present invention utilizes a wristband device that includes a plurality of sensors and a processor and a camera mounted to the wristband device to provide a gesture operated wrist-mounted camera system (wearable camera device).
  • the wearable camera device is an unobtrusive accessory-type wristband camera that can be worn by the user for 24 hours a day so that the user can take photos and videos at any time and with ease.
  • the wearable camera device is convenient (in accessibility, size, shape, weight, etc) and quickly produces high quality and framed camera shots and videos once user movements and gestures associated with various commands are detected by the wearable camera device.
  • Embodiments described herein can take the form of an entirely hardware implementation, an entirely software implementation, or an implementation containing both hardware and software elements.
  • Embodiments may be implemented in software, which includes, but is not limited to, application software, firmware, resident software, microcode, etc.
  • the steps described herein may be implemented using any suitable controller or processor, and software application, which may be stored on any suitable storage location or computer-readable medium.
  • the software application provides instructions that enable the processor to perform the functions described herein.
  • embodiments may take the form of a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system.
  • a computer-usable or computer-readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
  • the medium may be an electronic, magnetic, optical, electromagnetic, infrared, semiconductor system (or apparatus or device), or a propagation medium (non-transitory).
  • Examples of a computer-readable medium include a semiconductor or solid state memory, magnetic tape, a removable computer diskette a random access memory (RAM), a read-only memory (ROM), a rigid magnetic disk, and an optical disk.
  • Current examples of optical disks include DVD, compact disk-read-only memory (CD-ROM), and compact disk-read/write (CD-R/W).

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • User Interface Of Digital Computer (AREA)
  • Studio Devices (AREA)

Abstract

A system and method for capturing media are disclosed. In a first aspect, the system comprises a wristband device that includes at least one sensor and a camera coupled to the wristband device. The camera is controlled by at least one gesture determined using the at least one sensor. In a second aspect, the method comprises providing a wristband device that includes at least one sensor, coupling a camera to the wristband device, determining at least one gesture using the at least one sensor, and controlling the camera by using the at least one gesture.

Description

    CROSS REFERENCE
  • This application is a continuation of U.S. patent application Ser. No. 14/634,674, filed on Feb. 27, 2015, entitled GESTURE OPERATED WRIST MOUNT CAMERA SYSTEM which claims the benefit of U.S. Provisional Patent Application No. 61/946,622, filed on Feb. 28, 2014, entitled “GESTURE OPERATED WRIST MOUNTED CAMERA SYSTEM,” and U.S. Provisional Patent Application No. 62/101,289, filed on Jan. 8, 2015, entitled “GESTURE OPERATED WRIST MOUNTED CAMERA SYSTEM,” which are incorporated herein by reference in their entirety.
  • FIELD OF THE INVENTION
  • The present invention relates to wearable devices, and more particularly, to a wearable camera system.
  • BACKGROUND
  • Timing and framing are key factors in photography and videography that capture moments. People increasingly want to reduce the time it takes to capture moments. Conventional camera solutions require too much preparation or pre-preparation to take a shot or setup a shot. In addition, conventional solutions rely on mounts, optical and traditional viewfinders, and shutter buttons. Conventional camera solutions include traditional digital pocket cameras or digital single-lens reflex cameras (DSLRs), mounted cameras, smartphones, and wearable computers.
  • There is a strong need for solutions that effectively and efficiently reduce the time it takes to capture moments through photography and videography. The present invention addresses such a need.
  • SUMMARY OF THE INVENTION
  • A system and method for capturing media are disclosed. In a first aspect, the system comprises a wristband device that includes at least one sensor and a camera coupled to the wristband device. The camera is controlled by at least one gesture determined using the at least one sensor.
  • In a second aspect, the method comprises providing a wristband device that includes at least one sensor, coupling a camera to the wristband device, determining at least one gesture using the at least one sensor, and controlling the camera by using the at least one gesture.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying figures illustrate several embodiments of the invention and, together with the description, serve to explain the principles of the invention. One of ordinary skill in the art readily recognizes that the embodiments illustrated in the figures are merely exemplary, and are not intended to limit the scope of the present invention.
  • FIG. 1 illustrates a system for capturing media in accordance with an embodiment.
  • FIG. 2 illustrates a system for sensor data classification by a wearable camera device in accordance with an embodiment.
  • FIG. 3 illustrates a method for capturing media in accordance with an embodiment.
  • FIG. 4 illustrates a method for capturing media in accordance with another embodiment.
  • FIG. 5 illustrates a user point of view of a wearable camera device in accordance with an embodiment.
  • FIG. 6 illustrates a side view of the wearable camera device in accordance with an embodiment.
  • FIG. 7 illustrates a subject point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 8 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 9 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 10 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 11 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 12 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 13 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 14 illustrates a user point of view of the wearable camera device in accordance with an embodiment.
  • FIG. 15 illustrates a user point of view of a multi-wearable camera device system in accordance with an embodiment.
  • FIG. 16 illustrates a user interface display of a wearable camera device in accordance with an embodiment.
  • FIG. 17 illustrates a method for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment.
  • FIG. 18 illustrates a method for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment.
  • FIG. 19 illustrates a subject point of view of a wearable camera device in accordance with an embodiment.
  • DETAILED DESCRIPTION
  • The present invention relates to wearable devices, and more particularly, to a wearable camera system. The following description is presented to enable one of ordinary skill in the art to make and use the invention and is provided in the context of a patent application and its requirements. Various modifications to the preferred embodiment and the generic principles and features described herein will be readily apparent to those skilled in the art. Thus, the present invention is not intended to be limited to the embodiments shown but is to be accorded the widest scope consistent with the principles and features described herein.
  • Traditional camera solutions include digital pocket cameras (e.g., DSLRs), mounted cameras, smartphones, and camera eyeglasses. Traditional camera solutions typically utilize one of three methods for framing a shot. A first method utilizes optical viewfinders which suffer from the drawbacks including but not limited to requiring a user's eye to be in close proximity to an optical window which takes time and is inconvenient. A second method utilizes digital viewfinders which suffer from the drawbacks including but not limited to requiring a digital screen, having a screen size and weight that limits how portable any solution can be, requiring software to display and update the screen which can be slow, requiring power and extra hardware for the screen, and providing an unnatural interface that requires the user to look at a digital representation of what they can see with their own eyes in front of them. A third method utilizes no viewfinder so the user cannot frame a shot before taking it.
  • In all of the above methods, a user needs to retrieve a camera from storage, look through a viewfinder or struggle to frame a shot, and find and know how to operate the shutter button/mechanism. The camera is typically sized and weighted to be held in the user's hand. However, there is no standard shutter button/mechanism location across traditional camera solutions and retrieving the camera from storage and looking through a viewfinder takes time. These issues are problematic when a user wants to capture media (photos, videos, etc.) related to a moment very quickly.
  • A system and method in accordance with the present invention addresses the aforementioned issues by providing a wearable camera system (wearable camera device) that is gesture operated and mounted/worn on the wrist of the user. The gesture operated wearable camera device eliminates size, storage, retrieval, and operating knowledge constraints associated with traditional digital cameras. The user of the wearable camera device does not need to retrieve any device or place it into storage (because the device is already worn around the user's wrist), does not need to find and know how to operate the shutter button, and does not need to setup any device by holding or using a mount.
  • In addition, the gesture operated wearable camera device eliminates constraints associated with traditional mounted cameras. The user of the wearable camera device does not need to use a mount that requires additional time and effort and a receptacle and has a limited range of movement, does not need to predetermine the scope of the framing or use pre-mounting techniques, and does not need to rely on digital or optical viewfinders for framing.
  • In addition, the gesture operated wearable camera device eliminates constraints and the limited range of motion associated with traditional smartphone cameras. The user of the gesture operated camera system does not need to access or unlock (e.g., using a pattern or passcode) the device before capturing an image, does not need to utilize a specific photo application to capture the image, and does not need to use a device that that is not optimized for photo taking.
  • In addition, the gesture operated wearable camera device eliminates constraints associated with traditional camera eyeglasses. The user of the gesture operated wearable camera device does not need to utilize a lens fixed to the user's eye or any type of special eyewear, does not obscure the field of view with various actions, and does not include unintuitive camera triggers such as voice commands and winks.
  • The system and method in accordance with the present invention provides a wearable camera device that utilizes gesture recognition to allow for the user to take high quality photos/video/media quickly and efficiently. The photos and/or video are taken by easily framing the photo at different angles by simply using the user's hand, without the need for a viewfinder, or any contact with hardware in the palm or fingers of the user's hand, or the use of a second/free hand to trigger a shutter button. In one embodiment, the user of the wearable camera device raises an empty hand, gestures with the same hand as if taking a photo with a traditional camera, and captures a photo (or video) within a known framed area. The wearable camera device has the advantage of reducing the amount of time required to prepare, frame, and capture a media (photo/video).
  • In one embodiment, the wearable camera device comprises a wristband device (band component) and a camera that is either coupled to the wristband device (so that the user could potential upgrade or interchange the type of the camera) or integrated within the wristband device. The wristband device includes a plurality of embedded sensors including but not limited to microelectromechanical systems (MEMS) devices, gyroscopes, accelerometers, pressure sensors, optical sensors, biometric, electromagnetic sensors, and motion sensors that detect user movements and gestures and the wristband also includes a processor device/unit to analyze and classify the detected gestures.
  • In another embodiment, the wearable camera device comprises a glove-like device that is in contact with the user's fingers and a mounted camera coupled to the glove-like device. In this device, the wearable camera device does not include a wristband device portion and the sensor and processor and computing components are instead housed in the glove-like device that is worn by the user like a typical glove.
  • In one embodiment, the camera is small in size and lightweight. The camera includes a camera lens. In one embodiment, the camera lens is any of thumbnail sized, nano sized, the size of the tip of a pen, the size of a traditional camera lens located on a smartphone, and other small sizes. In another embodiment, the camera lens is smaller than thumbnail sized. In one embodiment, the camera is interchangeable with various types of lens so that the user can upgrade the wearable camera device. In one embodiment, the wristband device is interchanged with different colors and patterns and jewelry pieces to enable customizability.
  • In one embodiment, the camera with camera lens is coupled to the wristband device and positioned near the wrist in a manner that allows the camera lens to fluidly, automatically, and continuously follow the movement and/or rotation of the user's wrist. In one embodiment, the camera lens is flexibly fixed to the wristband device near the outer edge of the user's hand. This allows the camera lens to move in line with the wrist when rotated or when the hand is moved.
  • In another embodiment, the camera lens is rigidly fixed to a portion of the wristband device and is actuated electromechanically by electrical signals from the wristband device that are generated in response to wrist or hand movements that are detected by MEMS devices and sensors embedded within the wristband device.
  • The wristband device recognizes arm, wrist, hand, and finger gestures based on a variety of techniques by monitoring, filtering and detecting, classifying, and processing muscle movements, tendon movements, bone movements, wrist shape changes, hand shape shapes, finger shape changes, and/or bioelectromagnetics states, other user movements and changes to provide for sensor signal detection, feature extraction, and gesture recognition functionalities.
  • In one embodiment, the wristband device utilizes wrist contour biometrics and a contour mapping mechanism as a primary input for gesture recognition. In another embodiment, additional sensors that detect additional inputs and sensor data are utilized by the wristband device to determine the gestures and perform the gesture recognition. In one embodiment, firmware running on an embedded system in the wristband device can monitor, filter, feature extract, classify, and interpret recognized gestures, and then transmit the recognized gestures as camera firmware commands to the camera to control the camera firmware, hardware components, and electromechanical functions of the camera.
  • At a certain time point (N), the user's hand that is wearing the wristband device, the user's fingers, and the user's wrist are all oriented in particular orientations. When the hand, wrist, or fingers of the user are moved or change position, then the muscles/tendons contract and/or relax, and the bones and skin of the user also move or change/shift positions. These changes can result in physical anatomical changes on the surface of the wrist or cause bioelectromagnetics changes and the changes are recorded by the contour mapping mechanism to control the camera.
  • In one embodiment, the wristband device or band includes sensor arrays that are spaced apart from each other at predetermined distances and with known pitches so that the sensor arrays completely circumvent the user's wrist. In another embodiment, the wristband device includes sensor arrays that do not completely circumvent the user's wrist and instead are focused on a certain location or region. The sensor arrays either measure distance to the surface of the wrist or measure the pressure that the wrist exerts at each sensor position. Analog-to-digital converters (ADCs) then convert sensor signals detected by the sensor arrays to an array of values. The array of values that are collected from the user's wrist by the wristband device are a representation of the state of the user's wrist, hand, and fingers at the time point N.
  • The array of values are then filtered by the wristband device using a plurality of filters including but not limited to median filters and infinite impulse (IIR) filters to reduce/eliminate noise and motion artifacts, and extract/retain features that meet certain thresholds within a known tolerance criteria (i.e., distance or pressure measurement+/− some acceptable error at each sensor position). Once the wristband device filters the signals to extract features, the features are classified using sensor data classification.
  • In one embodiment, the wristband device includes a sensor data classification unit that includes pre-defined or pre-determined classifiers that contain features sets per each recognized gesture (e.g., particular arrangement of the user's wrist, hand, and/or fingers). In one embodiment, the pre-defined classifiers and feature sets are pre-loaded in the wristband device's firmware from reduced training sets that are previously collected from sufficiently large and diverse training population data.
  • In another embodiment, the pre-defined classifiers and feature sets are collected in real-time from a cloud-based database that stores data related to all of the users that are using the wearable camera device. In another embodiment, the pre-defined classifiers and feature sets are trained based on the current wearer/user of the wearable camera device over a predetermined time period. Training includes but is not limited to machine learning techniques, device feedback, and user feedback.
  • Once the wristband device classifies a set of extracted features as matching at least one known gesture from a plurality of recognized gestures (e.g., user gestures related to controlling the camera of the wearable camera device), firmware of the wristband device issues a command/callback to either the firmware of the wristband device to perform an action by the wristband device (e.g., dismissing an alert or notification, uploading data to cloud storage, etc.) or to the firmware of the camera to perform a camera-related action by the camera (e.g., taking a photo, controlling camera shutter, changing camera modes, actuating the camera lens, etc.).
  • In one embodiment, the wristband device detects an array of sensor data values associated with user movements at a frequency of K. Based on the detected array of values and subsequent feature extraction and gesture recognition, a sequence of commands at time point N are extracted from the recognized gestures and are classified at events N+1/K, N+1/K*2, N+1/K*3, etc. These events are processed continuously in a loop to provide constant command control to the wearable camera device. Other embodiments may use additional sensors including but not limited to Biometric electroencephalography (EEG) magnetoencephalography (MEG) sensors, and electromyography (EMG) sensors in combination with pressure and optical sensors to reduce noise, false positive features, or misclassification of gestures and commands.
  • In one embodiment, the wristband device monitors movements from the user's arm, wrist, hand, fingers, and thumb. When the user is ready to take a photo or video, the user's hand with the wearable camera device is raised and positioned to frame the photo/video. Once the user's hand is raised and the photo/video is framed by the user's hand, a photo/video (media) can be captured by a plurality of finger movements and gestures recognized by the wearable camera device.
  • In one embodiment, the index finger and thumb are extended, approximately 90 degrees from each other, effectively creating one corner of a camera frame. In another embodiment, varying angles and positions between the user's fingers or the index finger and thumb are possible to create the camera frame while increasing usability and minimizing potential user fatigue. For example, the user can hold up all four fingers and extend the thumb approximately 90 degrees from the four fingers to frame the camera or the user can merely hold up one finger to frame the photo/video. One of ordinary skill in the art readily recognizes that a plurality of user movements and gestures can be associated with a plurality of wearable camera device functions and that would be within the spirit and the scope of the present invention.
  • In the embodiment where the user has extended his/her index finger and thumb approximately 90 degrees from each other, the position of the shutter plane is always substantially parallel to the plane created by the index finger and the thumb—even when the user's wrist is rotated. In this embodiment, the wristband device can detect the extended index finger and thumb by monitoring, processing, classifying, and recognizing muscle, tendon, and wrist movements, wrist contours, hand shapes, and movements using feature extraction and gesture recognition to control the camera.
  • Once the wristband device has detected that the user has extended his/her index finger and thumb approximately 90 degrees from each other (or that the user has positioned his/her hand/fingers in another orientation associated with certain gestures), the wristband device transmits an instruction/command/call to the camera to open the shutter and wait in a ready state to take/capture a photo or video.
  • In one embodiment, the user gestures with the index finger and simulates a button press by slightly contracting the extended finger. In another embodiment, the user gestures with his/her finger to “flick” or “point” to the subject. The wristband device recognizes these gestures as a camera trigger gesture by monitoring, classifying, and extracting features from the finger and wrist movements using the plurality of embedded sensors that collects/detects muscle, tendon, bioelectromagnetics, and anatomical contour changes and the processor device that utilizes algorithms (e.g., machine learning classifiers) to analyze and classify the detected movements.
  • In this embodiment, the wristband device triggers the camera to capture a photo or video in response to the detected movements that are classified as a camera trigger gesture. In another embodiment, the wristband device detects other gestures for camera-related functions such as capturing photos/videos, changing/selecting operation modes, zooming in/out, etc.
  • In one embodiment, the operation modes include but are not limited to adjusting the lens position by rotating the wristband device around the wrist. In one embodiment, the wristband device has a light emitting user interface (UI) that is sized to be displayed on the user's wrist and that can be manipulated either by the detected user movements/gestures or by direct user touch gestures on a display screen of the UI. The UI can be either smaller than traditional video displays, similarly sized, or larger. The UI can be an LCD display, LED display, or another type of display unit.
  • In one embodiment, the wearable camera device or the wristband device itself includes a communication device (e.g., WiFi or Bluetooth receiver and transmitter) that includes any of WiFi and Bluetooth communication capabilities so that the sensor movement data that is detected and analyzed as well as the captured photos and videos can be wirelessly communicated to another device (e.g., smartphone, laptop, etc.) or to a cloud-computing storage system by the communication device. The captured photos and videos can be automatically transmitted (or transmitted according to a predetermined schedule) to the device or cloud-computing storage system so that the user's information is seamlessly backed up.
  • In one embodiment, the wristband device can be used in conjunction with a second wristband device on the user's opposing hand. The first and the second wristband devices can communicate different modes, operations, and data to each other and also work in conjunction as a multi-camera system. The multi-camera system can include additional gesture recognitions and features to enable advanced photo and video taking including but not limited to panorama photos/videos and 3D photos/videos.
  • The system and method in accordance with the present invention provide a gestured operated wrist-mounted camera system (wearable camera device) that is an unobtrusive accessory-type wristband camera that can be worn by the user for 24 hours a day so that the user can capture moments at any time or place. Therefore, the wearable camera device is convenient and quickly produces high quality and framed camera shots. The need for a physical viewfinder (either optical or digital), a shutter button, and command and control buttons is eliminated. Readying, framing, sizing, and shutter operations do not require physical buttons. The user can also personalize the wearable camera device by selecting a plurality of different styles, colors, designs, and patterns associated with the wristband.
  • The form factor of the wearable camera device provides camera hardware that is integrated into the wristband device so there is no need for costly hardware (physical or digital screen) or a viewfinder which reduces the footprint (size, power, complexity). The wristband device is form-fitting and adjustable and can include a small adjustable positioned camera and lens. In another embodiment, the wristband device includes a display unit that can serve as a viewfinder for added functionality.
  • In one embodiment, the camera is coupled to the wristband device at the junction of the pisiform and ulna. In another embodiment, the camera is coupled to the wristband device at or near the edge of an abductor digit minima. These positions allow for the plane of the field of view created naturally by the index finger and thumb (or other user's hand/finger orientations) to be parallel with the camera lens even when the wrist is maneuvered and rotated at any angle (eliminating any potential obstructions). In another embodiment, the position of the camera lens is electromechanically moveable to align with the movement of the hand and wrist.
  • The wearable camera device is able to monitor, recognize and interpret arm, wrist, hand, and finger gestures, movements, and changes by monitoring, detecting, processing, and classifying muscle movements, tendon movements, hand shapes, and wrist contours using feature extractions and gesture recognition. In one embodiment, the gestures that the wearable camera device can detect to initialize, ready the camera, and take photos and/or videos include but are not limited to a semicircle shape, circle shape, OK action, hang loose action, swipe left across index finger, swipe right across index finger, tap or double tap between fingers, tap or double tap across index finger, 90 degree framing, corner box framing, double view framing, framing, switching modes of operation (camera to video, etc.), traversing modes and options, selecting options, triggering a shutter, starting capture, motion capture, encoding, and zooming in/out.
  • The wristband device can include features outside of camera functionality, processing, and wireless communication including but not limited to a clock, timer, and Internet/email capabilities. In one embodiment, the wristband device includes an adjustable or rotatable (around the wrist) camera lens position and multiple lenses on the wristband (e.g., front, rear, facing, etc). In one embodiment, the wristband device includes perforations along the outside of the band that can emit light (e.g., light-emitting diodes or LEOs) to notify the user with different patterns, shapes, and designs that act as an interface display to provide feedback information to the user (e.g., photo taken, photo uploaded to cloud storage, error in taking photo, etc.). The LED display and patterns can be programmable by the user or predetermined.
  • In one embodiment, the wristband device is touch enabled to allow for touch gestures to be recognized directly on the wristband device (or the display unit of the wristband device) in addition to muscle, tendon, and bioelectromagnetics recognition. In one embodiment, the wristband device includes the ability to tap, drag, and flick (away) the emitted light based notifications and objects around the outer portion of the wristband. This allows the user to manipulate and adjust the position of the emitted display, to control the notifications once they are no longer relevant to the user, and to respond to certain notifications with various input, touch, and gesture responses.
  • In one embodiment, the wristband device includes the ability to manipulate modes of operation, options, and the interface display by performing a gesture with only one hand (palm free). In one embodiment, the wristband device includes at least one accelerometer to aid in framing the shots and with orientation. In one embodiment, the wristband device includes a vibration sensor that can vibrate as part of the interface display that provides feedback to the user (in addition to the LED notifications). In one embodiment, the wristband device includes a speaker with audio output as part of the interface display so that additional feedback (in audio format) can be provided.
  • In one embodiment, the wristband device includes a plurality of sensors that are embedded within and connected via circuitry to detect various data from the user and the user's surrounding environment. The plurality of sensors can include but are not limited to any of or any combination of MEMS devices, gyroscopes, accelerometers, torque sensors, weight sensors, pressure sensors, magnetometers, temperature sensors, light sensors, cameras, microphones, GPS, wireless detection sensors, altitude sensors, blood pressure sensors, heart rate sensors, biometric sensors, radio frequently identification (RFID), near field communication (NFC), mobile communication, Wi-Fi, strain gauges, fingerprint sensors, smell sensors, gas sensors, chemical sensors, color sensors, sound sensors, acoustic sensors, ultraviolet sensors, electric field sensors, magnetic field sensors, gravity sensors, wind speed sensors, wind direction sensors, compass sensors, geo-locator sensors, polarized light sensors, infrared emitter sensors, and photo-reflective sensors.
  • In one embodiment, the wristband device includes a processor device that analyzes the detected sensor data (from the plurality of sensors) using a sensor data classification unit that utilizes a plurality of algorithmic processes. The plurality of algorithmic processes can include but is not limited to any of or any combination of back propagation, bayes networks, machine learning, deep learning, neural networks, fuzzy mean max neural networks, hidden Markov chains, hierarchical temporal memory, k nearest neighbor (KNN), adaboot, and histogram analysis.
  • In one embodiment, the user utilizes a plurality of wristband devices worn on both hands. In this embodiment, wireless communication and synchronization between the plurality of wristband devices can provide multi-camera functionality including but not limited to any of new framing options, larger framing areas, 30 capability, and 360 degree capability. In one embodiment, the wearable camera device includes a capability of connecting to various networks (public or private).
  • To describe the features of the present invention in more detail, refer now to the following description in conjunction with the accompanying Figures.
  • FIG. 1 illustrates a system 100 for capturing media in accordance with an embodiment. The media can include photos, video, and/or other types of media. The system 100 is a wearable camera device that includes a wristband device 150 and a camera 134 coupled to the wristband device 150. The camera 134 comprises a lens, a barrel, and an actuator. In one embodiment, the wristband device 150 includes a plurality of components including but not limited to any of a power supply 102, a random access memory (RAM) 104, microcontroller (MCU) 106, read-only memory (ROM) 108, a clock 110, a storage/memory device 112 (including but not limited to Flash memory), a radio 114 (including but not limited to Bluetooth and WiFi), an antenna 116, a haptic/vibrator sensor 118, an accelerometer 120, a gyroscope 122, an analog-to-digital converter (ADC) 124, pre-amplifier/filter/amplifier/bias device 126, a plurality of external sensors 128 (e.g., EKG, EEG, MEG, EMG, pressure, optical, etc.), an in-system programming component (ISP) 130, and an internal sensor 132. In another embodiment, the embedded accelerometer 120 and gyroscope 122 are included in the plurality of external sensors. In another embodiment, the wristband device includes a processor, a memory device, an application, and a transmitter/receiver device.
  • The plurality of external sensors 128 detect user movements (e.g., muscle or tendon movements) and the plurality of components of the wristband device 150 determine a plurality of gestures by monitoring, classifying, and extracting features from the detected user movements. The plurality of gestures control various actions that are executed by either the wristband device 150 itself or the camera 134 including but not limited to taking pictures or videos, scrolling through various options, and scrolling through various modes.
  • FIG. 2 illustrates a system 200 for sensor data classification by a wearable camera device in accordance with an embodiment. The system 200 includes an application component 202, an operating system (OS) component 204, and a plurality of drivers 206. In one embodiment, the application component 202 includes inputs/outputs (I/O), a digitizer/classifier, a gesture detection component, a sensor data classification unit, and a commands component. In one embodiment, the OS 204 includes a scheduling/memory management component and a message passing component. In one embodiment, the each driver of a plurality of drivers 206 may include communication drivers, sensor drivers, general purpose I/O (GPIO) drivers, and file system (FS) drivers. In one embodiment, the system 200 detects user movements using the sensors of the wristband device and processes the detected user movements using the application component 202 and the sensor data classification unit to determine the gestures and associated commands that control the camera.
  • FIG. 3 illustrates a method 300 for capturing media in accordance with an embodiment. The media can include photos, video, and/or other types of media. The method 300 comprises providing a wristband device that includes at least one sensor, via step 302, coupling a camera to the wristband device, via step 304, determining at least one gesture using the at least one sensor, via step 306, and controlling the camera by using the at least one gesture, via step 308.
  • In one embodiment, the method further includes detecting, by the at least one sensor, user movements by using any of muscle, tendon, bioelectromagnetics, and anatomical contour changes (using the contour mapping mechanism) of a user. A processor of the wristband device then analyzes the user movements using the sensor data classification unit to extract features from the user movements and to determine at least one gesture from the extracted features using various classifiers. For example, if the detected user movement is determined to be the at least one gesture of a user extending an index finger and thumb approximately 90 degrees from each other, the wristband device will instruct/control the camera to open a shutter and await the photo/video capture.
  • Once the wristband device determines an additional gesture when the user simulates a button pressing by slightly contracting the index finger, the wristband device transmits an instruction to the camera to trigger the shutter thereby capturing the media (photo/video). In one embodiment, a communication device is coupled to both the wristband device and the camera, wherein the communication device transmits data from both the wristband device (e.g., gesture classifications) and the camera (e.g., photos and videos) to another device that comprises any of a smartphone, a laptop, a desktop, and a cloud-based server system.
  • FIG. 4 illustrates a method 400 for capturing media in accordance with another embodiment. The media can include photos, video, and/or other types of media. The method 400 comprises movements of a user using the wearable camera device occurring, via step 402, voltage sensors detecting the user movements, via step 404, the voltage sensors recording the detected user movements as a signal, via step 406, the signal passing through a pre-amplifier, via step 408, the pre-amplified signal passing through a plurality of filters to remove noise and additional motion artifacts, via step 410, the filtered signal passing through an amplifier, via step 412, and the amplified signal passing through an analog-to-digital converter (ADC), via step 414.
  • The filtered, amplified, and converted signal has features extracted from it, via step 416, and these feature extractions result in gesture determinations and associated instructions, callbacks, and notifications, via step 418. The callbacks and notifications are registered by the wearable camera device, via step 420, and the instructions or callbacks are transmitted from the wristband device to the camera for execution, via step 422, which sets up certain camera controls and functions (e.g., taking a photo/video), via step 424. The controls and functions are encoded, via step 426, which results in the triggering of the associated camera function (e.g., data capture), via step 428, and the wearable camera device then transmits the captured media, via step 430.
  • FIG. 5 illustrates a user point of view of a wearable camera device 500 in accordance with an embodiment. The wearable camera device 500 includes a wristband device 502 that is coupled to a camera (or the camera is embedded within the wristband device 502). The wristband device 502 also includes other hardware components and a muscle, tendon, finger gesture recognizer or sensor detection device. The wearable camera device 500 is shown from the user's point of view when framing a photo/video and includes a frame of view 504 that serves as the user's viewfinder (in place of traditional digital or optical viewfinders). In this embodiment, the wearable camera device 500 does not include a traditional optical or digital viewfinder which allows the user greater flexibility in taking nature photos and videos by using the frame of view 504.
  • In another embodiment, a viewfinder is displayed on a display unit/screen (e.g., LCD/LED screen) of the user interface (UI) of the wristband device. In this embodiment, as the user focuses on a subject using his/her fingers/hand to frame the subject, the user can verify the correct frame has been captured by checking the display unit of the wristband device that displays the signal from the camera lens.
  • In the diagram 500, the user's index finger and thumb are extended approximately 90 degrees from each other to frame the subject. In another embodiment, varying angles and positions between the user's fingers or the index finger and thumb are possible to create the camera frame while increasing usability and minimizing potential user fatigue. Once the user extends his/her fingers in this shape (index finger and thumb go degrees from each other), the wristband's internal sensors detect the user movements and shapes (orientation of the index finger relative to the thumb) as a muscle movement. The detected muscle movement is determined to be a certain gesture that readies the camera focus and shutter. The wearable camera device 500 then awaits another detected user movement that is determined to be a certain gesture that will trigger another camera action (such as taking the photo/video).
  • In another embodiment, the user wearing the wearable camera device faces his/her palm outwards and towards the subject with all four fingers raised (and possibly extending the thumb approximately 90 degrees from the four fingers or resting the thumb up against the index finger). In this embodiment, if the user lowered one or more of the four raised fingers, then the wristband device would detect user movements, determine a gesture from the detected user movements, and then transmit a command/instruction to the camera based upon the determined gesture to carry out a camera action including but not limited to triggering a camera shutter action. In another embodiment, the user wears the wearable camera device (or specifically the wristband device portion) at a position that is rotated 180 degrees or the opposite of the normal wearing position. In this embodiment, the camera lens is facing the user thereby allowing the user to take “selfie” style captures using similar user movements and gestures (e.g., lowering one of the fingers that are raised, etc.).
  • FIG. 6 illustrates a side view of the wearable camera device 600 in accordance with an embodiment. The wearable camera device 600 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 602 and a camera coupled to the wristband device 602. The camera includes a camera lens 604 focused on an object 608. The frame of view 606 is focused on the object 608. The user's index finger and thumb are in the same position (extended approximately 90 degrees from each other) and so the camera is once again in a ready position for when another user movement and associated gestured is determined.
  • FIG. 7 illustrates a subject point of view of the wearable camera device 700 in accordance with an embodiment. The wearable camera device 700 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 702 and a camera coupled to the wristband device 702. The camera includes a camera lens 704. The user's index finger and thumb are in the same position (extended approximately 90 degrees from each other) and so the camera is once again in a ready position for when another user movement and associated gestured is determined.
  • FIG. 8 illustrates a user point of view of the wearable camera device 800 in accordance with an embodiment. The wearable camera device 800 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 802 and a camera coupled to or integrated within the wristband device 802. The user's index finger and thumb have moved from the position of being extended approximately 90 degrees from each other to a button pressing simulation movement 806. In this embodiment, the button pressing simulation movement 806 is when the user's index finger is slightly lowered. In another embodiment, a different type of user movement can be associated with the button pressing simulation movement 806.
  • Once the user's index finger moves, the wristband device 502's sensors detect the user movement and the wristband device 502's internal components and sensor data classification unit determine that a specific gesture associated with the detected user movement has occurred. The determined gesture prompts the camera to take a picture/photo of an object 808 in the distance that is framed by a frame of view 804.
  • The wristband device of the wearable camera device detects a plurality of sensor signals and determines a plurality of gestures using the plurality of sensor signals. The plurality of sensor signals can represent either a single gesture (e.g., “clicking” motion of one finger, etc.) or a set of gestures in a specific sequence (e.g., sliding a finger to the left or right and tapping, etc.) that the firmware of the wristband device constitutes as a single action. Once the gesture is determined from the detected sensor signals, the firmware of the wristband device sends out an associated command to control the camera. FIGS. 9-11 represent additional sets of gestures that the wristband device can determine as a single action.
  • FIG. 9 illustrates a user point of view of the wearable camera device 900 in accordance with an embodiment. The wearable camera device 900 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 902 and a camera coupled to or integrated within the wristband device 902. The wristband device 902 has determined a gesture 904 that comprises the user tapping or double tapping the thumb. The gesture 904 triggers the wearable camera device 900 to display the target/subject of the photo/video on the wristband device 902. In this embodiment, the wristband device 902 includes a user interface display that can display the subject that is being photographed by the camera to ensure that the user has focused the camera correctly upon the target/subject that the user wants to photograph/video.
  • FIG. 10 illustrates a user point of view of the wearable camera device 1000 in accordance with an embodiment. The wearable camera device 1000 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1002 and a camera coupled to or integrated within the wristband device 1002. The wristband device 1002 has determined a gesture 1004 that comprises the user swiping left with the thumb. The gesture 1004 triggers the wearable camera device 1000 to change various modes of the camera and the mode change is displayed on the user interface of the wristband device 1002. In this embodiment, the wristband device 1002 includes a user interface display that can display the varying modes. In one embodiment, the modes are varied using alternating LED patterns. In another embodiment, the modes are varied and the text (e.g., “Photo Taking”, “Video Taking”, “Night-Time Photo”, etc.) is displayed on the user interface display.
  • FIG. 11 illustrates a user point of view of the wearable camera device 1100 in accordance with an embodiment. The wearable camera device 1100 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1102 and a camera coupled to or integrated within the wristband device 1102. The wristband device 1102 has determined a gesture 1104 that comprises the user swiping right with the thumb. The gesture 1104 triggers the wearable camera device to also change various modes of the camera (in the opposite direction as swiping left so essentially the user can scroll left and right through various options) and the mode change is displayed on the user interface of the wristband device 1102. In this embodiment, the wristband device 1102 includes a user interface display that can display the varying modes. In one embodiment, the modes are varied using alternating LED patterns. In another embodiment, the modes are varied and the text (e.g., “Photo Taking”, “Video Taking”, “Night-Time Photo”, etc.) is displayed on the user interface display.
  • FIG. 12 illustrates a user point of view of the wearable camera device 1200 in accordance with an embodiment. The wearable camera device 1200 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1202 and a camera coupled to or integrated within the wristband device 1202. The wristband device 1202 has determined a gesture 1204 that comprises the user holding his/her fingers in a “C” shape. The gesture 1204 triggers the wearable camera device 1200 to select the mode that the user has scrolled to (by swiping left or right). In another embodiment, the user can input various other hand gestures and shapes that are associated with mode selection.
  • FIG. 13 illustrates a user point of view of the wearable camera device 1300 in accordance with an embodiment. The wearable camera device 1300 has components similar to the wearable camera device 500 of FIG. 5 including a wristband device 1302 and a camera coupled to or integrated within the wristband device 1302. The wristband device 1302 has determined a gesture 1304 that comprises the user's fingers snapping. The gesture 1304 triggers an “OK” action to the wearable camera device 1300. In one embodiment, the “OK” action represents the user setting or acknowledging a particular setting or command that may be displayed to the user as feedback on the device interface output (e.g., LEOs, user interface display, etc.).
  • FIG. 14 illustrates a user point of view of the wearable camera device 1400 in accordance with an embodiment. The wearable camera device 1400 has components similar to the wearable camera device of FIG. 5 including a wristband device 1402 and a camera coupled to or integrated within the wristband device 1402. The wristband device 1402 has determined a gesture 1404 that comprises the user tapping or double tapping the index finger and the middle finger both to the thumb. The gesture 1404 triggers an “OK” action to the wearable camera device 1400. In one embodiment, the “OK” action represents the user setting or acknowledging a particular setting or command that may be displayed to the user as feedback on the device interface output (e.g., LEOs, user interface display, etc.). One of ordinary skill in the art readily recognizes that other user movements can be associated with an “OK” action and that would be within the spirit and scope of the present invention.
  • FIG. 15 illustrates a user point of view of a multi-wearable camera device system 1500 in accordance with an embodiment. The multi-wearable camera device system 1500 includes a first wearable camera device 1502 coupled to a second wearable camera device 1504. The first and the second wearable camera devices 1502-1504 enhance the user's ability to frame a target/subject 1506. In addition, the first and the second wearable camera devices 1502-1504 communicate with each other and enable the user to select from various dual wristband modes. In one embodiment, the multi-wearable camera device system 1500 syncs and stitches together image and video capture thereby creating panorama or wide angle capture. In another embodiment, the multi-wearable camera device system 1500 provides syncing and capturing of 30 image or video captures, 30 stereoscope vision, and 30 zooming.
  • FIG. 16 illustrates a user interface display of a wearable camera device 1600 in accordance with an embodiment. The wearable camera device 1600 includes a user interface display 1602. In one embodiment, the user interface display 1602 comprises LEOs that display various patterns such as the pattern 1604 based upon camera settings and notifications. In one embodiment, the LEOs are integrated via laser cut holes cut into the user interface display 1602. The user can use hand gestures (waiving across the user interface display 1602) or touch gestures (pressing on the user interface display 1602) to respond to the various notifications or delete them. One of ordinary skill in the art readily recognizes that the wearable camera device can include varying types of display units and user interfaces and that would be within the spirit and scope of the present invention.
  • FIG. 17 illustrates a method 1700 for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment. The media can include photos, video, and/or other types of media. The method 1700 represents two time points (N and N+1). The user of the wearable camera device starts in position A (with the index finger and thumb approximately 90 degrees from each other) at time point N and ends in position B (with the index finger slightly lowered in a “clicking” motion) at time point N+1. The wearable camera device detects the user's hand positioning and movements using a plurality of embedded sensors and determines various gestures from the detected positioning/movements, via step 1702.
  • To determine the gestures from the detected user movements, the wearable camera device utilizes a contour mapping mechanism that provides a contour map of the anatomical contours of the user's wrist that is wearing the wristband device of the wearable camera device, via step 1704. The wristband device (band and embedded sensors) position around the user's wrist is denoted by a solid line and the contour of the user's wrist is denoted by a dotted line. When the user's hand is in position A at time point N, the contour map of the user's wrist is in wrist contour shape one (1) and when the user's hand shifts in position from position A to position B at time point N+1, the contour map of the user's wrist is in wrist contour shape two (2).
  • After determining the change in the contour map via step 1704, the wristband device classifies the contour changes using feature extraction and associated classifiers from contour shapes, via step 1706. The key features of the classifier are plotted on a graph that depicts a distance in millimeters (mm) on the y-axis and that depicts the sensor position around the circumference of the wrist and based upon the contour map on the x-axis. When the user's hand is in position A at time point N, the plotted graph displays the wrist contour shape 1 in a first sensor position and when the user's hand is in position B at time point N+1, the plotted graph displays the wrist contour shape 2 in a second sensor position.
  • After classifying the sensor positions to determine the gesture, the wearable camera device the determined gesture is associated with a certain command that is then transmitted to the camera, via step 1708. For example, when the wearable camera device determines the first sensor position (that is associated with a first gesture), the camera receives a “camera ready” command (because that command is associated with the first gesture) and when the wearable camera device determines the second sensor position (after the user has changed orientation of his/her hand which is associated with a second gesture), the camera receives a “shutter press” command (because that command is associated with the second gesture) and the photo/video is captured by the wearable camera device. One of ordinary skill in the art readily recognizes that a variety of contour shapes and sensor positions can be associated with a variety of gestures and subsequently with a variety of commands and that would be within the spirit and scope of the present invention.
  • FIG. 18 illustrates a method 1800 for capturing media using gesture recognition by a wearable camera device in accordance with an embodiment. The media can include photos, video, and/or other types of media. The wearable camera device monitors the hand of a user wearing the wearable camera device (that includes a wristband device and a camera coupled to or integrated within the wristband device), via step 1802. When the user moves a portion of the hand, the fingers, and/or the wrist (user movements), the user movements cause muscle, skin, tendons, and bone to move as well allowing the wearable camera device to detect the user movements as sensor values, via step 1804, by using embedded sensors within the wristband device of the wearable camera device at a time point N, via step 1806. The detected sensor values are converted into a signal using an ADC and stored as an array of values (AV), via step 1808. After storage, filters (e.g., median filter) are applied to the array of values (AV), via step 1810. Steps 1802-1810 represent the “monitoring” phase of the method 1800 utilized by the wearable camera device.
  • The monitoring phase is repeated by the wearable camera device to create enough training data for certain gestures (G, G1, etc.), via step 1812. Once enough training data is created (or pre-downloaded into the wearable camera device), common features are identified for each gesture (G, G1, etc.) that are each associated with certain gesture classifiers (C, C1, etc.) that are each associated with certain actions (A, A1, etc.), via step 1814. The wearable camera device receives the determined gestures and creates a gesture classifier for each, via step 1816. Steps 1812-1816 represent the “classification” phase of the method 1800 utilized by the wearable camera device.
  • After the classification phase has been completed by the wearable camera device and enough training data has been created, via step 1816, the wearable camera device compares the filtered AV values to the gesture classifiers (created via step 1816) using matching methodologies including but not limited to machine learning and k-nearest neighbor algorithms (KNN), via step 1818. If the features from the filtered AV values match a certain gesture classifier, the wearable camera device calls the action (transmits the command to the camera) associated with that certain gesture classifier, via step 1820. For example, if filtered AV values (or sensor data) matches a certain gesture G1 that is associated with classifier C1 that is associated with action A1, then the wearable camera device will call and transmit a command/instruction to the camera for action A1. Steps 1818-1820 represent the “recognition” phase of the method 1800 utilized by the wearable camera device.
  • After the recognition phase has been completed by the wearable camera device, the action command or call (e.g., camera shutter) is received by the camera, via step 1822, and the camera firmware initiates camera encoding of the camera controls and functions, via step 1824. The action command or call is encoded, via step 1826, and the action is executed (e.g., the camera captures the image), via step 1828. Once the image (or video) is captured, the wearable camera device transmits the media to a local or remote storage, via step 1830, and a user can view the media, via step 1832. One of ordinary skill in the art readily recognizes that the wearable camera device can associated a plurality of sensor values with gestures, classifiers, and actions and that would be within the spirit and scope of the present invention.
  • FIG. 19 illustrates a subject point of view of a wearable camera device 1900 in accordance with an embodiment. The wearable camera device 1900 includes a wristband device 1902 (or band) that wraps around the wrist of the user and a camera 1904 that is coupled to or integrated within the wristband device 1902. The camera 1904 includes a camera lens 1906 and a camera sub assembly. In one embodiment, the camera lens 1906 is mechanically and flexibly affixed to the wristband device 1902 near the bottom of the user's palm. The flex fixing camera lens 1906 mechanically moves with the movements/gestures of the user's wrist and palm. In another embodiment, the camera lens 1906 is rigidly fixed to a portion of the wristband device 1902 that is controllable by electromechanical actuation based on movements/gestures.
  • The subject point of view represents the view from a subject that is looking straight at the camera lens 1906. The subject point of view shows the palm and various fingers (middle, ring, pinky) of the user. The wristband device 1902 includes a plurality of components including but not limited to sensors, display unit, hardware platform, battery, storage, and radio. In one embodiment, the wristband device 1902 rotates around the user's wrist enabling the camera lens 1906 to either face outward towards the subject or face inward towards the user for “selfie” picture capturing capability.
  • A system and method in accordance with the present invention discloses a wearable camera system (wearable camera device) for capturing media (photo/video). The wearable camera device comprises a wristband device that includes at least one sensor. In one embodiment, the wearable camera device also comprises a camera coupled to the wristband device. In another embodiment, the camera is integrated within the wristband device as one overall device. The camera is controlled by at least one gesture determined from user movements and sensor data detected by the at least one sensor.
  • The wristband device includes additional hardware components including but not limited to a processor that analyzes the user movements detected by the at least one sensor and a memory device that stores the various data (detected user movements, determined gestures, etc.). The at least one sensor detects the user movements by detecting any of muscle, tendon, bioelectromagnetics, and anatomical contour changes of the user. In one embodiment, the processor analyzes the detected user movements using a sensor data classification unit that utilizes filters and algorithms that extract features from the user movements to determine the at least one gesture. The sensor data classification unit utilizes any of back propagation, Bayes networks, neural networks, and machine learning to determine the user gestures.
  • In one embodiment, the at least one sensor is any of a gyroscope, an accelerometer, a pressure sensor, a temperature sensor, and a light sensor. In one embodiment, the camera is controlled without using any of an optical viewfinder or a digital viewfinder and instead uses a natural viewfinder created by the user's natural fingers. In one embodiment, the camera includes a lens that is small (e.g., thumbnail sized) and the lens is positioned to follow the rotational movement of the user. This ensures that the lens is never obstructed and is positioned on the user's wrist in a way that enables clear targeting of the photo/video/media subjects that the user wants to capture.
  • The wearable camera device can determine a plurality of gestures that trigger various camera actions. In one embodiment, the gesture includes a user extending an index finger and thumb approximately 90 degrees from each other to create one corner of a camera frame. In this embodiment, the camera shutter's plan is substantially parallel to a plane created by the extending of the index finger and thumb. In another embodiment, the user can determine various hand gestures and actions and correlate each of these hand gestures and actions to specific camera actions. In another embodiment, based upon a continuously updated database or software updates, the wearable camera device determines additional gestures and associated camera actions.
  • In one embodiment, once the user extends the index finger and thumb approximately 90 degrees from each other in a certain gesture and the wearable camera device determines the certain gesture, the wristband device of the wearable camera device instructs/controls/transmits a message to the camera to open the shutter and await a photo/video/media capture gesture from the user. In one embodiment, the user provides a photo/video/media capture gesture by simulating a button pressure. In another embodiment, the user provides a photo/video/media capture gesture by another gesture that is either predetermined or inputted by the user such as snapping of the fingers.
  • In one embodiment, the at least one gesture or plurality of gestures that the wearable camera device can determine includes any of initializing camera function, semicircle shapes, circle shapes, OK action, hang loose action, swipe left across index finger, swipe right across index finger, tap between fingers, double tap between fingers, witching modes, selecting options, zooming, triggering a shutter to start capture, and triggering a shutter to start motion capture.
  • In one embodiment, the wearable camera device includes a communication device (e.g., transmitter/receiver device) coupled to both the wristband device and the camera (or just one of either the wristband device and the camera). In this embodiment, the communication device transmits data from either or both the wristband device (e.g., user movements, gestures, etc.) and the camera (e.g., photos, videos, media, etc.) to another device that comprises any of a smartphone and a cloud-based server system. In one embodiment, the wristband device includes a user interface display that comprises a plurality of light emitting diodes (LEDs) that produces various patterns associated with camera actions and notifications and alerts for the user.
  • As above described, a system and method in accordance with the present invention utilizes a wristband device that includes a plurality of sensors and a processor and a camera mounted to the wristband device to provide a gesture operated wrist-mounted camera system (wearable camera device). The wearable camera device is an unobtrusive accessory-type wristband camera that can be worn by the user for 24 hours a day so that the user can take photos and videos at any time and with ease. The wearable camera device is convenient (in accessibility, size, shape, weight, etc) and quickly produces high quality and framed camera shots and videos once user movements and gestures associated with various commands are detected by the wearable camera device.
  • A system and method for operating a wrist-mounted camera system utilizing gestures has been disclosed. Embodiments described herein can take the form of an entirely hardware implementation, an entirely software implementation, or an implementation containing both hardware and software elements. Embodiments may be implemented in software, which includes, but is not limited to, application software, firmware, resident software, microcode, etc.
  • The steps described herein may be implemented using any suitable controller or processor, and software application, which may be stored on any suitable storage location or computer-readable medium. The software application provides instructions that enable the processor to perform the functions described herein.
  • Furthermore, embodiments may take the form of a computer program product accessible from a computer-usable or computer-readable medium providing program code for use by or in connection with a computer or any instruction execution system. For the purposes of this description, a computer-usable or computer-readable medium can be any apparatus that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
  • The medium may be an electronic, magnetic, optical, electromagnetic, infrared, semiconductor system (or apparatus or device), or a propagation medium (non-transitory). Examples of a computer-readable medium include a semiconductor or solid state memory, magnetic tape, a removable computer diskette a random access memory (RAM), a read-only memory (ROM), a rigid magnetic disk, and an optical disk. Current examples of optical disks include DVD, compact disk-read-only memory (CD-ROM), and compact disk-read/write (CD-R/W).
  • Although the present invention has been described in accordance with the embodiments shown, one of ordinary skill in the art will readily recognize that there could be variations to the embodiments and those variations would be within the spirit and scope of the present invention. Accordingly, many modifications may be made by one of ordinary skill in the art without departing from the spirit and scope of the appended claims.

Claims (20)

What is claimed is:
1. A system, comprising:
a wristband device, the wristband device including a sensor to determine gestures; and
a camera coupled to the wristband device, the camera being controlled by the gestures and being moveable along an outer surface of the wristband device, the gestures including a user extending an index finger and thumb approximately 90 degrees from each other.
2. The system of claim 1, wherein the wristband device includes a processor that detects signals and analyzes the signals using a contour mapping mechanism to generate a contour map.
3. The system of claim 2, wherein the signals are detected using at least one of muscle, tendon, bioelectromagnetics, or anatomical contour changes associated with user movements of the user.
4. The system of claim 2, wherein the contour map is analyzed by the processor using sensor data classification of the contour mapping mechanism to extract features, the gestures determined using the extracted features.
5. The system of claim 4, wherein the sensor data classification includes at least one of back propagation, bayes networks, neural networks, or machine learning.
6. The system of claim 1, wherein the sensor is at least one of a gyroscope, an accelerometer, a pressure sensor, a temperature sensor, or a light sensor.
7. The system of claim 1, wherein the camera is controlled by the gestures without using at least one of an optical viewfinder or a digital viewfinder.
8. The system of claim 3, wherein the camera includes a lens that is thumbnail sized.
9. The system of claim 3, wherein the user movements include at least one of rotating a wrist of the user or rotating an arm of the user.
10. The system of claim 1, wherein a plane of a shutter is substantially parallel to a plane created by the extending of the index finger and the thumb.
11. The system of claim 10, wherein in response to the user extending the index finger and the thumb approximately 90 degrees from each other, the processor of the wristband device transmits instructions to the camera to open the shutter.
12. The system of claim 11, wherein the processor of the wristband device determines that the user is simulating a button press when the index finger is contracted to a predetermined position thereby triggering the camera to capture media.
13. The system of claim 1, wherein the gestures include at least one of initializing camera function, semicircle shapes, circle shapes, OK action, hang loose action, swipe left across index finger, swipe right across index finger, tap between fingers, double tap between fingers, switching modes, selecting options, zooming, triggering a shutter to start capture, or triggering a shutter to start motion capture.
14. The system of claim 1, further comprising:
a communication device coupled to both the wristband device and the camera, the communication device transmitting data from both the wristband device and the camera to another device that comprises at least one of a smartphone or a server.
15. The system of claim 1, wherein the wristband device includes a light emitting user interface that produces notifications using light-emitting diodes (LEDs).
16. A method, comprising:
providing a wristband device, the wristband device including a camera and a sensor, the sensor configured to determine gestures, the camera being moveable along an outer surface of the wristband device;
determining a gesture comprising a user extending an index finger and thumb approximately 90 degrees from each other; and
controlling the camera using the gesture.
17. The method of claim 16, further comprising:
detecting signals using the sensor and at least one of muscle, tendon, or bioelectromagnetics changes associated with user movements of the user;
generating a contour map using the signals and a contour mapping mechanism;
analyzing the contour map using sensor data classification to extract features; and
determining the gestures using the extracted features.
18. The method of claim 17, further comprising:
in response to determining the gesture comprising the user extending the index finger and thumb approximately 90 degrees from each other, transmitting instructions to the camera to open a shutter.
19. The method of claim 18, further comprising:
determining that the user is simulating a button press by contracting the index finger to a predetermined position thereby triggering the camera to capture media.
20. The method of claim 16, further comprising:
coupling a communication device to both the wristband device and the camera, the communication device transmitting data from both the wristband device and the camera to another device that comprises at least one of a smartphone or a server.
US16/363,261 2014-02-28 2019-03-25 Gesture Operated Wrist Mounted Camera System Abandoned US20190220098A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US16/363,261 US20190220098A1 (en) 2014-02-28 2019-03-25 Gesture Operated Wrist Mounted Camera System
US17/690,754 US11861069B2 (en) 2014-02-28 2022-03-09 Gesture operated wrist mounted camera system
US18/503,762 US20240152215A1 (en) 2014-02-28 2023-11-07 Gesture Operated Wrist Mounted Camera System

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201461946622P 2014-02-28 2014-02-28
US201562101289P 2015-01-08 2015-01-08
US14/634,674 US10254843B2 (en) 2014-02-28 2015-02-27 Gesture operated wrist mounted camera system
US16/363,261 US20190220098A1 (en) 2014-02-28 2019-03-25 Gesture Operated Wrist Mounted Camera System

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US14/634,674 Continuation US10254843B2 (en) 2014-02-28 2015-02-27 Gesture operated wrist mounted camera system

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US17/690,754 Continuation US11861069B2 (en) 2014-02-28 2022-03-09 Gesture operated wrist mounted camera system

Publications (1)

Publication Number Publication Date
US20190220098A1 true US20190220098A1 (en) 2019-07-18

Family

ID=54009693

Family Applications (4)

Application Number Title Priority Date Filing Date
US14/634,674 Active US10254843B2 (en) 2014-02-28 2015-02-27 Gesture operated wrist mounted camera system
US16/363,261 Abandoned US20190220098A1 (en) 2014-02-28 2019-03-25 Gesture Operated Wrist Mounted Camera System
US17/690,754 Active US11861069B2 (en) 2014-02-28 2022-03-09 Gesture operated wrist mounted camera system
US18/503,762 Pending US20240152215A1 (en) 2014-02-28 2023-11-07 Gesture Operated Wrist Mounted Camera System

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US14/634,674 Active US10254843B2 (en) 2014-02-28 2015-02-27 Gesture operated wrist mounted camera system

Family Applications After (2)

Application Number Title Priority Date Filing Date
US17/690,754 Active US11861069B2 (en) 2014-02-28 2022-03-09 Gesture operated wrist mounted camera system
US18/503,762 Pending US20240152215A1 (en) 2014-02-28 2023-11-07 Gesture Operated Wrist Mounted Camera System

Country Status (4)

Country Link
US (4) US10254843B2 (en)
EP (1) EP3111303A4 (en)
JP (1) JP2017508229A (en)
WO (1) WO2015131157A1 (en)

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11275453B1 (en) 2019-09-30 2022-03-15 Snap Inc. Smart ring for manipulating virtual objects displayed by a wearable device
US11277597B1 (en) 2020-03-31 2022-03-15 Snap Inc. Marker-based guided AR experience
WO2022066578A1 (en) * 2020-09-28 2022-03-31 Snap Inc. Touchless photo capture in response to detected hand gestures
US20220334647A1 (en) * 2014-02-28 2022-10-20 Vikas Gupta Gesture Operated Wrist Mounted Camera System
US11520399B2 (en) 2020-05-26 2022-12-06 Snap Inc. Interactive augmented reality experiences using positional tracking
US11531402B1 (en) 2021-02-25 2022-12-20 Snap Inc. Bimanual gestures for controlling virtual and graphical elements
US11740313B2 (en) 2020-12-30 2023-08-29 Snap Inc. Augmented reality precision tracking and display
US11798429B1 (en) 2020-05-04 2023-10-24 Snap Inc. Virtual tutorials for musical instruments with finger tracking in augmented reality
US11861070B2 (en) 2021-04-19 2024-01-02 Snap Inc. Hand gestures for animating and controlling virtual and graphical elements
US11925863B2 (en) 2020-09-18 2024-03-12 Snap Inc. Tracking hand gestures for interactive game control in augmented reality
US12072406B2 (en) 2020-12-30 2024-08-27 Snap Inc. Augmented reality precision tracking and display
US12086324B2 (en) 2020-12-29 2024-09-10 Snap Inc. Micro hand gestures for controlling virtual and graphical elements
US12135840B2 (en) 2022-12-15 2024-11-05 Snap Inc. Bimanual gestures for controlling virtual and graphical elements

Families Citing this family (89)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10042422B2 (en) 2013-11-12 2018-08-07 Thalmic Labs Inc. Systems, articles, and methods for capacitive electromyography sensors
US20150124566A1 (en) 2013-10-04 2015-05-07 Thalmic Labs Inc. Systems, articles and methods for wearable electronic devices employing contact sensors
US10188309B2 (en) 2013-11-27 2019-01-29 North Inc. Systems, articles, and methods for electromyography sensors
US11921471B2 (en) 2013-08-16 2024-03-05 Meta Platforms Technologies, Llc Systems, articles, and methods for wearable devices having secondary power sources in links of a band for providing secondary power in addition to a primary power source
US9880632B2 (en) 2014-06-19 2018-01-30 Thalmic Labs Inc. Systems, devices, and methods for gesture identification
US9942222B1 (en) * 2014-09-02 2018-04-10 Amazon Technologies, Inc. Authentication with wearable device
CN104394312B (en) * 2014-10-23 2017-08-22 小米科技有限责任公司 Filming control method and device
US20160132642A1 (en) * 2014-11-06 2016-05-12 Raz Carmi Device and method for monitoring food intake
US20170090583A1 (en) * 2015-09-25 2017-03-30 Intel Corporation Activity detection for gesture recognition
US10630871B2 (en) * 2015-11-03 2020-04-21 Sam Salemnia Body mounted camera system
US10324494B2 (en) 2015-11-25 2019-06-18 Intel Corporation Apparatus for detecting electromagnetic field change in response to gesture
US20170177203A1 (en) * 2015-12-18 2017-06-22 Facebook, Inc. Systems and methods for identifying dominant hands for users based on usage patterns
US10782790B2 (en) 2015-12-22 2020-09-22 Intel Corporation System and method to collect gesture input through wrist tendon and muscle sensing
JP6727921B2 (en) * 2016-05-18 2020-07-22 ソニーモバイルコミュニケーションズ株式会社 Information processing device, information processing system, and information processing method
US10638316B2 (en) 2016-05-25 2020-04-28 Intel Corporation Wearable computer apparatus with same hand user authentication
US10185817B2 (en) * 2016-06-16 2019-01-22 International Business Machines Corporation Device security via swipe pattern recognition
DE112016007020T5 (en) * 2016-06-28 2019-03-21 Intel Corporation VIDEO WITH EMBEDDED GESTURE
US11034362B2 (en) * 2016-07-07 2021-06-15 Harman International Industries, Incorporated Portable personalization
WO2018022657A1 (en) 2016-07-25 2018-02-01 Ctrl-Labs Corporation System and method for measuring the movements of articulated rigid bodies
US11331045B1 (en) 2018-01-25 2022-05-17 Facebook Technologies, Llc Systems and methods for mitigating neuromuscular signal artifacts
US20190121306A1 (en) 2017-10-19 2019-04-25 Ctrl-Labs Corporation Systems and methods for identifying biological structures associated with neuromuscular source signals
US11179066B2 (en) 2018-08-13 2021-11-23 Facebook Technologies, Llc Real-time spike detection and identification
US11216069B2 (en) 2018-05-08 2022-01-04 Facebook Technologies, Llc Systems and methods for improved speech recognition using neuromuscular information
WO2018022658A1 (en) 2016-07-25 2018-02-01 Ctrl-Labs Corporation Adaptive system for deriving control signals from measurements of neuromuscular activity
WO2018022597A1 (en) 2016-07-25 2018-02-01 Ctrl-Labs Corporation Methods and apparatus for inferring user intent based on neuromuscular signals
US10687759B2 (en) 2018-05-29 2020-06-23 Facebook Technologies, Llc Shielding techniques for noise reduction in surface electromyography signal measurement and related systems and methods
CN110300542A (en) 2016-07-25 2019-10-01 开创拉布斯公司 Method and apparatus for predicting musculoskeletal location information using wearable automated sensors
US10772519B2 (en) 2018-05-25 2020-09-15 Facebook Technologies, Llc Methods and apparatus for providing sub-muscular control
JP6897037B2 (en) * 2016-09-15 2021-06-30 オムロン株式会社 Workability evaluation device
US10845885B2 (en) 2017-02-27 2020-11-24 International Business Machines Corporation Object scrolling and selection on a wearable computing device
CN106951109B (en) * 2017-03-31 2020-02-14 殷超 Method and device for acquiring hand gesture
JP6774367B2 (en) * 2017-04-11 2020-10-21 富士フイルム株式会社 Head-mounted display control device, its operation method and operation program, and image display system
WO2018217060A1 (en) * 2017-05-25 2018-11-29 Samsung Electronics Co., Ltd. Method and wearable device for performing actions using body sensor array
US20190007536A1 (en) * 2017-07-03 2019-01-03 Essential Products, Inc. Handheld writing implement form factor mobile device
US10481699B2 (en) * 2017-07-27 2019-11-19 Facebook Technologies, Llc Armband for tracking hand motion using electrical impedance measurement
US10462345B2 (en) 2017-08-11 2019-10-29 Essential Products, Inc. Deformable structure that compensates for displacement of a camera module of a camera accessory
WO2019060298A1 (en) 2017-09-19 2019-03-28 Neuroenhancement Lab, LLC Method and apparatus for neuroenhancement
US11717686B2 (en) 2017-12-04 2023-08-08 Neuroenhancement Lab, LLC Method and apparatus for neuroenhancement to facilitate learning and performance
EP3732672B1 (en) * 2017-12-27 2024-08-07 Palmplug, Inc. Wearable computing apparatus for augmented reality, virtual reality and artificial intelligence interactions, and methods relating thereto
US11318277B2 (en) 2017-12-31 2022-05-03 Neuroenhancement Lab, LLC Method and apparatus for neuroenhancement to enhance emotional response
US10523976B2 (en) 2018-01-09 2019-12-31 Facebook, Inc. Wearable cameras
US11907423B2 (en) 2019-11-25 2024-02-20 Meta Platforms Technologies, Llc Systems and methods for contextualized interactions with an environment
WO2019148002A1 (en) 2018-01-25 2019-08-01 Ctrl-Labs Corporation Techniques for anonymizing neuromuscular signal data
US11069148B2 (en) 2018-01-25 2021-07-20 Facebook Technologies, Llc Visualization of reconstructed handstate information
US11961494B1 (en) 2019-03-29 2024-04-16 Meta Platforms Technologies, Llc Electromagnetic interference reduction in extended reality environments
EP3742961A4 (en) 2018-01-25 2021-03-31 Facebook Technologies, Inc. Calibration techniques for handstate representation modeling using neuromuscular signals
US11481030B2 (en) 2019-03-29 2022-10-25 Meta Platforms Technologies, Llc Methods and apparatus for gesture detection and classification
US10460455B2 (en) 2018-01-25 2019-10-29 Ctrl-Labs Corporation Real-time processing of handstate representation model estimates
US10937414B2 (en) 2018-05-08 2021-03-02 Facebook Technologies, Llc Systems and methods for text input using neuromuscular information
CN112005198A (en) 2018-01-25 2020-11-27 脸谱科技有限责任公司 Hand state reconstruction based on multiple inputs
WO2019147958A1 (en) 2018-01-25 2019-08-01 Ctrl-Labs Corporation User-controlled tuning of handstate representation model parameters
US11150730B1 (en) 2019-04-30 2021-10-19 Facebook Technologies, Llc Devices, systems, and methods for controlling computing devices via neuromuscular signals of users
US10970936B2 (en) 2018-10-05 2021-04-06 Facebook Technologies, Llc Use of neuromuscular signals to provide enhanced interactions with physical objects in an augmented reality environment
US11493993B2 (en) 2019-09-04 2022-11-08 Meta Platforms Technologies, Llc Systems, methods, and interfaces for performing inputs based on neuromuscular control
WO2019173678A1 (en) * 2018-03-09 2019-09-12 Siemens Aktiengesellschaft Optimal hand pose tracking using a flexible electronics-based sensing glove and machine learning
US11364361B2 (en) 2018-04-20 2022-06-21 Neuroenhancement Lab, LLC System and method for inducing sleep by transplanting mental states
US10592001B2 (en) 2018-05-08 2020-03-17 Facebook Technologies, Llc Systems and methods for improved speech recognition using neuromuscular information
US11281302B2 (en) * 2018-05-18 2022-03-22 Steven Reynolds Gesture based data capture and analysis device and system
EP3807795A4 (en) 2018-06-14 2021-08-11 Facebook Technologies, LLC. User identification and authentication with neuromuscular signatures
US11045137B2 (en) 2018-07-19 2021-06-29 Facebook Technologies, Llc Methods and apparatus for improved signal robustness for a wearable neuromuscular recording device
US10802598B2 (en) * 2018-08-05 2020-10-13 Pison Technology, Inc. User interface control of responsive devices
US11099647B2 (en) * 2018-08-05 2021-08-24 Pison Technology, Inc. User interface control of responsive devices
WO2020047429A1 (en) 2018-08-31 2020-03-05 Ctrl-Labs Corporation Camera-guided interpretation of neuromuscular signals
WO2020056418A1 (en) 2018-09-14 2020-03-19 Neuroenhancement Lab, LLC System and method of improving sleep
WO2020061451A1 (en) 2018-09-20 2020-03-26 Ctrl-Labs Corporation Neuromuscular text entry, writing and drawing in augmented reality systems
EP3857342A4 (en) 2018-09-26 2021-12-01 Facebook Technologies, LLC. Neuromuscular control of physical objects in an environment
EP3886693A4 (en) 2018-11-27 2022-06-08 Facebook Technologies, LLC. Methods and apparatus for autocalibration of a wearable electrode sensor system
US10788897B1 (en) * 2019-01-15 2020-09-29 Facebook Technologies, Llc Systems and methods for sensing gestures via vibration-sensitive wearables donned by users of artificial reality systems
US10986287B2 (en) 2019-02-19 2021-04-20 Samsung Electronics Co., Ltd. Capturing a photo using a signature motion of a mobile device
US10905383B2 (en) 2019-02-28 2021-02-02 Facebook Technologies, Llc Methods and apparatus for unsupervised one-shot machine learning for classification of human gestures and estimation of applied forces
US11786694B2 (en) 2019-05-24 2023-10-17 NeuroLight, Inc. Device, method, and app for facilitating sleep
US11422623B2 (en) * 2019-10-23 2022-08-23 Interlake Research, Llc Wrist worn computing device control systems and methods
US12089953B1 (en) 2019-12-04 2024-09-17 Meta Platforms Technologies, Llc Systems and methods for utilizing intrinsic current noise to measure interface impedances
US11126267B2 (en) * 2019-12-19 2021-09-21 Giantplus Technology Co., Ltd Tactile feedback device and operation method thereof
US11914791B2 (en) * 2020-01-28 2024-02-27 Pison Technology, Inc. Gesture control using biopotential-based analog front end
US11157086B2 (en) * 2020-01-28 2021-10-26 Pison Technology, Inc. Determining a geographical location based on human gestures
US11199908B2 (en) * 2020-01-28 2021-12-14 Pison Technology, Inc. Wrist-worn device-based inputs for an operating system
US11397466B2 (en) * 2020-03-31 2022-07-26 Apple Inc. Skin-to-skin contact detection
US11397468B2 (en) 2020-03-31 2022-07-26 Apple Inc. Skin-to-skin contact detection
US20230260155A1 (en) * 2020-04-24 2023-08-17 Cornell University Deep continuous 3d hand pose tracking
US12045392B1 (en) * 2020-09-24 2024-07-23 Apple Inc. Method and device for defining custom hand gestures
CN114643826B (en) * 2020-12-21 2024-06-21 研能科技股份有限公司 Method for preventing and treating air pollution in vehicle
TWI811600B (en) * 2020-12-21 2023-08-11 研能科技股份有限公司 Solution for air pollution prevention in vehicles
US11868531B1 (en) 2021-04-08 2024-01-09 Meta Platforms Technologies, Llc Wearable device providing for thumb-to-finger-based input gestures detected based on neuromuscular signals, and systems and methods of use thereof
US11543892B2 (en) * 2021-04-28 2023-01-03 Google Llc Touch pressure input for devices
US11366529B1 (en) 2021-07-14 2022-06-21 Steven Reynolds Gesture based data capture and analysis device and system with states, confirmatory feedback and timeline
TWI824358B (en) * 2021-12-20 2023-12-01 研能科技股份有限公司 Method of notifying in-car air pollution
US11954263B1 (en) * 2022-09-26 2024-04-09 Pison Technology, Inc. Gesture control using biopotential wristband
WO2024180994A1 (en) * 2023-03-02 2024-09-06 パナソニックIpマネジメント株式会社 Wrist-worn sensor device

Citations (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040262342A1 (en) * 2003-06-26 2004-12-30 Pringle Christopher Robert Forearm camera mount
US20050094019A1 (en) * 2003-10-31 2005-05-05 Grosvenor David A. Camera control
US20090089682A1 (en) * 2007-09-27 2009-04-02 Rockwell Automation Technologies, Inc. Collaborative environment for sharing visualizations of industrial automation data
US20090110235A1 (en) * 2007-10-26 2009-04-30 Samsung Electronics Co., Ltd. System and method for selection of an object of interest during physical browsing by finger framing
US20090327171A1 (en) * 2008-06-26 2009-12-31 Microsoft Corporation Recognizing gestures from forearm emg signals
US20100199232A1 (en) * 2009-02-03 2010-08-05 Massachusetts Institute Of Technology Wearable Gestural Interface
US20120016960A1 (en) * 2009-04-16 2012-01-19 Gelb Daniel G Managing shared content in virtual collaboration systems
US20120249409A1 (en) * 2011-03-31 2012-10-04 Nokia Corporation Method and apparatus for providing user interfaces
US20120250947A1 (en) * 2011-03-30 2012-10-04 Gil Abramovich Apparatus and method for contactless high resolution handprint capture
US8292833B2 (en) * 2009-08-27 2012-10-23 Electronics And Telecommunications Research Institute Finger motion detecting apparatus and method
US20120281129A1 (en) * 2011-05-06 2012-11-08 Nokia Corporation Camera control
US20120293544A1 (en) * 2011-05-18 2012-11-22 Kabushiki Kaisha Toshiba Image display apparatus and method of selecting image region using the same
US20140028546A1 (en) * 2012-07-27 2014-01-30 Lg Electronics Inc. Terminal and control method thereof
US20140049417A1 (en) * 2012-08-20 2014-02-20 Playtabase, LLC Wireless motion activated command transfer device, system, and method
US20140140590A1 (en) * 2012-11-21 2014-05-22 Microsoft Corporation Trends and rules compliance with depth video
US20140145929A1 (en) * 2009-04-02 2014-05-29 Oblong Industries, Inc. Cross-user hand tracking and shape recognition user interface
US20140354537A1 (en) * 2013-05-29 2014-12-04 Samsung Electronics Co., Ltd. Apparatus and method for processing user input using motion of object
US20150063661A1 (en) * 2013-09-03 2015-03-05 Samsung Electronics Co., Ltd. Method and computer-readable recording medium for recognizing object using captured image
US20150138078A1 (en) * 2013-11-18 2015-05-21 Eyal Krupka Hand pose recognition using boosted look up tables
US9116666B2 (en) * 2012-06-01 2015-08-25 Microsoft Technology Licensing, Llc Gesture based region identification for holograms
US20150248235A1 (en) * 2014-02-28 2015-09-03 Samsung Electronics Company, Ltd. Text input on an interactive display
US20160119545A1 (en) * 2013-08-01 2016-04-28 Fujifilm Corporation Imaging module and electronic apparatus
US9390500B1 (en) * 2013-03-14 2016-07-12 Amazon Technologies, Inc. Pointing finger detection
US20160203360A1 (en) * 2015-01-13 2016-07-14 Google Inc. Systems and methods for performing actions in response to user gestures in captured images
US20160293143A1 (en) * 2013-11-22 2016-10-06 Sharp Kabushiki Kaisha Communication terminal, method of controlling communication terminal and control program
US9524028B2 (en) * 2013-03-08 2016-12-20 Fastvdo Llc Visual language for human computer interfaces
US9785247B1 (en) * 2014-05-14 2017-10-10 Leap Motion, Inc. Systems and methods of tracking moving hands and recognizing gestural interactions
US10133342B2 (en) * 2013-02-14 2018-11-20 Qualcomm Incorporated Human-body-gesture-based region and volume selection for HMD
US10185399B2 (en) * 2015-03-31 2019-01-22 Fujitsu Limited Image processing apparatus, non-transitory computer-readable recording medium, and image processing method
US10334233B2 (en) * 2014-07-09 2019-06-25 Lg Electronics Inc. Portable device that controls photography mode, and control method therefor

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10360685B2 (en) * 2007-05-24 2019-07-23 Pillar Vision Corporation Stereoscopic image capture with performance outcome prediction in sporting environments
US6953990B2 (en) * 2003-09-19 2005-10-11 Agilent Technologies, Inc. Wafer-level packaging of optoelectronic devices
US8107599B2 (en) * 2005-02-15 2012-01-31 Fastvdo, Llc Methods and apparatus for the composition and communication of digital composition coded multisensory messages (DCC MSMS)
US8428135B2 (en) * 2005-11-14 2013-04-23 Fastvdo, Llc Device and method for fast sub sample block-matching motion estimation in video encoders
US20140035805A1 (en) * 2009-04-02 2014-02-06 David MINNEN Spatial operating environment (soe) with markerless gestural control
US8520736B2 (en) * 2009-04-14 2013-08-27 Fastvdo, Llc Real-time superresolution and video transmission
US8861797B2 (en) 2010-11-12 2014-10-14 At&T Intellectual Property I, L.P. Calibrating vision systems
US9436286B2 (en) * 2011-01-05 2016-09-06 Qualcomm Incorporated Method and apparatus for tracking orientation of a user
US9042596B2 (en) * 2012-06-14 2015-05-26 Medibotics Llc Willpower watch (TM)—a wearable food consumption monitor
US20200348627A1 (en) * 2012-06-14 2020-11-05 Medibotics Llc Wrist-Worn Device with One or More Cameras and a Comfortable Arm Posture During Imaging
JP2015055662A (en) * 2013-09-10 2015-03-23 惠州市大亜湾永昶電子工業有限公司 Actuator unit
JP2017508229A (en) * 2014-02-28 2017-03-23 グプタ ヴィカスGUPTA, Vikas Wrist-mounted camera system operated by gestures
US20160078289A1 (en) * 2014-09-16 2016-03-17 Foundation for Research and Technology - Hellas (FORTH) (acting through its Institute of Computer Gesture Recognition Apparatuses, Methods and Systems for Human-Machine Interaction
US10095942B2 (en) * 2014-12-15 2018-10-09 Reflex Robotics, Inc Vision based real-time object tracking system for robotic gimbal control
US10372020B2 (en) * 2016-05-31 2019-08-06 iSquared Global, LLC Dual-function lighted wearable accessory

Patent Citations (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040262342A1 (en) * 2003-06-26 2004-12-30 Pringle Christopher Robert Forearm camera mount
US20050094019A1 (en) * 2003-10-31 2005-05-05 Grosvenor David A. Camera control
US20090089682A1 (en) * 2007-09-27 2009-04-02 Rockwell Automation Technologies, Inc. Collaborative environment for sharing visualizations of industrial automation data
US20090110235A1 (en) * 2007-10-26 2009-04-30 Samsung Electronics Co., Ltd. System and method for selection of an object of interest during physical browsing by finger framing
US20090327171A1 (en) * 2008-06-26 2009-12-31 Microsoft Corporation Recognizing gestures from forearm emg signals
US20100199232A1 (en) * 2009-02-03 2010-08-05 Massachusetts Institute Of Technology Wearable Gestural Interface
US20140145929A1 (en) * 2009-04-02 2014-05-29 Oblong Industries, Inc. Cross-user hand tracking and shape recognition user interface
US20120016960A1 (en) * 2009-04-16 2012-01-19 Gelb Daniel G Managing shared content in virtual collaboration systems
US8292833B2 (en) * 2009-08-27 2012-10-23 Electronics And Telecommunications Research Institute Finger motion detecting apparatus and method
US20120250947A1 (en) * 2011-03-30 2012-10-04 Gil Abramovich Apparatus and method for contactless high resolution handprint capture
US20120249409A1 (en) * 2011-03-31 2012-10-04 Nokia Corporation Method and apparatus for providing user interfaces
US20120281129A1 (en) * 2011-05-06 2012-11-08 Nokia Corporation Camera control
US20120293544A1 (en) * 2011-05-18 2012-11-22 Kabushiki Kaisha Toshiba Image display apparatus and method of selecting image region using the same
US9116666B2 (en) * 2012-06-01 2015-08-25 Microsoft Technology Licensing, Llc Gesture based region identification for holograms
US20140028546A1 (en) * 2012-07-27 2014-01-30 Lg Electronics Inc. Terminal and control method thereof
US20140049417A1 (en) * 2012-08-20 2014-02-20 Playtabase, LLC Wireless motion activated command transfer device, system, and method
US20140140590A1 (en) * 2012-11-21 2014-05-22 Microsoft Corporation Trends and rules compliance with depth video
US10133342B2 (en) * 2013-02-14 2018-11-20 Qualcomm Incorporated Human-body-gesture-based region and volume selection for HMD
US9524028B2 (en) * 2013-03-08 2016-12-20 Fastvdo Llc Visual language for human computer interfaces
US9390500B1 (en) * 2013-03-14 2016-07-12 Amazon Technologies, Inc. Pointing finger detection
US20140354537A1 (en) * 2013-05-29 2014-12-04 Samsung Electronics Co., Ltd. Apparatus and method for processing user input using motion of object
US20160119545A1 (en) * 2013-08-01 2016-04-28 Fujifilm Corporation Imaging module and electronic apparatus
US20150063661A1 (en) * 2013-09-03 2015-03-05 Samsung Electronics Co., Ltd. Method and computer-readable recording medium for recognizing object using captured image
US20150138078A1 (en) * 2013-11-18 2015-05-21 Eyal Krupka Hand pose recognition using boosted look up tables
US20160293143A1 (en) * 2013-11-22 2016-10-06 Sharp Kabushiki Kaisha Communication terminal, method of controlling communication terminal and control program
US20150248235A1 (en) * 2014-02-28 2015-09-03 Samsung Electronics Company, Ltd. Text input on an interactive display
US9785247B1 (en) * 2014-05-14 2017-10-10 Leap Motion, Inc. Systems and methods of tracking moving hands and recognizing gestural interactions
US10334233B2 (en) * 2014-07-09 2019-06-25 Lg Electronics Inc. Portable device that controls photography mode, and control method therefor
US20160203360A1 (en) * 2015-01-13 2016-07-14 Google Inc. Systems and methods for performing actions in response to user gestures in captured images
US10185399B2 (en) * 2015-03-31 2019-01-22 Fujitsu Limited Image processing apparatus, non-transitory computer-readable recording medium, and image processing method

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220334647A1 (en) * 2014-02-28 2022-10-20 Vikas Gupta Gesture Operated Wrist Mounted Camera System
US20240152215A1 (en) * 2014-02-28 2024-05-09 Vikas Gupta Gesture Operated Wrist Mounted Camera System
US11861069B2 (en) * 2014-02-28 2024-01-02 Vikas Gupta Gesture operated wrist mounted camera system
US20220113814A1 (en) 2019-09-30 2022-04-14 Yu Jiang Tham Smart ring for manipulating virtual objects displayed by a wearable device
US11275453B1 (en) 2019-09-30 2022-03-15 Snap Inc. Smart ring for manipulating virtual objects displayed by a wearable device
US11747915B2 (en) 2019-09-30 2023-09-05 Snap Inc. Smart ring for manipulating virtual objects displayed by a wearable device
US11277597B1 (en) 2020-03-31 2022-03-15 Snap Inc. Marker-based guided AR experience
US12108011B2 (en) 2020-03-31 2024-10-01 Snap Inc. Marker-based guided AR experience
US11798429B1 (en) 2020-05-04 2023-10-24 Snap Inc. Virtual tutorials for musical instruments with finger tracking in augmented reality
US12014645B2 (en) 2020-05-04 2024-06-18 Snap Inc. Virtual tutorials for musical instruments with finger tracking in augmented reality
US11520399B2 (en) 2020-05-26 2022-12-06 Snap Inc. Interactive augmented reality experiences using positional tracking
US12008153B2 (en) 2020-05-26 2024-06-11 Snap Inc. Interactive augmented reality experiences using positional tracking
US11925863B2 (en) 2020-09-18 2024-03-12 Snap Inc. Tracking hand gestures for interactive game control in augmented reality
US11546505B2 (en) 2020-09-28 2023-01-03 Snap Inc. Touchless photo capture in response to detected hand gestures
WO2022066578A1 (en) * 2020-09-28 2022-03-31 Snap Inc. Touchless photo capture in response to detected hand gestures
US12086324B2 (en) 2020-12-29 2024-09-10 Snap Inc. Micro hand gestures for controlling virtual and graphical elements
US11740313B2 (en) 2020-12-30 2023-08-29 Snap Inc. Augmented reality precision tracking and display
US12072406B2 (en) 2020-12-30 2024-08-27 Snap Inc. Augmented reality precision tracking and display
US11531402B1 (en) 2021-02-25 2022-12-20 Snap Inc. Bimanual gestures for controlling virtual and graphical elements
US11861070B2 (en) 2021-04-19 2024-01-02 Snap Inc. Hand gestures for animating and controlling virtual and graphical elements
US12135840B2 (en) 2022-12-15 2024-11-05 Snap Inc. Bimanual gestures for controlling virtual and graphical elements
US12141367B2 (en) 2023-11-03 2024-11-12 Snap Inc. Hand gestures for animating and controlling virtual and graphical elements

Also Published As

Publication number Publication date
US11861069B2 (en) 2024-01-02
EP3111303A4 (en) 2017-08-23
US20220334647A1 (en) 2022-10-20
US20150309582A1 (en) 2015-10-29
WO2015131157A1 (en) 2015-09-03
US10254843B2 (en) 2019-04-09
EP3111303A1 (en) 2017-01-04
JP2017508229A (en) 2017-03-23
US20240152215A1 (en) 2024-05-09

Similar Documents

Publication Publication Date Title
US11861069B2 (en) Gesture operated wrist mounted camera system
US11668938B2 (en) Wearable imaging device
US9886111B2 (en) Wearable electronic apparatus and acquisition control method
CN112118380B (en) Camera control method, device, equipment and storage medium
CN110506249B (en) Information processing apparatus, information processing method, and recording medium
WO2017036035A1 (en) Screen control method and device
US11782514B2 (en) Wearable device and control method thereof, gesture recognition method, and control system
KR20200031133A (en) Vein scanning device for automatic gesture and finger recognition
US20180123629A1 (en) Smart-ring methods and systems
CN105310659A (en) Apparatus and method for enhancing accuracy of a contactless body temperature measurement
US10979632B2 (en) Imaging apparatus, method for controlling same, and storage medium
CN116348832A (en) Gesture triggered augmented reality
CN108353129B (en) Photographing apparatus and control method thereof
US20190384419A1 (en) Handheld controller, tracking method and system using the same
US20160034252A1 (en) Smart device control
EP3767433A1 (en) Information processing device, information processing method, and program
KR102163996B1 (en) Apparatus and Method for improving performance of non-contact type recognition function in a user device
CN111831116A (en) Intelligent equipment interaction method based on PPG information
CN109117819B (en) Target object identification method and device, storage medium and wearable device
CN112114653A (en) Terminal device control method, device, equipment and storage medium
US20230359422A1 (en) Techniques for using in-air hand gestures detected via a wrist-wearable device to operate a camera of another device, and wearable devices and systems for performing those techniques
CN113439913A (en) Wearable equipment
WO2018076371A1 (en) Gesture recognition method, network training method, apparatus and equipment
US20170199578A1 (en) Gesture control method for interacting with a mobile or wearable device
US20240362796A1 (en) Image analysis to switch audio devices

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION