Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/2390776.2390785acmconferencesArticle/Chapter ViewAbstractPublication PagesmmConference Proceedingsconference-collections
research-article

Cooking gesture recognition using local feature and depth image

Published: 02 November 2012 Publication History

Abstract

In this paper, we propose a method combining visual local features and depth image information to recognize cooking gestures. We employ the feature calculation method[2] which used extended FAST detector and a compact descriptor CHOG3D to calculate visual local features. We pack the local features by BoW in frame sequences to represent the cooking gestures. In addition, the depth images of hands gestures are extracted and integrated spatio-temporally to represent the position and trajectory information of cooking gestures. The two kinds of features are used to describe cooking gestures, and recognition is realized by employing the SVM. In our method, we determine the gesture class for each frame in cooking sequences. By analyzing the results of frames, we recognize cooking gestures in a continue frame sequences of cooking menus, and find the temporal positions of the recognized gestures.

References

[1]
A. F. Bobick and J. W. Davis. The recognition of human movement using temporal templates. Transactions on Pattern Analysis and Machine Intelligence, 23(3):257--267, 2001
[2]
Y. Ji, A. Shimada, and R. Taniguchi. A compact 3d descriptor in roi for human action recognition. In IEEE TENCON, 2010.
[3]
B. Packer, K. Saenko, and D. Koller. A combined pose, object, and feature model for action understanding. In CVPR, 2012.
[4]
M. Rohrbach, S. Amin, M. Andriluka, and B. Schiele. A database for fine grained activity detection of cooking activities. In CVPR, 2012.
[5]
E. Rosten, R. Porter, and T. Drummond. Faster and better: A machine learning approach to corner detection. IEEE Trans. Pattern Analysis and Machine Intelligence, 32:105--119, Jan. 2010.
[6]
L. Rybok, S. Friedberger, U. D. Hanebeck, and R. Stiefelhagen. The kit robo-kitchen data set for the evaluation of view-based activity recognition systems. In IEEE-RAS International Conference on Humanoid Robots, 2011.

Cited By

View all
  • (2020)Arbitrary-view Human Action Recognition: A Varying-view RGB-D Action DatasetIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2020.2975845(1-1)Online publication date: 2020
  • (2018)A Large-scale RGB-D Database for Arbitrary-view Human Action RecognitionProceedings of the 26th ACM international conference on Multimedia10.1145/3240508.3240675(1510-1518)Online publication date: 15-Oct-2018
  • (2018)Privacy preserving recognition of object-based activities using near-infrared reflective markersPersonal and Ubiquitous Computing10.1007/s00779-017-1070-922:2(365-377)Online publication date: 1-Apr-2018
  • Show More Cited By

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Conferences
CEA '12: Proceedings of the ACM multimedia 2012 workshop on Multimedia for cooking and eating activities
November 2012
72 pages
ISBN:9781450315920
DOI:10.1145/2390776
  • General Chair:
  • Mutsuo Sano,
  • Program Chair:
  • Ichiro Ide
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 02 November 2012

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. cooking gestures
  2. depth feature
  3. gesture recognition
  4. local feature

Qualifiers

  • Research-article

Conference

MM '12
Sponsor:
MM '12: ACM Multimedia Conference
November 2, 2012
Nara, Japan

Acceptance Rates

Overall Acceptance Rate 20 of 33 submissions, 61%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)6
  • Downloads (Last 6 weeks)1
Reflects downloads up to 24 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2020)Arbitrary-view Human Action Recognition: A Varying-view RGB-D Action DatasetIEEE Transactions on Circuits and Systems for Video Technology10.1109/TCSVT.2020.2975845(1-1)Online publication date: 2020
  • (2018)A Large-scale RGB-D Database for Arbitrary-view Human Action RecognitionProceedings of the 26th ACM international conference on Multimedia10.1145/3240508.3240675(1510-1518)Online publication date: 15-Oct-2018
  • (2018)Privacy preserving recognition of object-based activities using near-infrared reflective markersPersonal and Ubiquitous Computing10.1007/s00779-017-1070-922:2(365-377)Online publication date: 1-Apr-2018
  • (2017)Modeling Restaurant Context for Food RecognitionIEEE Transactions on Multimedia10.1109/TMM.2016.261486119:2(430-440)Online publication date: 1-Feb-2017
  • (2017)Gesture recognition based on spatiotemporal histogram of oriented gradient variation2017 6th International Conference on Informatics, Electronics and Vision & 2017 7th International Symposium in Computational Medical and Health Technology (ICIEV-ISCMHT)10.1109/ICIEV.2017.8338581(1-4)Online publication date: Sep-2017
  • (2015)Robots in the HomeProceedings of the Tenth Annual ACM/IEEE International Conference on Human-Robot Interaction10.1145/2696454.2696465(319-326)Online publication date: 2-Mar-2015
  • (2015)Spatiotemporal auto-correlation of grayscale gradient with importance map for cooking gesture recognition2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR)10.1109/ACPR.2015.7486487(166-170)Online publication date: Nov-2015
  • (2015)An interactive tool for manual, semi-automatic and automatic video annotationComputer Vision and Image Understanding10.1016/j.cviu.2014.06.015131:C(88-99)Online publication date: 1-Feb-2015
  • (2015)Gesture recognition in cooking video based on image features and motion features using Bayesian network classifierEmerging Trends in Image Processing, Computer Vision and Pattern Recognition10.1016/B978-0-12-802045-6.00024-7(379-392)Online publication date: 2015
  • (2014)MimiCookProceedings of the 8th International Conference on Tangible, Embedded and Embodied Interaction10.1145/2540930.2540952(121-124)Online publication date: 16-Feb-2014
  • Show More Cited By

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media