Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/2858036.2858358acmconferencesArticle/Chapter ViewAbstractPublication PageschiConference Proceedingsconference-collections
research-article

M.Gesture: An Acceleration-Based Gesture Authoring System on Multiple Handheld and Wearable Devices

Published: 07 May 2016 Publication History

Abstract

Gesture-based interaction is still underutilized in the mobile context despite the large amount of attention it has been given. Using accelerometers that are widely available in mobile devices, we developed M.Gesture, a software system that supports accelerometer-based gesture authoring on single or multiple mobile devices. The development was based on a formative study that showed users' preferences for subtle, simple motions and synchronized, multi-device gestures. M.Gesture adopts an acceleration data space and interface components based on mass-spring analogy and combines the strengths of both demonstration-based and declarative approaches. Also, gesture declaration is done by specifying a mass-spring trajectory with planes in the acceleration space. For iterative gesture modification, multi-level feedbacks are provided as well. The results of evaluative studies have shown good usability and higher recognition performance than that of dynamic time warping for simple gesture authoring. Later, we discuss the benefits of applying a physical metaphor and hybrid approach.

Supplementary Material

suppl.mov (pn1600-file3.mp4)
Supplemental video
MP4 File (p2307-kim.mp4)

References

[1]
Daniel Ashbrook and Thad Starner. 2010. MAGIC: a motion gesture design tool. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI 10), 2159--2168.
[2]
Mehmet Aydın Baytaş, Yücel Yemez, and Oğuzhan Özcan. 2014. Hotspotizer: end-user authoring of mid-air gestural interactions. In Proceedings of the Nordic Conference on Human-Computer Interaction: Fun, Fast, Foundational (NordiCHI 14), 677--686.
[3]
Ari Y. Benbasat, and Joseph A. Paradiso. 2002. An inertial measurement framework for gesture recognition and applications. In Gesture and Sign Language in Human-Computer Interaction, Ipke Wachsmuth and Timo Sowa (eds.). Springer, 9--20.
[4]
Liwei Chan, Chi-Hao Hsieh, Yi-Ling Chen, Shuo Yang, Da-Yuan Huang, Rong-Hao Liang, and Bing-Yu Chen. 2015. Cyclops: Wearable and Single-Piece Full-Body Gesture Input Devices. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '15). 3001--3009.
[5]
Xiang 'Anthony' Chen, Tovi Grossman, Daniel J. Wigdor, and George Fitzmaurice. 2014. Duet: exploring joint interactions on a smart phone and a smart watch. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '14). 159--168.
[6]
Xiang 'Anthony' Chen, Nicolai Marquardt, Anthony Tang, Sebastian Boring, and Saul Greenberg. 2012. Extending a mobile device's interaction space through body-centric interaction. In Proceedings of the international conference on Human-computer interaction with mobile devices and services (MobileHCI '12). 151--160.
[7]
Pei-Yu (Peggy) Chi and Yang Li. 2015. Weave: Scripting Cross-Device Wearable Interaction. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '15). 3923--3932.
[8]
Andrew Crossan, John Williamson, Stephen Brewster, and Rod Murray-Smith. 2008. Wrist rotation for interaction in mobile contexts. In Proceedings of the international conference on Human computer interaction with mobile devices and services (MobileHCI '08). 435--438.
[9]
Allen Cypher. 1993. Watch what I do: programming by demonstration. MIT Press.
[10]
Tovi Grossman, Xiang Anthony Chen, and George Fitzmaurice. 2015. Typing on Glasses: Adapting Text Entry to Smart Eyewear. In Proceedings of the International Conference on Human-Computer Interaction with Mobile Devices and Services (MobileHCI '15). 144--152.
[11]
Björn Hartmann, Leith Abdulla, Manas Mittal, and Scott R. Klemmer. 2007. Authoring sensor-based interactions by demonstration with direct manipulation and pattern recognition. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '07). 145--154.
[12]
Lode Hoste, Brecht De Rooms, and Beat Signer. 2013. Declarative Gesture Spotting Using Inferred and Refined Control Points. In Proceedings of the International Conference on Pattern Recognition Applications and Methods (ICPRAM 2013).
[13]
Steven Houben and Nicolai Marquardt. 2015. WatchConnect: A Toolkit for Prototyping Smartwatch Centric Cross-Device Applications. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '15). 1247--1256.
[14]
Edwin L. Hutchins, James D. Hollan, and Donald A. Norman. 1985. Direct manipulation interfaces. Human-Computer Interaction, 1, 4: 311--338.
[15]
Ideum. 2013. CreativeML. Retrieved September, 2015 from http://www.creativeml.org
[16]
Ideum. 2014. Gesture Works. Retrieved September, 2015 from http://gestureworks.com
[17]
Ju-Whan Kim and Tek-Jin Nam. 2013. EventHurdle: supporting designers' exploratory interaction prototyping with gesture-based sensors. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '13). 267--276.
[18]
Kenrick Kin, Björn Hartmann, Tony DeRose, and Maneesh Agrawala. 2012. Proton: multitouch gestures as regular expressions. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '12). 2885--2894.
[19]
Kenrick Kin, Björn Hartmann, Tony DeRose, and Maneesh Agrawala. 2012. Proton++: a customizable declarative multitouch framework. In Proceedings of the ACM symposium on User interface software and technology (UIST '12). 477--486.
[20]
Sven Kratz and Michael Rohs. 2010. A $3 gesture recognizer: simple gesture recognition for devices equipped with 3D acceleration sensors. In Proceedings of the international conference on Intelligent user interfaces (IUI '10). 341--344.
[21]
Yang Li. 2010. Protractor: a fast and accurate gesture recognizer. In Proceesdings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '10). 2169--2172.
[22]
Jiayang Liu, Zhen Wang, Lin Zhong, Jehan Wickramasuriya, and Venu Vasudevan. 2009. uWave: Accelerometer-based personalized gesture recognition and its applications. In Proceedings of IEEE International Conference on Pervasive Computing and Communications (PerCom '09), 1--9.
[23]
Hao Lü and Yang Li. 2012. Gesture coder: a tool for programming multi-touch gestures by demonstration. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '12). 2875--2884.
[24]
Hao Lü and Yang Li. 2013. Gesture studio: authoring multi-touch interactions through demonstration and declaration. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '13). 257--266.
[25]
Sushmita Mitra and Tinku Acharya. 2007. Gesture recognition: A survey. IEEE Transactions on Systems, Man, and Cybernetics-Part C: Applications and Reviews, 37, 3: 311--324.
[26]
Miguel A. Nacenta, Yemliha Kamber, Yizhou Qiang, and Per Ola Kristensson. 2013. Memorability of pre-designed and user-defined gesture sets. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '13). 1099--1108.
[27]
Donald A. Norman and Stephen W. Draper. 1986. User Centered System Design; New Perspectives on Human Computer Interaction. L. Erlbaum Assoc. Inc.
[28]
Taiwoo Park, Jinwon Lee, Inseok Hwang, Chungkuk Yoo, Lama Nachman, and Junehwa Song. 2011. EGesture: a collaborative architecture for energy-efficient gesture recognition with hand-worn sensor and mobile devices. In Proceedings of the ACM Conference on Embedded Networked Sensor Systems (SenSys '11). 260--273.
[29]
Dean Rubine. 1991. Specifying gestures by example. SIGGRAPH Comput. Graph. 25, 4: 329--337.
[30]
Jaime Ruiz, Yang Li, and Edward Lank. 2011. User defined motion gestures for mobile interaction. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '11). 197--206.
[31]
Hiroaki Sakoe and Seibi Chiba. 1978. Dynamic programming algorithm optimization for spoken word recognition. IEEE Transactions on Acoustics, Speech and Signal Processing, 26, 1: 43--49.
[32]
Lucio Davide Spano, Antonio Cisternino, Fabio Paternò, and Gianni Fenu. 2013. GestIT: a declarative and compositional framework for multi-platform gesture definition. In Proceedings of the SIGCHI symposium on Engineering interactive computing systems (EICS '13). 187--196.
[33]
Jacob O. Wobbrock, Meredith Ringel Morris, and Andrew D. Wilson. 2009. User-defined gestures for surface computing. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems (CHI '09). 1083--1092.
[34]
Jacob O. Wobbrock, Andrew D. Wilson, and Yang Li. 2007. Gestures without libraries, toolkits or training: a $1 recognizer for user interface prototypes. In Proceedings of the ACM symposium on User interface software and technology (UIST '07). 159--168.

Cited By

View all
  • (2024)Transforming Everyday Objects into IoT Control Interfaces: Design and Evaluation of the 'e-Rings' SystemArchives of Design Research10.15187/adr.2024.11.37.5.2937:5(29-49)Online publication date: 30-Nov-2024
  • (2023)Brave New GES World: A Systematic Literature Review of Gestures and Referents in Gesture Elicitation StudiesACM Computing Surveys10.1145/363645856:5(1-55)Online publication date: 7-Dec-2023
  • (2023)SparseIMU: Computational Design of Sparse IMU Layouts for Sensing Fine-grained Finger MicrogesturesACM Transactions on Computer-Human Interaction10.1145/356989430:3(1-40)Online publication date: 10-Jun-2023
  • Show More Cited By

Index Terms

  1. M.Gesture: An Acceleration-Based Gesture Authoring System on Multiple Handheld and Wearable Devices

    Recommendations

    Comments

    Please enable JavaScript to view thecomments powered by Disqus.

    Information & Contributors

    Information

    Published In

    cover image ACM Conferences
    CHI '16: Proceedings of the 2016 CHI Conference on Human Factors in Computing Systems
    May 2016
    6108 pages
    ISBN:9781450333627
    DOI:10.1145/2858036
    Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

    Sponsors

    Publisher

    Association for Computing Machinery

    New York, NY, United States

    Publication History

    Published: 07 May 2016

    Permissions

    Request permissions for this article.

    Check for updates

    Author Tags

    1. acceleration space
    2. gesture authoring
    3. hybrid approach
    4. mass-spring visualization
    5. multi-device gesture

    Qualifiers

    • Research-article

    Funding Sources

    Conference

    CHI'16
    Sponsor:
    CHI'16: CHI Conference on Human Factors in Computing Systems
    May 7 - 12, 2016
    California, San Jose, USA

    Acceptance Rates

    CHI '16 Paper Acceptance Rate 565 of 2,435 submissions, 23%;
    Overall Acceptance Rate 6,199 of 26,314 submissions, 24%

    Upcoming Conference

    CHI 2025
    ACM CHI Conference on Human Factors in Computing Systems
    April 26 - May 1, 2025
    Yokohama , Japan

    Contributors

    Other Metrics

    Bibliometrics & Citations

    Bibliometrics

    Article Metrics

    • Downloads (Last 12 months)25
    • Downloads (Last 6 weeks)4
    Reflects downloads up to 17 Feb 2025

    Other Metrics

    Citations

    Cited By

    View all
    • (2024)Transforming Everyday Objects into IoT Control Interfaces: Design and Evaluation of the 'e-Rings' SystemArchives of Design Research10.15187/adr.2024.11.37.5.2937:5(29-49)Online publication date: 30-Nov-2024
    • (2023)Brave New GES World: A Systematic Literature Review of Gestures and Referents in Gesture Elicitation StudiesACM Computing Surveys10.1145/363645856:5(1-55)Online publication date: 7-Dec-2023
    • (2023)SparseIMU: Computational Design of Sparse IMU Layouts for Sensing Fine-grained Finger MicrogesturesACM Transactions on Computer-Human Interaction10.1145/356989430:3(1-40)Online publication date: 10-Jun-2023
    • (2023)SequenceSense: A Tool for Designing Usable Foot-Based Gestures Using a Sequence-Based Gesture RecognizerInternational Journal of Human-Computer Studies10.1016/j.ijhcs.2023.103035176(103035)Online publication date: Aug-2023
    • (2022)Exploration on Everyday Objects as an IoT Control InterfaceProceedings of the 2022 ACM Designing Interactive Systems Conference10.1145/3532106.3533472(1654-1668)Online publication date: 13-Jun-2022
    • (2022)Metaphoraction: Support Gesture-based Interaction Design with Metaphorical MeaningsACM Transactions on Computer-Human Interaction10.1145/351189229:5(1-33)Online publication date: 20-Oct-2022
    • (2020)CAPturAR: An Augmented Reality Tool for Authoring Human-Involved Context-Aware ApplicationsProceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology10.1145/3379337.3415815(328-341)Online publication date: 20-Oct-2020
    • (2019)GestureCardsProceedings of the ACM on Human-Computer Interaction10.1145/33311643:EICS(1-35)Online publication date: 13-Jun-2019
    • (2019)Turn it, Pour it, Twist itProceedings of the 5th International ACM In-Cooperation HCI and UX Conference10.1145/3328243.3328246(20-23)Online publication date: 1-Apr-2019
    • (2019)Charting Subtle Interaction in the HCI LiteratureProceedings of the 2019 CHI Conference on Human Factors in Computing Systems10.1145/3290605.3300648(1-15)Online publication date: 2-May-2019
    • Show More Cited By

    View Options

    Login options

    View options

    PDF

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    Figures

    Tables

    Media

    Share

    Share

    Share this Publication link

    Share on social media