Detecting and Visualizing Stops in Dance Training by Neural Network Based on Velocity and Acceleration
<p>Demonstration of flow of motion in Yosakoi Soran dance. From left to right in the first line: no stop, stops, no stop. The stops factor is incorporated into the continuous dance process as a mid-period.</p> "> Figure 2
<p>Overview of the proposed system. The dance motion data are recorded by MoCap (<a href="#sec2-sensors-22-05402" class="html-sec">Section 2</a>). The proposed system consists of three phases (<a href="#sec3-sensors-22-05402" class="html-sec">Section 3</a>). First, the motion features based on velocity and acceleration are calculated (<a href="#sec3dot1-sensors-22-05402" class="html-sec">Section 3.1</a>). Stops are detected by a neural network model (<a href="#sec3dot2-sensors-22-05402" class="html-sec">Section 3.2</a>). Stops are visualized using a humanoid 3D model via virtual reality spaces (<a href="#sec3dot3-sensors-22-05402" class="html-sec">Section 3.3</a>).</p> "> Figure 3
<p>Attachment of PN to the performer is presented with 18 small sensors at hand, arm, shoulder, leg, head, and waist that measure inertia, such as a gyroscope and an accelerometer. The relative positions among the sensors are measured, and the 3D positions of the sensors are obtained.</p> "> Figure 4
<p>Example of velocity transition.</p> "> Figure 5
<p>Example of acceleration transition.</p> "> Figure 6
<p>3D model for stop visualization system.</p> "> Figure 7
<p>Motion data playback.</p> "> Figure 8
<p>Transition of loss function.</p> "> Figure 9
<p>Comparative example of visualization timing of stops.</p> ">
Abstract
:1. Introduction
2. Recording of Dance Motion Data by MoCap
3. Detection and Visualization of Stops by NN Based on Velocity and Acceleration
3.1. Calculation of Motion Features
3.1.1. Calculation of Velocity
3.1.2. Calculation of Acceleration
3.1.3. Construction of Training Data Set for Stop Detection
3.2. Construction of a Stop Detection Model of the NN
3.3. Visualization of a Stop by Human-like 3D Model in Virtual Reality
4. Experimental Section
4.1. Verification of Stop Detection Accuracy
4.1.1. Verification of the Effectiveness of the Proposed Method
- The effectiveness of the feature data is confirmed by comparing the detection accuracy of the stops by the PM and Comp. 1.
- The effectiveness of stop detection by the NN is confirmed by comparing the detection accuracy of stops by the PM and Comp. 2.
- The effectiveness of combining feature data and NN in detecting stops is confirmed by comparing the accuracy of the PM and Comp. 3.
- We verify that the PM is more effective than traditional supervised learning-based methods by comparing the detection accuracy of stops by the PM and Comp. 4–Comp. 6.
4.1.2. Explanation of Evaluation Index
4.1.3. Results and Discussion
4.2. Confirmation of Visualization Timing of Stops
4.2.1. Confirmation Method
4.2.2. Results and Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
ICT | Information and Communication Technology |
AI | Artificial Intelligence |
MoCap | Motion capture |
NN | Neural network |
PN | Perception Neuron |
ReLU | Rectified Linear Unit |
VR | Virtual Reality |
UI | User Interfaces |
PM | Proposed method |
LSTM | Long Short-Term Memory |
TP | True Positive |
FP | False Positive |
FN | False Negative |
References
- Lindqvist, G. The relationship between play and dance. Res. Danc. Educ. 2001, 2, 41–52. [Google Scholar] [CrossRef]
- Alpert, P.T. The health benefits of dance. Home Health Care Manag. Pract. 2011, 23, 155–157. [Google Scholar] [CrossRef]
- Ward, S.A. Health and the power of dance. J. Phys. Educ. Recreat. Danc. 2008, 79, 33–36. [Google Scholar] [CrossRef]
- Huddy, A.; Stevens, K. The teaching artist: A model for university dance teacher training. Res. Danc. Educ. 2011, 12, 157–171. [Google Scholar] [CrossRef]
- Pedro, R.; Stevens, K.; Scheu, C. Creating a cultural dance community of practice: Building authentic Latin American dance experiences. Res. Danc. Educ. 2018, 19, 199–215. [Google Scholar] [CrossRef]
- Green, J. Power, service, and reflexivity in a community dance project. Res. Danc. Educ. 2000, 1, 53–67. [Google Scholar] [CrossRef]
- Olvera, A.E. Cultural dance and health: A review of the literature. Am. J. Health Educ. 2008, 39, 353–359. [Google Scholar] [CrossRef]
- Hast, D.E. Performance, transformation, and community: Contra dance in New England. Danc. Res. J. 1993, 25, 21–32. [Google Scholar] [CrossRef]
- Jackson, J.B. The opposite of powwow: Ignoring and incorporating the intertribal war dance in the Oklahoma stomp dance community. Plains Anthropol. 2003, 48, 237–253. [Google Scholar] [CrossRef]
- Van Rossum, J.H. The dance teacher: The ideal case and daily reality. J. Educ. Gift. 2004, 28, 36–55. [Google Scholar] [CrossRef] [Green Version]
- Hong, J.C.; Chen, M.L.; Ye, J.H. Acceptance of YouTube applied to dance learning. Int. J. Inf. Educ. Technol. 2020, 10, 7–13. [Google Scholar] [CrossRef] [Green Version]
- Akiba, F. YOSAKOI SORAN as a site of re-localization and its relationship to Japanese pop culture. In Proceedings of the 13th World Congress of the International Association for Semiotic Studies (IASS/AIS), Kaunas, Lithuania, 26–30 June 2017; pp. 653–661. [Google Scholar]
- Chan, J.C.; Leung, H.; Tang, J.K.; Komura, T. A virtual reality dance training system using motion capture technology. IEEE Trans. Learn. Technol. 2011, 4, 187–195. [Google Scholar] [CrossRef]
- Hachimura, K.; Kato, H.; Tamura, H. A prototype dance training support system with motion capture and mixed reality technologies. In Proceedings of the 13th IEEE International Workshop on Robot and Human Interactive Communication, Kurashiki, Japan, 22–24 September 2004; pp. 217–222. [Google Scholar]
- Shiratori, T.; Nakazawa, A.; Ikeuchi, K. Detecting dance motion structure using motion capture and musical information. In Proceedings of the 10th International Conference on Virtual Systems and Multimedia, Ogaki, Japan, 17–19 November 2004; Volume 4, pp. 1287–1296. [Google Scholar]
- Nakazawa, A.; Nakaoka, S.; Ikeuchi, K.; Yokoi, K. Imitating human dance motions through motion structure analysis. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Lausanne, Switzerland, 30 September–4 October 2002; Volume 3, pp. 2539–2544. [Google Scholar]
- Yoshimura, M.; Murasato, H.; Kai, T.; Kuromiya, A.; Yokoyama, K.; Hachimura, K. Analysis of Japanese dance movements using motion capture system. Syst. Comput. Jpn. 2006, 37, 71–82. [Google Scholar] [CrossRef]
- Solberg, R.T.; Jensenius, A.R. Optical or inertial? Evaluation of two motion capture systems for studies of dancing to electronic dance music. In Proceedings of the Systems, Man, and Cybernetics Conferences, Budapest, Hungary, 9–12 October 2016; pp. 469–474. [Google Scholar]
- Camurri, A.; El Raheb, K.; Even-Zohar, O.; Ioannidis, Y.; Markatzi, A.; Matos, J.M.; Morley-Fletcher, E.; Palacio, P.; Romero, M.; Sarti, A.; et al. WhoLoDancE: Towards a methodology for selecting motion capture data across different dance learning practice. In Proceedings of the 3rd International Symposium on Movement and Computing, Thessaloniki, Greece, 5–6 July 2016; pp. 1–2. [Google Scholar]
- Aristidou, A.; Stavrakis, E.; Charalambous, P.; Chrysanthou, Y.; Himona, S.L. Folk dance evaluation using laban movement analysis. ACM J. Comput. Cult. Herit. 2015, 8, 1–19. [Google Scholar] [CrossRef]
- Wang, Z. Modern social dance teaching approaches: Studying creative and communicative components. Think. Ski. Creat. 2022, 43, 100974. [Google Scholar] [CrossRef]
- Patrona, F.; Chatzitofis, A.; Zarpalas, D.; Daras, P. Motion analysis: Action detection, recognition and evaluation based on motion capture data. Pattern Recognit. 2018, 76, 612–622. [Google Scholar] [CrossRef]
- Zhang, Z.; Li, J.; Manikopoulos, C.; Jorgenson, J.; Ucles, J. HIDE: A hierarchical network intrusion detection system using statistical preprocessing and neural network classification. In Proceedings of the IEEE Workshop on Information Assurance and Security, St. Petersburg, Russia, 21–23 May 2001; Volume 85, p. 90. [Google Scholar]
- Kim, H.S.; Hong, N.; Kim, M.; Yoon, S.G.; Yu, H.W.; Kong, H.J.; Kim, S.J.; Chai, Y.J.; Choi, H.J.; Choi, J.Y.; et al. Application of a perception neuron® system in simulation-based surgical training. J. Clin. Med. 2019, 8, 124. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Akamatsu, Y.; Maeda, K.; Ogawa, T.; Haseyama, M. Classification of expert-novice level using eye tracking and motion data via conditional multimodal variational autoencoder. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 1360–1364. [Google Scholar]
- Box, G.E.; Pierce, D.A. Distribution of residual autocorrelations in autoregressive-integrated moving average time series models. J. Am. Stat. Assoc. 1970, 65, 1509–1526. [Google Scholar] [CrossRef]
- Agarap, A.F. Deep learning using rectified linear units (ReLU). arXiv 2018, arXiv:1803.08375. [Google Scholar]
- Memisevic, R.; Zach, C.; Pollefeys, M.; Hinton, G.E. Gated softmax classification. Adv. Neural Inf. Process. Syst. 2010, 23, 1603–1611. [Google Scholar]
- Kline, D.M.; Berardi, V.L. Revisiting squared-error and cross-entropy functions for training neural network classifiers. Neural Comput. Appl. 2005, 14, 310–318. [Google Scholar] [CrossRef]
- Lin, C.Y.; Yang, Z.H.; Zhou, H.W.; Yang, T.N.; Chen, H.N.; Shih, T.K. Combining leap motion with unity for virtual glove puppets. In Proceedings of the IEEE International Conference on Artificial Intelligence and Virtual Reality (AIVR), Taichung, Taiwan, 10–12 December 2018; pp. 251–255. [Google Scholar]
- Guo, H.; Sung, Y. Movement estimation using soft sensors based on Bi-LSTM and two-layer LSTM for human motion capture. Sensors 2020, 20, 1801. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Siami-Namini, S.; Tavakoli, N.; Namin, A.S. A comparison of ARIMA and LSTM in forecasting time series. In Proceedings of the 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 1394–1401. [Google Scholar]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef] [Green Version]
- Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
- Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.E.; Sheikh, Y. OpenPose: Realtime multi-person 2D pose estimation using part affinity fields. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 172–186. [Google Scholar] [CrossRef] [Green Version]
- Chen, C.H.; Ramanan, D. 3d human pose estimation= 2d pose estimation+ matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7035–7043. [Google Scholar]
- Wang, L.; Chen, Y.; Guo, Z.; Qian, K.; Lin, M.; Li, H.; Ren, J.S. Generalizing monocular 3d human pose estimation in the wild. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Seoul, Korea, 27 October–2 November 2019. [Google Scholar]
Types of Stops | Details |
---|---|
Short stop | Operation stop time is about 25 frames. |
Normal stop | Operation stop time is about 50 frames. |
Long stop | Operation is stopped for a while. |
UI | Details |
---|---|
Now Frame | Display the current frame. |
Load (Student) | Read the dance movement data of the expert dancer. |
Load (Teacher) | Read the performer’s dance movement data. |
Start | Start playback of dance movement data. |
Stop | Pause playback of dance movement data. |
Finish | End playback of dance movement data. |
0.5× | Play back dance movement data at 0.5× speed. |
−0.5× | Rewind dance movement data at 0.5× speed. |
1× | Play dance movement data at 1× speed |
−1× | Rewind dance movement data at 1× speed |
2× | Play back dance movement data at 2× speed. |
−2× | Rewind dance movement data at 2× speed. |
Subject | Gender | Age | Height | Dance Experience | No. of Samples |
---|---|---|---|---|---|
A | Male | 23 years old | 164 cm | 9 years | 128 |
B1 | Male | 23 years old | 168 cm | 15 years | 135 |
B2 | Male | 23 years old | 168 cm | 15 years | 138 |
B3 | Male | 23 years old | 168 cm | 15 years | 135 |
C | Female | 36 years old | 164 cm | 19 years | 143 |
Model | Data | Input Size | No. of Hidden Layers | No. of Hidden Nodes | Output Size | |
---|---|---|---|---|---|---|
PM | NN | Feature data | 4 | 1 | 16 | 4 |
Comp. 1 | NN | Time-series data | 102 | 1 | 128 | 4 |
Comp. 2 | LSTM [32] | Feature data | 2 × 2 | 1 | 16 | 4 |
Comp. 3 | LSTM [32] | Time-series data | 51 × 2 | 1 | 128 | 4 |
Model | Data | Input Size | Kernel | Output Size | ||
Comp. 4 | Nonlinear Support Vector Machine [33] | Feature data | 4 | Radial basis function | 4 | |
Model | Data | Input Size | No. of Neighbors | Output Size | ||
Comp. 5 | k-Nearest Neighbor [34] | Feature data | 4 | 5 | 1 | |
Model | Data | Input Size | No. of Trees in the Forest | Output Size | ||
Comp. 6 | Random Forest [35] | Feature data | 4 | 115 | 4 |
Training | Verification | Testing | |
---|---|---|---|
short stop | 50 | 10 | 10 |
normal stop | 51 | 12 | 12 |
long stop | 9 | 3 | 3 |
no stop | 363 | 78 | 78 |
Total | 473 | 103 | 103 |
Precision | Recall | F-Measure | |
---|---|---|---|
PM | 0.938 | 0.600 | 0.732 |
Comp. 1 | 1.000 | 0.400 | 0.571 |
Comp. 2 | 0.813 | 0.520 | 0.634 |
Comp. 3 | - | 0 | - |
Comp. 4 | 0.698 | 0.539 | 0.546 |
Comp. 5 | 0.715 | 0.472 | 0.536 |
Comp. 6 | 0.641 | 0.575 | 0.558 |
No. of Stop Detections | No. of Nondetections | No. of False Positives | |
---|---|---|---|
PM | 11 | 21 | 4 |
Comp. 1 | 6 | 26 | 1 |
Comp. 2 | 10 | 22 | 3 |
Comp. 3 | 0 | 32 | 0 |
Comp. 4 | 2 | 30 | 3 |
Comp. 5 | 11 | 21 | 12 |
Comp. 6 | 8 | 24 | 12 |
Literature | No. of Subjects/Dance Genres | Analysis Examples | Application for Motion Visualization |
---|---|---|---|
PM | 5 / Yosakoi (JPN) | NN-based stop detection | Highlighting a teacher and a student stop with VR |
Chan et al. [13] | 6 / Hip-hop and a-go-go (USA) | Motion matching from motion database | Highlighting incorrect movement joints with VR |
Hachimura et al. [14] | 5 / Street dance (USA) | - | Overlay of the computer graphics characteristics of a trainer with AR |
Shiratori et al. [15] | 2 / Aizu-bandaisan (JPN) | Segmentation of motion sequence based on the music rhythm | - |
Yoshimura et al. [17] | 5 / Fuji Musume (JPN) | Proposal of coordinate system considering local moving for motion tracking | - |
Aristidou et al. [20] | 3 / Bachatta dance (DMA) | Proposal of Laban Movement Analysis motion features for Laban | Only playback of tracked motion with VR |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Jin, Y.; Suzuki, G.; Shioya, H. Detecting and Visualizing Stops in Dance Training by Neural Network Based on Velocity and Acceleration. Sensors 2022, 22, 5402. https://doi.org/10.3390/s22145402
Jin Y, Suzuki G, Shioya H. Detecting and Visualizing Stops in Dance Training by Neural Network Based on Velocity and Acceleration. Sensors. 2022; 22(14):5402. https://doi.org/10.3390/s22145402
Chicago/Turabian StyleJin, Yuuki, Genki Suzuki, and Hiroyuki Shioya. 2022. "Detecting and Visualizing Stops in Dance Training by Neural Network Based on Velocity and Acceleration" Sensors 22, no. 14: 5402. https://doi.org/10.3390/s22145402