Nothing Special   »   [go: up one dir, main page]

skip to main content
10.1145/3501409.3501526acmotherconferencesArticle/Chapter ViewAbstractPublication PageseitceConference Proceedingsconference-collections
research-article

Design of a Novel Neural Network Compression Method for Tiny Machine Learning

Published: 31 December 2021 Publication History

Abstract

Traditional IoT processing data is sent from local devices to the cloud for processing, which has disadvantages such as low privacy, high latency, and low energy efficiency. These drawbacks can be effectively remedied by deploying the model for processing on devices at the "edge" of the cloud. In order to realize the data being processed directly on the cloud "edge" devices, the original machine learning algorithms need to be improved. One of the important steps is neural network compression.
In this paper, a neural network compression method for Tiny Machine Learning (TinyML) is proposed. The neural network is compressed by training a conventional neural network and then performing group convolution, pruning and asymmetric ternary quantization. In the next step, a model transformation is performed using TFLite to deploy it on embedded devices. With this novel neural network compression method, the size of the model can be greatly compressed with guaranteed accuracy. Consequently, the traditional machine learning is upgraded to TinyML, and finally a TinyML-based fall monitoring system for the elderly is built.

References

[1]
C.j.abate, Wo Mu. The future of machine learning_Daniel_Situnayake interview[J]. Microcontroller and Embedded Systems Applications, 2021, 21(6): 1--3, 6.
[2]
Wang L, Zhao YH, Yang GSH, et al. A review of deep neural network model compression techniques for embedded applications[J]. Journal of Beijing Jiaotong University, 2017, 41(6): 34--41.
[3]
Jin Lilei, Yang Wenzhu, Wang Sile, et al. A hybrid pruning method for compression of convolutional neural networks[J]. Small Microcomputer Systems, 2018, 39(12): 2596--2601.
[4]
Su Loach. Research and application of lightweight target detection algorithm based on deep learning[D]. South China University of Technology, 2020.
[5]
Ding J. Research on model compression and forward acceleration techniques for embedded deep neural networks [D]. University of Science and Technology of China, 2018.
[6]
Dehghanian M, Mosadegh MSM. Ternary Weighted Function and Beurling Ternary Banach Algebra L(1)(omega) (s) [J]. Abstract and Applied Analysis, 2011.
[7]
Iandola FN, Han S, Moskewicz MW, et al. Squeezenet: Alexnet-level Accuracy with 50x Fewer Parameters and <0.5mb Model Size[J], 2016.
[8]
Li D, Wang X, Kong D. Deeprebirth: Accelerating Deep Neural Network Execution on Mobile Devices[J], 2017.
[9]
Zhang X, Zhou X, Lin M, et al. Shufflenet: an Extremely Efficient Convolutional Neural Network for Mobile Devices[J], 2017.
[10]
Stauffer C, Grimson W E L. Adaptive background mixture models for real-time tracking[C]//Proceedings. 1999 IEEE computer society conference on computer vision and pattern recognition (Cat. No PR00149). IEEE, 1999, 2: 246--252.
[11]
Barnich O, Van Droogenbroeck M. ViBe: a powerful random technique to estimate the background in video sequences[C]//2009 IEEE international conference on acoustics, speech and signal processing. IEEE, 2009: 945--948.
[12]
Glorot x, Bengio-Y. Understanding the difficulty of training deep feedforwardneural networks [C]. Proceedings of the Thirteenth International Conference onArtificial Interlligence and Statistics, 2010, 249--256.
[13]
Chollet F. Xception: Deep Learning with Depthwise Separable Convolutions[C]//2017 {ieee} Conference on Computer Vision and Pattern Recognition ({cvpr}), [S.l.]: Institute of Electrical and Electronics Engineers Inc. 2017.
[14]
Ioffe S, Szegedy C. Batch normalization: Accelerating deep network training by reducing internal covariate shift[C]//International conference on machine learning. PMLR, 2015: 448--456.

Cited By

View all
  • (2024)A comprehensive review of model compression techniques in machine learningApplied Intelligence10.1007/s10489-024-05747-w54:22(11804-11844)Online publication date: 2-Sep-2024
  • (2022)Transfer Learning for Convolutional Neural Networks in Tiny Deep Learning EnvironmentsProceedings of the 26th Pan-Hellenic Conference on Informatics10.1145/3575879.3575984(145-150)Online publication date: 25-Nov-2022

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image ACM Other conferences
EITCE '21: Proceedings of the 2021 5th International Conference on Electronic Information Technology and Computer Engineering
October 2021
1723 pages
ISBN:9781450384322
DOI:10.1145/3501409
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 31 December 2021

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. Compression Neural Networks
  2. Fall Monitoring
  3. Tiny Machine Learning

Qualifiers

  • Research-article
  • Research
  • Refereed limited

Conference

EITCE 2021

Acceptance Rates

EITCE '21 Paper Acceptance Rate 294 of 531 submissions, 55%;
Overall Acceptance Rate 508 of 972 submissions, 52%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)25
  • Downloads (Last 6 weeks)3
Reflects downloads up to 18 Nov 2024

Other Metrics

Citations

Cited By

View all
  • (2024)A comprehensive review of model compression techniques in machine learningApplied Intelligence10.1007/s10489-024-05747-w54:22(11804-11844)Online publication date: 2-Sep-2024
  • (2022)Transfer Learning for Convolutional Neural Networks in Tiny Deep Learning EnvironmentsProceedings of the 26th Pan-Hellenic Conference on Informatics10.1145/3575879.3575984(145-150)Online publication date: 25-Nov-2022

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Media

Figures

Other

Tables

Share

Share

Share this Publication link

Share on social media