skip to main content
10.1145/3575879.3575984acmotherconferencesArticle/Chapter ViewAbstractPublication PagespciConference Proceedingsconference-collections
research-article
Open Access

Transfer Learning for Convolutional Neural Networks in Tiny Deep Learning Environments

Published:29 March 2023Publication History

ABSTRACT

Tiny Machine Learning (TinyML) and Transfer Learning (TL) are two widespread methods of successfully deploying ML models to resource-starving devices. Tiny ML provides compact models, that can run on resource-constrained environments, while TL contributes to the performance of the model by using pre-existing knowledge. So, in this work we propose a simple but efficient TL method, applied to three types of Convolutional Neural Networks (CNN), by retraining more than the last fully connected layer of a CNN in the target device, and specifically one or more of the last convolutional layers. Our results shown that our proposed method (FxC1) achieves about increase in accuracy and increase in convergence speed, while it incurs a bit larger energy consumption overhead, compared to two baseline techniques, namely one that retrains the last fully connected layer, and another that retrains the whole network.

References

  1. Colby R. Banbury, Vijay Janapa Reddi, Maximilian Lam, William Fu, Amin Fazel, Jeremy Holleman, Xinyuan Huang, Robert Hurtado, David Kanter, Anton Lokhmotov, David A. Patterson, Danilo Pau, Jae sun Seo, Jeff Sieracki, Urmish Thakker, Marian Verhelst, and Poonam Yadav. 2020. Benchmarking TinyML Systems: Challenges and Direction. ArXiv abs/2003.04821(2020).Google ScholarGoogle Scholar
  2. PUNEET BANSAL. [n. d.]. Intel Image Classification. https://www.kaggle.com/datasets/puneet6060/intel-image-classification/discussionGoogle ScholarGoogle Scholar
  3. Andreas Chouliaras, Evangelia Fragkou, and Dimitrios Katsaros. 2021. Feed Forward Neural Network Sparsificationwith Dynamic Pruning. In 25th Pan-Hellenic Conference on Informatics (Volos, Greece) (PCI 2021). Association for Computing Machinery, New York, NY, USA, 12–17. https://doi.org/10.1145/3503823.3503826Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Oscar Day and Taghi M. Khoshgoftaar. 2017. A survey on heterogeneous transfer learning. Journal of Big Data 4(2017), 1–42.Google ScholarGoogle ScholarCross RefCross Ref
  5. Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li, and Li Fei-Fei. 2009. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition. Ieee, 248–255.Google ScholarGoogle Scholar
  6. Simone Disabato and Manuel Roveri. 2020. Incremental On-Device Tiny Machine Learning. Proceedings of the 2nd International Workshop on Challenges in Artificial Intelligence and Machine Learning for Internet of Things (2020).Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Evangelia Fragkou, Marianna Koultouki, and Dimitrios Katsaros. 2022. Model reduction of feed forward neural networks for resource-constrained devices. Applied Intelligence(2022).Google ScholarGoogle Scholar
  8. Yaroslav Ganin and Victor S. Lempitsky. 2015. Unsupervised Domain Adaptation by Backpropagation. ArXiv abs/1409.7495(2015).Google ScholarGoogle Scholar
  9. Yaroslav Ganin, E. Ustinova, Hana Ajakan, Pascal Germain, H. Larochelle, François Laviolette, Mario Marchand, and Victor S. Lempitsky. 2016. Domain-Adversarial Training of Neural Networks. In J. Mach. Learn. Res.Google ScholarGoogle Scholar
  10. Sedigh Ghamari, Koray Ozcan, Thu Dinh, Andrey Melnikov, Juan Carvajal, Jan Ernst, and Sek M. Chai. 2021. Quantization-Guided Training for Compact TinyML Models. ArXiv abs/2103.06231(2021).Google ScholarGoogle Scholar
  11. Lennart Heim, Andreas Biri, Zhongnan Qu, and Lothar Thiele. 2021. Measuring what Really Matters: Optimizing Neural Networks for TinyML. ArXiv abs/2104.10645(2021).Google ScholarGoogle Scholar
  12. Kavya Kopparapu and Eric Lin. 2021. TinyFedTL: Federated Transfer Learning on Tiny Devices. ArXiv abs/2110.01107(2021).Google ScholarGoogle Scholar
  13. Alex Krizhevsky, Vinod Nair, and Geoffrey Hinton. [n. d.]. CIFAR10 and CIFAR100. https://www.cs.toronto.edu/ kriz/cifar.htmlGoogle ScholarGoogle Scholar
  14. Jisu Kwon and Daejin Park. 2021. Toward Data-Adaptable TinyML using Model Partial Replacement for Resource Frugal Edge Device. The International Conference on High Performance Computing in Asia-Pacific Region (2021).Google ScholarGoogle ScholarDigital LibraryDigital Library
  15. Liangzhen Lai, Naveen Suda, and Vikas Chandra. 2018. CMSIS-NN: Efficient Neural Network Kernels for Arm Cortex-M CPUs. ArXiv abs/1801.06601(2018).Google ScholarGoogle Scholar
  16. Josen Daniel De Leon and Rowel Atienza. 2022. Depth Pruning with Auxiliary Networks for Tinyml. ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)(2022), 3963–3967.Google ScholarGoogle Scholar
  17. Yingling Li, Zhipeng Li, Tianxing Zhang, Peng Zhou, Siyin Feng, and Kunqin Yin. 2021. Design of a Novel Neural Network Compression Method for Tiny Machine Learning. Proceedings of the 2021 5th International Conference on Electronic Information Technology and Computer Engineering(2021).Google ScholarGoogle ScholarDigital LibraryDigital Library
  18. Edgar Liberis, Lukasz Dudziak, and Nicholas D. Lane. 2021. μNAS: Constrained Neural Architecture Search for Microcontrollers. Proceedings of the 1st Workshop on Machine Learning and Systems (2021).Google ScholarGoogle ScholarDigital LibraryDigital Library
  19. Wei Yang Bryan Lim, Nguyen Cong Luong, Dinh Thai Hoang, Yutao Jiao, Ying-Chang Liang, Qiang Yang, Dusit Tao Niyato, and Chunyan Miao. 2020. Federated Learning in Mobile Edge Networks: A Comprehensive Survey. IEEE Communications Surveys & Tutorials 22 (2020), 2031–2063.Google ScholarGoogle Scholar
  20. Mingsheng Long, Yue Cao, Jianmin Wang, and Michael I. Jordan. 2015. Learning Transferable Features with Deep Adaptation Networks. ArXiv abs/1502.02791(2015).Google ScholarGoogle Scholar
  21. Sinno Jialin Pan and Qiang Yang. 2010. A Survey on Transfer Learning. IEEE Transactions on Knowledge and Data Engineering 22, 10(2010), 1345–1359. https://doi.org/10.1109/TKDE.2009.191Google ScholarGoogle ScholarDigital LibraryDigital Library
  22. Dimitrios Papakostas, Theodoros Kasidakis, Evangelia Fragkou, and Dimitrios Katsaros. 2021. Backbones for Internet of Battlefield Things. In 2021 16th Annual Conference on Wireless On-demand Network Systems and Services Conference (WONS). 1–8. https://doi.org/10.23919/WONS51326.2021.9415560Google ScholarGoogle Scholar
  23. Haoyu Ren, Darko Anicic, and Thomas A. Runkler. 2021. TinyOL: TinyML with Online-Learning on Microcontrollers. 2021 International Joint Conference on Neural Networks (IJCNN) (2021), 1–8.Google ScholarGoogle Scholar
  24. Muhammad Akmal Shafique, Theocharis Theocharides, Vijay Janapa Reddy, and Boris Murmann. 2021. TinyML: Current Progress, Research Challenges, and Future Roadmap. 2021 58th ACM/IEEE Design Automation Conference (DAC) (2021), 1303–1306.Google ScholarGoogle ScholarDigital LibraryDigital Library
  25. Baochen Sun and Kate Saenko. 2016. Deep CORAL: Correlation Alignment for Deep Domain Adaptation. In ECCV Workshops.Google ScholarGoogle Scholar
  26. Maxim Zemlyanikin, Alexander Smorkalov, Tatiana Khanova, Anna Petrovicheva, and Grigory Serebryakov. 2019. 512KiB RAM Is Enough! Live Camera Face Recognition DNN on MCU. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW) (2019), 2493–2500.Google ScholarGoogle ScholarCross RefCross Ref
  27. Fuzhen Zhuang, Zhiyuan Qi, Keyu Duan, Dongbo Xi, Yongchun Zhu, Hengshu Zhu, Hui Xiong, and Qing He. 2021. A Comprehensive Survey on Transfer Learning. Proc. IEEE 109, 1 (2021), 43–76. https://doi.org/10.1109/JPROC.2020.3004555Google ScholarGoogle Scholar

Index Terms

  1. Transfer Learning for Convolutional Neural Networks in Tiny Deep Learning Environments

    Recommendations

    Comments

    Login options

    Check if you have access through your login credentials or your institution to get full access on this article.

    Sign in
    • Published in

      cover image ACM Other conferences
      PCI '22: Proceedings of the 26th Pan-Hellenic Conference on Informatics
      November 2022
      414 pages

      Copyright © 2022 ACM

      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      • Published: 29 March 2023

      Permissions

      Request permissions about this article.

      Request Permissions

      Check for updates

      Qualifiers

      • research-article
      • Research
      • Refereed limited

      Acceptance Rates

      Overall Acceptance Rate190of390submissions,49%
    • Article Metrics

      • Downloads (Last 12 months)77
      • Downloads (Last 6 weeks)14

      Other Metrics

    PDF Format

    View or Download as a PDF file.

    PDF

    eReader

    View online with eReader.

    eReader

    HTML Format

    View this article in HTML Format .

    View HTML Format