Exploring Cell-Based Neural Architectures for Embedded Systems

Open Access
Authors
Publication date 2021
Host editors
  • M. Kamp
  • I. Koprinska
  • A. Bibal
  • T. Bouadi
  • B. Frénay
  • L. Galárraga
  • J. Oramas
  • L. Adilova
Book title Machine Learning and Principles and Practice of Knowledge Discovery in Databases
Book subtitle International Workshops of ECML PKDD 2021, virtual event, September 13-17, 2021 : proceedings
ISBN
  • 9783030937355
  • 9783030937379
ISBN (electronic)
  • 9783030937362
Series Communications in Computer and Information Science
Event Machine Learning and Principles and Practice of Knowledge Discovery in Databases
Volume | Issue number I
Pages (from-to) 363–374
Publisher Cham: Springer
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
Neural Architectures Search (NAS) methodologies, which automatically discover state-of-the-art neural networks, have seen a growing interest in recent years. One particular group of NAS methodologies searches for small sub-networks called cells, which are then linearly connected to form the complete neural network. The composition of the final neural network, established through the width of the cells and the depth of the connections, is manually designed while being influenced by the available GPU memory. Typically, the hardware architectures targeted in NAS research are powerful, high-end GPUs. Hence, the attention is on creation of a large neural network that will still fit in the GPU, in turn leading to a very high accuracy for the given task. In direct contrast, we exploit the inherent flexibility of cells to create smaller neural networks, with the intention to study their behaviour on resource-constrained embedded systems. We use the cells discovered from Stochastic Neural Architecture Search (SNAS), to explore the effect that the composition of the cell has on various metrics, namely, the number of parameters, accuracy, latency and power usage. The last two metrics are measured on NVIDIA Jetson Nano, an embedded AI computing platform with a small GPU with mere 4GB on-chip memory. When comparing results of our exploration to the original SNAS architecture’s with 97.02% accuracy for the CIFAR-10 dataset, one particular architecture, with only a tenth of original parameters, achieved an accuracy of 96.14%, notably with 15% lower power consumption and ≈3x faster inference time. Furthermore, this model outperforms other architectures, which are designed for edge devices, specifically to reduce the model size. Thus demonstrating that cell-based architectures, with adequate composition, provide efficient models to be deployed on resource-constrained edge devices.
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-3-030-93736-2_28
Downloads
978-3-030-93736-2_28 (Final published version)
Permalink to this page
Back