Skip to main content
Article
EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for Mobile Vision Applications
arXiv
  • Muhammad Maaz, Mohamed bin Zayed University of Artificial Intelligence
  • Abdelrahman Shaker, Mohamed bin Zayed University of Artificial Intelligence
  • Hisham Cholakkal, Mohamed bin Zayed University of Artificial Intelligence
  • Salman Khan, Australian National University, Australia & Mohamed bin Zayed University of Artificial Intelligence
  • Syed Waqas Zamir, Inception Institute of Artificial Intelligence
  • Rao Anwer, Mohamed bin Zayed University of Artificial Intelligence
  • Fahad Shahbaz Khan, Mohamed bin Zayed University of Artificial Intelligence
Document Type
Article
Abstract

In the pursuit of achieving ever-increasing accuracy, large and complex neural networks are usually developed. Such models demand high computational resources and therefore cannot be deployed on edge devices. It is of great interest to build resource-efficient general purpose networks due to their usefulness in several application areas. In this work, we strive to effectively combine the strengths of both CNN and Transformer models and propose a new efficient hybrid architecture EdgeNeXt. Specifically in EdgeNeXt, we introduce split depth-wise transpose attention (SDTA) encoder that splits input tensors into multiple channel groups and utilizes depth-wise convolution along with self-attention across channel dimensions to implicitly increase the receptive field and encode multi-scale features. Our extensive experiments on classification, detection and segmentation tasks, reveal the merits of the proposed approach, outperforming state-of-the-art methods with comparatively lower compute requirements. Our EdgeNeXt model with 1.3M parameters achieves 71.2% top-1 accuracy on ImageNet-1K, outperforming MobileViT with an absolute gain of 2.2% with 28% reduction in FLOPs. Further, our EdgeNeXt model with 5.6M parameters achieves 79.4% top-1 accuracy on ImageNet-1K. The code and models are publicly available at https://t.ly/_Vu9. © 2022, CC BY.

DOI
10.48550/arXiv.2206.10589
Publication Date
6-21-2022
Keywords
  • Computer vision,
  • Convolutional neural networks,
  • Application area,
  • CNN models,
  • Complex neural networks,
  • Computational resources,
  • Hybrid architectures,
  • Mobile vision,
  • Multiple channels,
  • Resource-efficient,
  • Transformer modeling,
  • Vision applications
Comments

Preprint: arXiv

Archived with thanks to arXiv

Preprint License: CC by 4.0

Uploaded 15 July 2022

Citation Information
M. Maaz, et al, "EdgeNeXt: Efficiently Amalgamated CNN-Transformer Architecture for Mobile Vision Applications", 2022, arXiv:2206.10589