An Energy-Efficient Near-Memory Computing Architecture for CNN Inference at Cache Level

Masoud Nouripayam, Arturo Prieto, Vignajeth Kuttuva Kishorelal, Joachim Rodrigues

Research output: Chapter in Book/Report/Conference proceedingPaper in conference proceedingpeer-review

Abstract

A non-von Neumann Near-Memory Computing architecture, optimized for CNN inference in edge computing, is integrated in the cache memory sub-system of a microcontroller unit. The NMC co-processor is evaluated using an 8-bit fixed-point quantized CNN model, and achieves an accuracy of 98% on the MNIST dataset. A full inference of the CNN model executed on the NMC processor, demonstrates an improvement of more than 34× in performance, and 28× in energy-efficiency, compared to the baseline scenario of a conventional single-core processor. The design achieves a performance of 1.39 GOPS (at 200 MHz) and an energy-efficiency of 49 GOPS/W, with negligible area overhead of less than 1%.
Original languageEnglish
Title of host publication2021 28th IEEE International Conference on Electronics, Circuits, and Systems (ICECS)
PublisherIEEE - Institute of Electrical and Electronics Engineers Inc.
Pages1-4
ISBN (Electronic)978-1-7281-8281-0
ISBN (Print)978-1-7281-9493-6
DOIs
Publication statusPublished - 2021 Dec

Subject classification (UKÄ)

  • Computer Engineering
  • Computer Systems

Fingerprint

Dive into the research topics of 'An Energy-Efficient Near-Memory Computing Architecture for CNN Inference at Cache Level'. Together they form a unique fingerprint.

Cite this