IMCA: An Efficient In-Memory Convolution Accelerator

Hasan Erdem Yantir, Ahmed Eltawil, Khaled N. Salama

Research output: Contribution to journalArticlepeer-review

7 Scopus citations


Traditional convolutional neural network (CNN) architectures suffer from two bottlenecks: computational complexity and memory access cost. In this study, an efficient in-memory convolution accelerator (IMCA) is proposed based on associative in-memory processing to alleviate these two problems directly. In the IMCA, the convolution operations are directly performed inside the memory as in-place operations. The proposed memory computational structure allows for a significant improvement in computational metrics, namely, TOPS/W. Furthermore, due to its unconventional computation style, the IMCA can take advantage of many potential opportunities, such as constant multiplication, bit-level sparsity, and dynamic approximate computing, which, while supported by traditional architectures, require extra overhead to exploit, thus reducing any potential gains. The proposed accelerator architecture exhibits a significant efficiency in terms of area and performance, achieving around 0.65 GOPS and 1.64 TOPS/W at 16-bit fixed-point precision with an area less than 0.25 mm².
Original languageEnglish (US)
Pages (from-to)1-14
Number of pages14
JournalIEEE Transactions on Very Large Scale Integration (VLSI) Systems
StatePublished - 2021

Bibliographical note

KAUST Repository Item: Exported on 2021-01-21


Dive into the research topics of 'IMCA: An Efficient In-Memory Convolution Accelerator'. Together they form a unique fingerprint.

Cite this