This work presents AIDA, an associative in-memory deep learning processor for edge devices. An associative processor is a massively parallel non-von Neumann accelerator that uses memory cells for computing; the bulk of data is never transferred outside the memory arrays for external processing. AIDA utilizes a dynamic content addressable memory for both data storage and processing, and benefits from sparsity and limited arithmetic precision, typical in modern deep neural networks (DNNs). The novel in-data processing implementation designed for the AIDA accelerator achieves a speedup of 270× over an advanced CPU at more than three orders-of-magnitude better energy efficiency.
AIDA: Associative In-memory Deep learning Accelerator
Garzon Esteban
;Lanuzza M.;
2022-01-01
Abstract
This work presents AIDA, an associative in-memory deep learning processor for edge devices. An associative processor is a massively parallel non-von Neumann accelerator that uses memory cells for computing; the bulk of data is never transferred outside the memory arrays for external processing. AIDA utilizes a dynamic content addressable memory for both data storage and processing, and benefits from sparsity and limited arithmetic precision, typical in modern deep neural networks (DNNs). The novel in-data processing implementation designed for the AIDA accelerator achieves a speedup of 270× over an advanced CPU at more than three orders-of-magnitude better energy efficiency.I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.