Sunteți pe pagina 1din 1

Special Session 1: In-Memory Processing for Future Electronics GLSVLSI ’19, May 9–11, 2019, Tysons Corner, VA,

LSVLSI ’19, May 9–11, 2019, Tysons Corner, VA, USA.

Digital and Analog-Mixed-Signal In-Memory Processing in


CMOS SRAM
Akhilesh Jaiswal, Amogh Agrawal, Indranil Chakraborty, Mustafa Fayez Ali, and Kaushik Roy
School of Electrical and Computer Engineering, Purdue University
West Lafayette, IN, USA
{jaiswal,agrawa64,ichakra,ali102,kaushik}@purdue.edu

ACM Reference Format: rows, simultaneously, and read out a desired Boolean func-
Akhilesh Jaiswal, Amogh Agrawal, Indranil Chakraborty, Mustafa Fayez Ali, tion of the constituent words through modified sensing cir-
and Kaushik Roy . 2019. Digital and Analog-Mixed-Signal In-Memory Pro- cuits [1]-[2].
cessing in CMOS SRAM . In Great Lakes Symposium on VLSI 2019 (GLSVLSI (2) Analog-Mixed-Signal Binary Dot-Products: Binary dot
’19), May 9–11, 2019, Tysons Corner, VA, USA. ACM, New York, NY, USA, products comprise of bit-wise XNOR followed by population-
1 page. https://doi.org/10.1145/3299874.3319449
count (pop-count: counting the number of ‘1’s’). The usual
approach here is to enable bulk-bit-wise operation and gen-
1 INTRODUCTION erate an analog voltage that is proportional to the pop-count
Historically, Moore’s Law, has been the key driving force for the in the resultant bit-stream. By sensing the analog voltage
ever increasing compute power on silicon chips. Architecturally, a one can estimate the number of ‘1’s’ thereby accomplishing
key factor fueling the improvement in computing efficiency is due binary dot product operation [5]-[6].
to the availability of large on-chip memory storage in form of SRAM (3) Analog-Mixed-Signal Multi-bit Dot Products: Multi-bit
arrays as a result of shrinking transistor sizes. The historic trend of dot products require analog element-wise multiplication of
smaller transistors and hence larger on-chip memories is slowing two vectors followed by summation of the elements of the
down recently. In sharp contrast to this slowdown, the demand for resulting vector. This summation of multiplications can be
high throughput computations and associated energy-efficiency accomplished in 8T SRAM cells through current-mode com-
has escalated considerably in recent times. Researchers are in active putations wherein input vector voltages can be multiplied by
pursuit of exploring alternate means to improve the computation effective conductance of the transistors through Ohm’s Law
throughput while also lowering the required energy consumption. and the resulting currents can be summed up in accordance
‘In-Memory computing’ has emerged as a promising candidate to to Kirchoff’s Current Law [8].
achieve both increased throughput along with decrease in energy Finally, due to the inherent approximate nature of analog pro-
consumption. SRAM based in-memory computations are of special cessing, analog-mixed-signal schemes are suitable only for class of
interest due to the pervasive presence of large SRAM sub-arrays in error-resilient applications as in machine learning. In contrast, digi-
state-of-the-art processors. One could, therefore, think of these in- tal in-memory computing can cater to a wider class of applications
memory enabled SRAM arrays as on-chip compute engines, thereby, including both error-resilient as well as those that require accurate
exposing the end application to massive levels of parallelism. computations.
2 SRAM BASED IN-MEMORY COMPUTING 3 CONCLUSION
Several proposals achieving various levels of compute parallelism In this paper, we have highlighted three representative approaches
within SRAM cells have been presented in the literature in recent with respect to in-memory computing in SRAM arrays ranging
times [1]-[8]. Interestingly, both digital [1]-[2] as well as analog- from strictly digital simple computations as in bulk bit-wise op-
mixed-signal approaches [3]-[8] have been explored extensively. erations to massively parallel analog-mixed-signal computations
In general, digital approaches lead to accurate computations at the involving accelerated multi-bit dot products estimations. It is im-
cost of reduced compute throughput. On the other hand, analog portant to note that the level of parallelism, the complexity of the
approaches lead to higher levels of data parallelism at the cost of required peripheral circuits and the approximation in computa-
approximate computations due to the inherent nature of analog- tions increases when moving from digital to analog-mixed-signal
processing. Few representative strategies geared toward SRAM in-memory computations.
based in-memory computations are as follows:
(1) Digital Bulk-Bit-wise Computations: The generic approach ACKNOWLEDGMENT
toward bulk-bit-wise operations is to enable two memory C-BRIC, one of six centers in JUMP, sponsored by SRC, DARPA,
Permission to make digital or hard copies of part or all of this work for personal or
NSF, Intel Corporation and Vannevar Bush Fellowship.
classroom use is granted without fee provided that copies are not made or distributed
for profit or commercial advantage and that copies bear this notice and the full citation REFERENCES
on the first page. Copyrights for third-party components of this work must be honored.
For all other uses, contact the owner/author(s). [1] S. Jeloka, et. al. JSSCC, 2018. [2] A. Aggrawal et. al. TCAS-I, 2018.
GLSVLSI ’19, May 9–11, 2019, Tysons Corner, VA, USA [3] S.K. Gonugondla, et. al. ISSCC, 2018. [4] J. Zhang et. al. JSSCC,
© 2019 Copyright held by the owner/author(s).
ACM ISBN 978-1-4503-6252-8/19/05.
2017. [5] A. Biswas, et. al. ISSCC, 2018. [6] A. Agrawal et. al. arxiv,
https://doi.org/10.1145/3299874.3319449 2018. [7] Z. Jiang et. al. VLSI, 2018. [8] A. Jaiswal, et. al. arxiv, 2018.

371

S-ar putea să vă placă și