Horowitz, M. Computing’s Energy Problem (And What We Can Do About It). In Proc. International Solid State Circuit Conference (ISSCC) 10-14 (IEEE, 2014).
Keckler, SW, Dally, WJ, Khailany, B., Garland, M. & Glasco, D. GPUs and the future of parallel computing. IEEE micro 31, 7-17 (2011).
Lied, J. et al. An 11.5 TOPS/W 1024-MAC dual-core sparsity-aware butterfly neural processing engine in a flagship 8nm mobile SoC. In 2019 IEEE Int. Solid-State Circuits Conference Digest of Technical Papers (ISSCC) 130-131 (IEEE, 2019).
Sebastian, A. et al. Storage devices and applications for in-memory computing. nat. Nanotechnology. fifteen, 529–544 (2020).
Wang, Z. et al. Resistance switching materials for information processing. nat. Rev Mater. 5, 173–195 (2020).
Ielmini, D. & Wong, HP In-Memory Computing with Resistive Switching Devices. nat. Electron. 1, 333–343 (2018).
Verma, N. et al. In-Memory Computing: Advances and Prospects. IEEE Solid State Circuits Mag. 11, 43–55 (2019).
Woo, J. et al. Improved synaptic behavior under identical pulses with AlOx/HfO2 two-layer RRAM array for neuromorphic systems. IEEE Electron Device Lett. 37, 994–997 (2016).
Yao, P. et al. Face classification with electronic synapses. nat. commune 8th, 15199 (2017).
Wu, H. et al. Device and circuit optimization of RRAM for neuromorphic computing. In 2017 IEEE International Electron Devices Meeting 11.5.1-11.5.4 (IEEE, 2017).
Li, C. et al. Efficient and Self-Adaptive In Situ Learning in Multilayer Memristor Neural Networks. nat. commune 9, 2385 (2018).
Chen, W. et al. CMOS-integrated memristive non-volatile computing-in-memory for AI edge processors. nat. Electron. 2, 420–428 (2019).
Yao, P. et al. Fully hardware implemented memristor convolutional neural network. nature 577, 641–646 (2020).
Le Gallo, M. et al. Mixed precision in-memory computing. nat. Electron. 1, 246–253 (2018).
Ambrogio, S. et al. Accelerated neural network training with equivalent precision using analog memory. nature 558, 60–67 (2018).
Merrikh-Bayat, F. et al. High performance mixed signal neurocomputing with nanoscale floating gate memory cell arrays. IEEE Trans Neural Network. To learn. system 29, 4782-4790 (2018).
Wang, P. et al. Three-dimensional NAND flash for vector matrix multiplication. IEEE Trans. VLSI Syst. 27, 988–991 (2019).
Xiang, Y. et al. Efficient and Robust Spike Driven Deep Convolutional Neural Networks based on NOR Flash Computing Array. IEEE Trans. electron developer 67, 2329–2335 (2020).
Lin, Y.-Y. et al. A novel vector matrix multiplication architecture with voltage accumulation using a resistor shunt floating gate flash memory device for low power, high density neural network applications. In 2018 IEEE International Electron Devices Meeting 2.4.1−2.4.4 (IEEE, 2018).
Song, YJ et al. Demonstration of highly manufacturable STT MRAM embedded in 28nm logic. In 2018 IEEE International Electron Devices Meeting 18.2.1-18.2.4 (IEEE, 2018).
Lee, YK et al. Embedded STT MRAM in 28nm FDSOI logic process for industrial MCU/IoT application. In 2018 IEEE Symposium on VLSI Technology 181-182 (IEEE, 2018).
Wei, L. et al. A 7MB STT MRAM in 22FFL FinFET technology with 4ns read capture time at 0.9V using write-verify-write scheme and offset cancellation capture technique. In 2019 IEEE Int. Solid-State Circuits Conference Digest of Technical Papers 214-216 (IEEE, 2019).
LeCun Y, Bengio Y & Hinton G Deep Learning. nature 521, 436–444 (2015).
Yu, S. Neuro-Inspired Computing with New Non-Volatile Memory. Proc.IEEE 106, 260–285 (2018).
Patil, AD et al. An MRAM-based deep-in-memory architecture for deep neural networks. In 2019 International IEEE Symposium on Circuits and Systems (IEEE, 2019).
Zabihi, M. et al. In-memory processing on the spintronics CRAM: from hardware design to application mapping. IEEE Trans. Calculation. 68, 1159–1173 (2019).
Kang, SH Embedded STT-MRAM for low-power and low-cost mobile systems. In 2014 IEEE Symposium on VLSI Technology (IEEE, 2014).
Zeng, ZM et al. Effect of resistivity area product on spin-transfer switching in MgO-based magnetic tunnel junction memory cells. appl. physics Latvian. 98, 072512 (2011).
Kim, H. & Kwon, S.-W. Full-precision neural network approximation based on time-domain binary MAC operations. U.S. Patent 17/085,300.
Hung, J.-M. et al. Challenges and trends in the development of non-volatile memory-capable computer chips for intelligent edge devices. IEEE Trans. electron developer 67, 1444–1453 (2020).
Jiang, Z., Yin, S., Seo, J. & Seok, M. C3SRAM: an in-memory computing SRAM macro based on a robust capacitive-coupling computing mechanism. IEEE J. Solid State Circuits 55, 1888–1897 (2020).
Hubara, I. et al. Binary Neural Networks. In Advances in neural information processing systems 4107-4115 (NeurIPS, 2016).
Rastegari, M., Ordonez, V., Redmon, J. & Farhadi, A. XNOR-Net: ImageNet classification using binary convolutional neural networks. In 2016 European Conference on Computer Vision 525-542 (2016).
Lin, X., Zhao, C. & Pan, W. Towards an accurate binary convolutional neural network. In Advances in neural information processing systems 345-353 (NeurIPS, 2017).
Zhuang, B. et al. Structured binary neural networks for accurate image classification and semantic segmentation. In 2019 IEEE Conference on Computer Vision and Pattern Recognition 413-422 (IEEE, 2019).
Shafiee, A. et al. ISAAC: a convolutional neural network accelerator with in situ analog arithmetic in crossbars. In 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture 14-26 (IEEE, 2016).
Liu, B. et al. Digitally assisted noise reduction training for memristor crossbar based analog neuromorphic calculator. In 2013 50th ACM/EDAC/IEEE Design Automation Conference 1-6 (IEEE, 2013).
Wu B, Iandola F, Jin PH & Keutzer K SqueezeDet: Unified, Small, Low Power, Fully Convolutional Neural Networks for Real-Time Object Detection for Autonomous Driving. In 2017 IEEE Conference on Computer Vision and Pattern Recognition 129-137 (IEEE, 2017).
Ham, D., Park, H., Hwang, S. & Kim, K. Neuromorphic electronics based on brain copy-paste. nat. Electron. 4, 635–644 (2021).
Wang, P. et al. Two-level quantization for low-bit neural networks. In 2018 IEEE Conference on Computer Vision and Pattern Recognition 4376-4384 (IEEE, 2018).