当前位置: X-MOL 学术Front. Comput. Neurosci. › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
Towards Iso-Accuracy on Transformer-based Deep Neural Networks with Analog Memory Devices
Frontiers in Computational Neuroscience ( IF 3.2 ) Pub Date : 2021-05-14 , DOI: 10.3389/fncom.2021.675741
Katie Spoon 1 , Hsinyu Tsai 1 , An Chen 1 , Malte J Rasch 2 , Stefano Ambrogio 1 , Charles Mackin 1 , Andrea Fasoli 1 , Alexander M Friz 1 , Pritish Narayanan 1 , Milos Stanisavljevic 3 , Geoffrey W Burr 1
Affiliation  

Recent advances in deep learning have been driven by everincreasing model sizes, with networks growing to millions or even billions of parameters. Such enormous models call for fast and energy-efficient hardware accelerators. We study the potential of Analog AI accelerators based on Non-Volatile Memory, in particular Phase Change Memory, for iso-accurate inference of natural language processing applications. We demonstrate a path to iso-accuracy for the GLUE benchmark on BERT (Bidirectional Encoder Representations from Transformers), by combining noise-aware training to combat inherent PCM drift and noise sources, together with reduced-precision digital attention-block computation down to INT6.

中文翻译:

在具有模拟存储设备的基于变压器的深度神经网络上实现等精度

不断增长的模型规模推动了深度学习的最新进展,网络规模增长到数百万甚至数十亿。如此庞大的模型需要快速且节能的硬件加速器。我们研究了基于非易失性存储器(尤其是相变存储器)的模拟AI加速器对自然语言处理应用程序进行等精度推断的潜力。我们通过结合噪声感知训练来对抗固有的PCM漂移和噪声源以及降低精度的数字注意块计算(直至INT6),证明了BERT(来自变压器的双向编码器表示)上GLUE基准测试的等精度途径。 。
更新日期:2021-05-14
down
wechat
bug