计算机科学
注册存储器
交错存储器
半导体存储器
平面存储模型
静态随机存取存储器
通用存储器
非易失性随机存取存储器
内存计算
嵌入式系统
内存管理
内存带宽
内存刷新
计算机存储器
动态随机存取存储器
德拉姆
内存映射
计算机体系结构
并行计算
计算机硬件
作者
Dimin Niu,Shuangchen Li,Yuhao Wang,Wei Han,Zhe Zhang,Yijin Guan,Tianchan Guan,Fei Sun,Fei Xue,Lide Duan,Yuanwei Fang,Hongzhong Zheng,Xiping Jiang,Song Wang,Fengguo Zuo,Yubing Wang,Bing Yu,Qiwei Ren,Yuan Xie
标识
DOI:10.1109/isscc42614.2022.9731694
摘要
The era of AI computing brings significant challenges to traditional computer systems. As shown in Fig. 29.1.1, while the AI model computation requirement increases 750x every two years, we only observe a very slow-paced improvement of memory system capability in terms of both capacity and bandwidth. There are many memory-bound applications, such as natural language processing, recommendation systems, graph analytics, graph neural networks, as well as multi-task online inference, that become dominating AI applications in modern cloud datacenters. Current primary memory technologies that power AI systems and applications include on-chip memory (SRAM), 2.5D integrated memory (HBM [1]), and off-chip memory (DDR, LPDDR, or GDDR SDRAM). Although on-chip memory enjoys low energy access compared to off-chip memory, limited on-chip memory capacity prevents the efficient adoption of large AI models due to intensive and costly off-chip memory access. In addition, the energy consumption of data movement of off-chip memory solutions (HBM and DRAM) is several orders of magnitude larger than that of on-chip memory, bringing the well-known “memory wall [2]“problem to AI systems. Process-near-memory (PNM) and computing-in-memory (CIM) have become promising candidates to tackle the “memory wall” problem in recent years.
科研通智能强力驱动
Strongly Powered by AbleSci AI