麻豆影视文化传媒在线看|果冻传媒91制片厂麻豆|性色网站|国产成人吃瓜网|麻豆文化传媒百度云|韩国黄色一级黄色片|成人电影区|糖心vlog是真的吗|黄瓜视频丝瓜视频香蕉视频|国产精品视频在一区鲁鲁,性感丰满美乳巨乳,蜜桔影院91制片厂,爱豆传媒陈可心作品名字

Home>LATEST NEWS

Tsinghua Reconfigurable Computing Team proposes a new approach to Memory Optimization for AI Chips, significantly improving their energy efficiency

Tsinghua Reconfigurable Computing Team proposes a new approach to Memory Optimization for AI Chips, significantly improving their energy efficiency


The 45th International Symposium on Computer Architecture (ISCA) was held in Los Angeles, California, from June 2nd to June 6th. Fengbin Tu, PhD student at the Institute of Microelectronics from Tsinghua University, gave an oral presentation titled “RANA: Towards Efficient Neural Acceleration with Refresh-Optimized Embedded DRAM” on June 4th. His approach to this work can significantly improve the energy efficiency of Artificial Intelligence (AI) Chips.

Fengbin Tu, PhD student at the Institute of Microelectronics, gave an oral presentation at ISCA 2018.

ISCA is the top conference for computer architecture. Altogether, 378 papers were submitted this year, and only 64 of them were accepted, an acceptance rate of 16.9%. Fengbin Tu et al.’s work is the only paper first-authored by a Chinese research team (the Tsinghua Reconfigurable Computing Team) in ISCA 2018. Shouyi Yin, Associate Professor at the Institute of Microelectronics, is the corresponding author, and Fengbin Tu is the first author. The paper is also co-authored by Prof. Shaojun Wei, Prof. Leibo Liu and Weiwei Wu (Master’s student) at the Institute of Microelectronics, Tsinghua University.

The growing size of deep neural networks (DNNs) requires large amounts of on-chip storage. In many DNN accelerators, their limited on-chip memory capacity causes massive off-chip memory access and leads to very high system energy consumption. So memory optimization is an important problem in AI chip design. Tsinghua Reconfigurable Computing Team proposed a Retention-Aware Neural Acceleration (RANA) framework for DNN accelerators to save total system energy consumption with refresh-optimized embedded DRAM (eDRAM). The RANA framework includes three levels of techniques: a retention-aware training method, a hybrid computation pattern and a refresh-optimized eDRAM controller. At the training level, DNN’s error resilience is exploited in training to improve eDRAM’s tolerable retention time. At the scheduling level, RANA assigns to each DNN layer a computation pattern that consumes the lowest energy. At the architecture level, a refresh-optimized eDRAM controller is proposed to alleviate unnecessary refresh operations. Owing to the RANA framework, 99.7% eDRAM refresh operations can be removed with negligible performance and accuracy loss. Compared with the conventional SRAM-based DNN accelerator, an eDRAM-based DNN accelerator strengthened by RANA can save 41.7% off-chip memory access and 66.2% system energy consumption, with the same area cost. The proposed framework significantly improves the energy efficiency of AI Chips.


 
RANA: Retention-aware neural acceleration framework.

The Tsinghua Reconfigurable Computing Team has designed a series of reconfigurable architecture based AI Chips (Thinker I, Thinker II, Thinker S) in recent years, which have attracted many researchers and engineers from all over the world. Their work was presented at ISCA 2018, and significantly improves the energy efficiency of AI Chips through memory optimization and software-hardware co-design techniques, which promotes a new way for AI Chips’ architectural development.

Editors: John Olbrich, Zhu Lvhe

Copyright 2001-2021 news.tsinghua.edu.cn. All rights reserved.