谷歌浏览器插件
订阅小程序
在清言上使用

Integer Arithmetic-Based and Activation-Aware GELU Optimization for Vision Transformer

Zihan Zou, Chen Zhang, Shikuang Chen, Hui Kou,Bo Liu

2024 Conference of Science and Technology for Integrated Circuits (CSTIC)(2024)

引用 0|浏览4
暂无评分
摘要
The implementation for non-linear activation functions like GELU in Transformer-based model is an challenging problem, especially in pursuit of energy-efficiency and high accuracy. To address this challenge, this paper presents a hardware-friendly optimization method for GELU deployment, an activation-aware strategy with integer arithmetic-based GELU calculation. This method detects activation magnitudes, approximates small and extremely large activation values with linear functions, while a more exact polynomial approximation function is applied to other values to ensure model accuracy. Implemented and evaluated under 22nm CMOS technology, the proposed design can improve the energy-efficiency by 2.14x and area-efficiency by 1.41×, compared with contemporary designs, while the accuracy loss is less than 1 % in Vit, Deit and Swin models.
更多
查看译文
关键词
Vision Transformer,Linear Function,Accuracy Loss,Contemporary Design,Root Mean Square Error,Convolutional Neural Network,Feature Values,PyTorch,Linear Approximation,Multilayer Perceptron,Polynomial Regression,Transformer Model,Approximate Interval,Sign Bit
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要