Your Location:
Home >
Browse articles >
Minimizing transformer inference overhead using controlling element on Shenwei AI accelerator
Regular Papers | Updated:2025-05-06
    • Minimizing transformer inference overhead using controlling element on Shenwei AI accelerator

    • 使用申威人工智能加速器的控制单元最小化Transformer推理开销
    • In the field of natural language processing, this study addresses the computational overhead challenge in transformer models. Expert researchers developed a three-tier scheduling framework and zero-copy memory management technique, laying a foundation for optimizing transformer models and improving inference efficiency on AI accelerators.
    • Frontiers of Information Technology & Electronic Engineering   Vol. 26, Issue 4, Pages: 605-622(2025)
    • DOI:10.1631/FITEE.2400453    

      CLC: TP181
    • Received:28 May 2024

      Revised:25 August 2024

      Published:2025-04

    Scan QR Code

  • Yulong ZHAO, Chunzhi WU, Yizhuo WANG, et al. Minimizing transformer inference overhead using controlling element on Shenwei AI accelerator[J]. Frontiers of information technology & electronic engineering, 2025, 26(4): 605-622. DOI: 10.1631/FITEE.2400453.

  •  
  •  

0

Views

0

Downloads

0

CSCD

>
Alert me when the article has been cited
Submit
Tools
Download
Export Citation
Share
Add to favorites
Add to my album

Related Articles

No data

Related Author

No data

Related Institution

No data
0