Your Location:
Home >
Browse articles >
TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision Transformer quantization
Regular Papers | Updated:2026-02-06
    • TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision Transformer quantization

    • TP-ViT:面向视觉Transformer量化的截断均匀对数量化器与渐进式比特衰减重建方法
    • Vision Transformers (ViTs) have made significant strides in AI-based computer vision applications. However, deploying ViTs on edge devices is challenging due to their high computational and memory demands. To tackle this, researchers have developed the TP-ViT framework, which includes a truncated uniform-log2 quantizer and a bit-decline optimization strategy. These innovations significantly reduce quantization errors and maintain model performance, even under extreme low-bit conditions. Experiments show TP-ViT outperforms state-of-the-art methods, especially in 3-bit quantization, achieving a 6.18 percentage points improvement in top-1 accuracy for ViT-small. This advancement paves the way for more efficient ViT deployment on edge hardware.
    • ENGINEERING Information Technology & Electronic Engineering   Vol. 27, Issue 1, Pages: 1-12(2026)
    • DOI:10.1631/ENG.ITEE.2025.0081    

      CLC: TP391.4
    • Received:13 October 2025

      Revised:2026-01-15

      Published:2026-01

    Scan QR Code

  • Xichuan ZHOU, Sihuan ZHAO, Rui DING, et al. TP-ViT: truncated uniform-log2 quantizer and progressive bit-decline reconstruction for vision Transformer quantization[J]. ENGINEERING Information Technology & Electronic Engineering, 2026, 27(1): 1-12. DOI: 10.1631/ENG.ITEE.2025.0081.

  •  
  •  

0

Views

0

Downloads

0

CSCD

>
Alert me when the article has been cited
Submit
Tools
Download
Export Citation
Share
Add to favorites
Add to my album

Related Articles

S3Det: a fast object detector for remote sensing images based on artificial to spiking neural network conversion
Prompt learning in computer vision: a survey
UI layers merger: merging UI layers via visual learning and boundary prior
A modified YOLOv4 detection method for a vision-based underwater garbage cleaning robot&
Representation learning via a semi-supervised stacked distance autoencoder for image classification

Related Author

Mingqian SUN
Xiaofeng QI
Yanzhao GAO
Guangwei XIE
Fan ZHANG
Li CHEN
Yiming LEI
Jingqi LI

Related Institution

School of Cyber Science and Engineering, Southeast University
Computation and Artificial Intelligence Innovative College, Fudan University
National Digital Switching System Engineering & Technological R&D Center
Shanghai Center for Brain Science and Brain-Inspired Technology
MOE Frontiers Center for Brain Science, Fudan University
0