Fine- and Coarse-Granularity Hybrid Self-Attention for Efficient BERT

Created by MG96

External Public cs.CL

Statistics

Citations
0
References
0
Last updated
Loading...
Authors

Jing Zhao Yifan Wang Junwei Bao Youzheng Wu Xiaodong He
Project Resources

Name Type Source Actions
ArXiv Paper Paper arXiv
GitHub Repository Code Repository GitHub
Abstract

Transformer-based pre-trained models, such as BERT, have shown extraordinary success in achieving state-of-the-art results in many natural language processing applications. However, deploying these models can be prohibitively costly, as the standard self-attention mechanism of the Transformer suffers from quadratic computational cost in the input sequence length. To confront this, we propose FCA, a fine- and coarse-granularity hybrid self-attention that reduces the computation cost through progressively shortening the computational sequence length in self-attention. Specifically, FCA conducts an attention-based scoring strategy to determine the informativeness of tokens at each layer. Then, the informative tokens serve as the fine-granularity computing units in self-attention and the uninformative tokens are replaced with one or several clusters as the coarse-granularity computing units in self-attention. Experiments on GLUE and RACE datasets show that BERT with FCA achieves 2x reduction in FLOPs over original BERT with <1% loss in accuracy. We show that FCA offers a significantly better trade-off between accuracy and FLOPs compared to prior methods.

Note:

No note available for this project.

No note available for this project.
Contact:

No contact available for this project.

No contact available for this project.