Key points are not available for this paper at this time.
Transformer is a machine learning model based on attention mechanism, which is widely used. When the Transformer model was first proposed, it gradually developed many variants and was promoted and applied in many fields, becoming an important research part in the areas of deep learning. However, the critical attention mechanism of Transformers has issues such as square complexity that affect computational speed and data processing efficiency. In order to meet the needs of data processing and related computing, there have been endless efforts to improve the attention mechanism in Transformers in different work areas. This article mainly provides a brief overview of the recent research progress on the attention mechanism in Transformers. Select representative studies from several directions of attention improvement work to introduce, in order to explore the latest research trends in its improvement work and lay a foundation for pointing out potential research directions for future research work and further improving the performance of Transformers.
Building similarity graph...
Analyzing shared references across papers
Loading...
Yuzhong Chen
Hongren Pu
Qu Yang
Applied and Computational Engineering
Sichuan University
Taiyuan University of Technology
Xi’an Jiaotong-Liverpool University
Building similarity graph...
Analyzing shared references across papers
Loading...
Chen et al. (Fri,) studied this question.
www.synapsesocial.com/papers/68e73ec2b6db6435876b81c9 — DOI: https://doi.org/10.54254/2755-2721/47/20241291