Earticle

다운로드

Reversing Attention Mechanisms in Transformers to Improve Object Detection Performance

원문정보

초록

영어
Recent advancements in object detection increasingly leverage end-to-end transformer architectures. However, many studies in this domain have applied transformer structures, originally designed for natural language processing, directly to object detection models. This direct application can lead to issues such as skipping self-attention in first decoder layer and the prediction of duplicate objects during training. In this study, we propose a novel approach to address these challenges by reversing the attention order in the transformer decoder from the self-cross to a cross-self structure. This modification structurally prevents the initial attention skip and mitigates the issue of predicting the same object multiple times by delaying the implementation of self-attention. Experimental results demonstrate that reversing the attention order in the decoder improves both the training loss and test performance across all stages of the learning process.

목차

Abstract
I. INTRODUCTION
II. METHODOLOGY
A. RT-DETR
B. Reversing Attention Mechanisms
C. Experimental Results
III. CONCLUSION
ACKNOWLEDGMENT
REFERENCES

저자

  • Chan-Young Choi [ School of Computing Gachon University ]
  • Sung-Yoon Ahn [ School of Computing Gachon University ]
  • Abrar Alabdulwahab [ School of Computing Gachon University ]
  • Joo-Hee Oh [ School of Computing Gachon Gachon University ]
  • Sang-Woong Lee [ School of Computing Gachon University ] Corresponding Author

참고문헌

자료제공 : 네이버학술정보

    간행물 정보

    • 간행물
      한국차세대컴퓨팅학회 학술대회
    • 간기
      반년간
    • 수록기간
      2021~2025
    • 십진분류
      KDC 566 DDC 004