DETR with Additional Object Instance-Specific Features for Encoder

Yao Wang, Jong Eun Ha

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

1 Scopus citations

Abstract

This paper focuses on the process of developing from a convolutional neural network (CNN)-based target detection method to a transformer-based DETR paradigm-based target detection method. DETR adopts a Transformer-based end-to-end detection method, and it does not use the traditional anchor box and non-maximum suppression by transforming target detection into a set prediction problem. DETR has shown competitive results on public datasets and brought new ideas and methods to the field of object detection. We observed that DETR and DETE-like models include backbone and encoder that have same effect on the image, that is, they both did the same feature extraction function. We propose to add additional embedding module, which represents the full class information, and establishes global attention between feature tokens to provide prior knowledge for the extractor.

Original languageEnglish
Title of host publication23rd International Conference on Control, Automation and Systems, ICCAS 2023
PublisherIEEE Computer Society
Pages238-240
Number of pages3
ISBN (Electronic)9788993215274
DOIs
StatePublished - 2023
Event23rd International Conference on Control, Automation and Systems, ICCAS 2023 - Yeosu, Korea, Republic of
Duration: 17 Oct 202320 Oct 2023

Publication series

NameInternational Conference on Control, Automation and Systems
ISSN (Print)1598-7833

Conference

Conference23rd International Conference on Control, Automation and Systems, ICCAS 2023
Country/TerritoryKorea, Republic of
CityYeosu
Period17/10/2320/10/23

Keywords

  • Deep Learning
  • DETR
  • Object Detection
  • Transformer

Fingerprint

Dive into the research topics of 'DETR with Additional Object Instance-Specific Features for Encoder'. Together they form a unique fingerprint.

Cite this