
@Article{cmc.2023.044284,
AUTHOR = {Jian Wei, Qinzhao Wang, Zixu Zhao},
TITLE = {Interactive Transformer for Small Object Detection},
JOURNAL = {Computers, Materials \& Continua},
VOLUME = {77},
YEAR = {2023},
NUMBER = {2},
PAGES = {1699--1717},
URL = {http://www.techscience.com/cmc/v77n2/54836},
ISSN = {1546-2226},
ABSTRACT = {The detection of large-scale objects has achieved high accuracy, but due to the low peak signal to noise ratio (PSNR), fewer distinguishing features, and ease of being occluded by the surroundings, the detection of small objects, however, does not enjoy similar success. Endeavor to solve the problem, this paper proposes an attention mechanism based on cross-Key values. Based on the traditional transformer, this paper first improves the feature processing with the convolution module, effectively maintaining the local semantic context in the middle layer, and significantly reducing the number of parameters of the model. Then, to enhance the effectiveness of the attention mask, two Key values are calculated simultaneously along Query and Value by using the method of dual-branch parallel processing, which is used to strengthen the attention acquisition mode and improve the coupling of key information. Finally, focusing on the feature maps of different channels, the multi-head attention mechanism is applied to the channel attention mask to improve the feature utilization effect of the middle layer. By comparing three small object datasets, the plug-and-play interactive transformer (IT-transformer) module designed by us effectively improves the detection results of the baseline.},
DOI = {10.32604/cmc.2023.044284}
}



