收录:
摘要:
Current mainstream image captioning models are based on the encoder-decoder framework with multi-head attention, which commonly employs grid image features as the input and has shown superior performance. However, self-attention in the encoder only models the visual relations of fixed-scale grid features, the multi-head attention mechanism is not fully exploited to capture diverse information for more efficient feature representation, thus affecting the quality of the generated captions. To solve this problem, we propose a novel Scale-aware Multi-head Information Aggregation (SMIA) model for image captioning. SMIA introduces multi-scale visual features to improve the feature representation from the perspective of attention heads. Specifically, a scale expansion algorithm is proposed to extract multi-scale visual features. Then, for different heads of the multi-head attention, different high-scale features are integrated into the fixed low-scale grid features to capture diverse and richer information. In addition, different high-scale features are introduced for shallow and deep layers of encoder to further improve the feature representation. Besides, SMIA is flexible to combine with existing Transformer models to further improve performance. Experimental results on the MS COCO dataset demonstrate the effectiveness of our proposed SMIA. © 2024 Technical Committee on Control Theory, Chinese Association of Automation.
关键词:
通讯作者信息:
电子邮件地址:
来源 :
ISSN: 1934-1768
年份: 2024
页码: 8771-8777
语种: 英文
归属院系: