It is widely shared that capturing relationships among multi-modality features would be helpful for representing and ultimately describing an image. In this paper, we present a novel Intra- and Inter-modality visual Relation Transformer to improve connections among visual features, termed I2RT. Firstly, we propose Relation Enhanced Transformer Block (RETB) for image feature learning, which strengthens intra-modality visual relations among objects. Moreover, to bridge the gap between inter-modality feature representations, we align them explicitly via Visual Guided Alignment (VGA) module. Finally, an end-to-end formulation is adopted to train the whole model jointly. Experiments on the MS-COCO dataset show the effectiveness of our model, leading to improvements on all commonly used metrics on the "Karpathy"test split. Extensive ablation experiments are conducted for the comprehensive analysis of the proposed method.
|Original language||English (US)|
|Title of host publication||MM 2020 - Proceedings of the 28th ACM International Conference on Multimedia|
|Publisher||Association for Computing Machinery, Inc|
|Number of pages||9|
|State||Published - Oct 12 2020|