谷歌浏览器插件
订阅小程序
在清言上使用

Cross-Modal Remote Sensing Image-Audio Retrieval With Adaptive Learning for Aligning Correlation.

IEEE Trans. Geosci. Remote. Sens.(2024)

引用 0|浏览4
暂无评分
摘要
An important challenge that existing work has yet to address is the relatively small differences in audio representations compared to the rich content provided by remote sensing images, making it easy to overlook certain details in the images. This imbalance in information between modalities poses a challenge in maintaining consistent representations. In response to this challenge, we propose a novel cross-modal RSIA retrieval method called Adaptive Learning for Aligning Correlation (ALAC). ALAC integrates region-level learning into image annotation through a region-enhanced learning attention module. By collaboratively suppressing features at different region levels, ALAC is able to provide a more comprehensive visual feature representation. Additionally, a novel adaptive knowledge transfer strategy has been proposed, which guides the learning process of the frontend network using aligned feature vectors. This approach allows the model to adaptively acquire alignment information during the learning process, thereby facilitating better alignment between the two modalities. Finally, to better utilize mutual information between different modalities, we introduce a plug-and-play result rerank module. This module optimizes the similarity matrix by using retrieval mutual information between modalities as weights, significantly improving retrieval accuracy. Experimental results on four RSIA datasets demonstrate that ALAC outperforms other methods in retrieval performance. Compared to state-of-the-art methods, improvements of 1.49%, 2.25%, 4.24% and 1.33% were respectively achieved by ALAC. The codes are accessible at https://github.com/huangjh98/ALAC.
更多
查看译文
关键词
Cross-modal remote sensing retrieval,region-enhanced learning attention,adaptive learning,knowledge transfer,mutual information
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要