A MultiModal Feature Fusion Network for Building Extraction with Very High-resolution Remote Sensing Image and LiDAR data

IEEE Transactions on Geoscience and Remote Sensing(2024)

引用 0|浏览1
暂无评分
摘要
Building extraction from remote sensing images is extremely important for urban planning, land-cover change analysis, disaster monitoring and so on. With the growing diversity in building features, shape, and texture, coupled with frequent occurrences of shadowing and occlusion, the use of high-resolution remote sensing image (HRI) alone has limitations in building extraction. Therefore, feature fusion using multisource data has gradually become one of the most popular. However, the unique characteristics and noise issues make it difficult to achieve effective fusion and utilization. So it is very challenging to realize the full fusion of multisource data to achieve complementary advantages. In this paper, we propose an end-to-end multimodal feature fusion building extraction network based on segformer, which utilizes the fusion of HRI and LiDAR data to realize the building extraction. Firstly, we utilize the segformer encoder to break through the limitations of the traditional convolutional neural network with restricted receptive field so as to achieve effective feature extraction of complex building. In addition, we propose a cross-modal feature fusion (CMFF) method utilizing the self-attention mechanism to ensure the fusion of multisource data. In the decoder part, we propose a multi-scale up-sampling decoder (MSUD) strategy to achieve full fusion of multi-level features. As demonstrated by experiments on three datasets, our model shows better performance than several multisource building extraction and semantic segmentation models. The IoU for buildings on the three datasets reach 91.80%, 93.03%, and 84.59%. Subsequent ablation experiments further validate the effectiveness of each strategy.
更多
查看译文
关键词
Building extraction,Segformer network,Multisource fusion
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要