skip to main content
Language:
Search Limited to: Search Limited to: Resource type Show Results with: Show Results with: Search type Index

A Hybrid Attention-Aware Fusion Network (HAFNet) for Building Extraction from High-Resolution Imagery and LiDAR Data

Remote sensing (Basel, Switzerland), 2020-11, Vol.12 (22), p.3764 [Peer Reviewed Journal]

2020. This work is licensed under http://creativecommons.org/licenses/by/3.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License. ;ISSN: 2072-4292 ;EISSN: 2072-4292 ;DOI: 10.3390/rs12223764

Full text available

Citations Cited by
  • Title:
    A Hybrid Attention-Aware Fusion Network (HAFNet) for Building Extraction from High-Resolution Imagery and LiDAR Data
  • Author: Zhang, Peng ; Du, Peijun ; Lin, Cong ; Wang, Xin ; Li, Erzhu ; Xue, Zhaohui ; Bai, Xuyu
  • Subjects: Architecture ; attention mechanism ; building extraction ; Buildings ; Classification ; deep learning ; Design ; Digital media ; High resolution ; high-resolution imagery (HRI) ; Image resolution ; Labeling ; Learning ; Lidar ; light detection and ranging (LiDAR) ; Machine learning ; multimodal data fusion ; Semantics
  • Is Part Of: Remote sensing (Basel, Switzerland), 2020-11, Vol.12 (22), p.3764
  • Description: Automated extraction of buildings from earth observation (EO) data has long been a fundamental but challenging research topic. Combining data from different modalities (e.g., high-resolution imagery (HRI) and light detection and ranging (LiDAR) data) has shown great potential in building extraction. Recent studies have examined the role that deep learning (DL) could play in both multimodal data fusion and urban object extraction. However, DL-based multimodal fusion networks may encounter the following limitations: (1) the individual modal and cross-modal features, which we consider both useful and important for final prediction, cannot be sufficiently learned and utilized and (2) the multimodal features are fused by a simple summation or concatenation, which appears ambiguous in selecting cross-modal complementary information. In this paper, we address these two limitations by proposing a hybrid attention-aware fusion network (HAFNet) for building extraction. It consists of RGB-specific, digital surface model (DSM)-specific, and cross-modal streams to sufficiently learn and utilize both individual modal and cross-modal features. Furthermore, an attention-aware multimodal fusion block (Att-MFBlock) was introduced to overcome the fusion problem by adaptively selecting and combining complementary features from each modality. Extensive experiments conducted on two publicly available datasets demonstrated the effectiveness of the proposed HAFNet for building extraction.
  • Publisher: Basel: MDPI AG
  • Language: English
  • Identifier: ISSN: 2072-4292
    EISSN: 2072-4292
    DOI: 10.3390/rs12223764
  • Source: ROAD: Directory of Open Access Scholarly Resources
    ProQuest Central
    DOAJ Directory of Open Access Journals

Searching Remote Databases, Please Wait