the Creative Commons Attribution 4.0 License.
the Creative Commons Attribution 4.0 License.
BRIGHT: A globally distributed multimodal building damage assessment dataset with very-high-resolution for all-weather disaster response
Abstract. Disaster events occur around the world and cause significant damage to human life and property. Earth observation (EO) data enables rapid and comprehensive building damage assessment, an essential capability crucial in the aftermath of a disaster to reduce human casualties and inform disaster relief efforts. Recent research focuses on developing artificial intelligence (AI) models to accurately map unseen disaster events, mostly using optical EO data. These solutions based on optical data are limited to clear skies and daylight hours, preventing a prompt response to disasters. Integrating multimodal EO data, particularly combining optical and synthetic aperture radar (SAR) imagery, makes it possible to provide all-weather, day-and-night disaster responses. Despite this potential, the lack of suitable benchmark datasets has constrained the development of robust multimodal AI models. In this paper, we present a Building damage assessment dataset using veRy-hIGH-resoluTion optical and SAR imagery (BRIGHT) to support AI-based all-weather disaster response. To the best of our knowledge, BRIGHT is the first open-access, globally distributed, event-diverse multimodal dataset specifically curated to support AI-based disaster response. It covers five types of natural disasters and two types of human-made disasters across 14 regions worldwide, focusing on developing countries where external assistance is most needed. The dataset's optical and SAR images with spatial resolutions between 0.3 and 1 meters provide detailed representations of individual buildings, making it ideal for precise damage assessment. We train seven advanced AI models on BRIGHT to validate transferability and robustness. Beyond that, it also serves as a challenging benchmark for a variety of tasks in real-world disaster scenarios, including unsupervised domain adaptation, semi-supervised learning, unsupervised multimodal change detection, and unsupervised multimodal image matching. The experimental results serve as baselines to inspire future research and model development. The dataset (Chen et al., 2025), along with the code and pretrained models, is available at https://github.com/ChenHongruixuan/BRIGHT and will be updated as and when a new disaster data is available. BRIGHT also serves as the official dataset for the 2025 IEEE GRSS Data Fusion Contest Track II. We hope that this effort will promote the development of AI-driven methods in support of people in disaster-affected areas.
- Preprint
(32165 KB) - Metadata XML
- BibTeX
- EndNote
Status: final response (author comments only)
-
RC1: 'Comment on essd-2025-269', Anonymous Referee #1, 11 Jul 2025
This paper introduces BRIGHT, a novel and timely benchmark dataset for building damage assessment using multimodal high-resolution optical and SAR imagery. Covering 14 globally distributed disaster events, BRIGHT provides pixel-level damage annotations for over 384,000 buildings. The dataset is designed to facilitate AI-based disaster response research, particularly in challenging all-weather conditions. The authors also benchmark a suite of machine learning and deep learning models on multiple tasks. The authors provided detailed documents and descriptions, making the data, related source code, and pretrained weights of models easy to understand and use.
In summary, this is quite interesting and solid work. I’d like to recommend the acceptance of this work since it represents an important contribution to Earth observation and disaster response communities. Yet before acceptance, several clarifications and refinements are suggested.
- The manuscript would benefit from deeper exploration of what the models learn from multimodal fusion. Specifically, what roles do optical images play in multimodal building damage assessment? Is it beyond just building footprint localization? On the other words, are the features extracted from optical imagery actively compared with SAR representations? Some discussion (e.g., based on CAMs in Fig. 7) is provided but can be more explicitly elaborated.
- The manuscript makes extensive evaluation of supervised and unsupervised change detection models, but the conceptual and methodological relationship between building damage assessment and generic change detection remains unclear, which is largely implied rather than discussed. An explicit and clearer explanation would be great for readers who lack of related background.
- Since UMCD methods underperform, consider including a random guessing baseline for reference. This would contextualize the difficulty of BRIGHT and help readers understand the performance floor under UMCD setup.
- While Table 1 offers a comprehensive comparison of datasets, several datasets seem relevant and should be included to enhance its completeness, like CRASAR-U-DROIDs [arXiv:2407.17673] and Noto-Earthquake building damage dataset [10.5194/essd-2024-363].
- The paper describes careful multimodal alignment but omits the software used, e.g., ENVI, ArcGIS, or QGIS. Please provide related details.
- Appendix G includes important new experimental setups and evaluation methods for UMCD. However, too much content is composed together now. It is not easy for people to grasp information. Adding section subtitles could improve readability.
- 8: Please specify in the figure or caption that the values represent average ± standard deviation across models.
- 10: Add a note in the caption to clarify that each dot corresponds to performance on a single test event under cross-event transfer.
- Typo in Table 7: “Object-based major voting” should be corrected to “Object-based majority voting”.
- Clarify the meaning of “–” symbols in Table 11. Do they indicate missing data or inapplicability? This should be stated explicitly.
- “ML” should be defined on its first use and consistently used thereafter instead of alternating with [machine learning].
- Standardize currency formatting (e.g., USD vs. US$).
- Define abbreviations such as IGN and GSI when first mentioned as data providers.
- The format of references should be standardized. Some of these entries use abbreviations for journals, while others have full titles.
Citation: https://doi.org/10.5194/essd-2025-269-RC1 -
AC1: 'Reply on RC1', Hongruixuan Chen, 24 Jul 2025
Thank you so much for the instructive and constructive comments for our paper! We have updated a point by point response attached as a supplement for your reference.
-
RC2: 'Reply on AC1', Anonymous Referee #1, 24 Jul 2025
Thanks for the author's responses. This revision has significantly improved the quality and soundness of their work. My major concerns have been fully addressed. I recommend accepting this paper.
Citation: https://doi.org/10.5194/essd-2025-269-RC2 -
AC2: 'Reply on RC2', Hongruixuan Chen, 26 Jul 2025
Thank you once again for your insightful comments, which have greatly helped us improve our manuscript. We will upload the revised version, based on your suggestions and as reflected in our responses, as the Final Response after the open discussion.
Citation: https://doi.org/10.5194/essd-2025-269-AC2
-
AC2: 'Reply on RC2', Hongruixuan Chen, 26 Jul 2025
-
RC2: 'Reply on AC1', Anonymous Referee #1, 24 Jul 2025
-
RC3: 'Comment on essd-2025-269', Anonymous Referee #2, 26 Aug 2025
This manuscript introduces the BRIGHT dataset, which is a first open building damage assessment dataset with global coverage, multi-hazard scenarios, multimodal imagery (Optical and SAR), and sub-meter resolution. The paper systematically describes data collection, annotation, and quality control methods, and validates the dataset with multiple deep learning models, including cross-disaster transfer (zero-shot and one-shot), semi-supervised, and unsupervised approaches. The dataset demonstrates clear novelty and practical value, and it is of significant importance for advancing research and applications in disaster emergency response, remote sensing, and artificial intelligence. Generally, the paper is well structured, logically clear, with detailed results and strong value in terms of data sharing. Although the manuscript is rich in content, there are still details that require improvement, and I recommend appropriate revisions.
Major Comments
- The explanation of annotation consistency and reliability remains insufficient. Although the authors state that the data annotations were obtained from multiple institutions such as Copernicus EMS, UNOSAT, and FEMA and then refined manually, there may be inconsistencies in how different institutions define “damaged” and “destroyed.” This could affect the consistency of annotations across disaster scenarios. It is therefore necessary to further elaborate on the process of unifying annotations, provide more detail on the manual refinement procedures.
- The treatment of class imbalance is not sufficient. Figure 5(d) shows that intact buildings account for over 80%, while destroyed buildings account for less than 7%. This severe imbalance directly affects the accuracy of recognizing destroyed classes. Although the authors employed the Lovasz loss function to partially alleviate the issue, this is still not enough to solve the problem. Is this imbalance one of the reasons for the relatively low performance of the subsequent experimental results?
- The discussion on cross-disaster generalization needs to be strengthened. Table 6 shows that in different disaster types, certain events perform particularly poorly. In particular, the mIoU values are the lowest for explosion and chemical accident events such as Bata-EP-2021 and Kyaukpyu-CC-2023, while earthquake events such as Morocco-EQ-2023 and Noto-EQ-2024 also remain highly challenging. This indicates that the models face significant difficulties in handling highly destructive, structurally complex, and spatially heterogeneous disaster scenarios. The authors should analyze these challenges in more depth, such as the heterogeneity and extreme local variations in explosion damage, the diversity of collapse patterns in earthquake events, and the limitations of SAR data in capturing fine-grained details. It is also recommended to provide typical error cases and compare model errors across disaster types to better illustrate the shortcomings in generalization.
- The manuscript currently lacks a comparison with optical-only baselines, which is crucial to highlight the value of multimodal methods. Readers may question whether the inclusion of SAR brings significant benefits and whether the additional cost of multimodality is justified. To avoid such doubts, I suggest adding experiments with optical-only inputs and comparing them with Optical+SAR results. This would further emphasize the unique value of the BRIGHT dataset and provide stronger evidence for the necessity of multimodal fusion.
- The discussion of limitations and future directions is insufficient. At present, the conclusion mainly emphasizes the dataset’s contributions, but it does not address its shortcomings in detail. It is suggested to include a separate subsection summarizing the limitations, such as the use of single-polarization SAR, the lack of time-series data, and the fact that most disaster events are concentrated after 2020.
Minor Comments
- The description of study areas and disaster events is somewhat redundant.
- The terminology for “one-shot” may not be accurate. The authors describe it as using “a small number of labeled samples,” which may be better defined as “few-shot.” Since these concepts are borrowed from previous work, it is suggested to cite the corresponding references.
- At line 420, the authors state that SAR is not sensitive to fine structural changes. Would this limitation reduce the value of multimodality in certain scenarios?
Citation: https://doi.org/10.5194/essd-2025-269-RC3 -
AC3: 'Reply on RC3', Hongruixuan Chen, 01 Sep 2025
Thank you so much for the instructive and constructive comments for our paper! We have updated a point by point response attached as a supplement for your reference.
-
RC4: 'Reply on AC3', Anonymous Referee #2, 02 Sep 2025
I appreciate the authors’ careful revisions. The manuscript has been substantially improved in quality and clarity. All of my previous concerns have been satisfactorily addressed. I therefore recommend acceptance of this paper.
Citation: https://doi.org/10.5194/essd-2025-269-RC4 -
AC4: 'Reply on RC4', Hongruixuan Chen, 02 Sep 2025
Thank you once again for your insightful comments, which have greatly helped us improve our manuscript. We will upload the revised version, based on your suggestions and as reflected in our responses, as the Final Response after the open discussion.
Citation: https://doi.org/10.5194/essd-2025-269-AC4
-
AC4: 'Reply on RC4', Hongruixuan Chen, 02 Sep 2025
-
RC4: 'Reply on AC3', Anonymous Referee #2, 02 Sep 2025
-
AC5: 'Final response to referees' comments on essd-2025-269', Hongruixuan Chen, 09 Sep 2025
Again, we would like to express our sincere gratitude to both reviewers for their professional, insightful comments and constructive suggestions.
This document summarizes our final responses to the two reviewers' comments. The content is consistent with our replies to two reviewers during the open discussion phase. We have ensured that all in this final response are fully synchronized with the latest revised version of the manuscript.
Data sets
BRIGHT Hongruixuan Chen, Jian Song, Olivier Dietrich, Clifford Broni-Bediako, Weihao Xuan, Junjue Wang, Xinlei Shao, Yimin Wei, Junshi Xia, Cuiling Lan, Konrad Schindler, Naoto Yokoya https://zenodo.org/records/14619797
Model code and software
BRIGHT Hongruixuan Chen, Jian Song, Olivier Dietrich, Clifford Broni-Bediako, Weihao Xuan, Junjue Wang, Xinlei Shao, Yimin Wei, Junshi Xia, Cuiling Lan, Konrad Schindler, Naoto Yokoya https://github.com/ChenHongruixuan/BRIGHT
Viewed
HTML | XML | Total | BibTeX | EndNote | |
---|---|---|---|---|---|
947 | 375 | 44 | 1,366 | 30 | 43 |
- HTML: 947
- PDF: 375
- XML: 44
- Total: 1,366
- BibTeX: 30
- EndNote: 43
Viewed (geographical distribution)
Country | # | Views | % |
---|
Total: | 0 |
HTML: | 0 |
PDF: | 0 |
XML: | 0 |
- 1
Cited
2 citations as recorded by crossref.
- Deep Learning-Based Detection and Assessment of Road Damage Caused by Disaster with Satellite Imagery J. Cha et al. 10.3390/app15147669
- DamageScope: An Integrated Pipeline for Building Damage Segmentation, Geospatial Mapping, and Interactive Web-Based Visualization S. Al Shafian et al. 10.3390/rs17132267