Publications about the project
Synthetic Images Flood Scenario
Synthetic RGB image dataset depicting 766 images of floods in varied scenarios generated with AI methods.
This data can be used to train/evaluate flood segmentation models on RGB images.
FloodsScenarios/
├── images/ (image_00001_.jpg → image_00766_.jpg)
├── annotated_images/ (image_00001_.png → image_00766_.png)
└── annotated_masks/ (image_00001_.png → image_00766_.png)
Synthetic Images Fire Scenario
Synthetic RGB image dataset depicting 600 images of forest fires in varied scenarios generated with AI methods.
This data can be used to train/evaluate fire detection models on RGB images.
FireScenarios/
├── images/ (image_00001_.jpg → image_00600_.jpg)
├── annotated_images/ (image_00001_.jpg → image_00600_.jpg)
└── labels_filtered/ (image_00001_.txt → image_00600_.txt)
TEMA AIIA_wildfire Dataset
The AIIA_wildfire dataset is a collection of 2,237 natural disaster images designed for semantic segmentation, focusing on burnt areas, smoke, and fire. It aggregates and standardizes images from three distinct sources: the BLAZE classification dataset (a subset of which we annotated), KAHY trials, and RAS. The dataset is organized by source (BLAZE1, KAHY, RAS), each with standard train/val splits containing .jpg images and corresponding .png label masks. Labels follow a four-class hierarchy (0: background, 1: burnt, 2: smoke, 3: fire). The final composition is 985 images from BLAZE (https://aiia.csd.auth.gr/blaze-fire-classification-segmentation-dataset/) (655 annotated), 584 from KAHY, and 668 from RAS, split into 1,528 training and 655 validation images almost a 70 – 30% split.
Details on acquiring the dataset can be found here.
TEMA AIIA_flood Dataset
The dataset for the flood binary segmentation task comprises 720 images consolidated from two sources: D.MALIAN_1 and BRK_1 trials. It is structured into distinct directories for each source (dmalian_1 and brk_1), each containing standard train and val splits with separate folders for images (.jpg) and labels (.png). The annotation masks are binary, where pixels are labeled as 0 for background and 1 for floodwater. The total split consists of 428 training images (198 from D.MALIAN_1 and 230 from BRK_1) and 292 validation images (140 from D.MALIAN_1 and 152 from BRK_1), resulting in an approximate 60% - 40% training-validation distribution.
Mora information about the dataset can be found here.
AUW Dataset
The sample dataset, called the AUTH-Unreal-Wildfire (AUW) dataset, is a synthetic collection created to advance deep learning for wildfire segmentation. It addresses the critical challenge of obtaining accurately annotated training data in natural disaster management by using a novel, open-source pipeline built with the AirSim simulator. This pipeline uniquely integrates a custom particle segmentation camera and Procedural Content Generation (PCG) tools to produce photorealistic wildfire images paired with precise pixel-level segmentation masks—a feature previously difficult to achieve since fire assets are typically particle-based without a defined 3D mesh. The dataset consists of 1,500 training and 200 test images and was specifically designed to train and evaluate state-of-the-art segmentation models like PIDNet, both on its own and as a data augmentation resource to enhance performance on real-world wildfire imagery.
If someone use it, cite with this "E. Spatharis, C. Papaioannidis, V. Mygdalis and I. Pitas, "UNREALFIRE: A synthetic dataset creation pipeline for annotated fire imagery in Unreal Engine", IEEE International Conference on Image Processing (ICIP), Workshop on Bridging the Gap: Advanced Data Processing for Natural Disaster Management – Integrating Visual and Non-Visual Insights, Anchorage, Alaska, USA, 13-17 September, 2025"
Details on acquiring the dataset can be found here
Geo-social media and AI for Early WarningA data source for multifaceted spatiotemporal information
Leveraging Collective Knowledge for Forest Fire Classification
This paper presents a novel Fire Classification Multi-Agent (FCMA) framework that utilizes peer-to-peer learning and distributed learning techniques to disseminate knowledge within the agent community. Furthermore, we define and introduce the architecture of a Deep Neural Network (DNN) agent, which can infinitely interact with other DNN agents and the external environment upon deployment. The FCMA framework is suitable for natural disaster management systems where multiple agents are required to run autonomously and foster the community’s knowledge. The FCMA provides two options for knowledge transfer, a peer-to-peer and a federated one. The experimental results display the effective knowledge transfer using both options and also compare the two options with each other in a forest fire classification setting.
An Aspect-Based Emotion Analysis Approach on Wildfire-Related Geo-Social Media Data—A Case Study of the 2020 California Wildfires
Natural disasters like wildfires pose significant threats to communities, which necessitates timely and effective disaster response strategies. While Aspect-based Sentiment Analysis (ABSA) has been widely used to extract sentiment-related information at the sub-sentence level, the corresponding field of Aspect-based Emotion Analysis (ABEA) remains underexplored due to dataset limitations and the increased complexity of emotion classification. In this study, we used EmoGRACE, a fine-tuned BERT-based model for ABEA, which we applied to georeferenced tweets of the 2020 California wildfires. The results for this case study reveal distinct spatio-temporal emotion patterns for wildfire-related aspect terms, with fear and sadness increasing near wildfire perimeters. This study demonstrates the feasibility of tracking emotion dynamics across disaster-affected regions and highlights the potential of ABEA in real-time disaster monitoring. The results suggest that ABEA can provide a nuanced understanding of public sentiment during crises for policymakers.