Presentation type:
ESSI – Earth & Space Science Informatics

Data are a fundamental building block of science. Ever-increasing volumes and diversity of data are allowing us to solve ever more complicated research questions; yet they are also creating new challenges around efficient data management and storage. This talk focuses on geochemical data, that are relatively low in volume compared to other Earth System Science disciplines, but are highly diverse due to the large range of materials analysed and analytical techniques employed. Modern geochemical research increasingly draws on large compilations of data previously collected by multiple authors using multiple analytical methods, over years and decades. Harmonising data from such diverse sources, and ensuring consistency and comparable data quality, is a non-trivial task that requires significant investment of time and resources. As a consequence, data compilations are increasingly published in high-ranking journals. Yet often they are singular, one-time efforts for specific projects by individual authors that quickly become outdated and lose relevance. In contrast, curated synthesis databases, such as the GEOROC database for igneous geochemical rock and mineral compositions, are continuously being updated and can offer long-term consistent curation over decades. By providing free access to, and customisable search of, their comprehensive data and metadata collections, they enable the compilation of a diverse range of smaller, targeted datasets that can form the basis of many different research projects across multiple (sub)disciplines. Long-term synthesis databases are an invaluable resource for the geochemical and broader scientific community. However, despite their broad relevance and usage, many such community databases struggle to secure the required resources for database maintenance and continuous technical developments to cater to changing scientific demands. This burden can be partly alleviated through integration of databases with curated, domain data repositories. Data harmonisation is greatly aided by adherence to best practices and standards during data publication. Repositories that publish curated, discipline-specific datasets, therefore, play an important role in ensuring new analyses are sufficiently well documented to allow quality assessment and reuse by third parties. They also support data rescue and the alignment of legacy data with modern data requirements. These standards and best practices should in turn be developed based on community expertise and consensus, which requires international collaboration. In geochemistry, data providers and services from three different continents formed the OneGeochemistry initiative. OneGeochemistry promotes exchange and agreement on minimum common variables between researchers from all geochemical sub-disciplines and the more than 15 international societies, associations and science unions that govern different types of geochemical data. As a participant in the WorldFAIR project, OneGeochemistry aims to reconcile cross-domain solutions for data interoperability with domain-specific geochemical requirements. The implementation of geochemical data standards in repositories, and their broad adoption by the geochemical community, will enhance the value of data and services provided by synthesis databases, which will lead to better access to comprehensive data compilations and, ultimately, better science.

How to cite: Klöcking, M.: The importance of curated domain repositories and synthesis databases as evolving community resources for modern Earth System Science research, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9936, https://doi.org/10.5194/egusphere-egu25-9936, 2025.

Earth System Science datasets have been acquired for centuries across five broad spheres: geosphere, cryosphere, hydrosphere, biosphere and atmosphere. They vary from human observations to sensor-derived measurements ranging from nanoscale laboratory data to large-volume petascale datasets collected remotely by satellites, drones, etc. Across all spheres most datasets have their roots in three core disciplines: Geology, Geophysics and Geochemistry. Today we are generating unprecedented volumes of data and when combined with computer capacity, now at exascale, our capability to integrate and analyse data should be unparalleled.

Digital data repositories emerged around 1980 and the internet soon after. Initially data was shared by shipping on hard media. The internet soon enabled globally data sharing data, including by web services (e.g., OneGeology In 2008). Multiple global data sharing networks were envisioned, but few moved beyond those that proposed them. Machine-to-machine data sharing is still a challenge. Many spheres cannot utilise the existing capacity of computers, including the full potential of AI applications, because these cannot read the volumes of available data. 

History has repeatedly shown that revolutionary infrastructures can take decades to realise their full potential and change from being a new way of doing things to multiple ways of doing new things. 

The FAIR principles were specifically designed to increase machine-to-machine interoperability of data: they are the blueprint of WHAT needs to be done but the HOW will involve rethinking 3 key steps. 

Firstly, shift the onus on aggregating data from the consumer to repositories capable of implementing discipline-centric FAIR (meta)data standards. 

Secondly, as recommended by the WorldFAIR Second Policy Brief to the European Open Science Cloud (EOSC), change from a bibliographic approach to data stewardship to one of data engineering, where richer and more comprehensive standardised (meta)data at the datum level enables machine-to-machine access of specific variables of interest across multiple disciplinary datasets. Take a more holistic approach to standards development (e.g., Observation, Measurement and Samples Standard (ISO 19156:2023)) and identify common universals across disciplines (e.g., time, place, units of measure). Initiatives like OneGeology and Geochemistry and hopefully soon OneGeophysics can support higher--level discipline centric (meta)data standards. Standards coordination groups (e.g., CODATA, Research Data Alliance) are critical. PIDS at the object level will be essential.

Thirdly, prioritise which datasets are made fully FAIR compliant and fund their curation in repositories that offer discipline based curation. The 2019 Beijing Declaration on Research Data notes that ‘publicly funded research data should be interoperable, and preferably without further manipulation or conversion, to facilitate their broad reuse in scientific research’. The myriad of data products generated from these primary data sources can go to generalist and institutional repositories.

Revolutionary infrastructures do take time to realise their full potential. It is nearly 25 years since the early experiments using the internet to globally network data repositories. The WorldFAIR Second EOSC Policy Brief emphasises that the change to machine-actionable FAIR data ‘is one of a magnitude which will necessitate considerable resourcing, investment, and upskilling; but it will also achieve significant benefits, including creating a digitally integrated Earth to support sustainable development of our planet.

How to cite: Wyborn, L.: Rethinking HOW We Create Global Networks of Earth and Environmental Datasets to Maximise Their Potential to Underpin Integrated Research for a Sustainable Planet., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16237, https://doi.org/10.5194/egusphere-egu25-16237, 2025.

EGU25-121 | ECS | Posters virtual | VPS19 | Highlight

A Novel Q-Switch Technique for  Borehole NMR Measurement 

Sihui Luo, Xin Li, Huiju Yu, Zhengduo Wang, Tianyu Xing, Zhihao Long, Cheng Che, Guangzhi Liao, and Lizhi Xiao

Nuclear Magnetic Resonance (NMR) is a crucial logging technique for the unconventional and complex reservoir evaluation. However, the echo spacing is always an issue of borehole NMR measurement, which limits the performance of NMR tools to acquire the short relaxation components.

In this abstract, we proposed a novel Q-Switch technique aiming at breaking through the limitation of dead-time of borehole NMR logging tool, and to achieve much shorter echo spacing. Instead of using resistors of larger resistance in parallel with the radio-frequency (RF) coil to reduce the active dead-time, an inductive coupling circuit was introduced to decrease the ringing-down time significantly after transmitting the RF pulses with high voltage. The Q-Switch circuit consists of inductive coupling coil, capacitors, resistors and active high-voltage MOSFETs. The ringing-down time of RF system was decreased by at least 10 times compared to the system without using proposed Q-switch scheme, leading to echo spacing lower to 0.3 ms under the condition with resonant frequency lower to 500 kHz.

Both simulations and experiments were in great agreements, validating the feasibility and efficiency of proposed Q-switch scheme, and proved to be promising in the borehole NMR applications.

How to cite: Luo, S., Li, X., Yu, H., Wang, Z., Xing, T., Long, Z., Che, C., Liao, G., and Xiao, L.: A Novel Q-Switch Technique for  Borehole NMR Measurement, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-121, https://doi.org/10.5194/egusphere-egu25-121, 2025.

EGU25-4077 | ECS | Posters virtual | VPS19

Multiscale model coupling for watershed-scale contaminant transport modeling from point sources in Savannah River Site 

Kazuyuki Sakuma, Haruko Wainwright, Zexuan Xu, Angelique Lawrence, and Pieter Hazenberg

Soil and groundwater contamination at some sites impacts downstream populations when contaminants migrate from groundwater to rivers. Predictive modeling is challenging since it is required to include detailed subsurface structure and groundwater flow models within the site, as well as watershed-scale models for large-scale transport. Now that climate change impacts are major concerns at many sites, it is important to have the capability to represent the water balance change and its impact on contaminant transport both at the site and watershed scale in a consistent manner. This study introduces a new simulation framework to couple a detailed 2D site/hillslope-scale groundwater model to the 3D watershed-scale model to describe contaminant transport from groundwater to river water within the catchment. Within the site, we estimate the contaminant discharges to the river from contaminant sources based on the Richards equation and advection-dispersion equation. The discharges are then applied as the boundary conditions to the watershed-scale model considering the width of the 2D site/hillslope-scale groundwater model and recharge rates for both models.

We demonstrate and validate our framework based on the tritium concentration datasets in surface water and groundwater collected at the Savannah River Site F-Area. Results show that the method can successfully reproduce the contaminant concentration time series in river water.

How to cite: Sakuma, K., Wainwright, H., Xu, Z., Lawrence, A., and Hazenberg, P.: Multiscale model coupling for watershed-scale contaminant transport modeling from point sources in Savannah River Site, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4077, https://doi.org/10.5194/egusphere-egu25-4077, 2025.

EGU25-4641 | Posters virtual | VPS19 | Highlight

Size distribution of elemental components in atmospheric particulates from a typical industrial and mining city of Central China 

Hongxia Liu, Jiaquan Zhang, Changlin Zhan, Shan Liu, Ting Liu, and Wensheng Xiao

As one of crucial factor in atmospheric particulate matter, elemental components exhibit distinct distribution features within different particle size ranges. Crustal elements (such as Al, Si, Fe, Ca, Mg) are primarily concentrated in coarse particulate matter, whereas elements originating from anthropogenic pollution sources (such as heavy metal elements including Pb, Zn, Cd, As, Cr) are more frequently distributed in fine particulate matter. Furthermore, some specific elements may also exhibit peak concentrations in particular particle size, which is closely related to their sources and formation processes. In recent years, there are still some challenges and deficiencies. Further research is needed on the particle size distribution characteristics of complex pollution sources (such as industrial emissions and traffic emissions). Additionally, there is a need to enhance the understanding of the transformation mechanisms and health effects of elemental components within particulate matter. This study selected a typical industrial and mining city to investigate particle size distribution characteristics of elemental components in atmospheric particulate matter. Anderson Eight-Stage Particulate Impactor Sampler was used to collect atmospheric particulate matter in the urban area of Huangshi during winter and summer. Nine particle size range samples were obtained spanning from 0 to 0.4 µm, 0.4 to 0.7 µm, 0.7 to 1.1 µm, 1.1 to 2.1 µm, 2.1 to 3.3 µm, 3.3 to 4.7 µm, 4.7 to 5.8 µm, 5.8 to 9.0 µm, and 9.0 to 10 µm. Energy Dispersive X-Ray Fluorescence Spectrometry (ED-XRF) was employed to determine the concentrations of 17 elemental components, including S, Cl, K, Ca, Ti, V, Cr, Mn, Fe, Co, Ni, Cu, Zn, Rb, Sb, Ba, and Pb. Elements Ca, S, Fe, K, Zn, Ba, and Pb were identified as the primary pollutants during the sampling period. All the elemental concentrations exhibited distinct seasonal variations, demonstrating higher levels in winter compared to summer. Each element demonstrated distinct particle size distribution characteristics with peak concentrations for most elements occurring in the 5.8 to 9.0 µm range and peaks for the remaining elements in the 0.4 to 1.1 µm range. The highest elemental concentrations in both summer and winter were mainly distributed in the 5.8 to 9.0 µm and 0.7 to 1.1 µm size ranges. In summer, most elemental concentrations were negatively correlated with relative humidity. However, in winter, there was no significant correlation with relative humidity. Rainfall had a certain scavenging effect on elements but was also influenced by other meteorological factors. Element S had the highest enrichment factor values in both summer and winter. Element Cl was highly enriched in finer particle size fractions in both seasons. Most elements were slightly enriched across all particle size fractions. Principal component analysis further identified the main sources as soil dust and wind-blown sand, coal combustion, vehicle exhaust emissions, biomass burning, mining and construction activities, and other pollution sources. These findings contribute to the formulation of effective pollution control measures and the protection of public health.

How to cite: Liu, H., Zhang, J., Zhan, C., Liu, S., Liu, T., and Xiao, W.: Size distribution of elemental components in atmospheric particulates from a typical industrial and mining city of Central China, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4641, https://doi.org/10.5194/egusphere-egu25-4641, 2025.

EGU25-7004 | Posters virtual | VPS19

Industrial High Performance Computing Scalable and FAIR Workflow Opportunities for EO Operations Processing, Operations, and Archiving 

Caroline Ball, Mark Chang, James Cruise, Camille de Valk, and Venkatesh Kannan
The computational demands of Sentinel data processing, archiving, and dissemination require scalable, efficient, and innovative solutions. While cloud computing-based services currently address these needs, integrating High-Performance Computing (HPC) systems into specific workflows could unlock a new level of industrial-scale capabilities. These include reduced processing times, faster data turnaround, and lower CO2 emissions. Leveraging HPC as a service allows for optimized data storage and access, enabling long-term strategies that prioritize essential data products and enhance operational efficiency.
Next-generation Quantum Computing (QC) holds the potential to redefine Earth Observation (EO) workflows by offering breakthroughs in solving complex optimization problems. As an operational service, QC could deliver significant cost and energy savings, provided that workflows can be seamlessly adapted to quantum-compatible infrastructures.
This presentation focuses on the evolution of HPC and QC technologies from research-driven concepts to industrial solutions, highlighting their maturity and applicability as services. We will explore the tangible benefits, associated costs, and pathways to operationalize these technologies for Level-0 to Level-2 data processing, operations, and archiving in support of current and future Sentinel missions.  We examine, at a high level, how artificial intelligence (AI) can provide a solution to hybrid HPC-QC challenges for EO data processing.

How to cite: Ball, C., Chang, M., Cruise, J., de Valk, C., and Kannan, V.: Industrial High Performance Computing Scalable and FAIR Workflow Opportunities for EO Operations Processing, Operations, and Archiving, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7004, https://doi.org/10.5194/egusphere-egu25-7004, 2025.

Severe convection, including thunderstorms and related phenomena like flash flooding, hail, and strong winds, can have significant socioeconomic impacts. Nowcasting, which provides real-time, short-term predictions, is vital for issuing timely warnings to mitigate these impacts. Satellite imagery is essential for monitoring convection and offering accurate predictions of storm evolution, thereby enhancing early warning systems. Ensemble forecasting, which generates multiple potential scenarios, helps better quantify uncertainties in nowcasting. However, most ensemble forecasting methods are computationally intensive and typically do not incorporate satellite images directly. The Analog Ensemble (AnEn) method, a lower-cost ensemble approach, identifies similar past weather events based on forecast data. For a given time and location, the AnEn method identifies analogs from past model predictions that are similar to current forecast conditions. Then their associated observations are used as ensemble members. Despite its advantages, AnEn struggles with locality and is sensitive to the choice of similarity metrics. This study presents an improved AnEn system that replaces forecast archives with satellite images to identify analogs of convective conditions. The system utilizes pretrained deep learning algorithms (VGG16, Xception, and Inception-ResNet) to assess image similarity. The training dataset consists of daily convection satellite images from EUMETSAT for the period 2020-2023, and the domain covers 40°N to 20°S and -20°W to 4°E. The year 2024 is used for testing, with ERA5 reanalysis of total precipitation as the verification ground-state. For a present convection satellite image this image is encoded and compared to all past encoded images of the training period using different metrics. The most similar images to the current one are then selected and their associated ERA5 total precipitation reanalysis are considered the members or our ensemble. Preliminary results indicate an average maximum precipitation anomaly of 15 mm between the analog ensemble mean and the current reanalysis, showing that the proposed system offers promising improvements in short-term forecasting.

Key words: Convection; Ensemble Forecasting; Deep Learning; VGG; Xception; ResNet; Analog Ensemble; Morocco; Nowcasting; EUMETSAT; ERA5; Morocco; Satellite Images; Remote Sensing;

How to cite: Alaoui, B., Bounoun, C., and Bari, D.: Leveraging Pretrained Deep Learning Models to Extract Similarities for the Analog Ensemble Method Applied to Convection Satellite Imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7258, https://doi.org/10.5194/egusphere-egu25-7258, 2025.

Amid global environmental degradation, understanding the spatiotemporal dynamics and trade-offs of ecosystem services (ESs) under varying land-use scenarios is critical for advancing the sustainable development of social–ecological systems. This study analyzed the Chaohu Lake Basin (CLB), focusing on four scenarios: natural development (ND), economic priority (ED), ecological protection (EP), and sustainable development (SD). Using the PLUS model and multi-objective genetic algorithm (MOGA), land-use changes for 2030 were simulated, and their effects on ESs were assessed quantitatively and qualitatively. The ND scenario led to significant declines in cropland (3.73%) and forest areas (0.18%), primarily due to construction land expansion. The EP scenario curbed construction land growth, promoted ecosystem recovery, and slightly increased cropland by 0.05%. The SD scenario achieved a balance between ecological and economic goals, maintaining relative stability in ES provision. Between 2010 and 2020, construction land expansion, mainly concentrated in central Hefei City, led to a marked decline in habitat quality (HQ) and landscape aesthetics (LA), whereas water yield (WY) and soil retention (SR) improved. K-means clustering analysis identified seven ecosystem service bundles (ESBs), revealing significant spatial heterogeneity. Bundles 4 through 7, concentrated in mountainous and water regions, offered high biodiversity maintenance and ecological regulation. In contrast, critical ES areas in the ND and ED scenarios faced significant encroachment, resulting in diminished ecological functions. The SD scenario effectively mitigated these impacts, maintaining stable ES provision and ESB distribution. This study highlights the profound effects of different land-use scenarios on ESs, offering insights into sustainable planning and ecological restoration strategies in the CLB and comparable regions.

How to cite: Jin, A.: Ecosystem Services Trade-Offs in the Chaohu Lake Basin Based on Land-Use Scenario Simulations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7785, https://doi.org/10.5194/egusphere-egu25-7785, 2025.

EGU25-7921 | ECS | Posters virtual | VPS19

Integrating high-resolution satellite and multispectral drone Imagery for monitoring vegetation in the Chaschoc-Sejá lagoon system 

Jacob Nieto, Nelly Lucero Ramírez Serrato, Alejandro Romero Herrera, Candelario Peralta Carreta, Graciela Herrera Zamarrón, Mario Alberto Hernández Hernández, Guillermo de Jesús Hernández García, Selene Olea Olea, Erick Morales Casique, and Alejandra Cortez Silva

Seasonal ecosystems play a crucial role in environmental regulation and biodiversity by hosting complex ecological dynamics that vary with climatic conditions. The Chaschoc-Sejá wetlands are a key example of such systems in southeastern Mexico. The interaction between the lagoon system and the Usumacinta River is highly dynamic; during the rainy season, the lagoons increase in volume, reaching depths of 8 to 10 meters. However, the lagoons completely dry up during the dry season, leaving vegetation at the surface level. 

This project aims to analyze the dynamics of vegetation cover in this environment by comparing high-resolution satellite images (Planet, 3 m) and ortho-mosaics generated with a DJI Mavic 3 Multispectral drone (10 cm). By combining these datasets, we aim to improve our previous vegetation maps and obtain a more accurate and detailed assessment of the Chaschoc-Sejá Lagoon system. Understanding vegetation patterns at a larger scale during specific periods and the variations in plant life within the lagoon and along its shores is a key focus.

 

Data processing involved classifying vegetation cover and identifying seasonal changes using indices such as NDVI and NDWI. We also generated 3D models to estimate vegetation height. Results show that integrating both techniques significantly improves spatial resolution and temporal accuracy in monitoring these ecosystems. This study provides essential tools for managing seasonal systems and their conservation in the face of climatic and anthropogenic factors. This monitoring will aid in understanding vegetation status, identifying plant species, and contributing to managing and preserving the lagoon system.

How to cite: Nieto, J., Ramírez Serrato, N. L., Romero Herrera, A., Peralta Carreta, C., Herrera Zamarrón, G., Hernández Hernández, M. A., Hernández García, G. D. J., Olea Olea, S., Morales Casique, E., and Cortez Silva, A.: Integrating high-resolution satellite and multispectral drone Imagery for monitoring vegetation in the Chaschoc-Sejá lagoon system, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7921, https://doi.org/10.5194/egusphere-egu25-7921, 2025.

Benevento Province, located in the Campania region of Italy, may experience environmental quality impacts from neighboring developed areas such as Naples and Caserta. Previous studies have suggested that some agricultural chemicals from Naples, such as hexachlorobenzene, may be transported through the air to rural areas of Benevento. Additionally, high concentrations of polycyclic aromatic hydrocarbons (PAHs) have been detected in Naples and Caserta, making Benevento Province a potential PAH "sink." This study systematically investigated the occurrence of PAHs in soil from Benevento Province, southern Italy, and their correlations with environmental factors, soil-air exchange processes, and health risks. Over 95% of sampling sites exhibited ∑16PAHs concentrations at non-polluted levels (9.50-1188 ng/g, mean = 55.0 ± 152 ng/g), and four-ring PAHs were the dominant pollutants contributing to 28.3% of ∑16PAHs. The spatial distribution of PAHs presented significant heterogeneity, with hotspots concentrated near landfills. The results of Positive Matrix Factorization (PMF) model showed that the main sources of PAHs were vehicle emissions, coal/biomass combustion, and petroleum products volatilization/leakage, contributing 42.2%, 40.2%, and 17.6%, respectively. Most of PAHs correlated significantly with total organic carbon in the soil and population density, while only Benzo(b)fluoranthene (BbF) showed a significantly negative correlation with pH. The mass inventory of ∑16PAHs ranged from 0.94 to 29.4 tons, averaging 2.45 tons. The synergistic effects of pollution hotspots and the persistent accumulation of PAHs in the soil suggested that the soil might act as a secondary source of PAHs. Toxicity equivalent and probabilistic risk assessments indicated that health risks from PAHs remained within acceptable limits.

How to cite: Qu, C. and Pu, C.: Investigation of polycyclic aromatic hydrocarbons in the soils of Benevento Province, Italy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10723, https://doi.org/10.5194/egusphere-egu25-10723, 2025.

EGU25-13330 | ECS | Posters virtual | VPS19

Advancing and Supporting FAIR Principle Adoption through Innovative Social Infrastructure Tools 

Leonie Raijmakers, Edvard Hübinette, Elianna DeSota, Sina Iman, and Philipp Koellinger

The adoption of the FAIR data principles has revolutionised data management in Earth System Sciences (ESS), yet challenges persist in achieving true machine-actionability and comprehensive implementation. 

DeSci Labs introduces two innovative tools—Decentralised Persistent Identifiers (DPIDs) and the CODEX protocol—to address the barriers to FAIR data practice implementation in general; whilst fostering widespread uptake of FAIR principles in the ESS community in particular through involvement in the FAIR2Adapt consortium.

DPIDs are globally unique persistent identifiers based on, and linked directly to, the content of the files it refers to. Each version of every file, regardless of type, is automatically assigned a cryptographic fingerprint, ensuring deterministic resolution and transparent versioning. The DPID has been specifically designed to support FAIR digital research objects. The flexibility to alias DPIDs with existing systems (e.g., DOIs) and programmatic publishing capabilities via NodesLib enhances interoperability while preserving data sovereignty.

The CODEX protocol, an open scholarly infrastructure, further complements this by enabling the storage and retrieval of FAIR digital research objects via a decentralised peer-to-peer network (IPFS). This architecture allows multiple copies of the same content to be stored by different network participants using the same PID. By empowering researchers to collaborate within an open-state repository, the protocol minimises reliance on centralised actors, ensuring long-term accessibility, data integrity, and transparency. Its modular design facilitates diverse gateway applications, maximising participation and reducing barriers to entry.

These tools address core challenges in FAIR adoption by providing robust, scalable, and interoperable solutions tailored to the ESS community. By integrating DPIDs and CODEX into data workflows, researchers can enhance data reusability, improve the provenance of research outputs, and safeguard the collective scientific record. This presentation explores how these technologies can catalyse the next decade of FAIR data practices in ESS, fostering trust, reproducibility, and innovation.

How to cite: Raijmakers, L., Hübinette, E., DeSota, E., Iman, S., and Koellinger, P.: Advancing and Supporting FAIR Principle Adoption through Innovative Social Infrastructure Tools, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13330, https://doi.org/10.5194/egusphere-egu25-13330, 2025.

EGU25-13917 | ECS | Posters virtual | VPS19

Automated Mineral Grain Extraction for Geometallurgical Studies Using Segment Anything Model (SAM) and Core Scanning Techniques 

Yuanzhi Cai, Ryan Manton, and Morgan Williams

In mineral exploration and geometallurgical studies, accurately segmenting mineral grains from core scanning datasets may be used to predict metal recovery. This study introduces the application of the Segment Anything Model (SAM), a cutting-edge deep learning tool, to automate the segmentation and extraction of mineral grains from Laser-Induced Breakdown Spectroscopy (LIBS) and hyperspectral core scanning datasets. SAM demonstrates high efficiency and precision in identifying mineral grains, forming the foundation for downstream analyses, including the evaluation of mineral associations, grain size distribution, and other key geometallurgical metrics. Through case studies on pegmatite deposits, this research showcases the potential of SAM to address challenges posed by mineralogically complex ore. By enabling detailed mineralogical characterisation and advancing geometallurgical methods, SAM-based grain extraction presents a transformative tool for supporting sustainable and efficient mining practices.

How to cite: Cai, Y., Manton, R., and Williams, M.: Automated Mineral Grain Extraction for Geometallurgical Studies Using Segment Anything Model (SAM) and Core Scanning Techniques, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13917, https://doi.org/10.5194/egusphere-egu25-13917, 2025.

EGU25-14821 | ECS | Posters virtual | VPS19

Simulation of Monthly Global Sea Surface Temperature Data using Ensemble GAN Model 

Deepayan Chakraborty and Adway Mitra

Synthetic data has become an indispensable tool in climate science, offering extensive spatio-temporal
coverage to address data limitations in both current and future scenarios. Such synthetic data, derived
from climate simulation models, must exhibit statistical consistency with observational datasets to ensure
their utility. Among global climate simulation initiatives, the Coupled Model Intercomparison Project
Phase 6 (CMIP6) represents the latest and most comprehensive suite of General Circulation Models
(GCMs). However, the substantial High Performance Computing (HPC) resources required for these
physics-based models limit their accessibility to a broader research community. In response, genera-
tive machine learning models have emerged as a promising alternative for simulating climate data with
reduced computational demands.
This study introduces an ensemble model based on the Pix2Pix conditional Generative Adversarial
Network (cGAN) to generate high-resolution spatio-temporal maps of monthly global Sea Surface Tem-
perature (SST) with significantly lower computational cost and time. The proposed model comprises two
components: the GAN, which produces simulated SST climatology data , and the Predictor, which is
trained with the variability of the data that forecasts SST anomaly for the subsequent month using the
output data from the previous month. Both components contain the same architecture, but the training
processes are different. The predictor model can be fine-tuned with observed data for some epochs to
adopt its domain.
The ensemble model was calibrated with monthly SST observations from the COBE dataset as in-
put and output. The Empirical Orthogonal Functions (EOF) shows the model’s ability to simulate the
variabilty of the observed data. The model’s performance was evaluated using the temporal Pearson cor-
relation coefficient and mean squared error (MSE). Results demonstrate that the ensemble cGAN model
generates maps with statistical characteristics closely matching those of CMIP6 simulations and obser-
vations, achieving a mean temporal correlation coefficient around 0.5 and an MSE around 1.13 for both
cases.

How to cite: Chakraborty, D. and Mitra, A.: Simulation of Monthly Global Sea Surface Temperature Data using Ensemble GAN Model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14821, https://doi.org/10.5194/egusphere-egu25-14821, 2025.

EGU25-14839 | ECS | Posters virtual | VPS19

Leveraging MAUNet for Bias Correction of TRMM Precipitation Estimates 

Sumanta Chandra Mishra Sharma and Adway Mitra

Deep neural networks have revolutionized various fields due to their remarkable adaptability, enabling them to address related tasks through retraining and transfer learning. These capabilities make them invaluable tools for diverse applications, including climate and hydrological modeling. In an earlier work (Mishra Sharma et al., 2024), we introduced a novel neural network architecture, the Max-Average U-Net (MAUNet), which leverages Max-Average Pooling to downscale gridded precipitation data to higher spatial resolutions. The model demonstrated significant improvements in resolving finer-scale precipitation features, making it well-suited for climate data applications.

In this study, we utilized the MAUNet architecture to tackle the critical task of bias correction in satellite-based precipitation estimates. Bias correction is essential for improving the reliability of precipitation data derived from satellite missions, which often exhibit systematic discrepancies compared to ground-based measurements. Specifically, we focused on correcting biases in precipitation estimates from the Tropical Rainfall Measuring Mission (TRMM) by calibrating them against high-resolution, ground-based gridded datasets from the India Meteorological Department (IMD).

Our experimental results reveal that MAUNet effectively reduces biases in TRMM precipitation estimates, achieving significantly improved agreement with ground truth data. This success is attributed to the model’s robust feature extraction and reconstruction capabilities, which enable it to learn and correct systematic errors in satellite data. The findings also highlight the potential of advanced neural network architectures in addressing bias correction challenges.

This work underscores the utility of deep learning architectures in precipitation modeling, contributing to broader goals of improving the spatial distribution of precipitation estimates. By bridging the gap between satellite observations and ground truth, the MAUNet model offers a comprehensive solution for enhancing precipitation datasets, with significant implications for climate research, hydrological studies, and policy planning.

How to cite: Mishra Sharma, S. C. and Mitra, A.: Leveraging MAUNet for Bias Correction of TRMM Precipitation Estimates, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14839, https://doi.org/10.5194/egusphere-egu25-14839, 2025.

EGU25-15216 | Posters virtual | VPS19

A relevant accessible and interoperable geotechnical data tool to support the landslide risk management 

Graziella Emanuela Scarcella, Luigi Aceto, and Giovanni Gullà

The rising frequency and severity of landslides, exacerbated by the effects of climate change and human development in unstable areas, call for effective risk management strategies. In this context, a systematic collection of all the available data regarding geotechnical aspects, in particular geomaterial parameters, results plays a crucial role, providing a decisive contribution to define strategies for sustainable landslide risk management.

In this work, we present the translation of a geotechnical database to the aims of the project Tech4You Innovation Ecosystem – Goal 1 - Pilot Project 1, useful to identify the typical landslide scenarios, to identify sufficient knowledge for the definition of the geotechnical model and geomaterials typing in similar geo-environmental contexts. The database contains the results of laboratory tests carried out in the past by researchers at CNR IRPI in Rende, relating to 11 sites in Calabria, of which 10 in the Province of Catanzaro and 1 in the Province of Vibo Valentia.  For each site, geotechnical characterisation data of the geomaterials, which represent a key cognitive element, were grouped by type of laboratory test (grain size, indices, Atterberg limits, oedometric, direct shear and triaxial tests). We uploaded these data to validate a tool, named GeoDataTech vers. 2.0, which is an update of a previous version. In particular, we have tested the correct functioning (display, query, extraction data) with a significant sample of data. GeoDataTech vers. 2.0 can manage 2399 laboratory tests to date: 61 oedometric tests, 636 grain size, 537 indices, 78 Atterberg limits, 454 specific gravity, 512 direct shear tests and 121 triaxial tests.

This tool will be available to a wide range of stakeholders (researchers, professionals, territorial administrations, public bodies and citizens) allowing us to acquire, interrogate, export data and to upload their own files to integrate them into the database of the tool, performing advanced analyses with reference to the typification of geomaterials. By enabling the sharing of such data between researchers, practitioners and public institutions, the geotechnical tool will contribute significantly to improving disaster prevention strategies, in particular with regard to the reduction of landslide risks, thereby responding to the growing demand for accessible and interoperable data networks that increase synergic interdisciplinary research on topics such as landslide hazard.

This work was funded by the Next Generation EU—Italian NRRP, Mission 4, Component 2, Investment 1.5, call for the creation and strengthening of ‘Innovation Ecosystems’, building ‘Territorial R&D Leaders’ (Directorial Decree n. 2021/3277)—project Tech4You—Technologies for climate change adaptation and quality of life improvement, n. ECS0000009. This work reflects only the authors' views and opinions, neither the Ministry for University and Research nor the European Commission can be considered responsible for them.

How to cite: Scarcella, G. E., Aceto, L., and Gullà, G.: A relevant accessible and interoperable geotechnical data tool to support the landslide risk management, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15216, https://doi.org/10.5194/egusphere-egu25-15216, 2025.

EGU25-15381 | ECS | Posters virtual | VPS19

Prediction of landuse landcover using CA-Markov model for the valley regions of Manipur, India 

Maisnam Nongthouba, Bakimchandra Oinam, and Khwairakpam Sachidananda

Changes in land use and cover (LULC) serve as critical indicators of socioeconomic and environmental shifts induced by both natural and man-made factors. This assessment was carried out in the Imphal valley region to forecast changes in land use and land cover. In order to examine the spatiotemporal distributions of LULC, the LULC Classification was analysed using Landsat images from 2007, 2014, and 2017. The CA-Markov Chain model was used to simulate the future LULC for the year 2030 of Imphal valley region based on these the past LULCs. The model result showed that wetland herbaceous will decline by 3.3% and settlement area will expand by 28.71%. The Imphal city area is where the majority of the expanding settlement area is located. As a vital resource for future planning initiatives, this study suggests planners, environmentalists, and decision-makers to prioritise sustainable practices and make appropriate decisions for the sustainability of the region.

How to cite: Nongthouba, M., Oinam, B., and Sachidananda, K.: Prediction of landuse landcover using CA-Markov model for the valley regions of Manipur, India, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15381, https://doi.org/10.5194/egusphere-egu25-15381, 2025.

EGU25-16092 | Posters virtual | VPS19

Transforming GNSS Data into FAIR Digital Objects 

Carine Bruyninx, Anna Miglio, Andras Fabian, Juliette Legrand, Eric Pottiaux, and Fikri Bamahry

GNSS (Global Navigation Satellite System) data play a crucial role in both scientific research and practical applications. GNSS datasets are used to monitor atmospheric conditions, tectonic plate movements, and Earth deformation, providing valuable insights for geodetic and geophysical studies. Although widely accessible, GNSS data often lacks the necessary structure and metadata for effective reuse, particularly for data-driven research based on machine learning. To address these challenges, we applied the FAIR (Findable, Accessible, Interoperable, Reusable) data principles to GNSS RINEX observation files hosted by the EUREF Historical Data Centre (EUREF-HDC).

The EU action plan “Turning FAIR into Reality” introduced the concept of FAIR Digital Objects (FDOs), emphasizing the need for Persistent Identifiers (PIDs) and rich, standardized metadata to ensure data can be reliably found, accessed, utilized, and cited. Building on this foundation, we developed a multi-layered FDO structure centered on GNSS RINEX data. Given the established nature of the EUREF-HDC repository, we adapted the FDO concept by prioritizing structured metadata, followed by persistent identifiers and robust (meta)data access procedures.

To implement this approach, we designed the GNSS-DCAT-AP metadata schema, assigned PIDs to both data and metadata, and developed web services enabling humans and machines alike to seamless search, retrieve, and download (meta)data. The effectiveness of our solution was evaluated using the FAIRsFAIR Data Object Assessment Metrics, demonstrating a significant improvement in FAIR compliance.

This work showcases the feasibility of transforming GNSS RINEX data into FAIR Digital Objects and could provide a practical roadmap for other geospatial data repositories seeking alignment with FAIR principles.

How to cite: Bruyninx, C., Miglio, A., Fabian, A., Legrand, J., Pottiaux, E., and Bamahry, F.: Transforming GNSS Data into FAIR Digital Objects, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16092, https://doi.org/10.5194/egusphere-egu25-16092, 2025.

EGU25-16363 | ECS | Posters on site | GI2.4

Hybrid Machine Learning approach for Tropical Cyclones Detection 

Davide Donno, Gabriele Accarino, Donatello Elia, Enrico Scoccimarro, and Silvio Gualdi

Tropical Cyclones (TCs) are among the most impactful weather phenomena, with climate change intensifying their duration and strength, posing significant risks to ecosystems and human life. Accurate TC detection, encompassing localization and tracking of TC centers, has become a critical focus for the climate science community. 

Traditional methods often rely on subjective threshold tuning and might require several input variables, thus making the tracking computationally expensive. We propose a cost-effective hybrid Machine Learning (ML) approach consisting in splitting the TC detection into two separate sub-tasks: localization and tracking. The TC task localization is fully data-driven: multiple Deep Neural Networks (DNNs) architectures have been explored to localize TC centers using a different set of input fields related to the cyclo-genesis, aiming also at reducing the number of input drivers required for detection. A neighborhood matching algorithm is then applied to join previously localized TC center estimates into potential trajectories over time. 

We train the DNNs on 40 years of ERA5 reanalysis data and International Best Track Archive for Climate Stewardship (IBTrACS) records across the East and West North Pacific basins. The hybrid approach is then compared with four state-of-the-art deterministic trackers (namely OWZ, TRACK, CNRM and UZ), reporting comparable or even better results in terms of Probability of Detection and False Alarm Rate, additionally capturing the interannual variability and spatial distribution of TCs in the target domain. 

The resulting hybrid ML model represents the core component of a Digital Twin (DT) application implemented in the context of the EU-funded interTwin project.

How to cite: Donno, D., Accarino, G., Elia, D., Scoccimarro, E., and Gualdi, S.: Hybrid Machine Learning approach for Tropical Cyclones Detection, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16363, https://doi.org/10.5194/egusphere-egu25-16363, 2025.

The STARS4Water project addresses the critical need to understand the impacts of climate change and anthropogenic activities on freshwater availability and ecosystem resilience at the river basin scale. By developing innovative data services and models tailored to stakeholder needs, the project will improve decision-making processes for sustainable water resource management. A distinctive feature of STARS4Water is its focus on co-creating solutions with local stakeholders using a living lab approach, ensuring that newly developed tools remain relevant and usable beyond the life of the project.

 

This extension of the original project—funded with a special grant from Unitatea Executivă pentru Finanțarea Învățământului Superior, a Cercetării, Dezvoltării și Inovării (UEFISCDI) from Romania—focuses on a detailed change detection analysis to monitor and quantify land cover transformations in the emblematic Danube Delta region. The objective is to assess how environmental and anthropogenic changes have influenced this ecologically significant wetland over several decades. To achieve this, a comprehensive database of multispectral satellite images from the Landsat archive, spanning from 1985 to 2023, will be constructed. The long-term dataset enables a detailed temporal analysis, important for detecting land cover dynamics over time.

 

The methodology involves several key phases: (1) data collection and preprocessing of Landsat satellite images to correct errors and align imagery for consistent comparative analysis; (2) sampling and training a deep learning model using convolutional neural network (CNN) architectures, to classify various land cover types; (3) performing land cover classification on the processed images using the trained model, followed by accuracy assessment; and (4) conducting a comprehensive change detection analysis to quantify and interpret the observed transformations in land use and land cover.

 

The results of this analysis will deliver important knowledge on the long-term dynamics of the Danube Delta landscape, highlighting critical changes with implications for biodiversity, water management and ecosystem services. This approach will support adaptive ecosystem management and contribute to the scientific understanding of climate-related and anthropogenic changes in fragile wetland ecosystems.

 

Acknowlegments

This work was supported by a grant of the Ministry of Research, Innovation and Digitization, CNCS/CCCDI - UEFISCDI, project number PN-IV-P8-8.1-PRE-HE-ORG-2023-0094, within PNCDI IV.

How to cite: Scrieciu, A. and Toma, A.: Monitoring Long-Term Land Cover Transformations in the Danube Delta using Landsat Satellite Imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16549, https://doi.org/10.5194/egusphere-egu25-16549, 2025.

EGU25-16634 | Posters virtual | VPS19

Finetuning and Benchmarking an AI Foundation Model for Cloud Gap Imputation  

Tadie Birihan Medimem, Gabriele Padovani, Takuya Kurihana, Ankur Kumar, Farid Melgani, Valentine G Anantharaj, and Sandro Luigi Fiore

Abstract: Cloud cover poses a significant obstacle in harnessing multi-spectral satellite imagery for various earth observation applications including disaster response, land use and land cover mapping. To address this issue, this study investigates the potential of Prithvi WxC foundation model (Johannes Schmude et al., 2024), a deep learning architecture designed for weather and climate applications, to perform cloud gap imputation. By leveraging its ability to capture atmospheric dynamics and predict missing data, Prithvi WxC offers a promising solution.

The primary objective is to assess the accuracy and efficiency of Prithvi WxC in reconstructing cloudy pixels in Moderate Resolution Imaging Spectroradiometer (MODIS) surface reflectance; MOD09 (Eric Vermote, 2015). MOD09 data provides valuable information about earth surface, cloud cover and atmospheric conditions, which are instrumental in informing the Prithvi WxC model during the finetuning and imputation process.

This research evaluates the Prithvi WxC foundation model for cloud gap imputation applications and benchmarks its performance against other foundation models, such as Prithvi EO (Jakubik et al., 2023, 2024). The process begins with preprocessing the MOD09 dataset, filtering out missing and cloudy pixels to create clean visible patches, while real-world cloudy patches are used as masks. The preprocessed data is then resampled to align with the temporal and spatial resolution requirements of both the Prithvi WxC and Prithvi EO foundation models. Through rigorous fine-tuning strategies, these models learn to reconstruct the masked regions, effectively filling the gaps caused by cloud cover. Finally, the fine-tuned foundation models are benchmarked using quantitative metrics, such as the Structural Similarity Index Measure (SSIM) and Mean Absolute Error (MAE), complemented by qualitative visual analysis.

This research explores the potential of Prithvi WxC foundation model, pre-trained on extensive weather and climate data, to improve cloud gap imputation in satellite imagery, and subsequently benchmarks it against earth observation foundation models, such as Prithvi EO. Through this evaluation, we aim to enhance scientific understanding via multi-modality and sensor-independent approaches.

 References

Johannes Schmude, Sujit Roy, Will Trojak, Johannes Jakubik, Daniel Salles Civitarese, Shraddha Singh, Julian Kuehnert, Kumar Ankur, Aman Gupta, Christopher E Phillips, Romeo Kienzler, Daniela Szwarcman, Vishal Gaur, Rajat Shinde, Rohit Lal, Arlindo Da Sil: Prithvi WxC: Foundation Model for Weather and Climate." arXiv preprint arXiv:2409.13598, 2024.

C. Roger, E. F. Vermote, J. P. Ray: https://modis-land.gsfc.nasa.gov/pdf/MOD09_UserGuide_v1.4.pdf. NASA, MODIS Surface Reflectance User’s Guide, Collection 6, 2015.

Daniela Szwarcman, Sujit Roy, Paolo Fraccaro, Þorsteinn Elí Gíslason, Benedikt Blumenstiel, Rinki Ghosal, Pedro Henrique de Oliveira, Joao Lucas de Sousa Almeida, Rocco Sedona, Yanghui Kang, Srija Chakraborty, Sizhe Wang, Ankur Kumar, Myscon Truong, Denys: Prithvi-EO-2.0: A Versatile Multi-Temporal Foundation Model for Earth Observation Applications. https://arxiv.org/abs/2412.02732, 2024.

Johannes Jakubik, Sujit Roy, C. E. Phillips, Paolo Fraccaro, Denys Godwin, Bianca Zadrozny, Daniela Szwarcman, Carlos Gomes, Gabby Nyirjesy, Blair Edwards, Daiki Kimura, Naomi Simumba, Linsong Chu, S. Karthik Mukkavilli, Devyani Lambhate, Kamal Das, Ranji: Foundation Models for Generalist Geospatial Artificial Intelligence, 2023.

Eric Vermote: MOD09 MODIS/Terra L2 Surface Reflectance, 5-Min Swath 250m, 500m, and 1km. NASA LP DAAC., NASA GSFC and MODAPS SIPS, NASA, http://doi.org/10.5067/MODIS/MOD09.061, 2015.

How to cite: Medimem, T. B., Padovani, G., Kurihana, T., Kumar, A., Melgani, F., Anantharaj, V. G., and Fiore, S. L.: Finetuning and Benchmarking an AI Foundation Model for Cloud Gap Imputation , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16634, https://doi.org/10.5194/egusphere-egu25-16634, 2025.

EGU25-18822 | Posters virtual | VPS19

Visualizing a climate and disaster resilience taxonomy from research evidence: scaling and accelerating knowledge interoperability 

Sukaina Bharwani, Rosie Witton, Kate Williamson, and Ruth Butterfield

The urgency of the climate crisis and the need to accelerate learning and climate action requires that we build on previous knowledge, rather than replicating it. There is an abundance of knowledge on climate change adaptation and mitigation dispersed across websites, projects, platforms, and documents. There is either too much information that is not easily discoverable (sitting in silos) or it is too technical or complex, and not ‘usable’ or fit for purpose in terms (e.g. language or format). Both issues cause redundancy and sometimes replication of work, wasting resources. In the worst case, they can also cause unintended consequences such as maladaptation, or increased vulnerability. However, the issue is not a lack of information, but rather how to organise and connect such knowledge to allow people to discover what already exists and put it to effective use. As such, our goal is to make climate action knowledge findable, accessible, interoperable and reusable (FAIR) and reduce climate change knowledge silos. The recently awarded FAIR2Adapt Project aims to establish a comprehensive FAIR and open data framework for CCA and to demonstrate the impact of FAIR data on CCA strategies. By making CCA data FAIR, FAIR2Adapt will accelerate adaptation actions so that they are visible, understandable, and actionable for various purposes and different types of stakeholders. FAIR taxonomies are one approach to help tackle this issue by making climate change knowledge FAIR and by ensuring, that going forward, platforms have a way to make their knowledge FAIR and thus more reusable by the climate change community. 


The Climate Connectivity Hub and Taxonomy seek to visualize and connect online platform data (e.g. Cordis, Climate-ADAPT, weADAPT, PreventionWeb) to increase discoverability, interoperability and a shared understanding of the research results and their potential application in future policy, research and practice. It builds on past knowledge to scale up and accelerate climate action whilst also identifying key knowledge gaps. This presentation will show that: 1) taxonomies are useful supporting the interoperability of online climate knowledge and can usefully emerge from combined expert and machine learning of project results (e.g. Cordis); 2) shared vocabularies and different interpretations of language and terminology add value to project planning and implementation ; and, 3) the visualization of these elements for decision-makers, planners, researchers, policy makers, etc. can help to enable and scale accelerated climate action. 

How to cite: Bharwani, S., Witton, R., Williamson, K., and Butterfield, R.: Visualizing a climate and disaster resilience taxonomy from research evidence: scaling and accelerating knowledge interoperability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18822, https://doi.org/10.5194/egusphere-egu25-18822, 2025.

EGU25-18933 | Posters virtual | VPS19

Deep Learning-based Spatial-Spectral Analysis for Peatland Degradation characterization 

Harsha Vardhan Kaparthi and Alfonso Vitti

The study explores using advanced deep learning (DL) techniques for spatial-spectral analysis to detect and map peatland degradation at a granular level. Peatlands, vital carbon sinks in global ecosystems, face degradation threats that demand precise and scalable monitoring solutions. Our method combines convolutional neural networks (CNNs), fully convolutional networks (FCNs), and 3D CNNs to examine complex spatial-spectral patterns in SAR, multispectral, and hyperspectral sensor data (e.g., Sentinel-1, Sentinel-2, PRISMA) over the temperate peatland study area of the Monte Bondone region (Latitude: 46°00’48.6” N, Longitude: 11°03'14.6” E), covering an area of 40 hectares as shown in the figures.

CNNs capture spatial relationships between precipitation, temperature, vegetation, soil, and moisture, offering a detailed view of peatland composition. Using multi-dimensional, gridded data from meteorological stations and remote sensing images, CNNs identify patterns affecting peatland health. Fully Convolutional Networks (FCNs) help with spectral unmixing, isolating land cover components at the pixel level, which aids in detecting vegetation degradation and understanding ecosystem changes.

3D CNNs incorporate temporal data to classify Peatland landscapes into different degradation states. The model identifies changes over time, distinguishing between healthy, partially degraded, and fully degraded regions. Deep clustering models also classify peatland areas into degradation states, revealing trends without labeled data.

This deep learning framework supports accurate degradation mapping through spatial-spectral feature extraction, providing precise, pixel-level information to aid ecosystem management and conservation. It helps monitor peatland health and assess environmental changes across diverse landscapes.

How to cite: Kaparthi, H. V. and Vitti, A.: Deep Learning-based Spatial-Spectral Analysis for Peatland Degradation characterization, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18933, https://doi.org/10.5194/egusphere-egu25-18933, 2025.

EGU25-19489 | ECS | Posters virtual | VPS19

Model-Agnostic Meta-Learning for Data Integration Across Heterogeneous Hydrological Datasets 

Asma Slaimi and Michael Scriney

Integrating heterogeneous hydrological datasets remains a significant challenge in environmental modelling due to variations in feature spaces, data distributions, and temporal and spatial scales across sources. This study introduces a Model-Agnostic Meta-Learning (MAML) approach to address the challenge of integrating heterogeneous hydrological datasets, leveraging a collection of datasets compiled from diverse sources. These datasets, characterized by varying features, distributions, and temporal and spatial scales, provide an ideal basis for evaluating MAML's ability to handle real-world data heterogeneity.

MAML’s unique capability to learn shared representations across datasets with minimal feature overlap and significant variability allows it to effectively transfer knowledge between subsets, offering a flexible and scalable solution for integrating hydrological data with diverse characteristics.

The proposed approach trains a base model on one subset of the data while utilizing MAML's meta-learning capabilities to adapt and transfer knowledge to other subsets with differing feature distributions. To test the model's adaptability, we simulate scenarios with varying degrees of feature overlap. Model performance is assessed using metrics such as mean squared error (MSE), both before and after fine-tuning on unseen data subsets.

Preliminary results demonstrate that MAML effectively learns shared representations across datasets, achieving significant improvements in prediction accuracy. Fine-tuning further enhances the model's adaptability, particularly for datasets with minimal feature overlap. These findings highlight MAML's potential as a powerful and flexible tool for integrating and predicting across heterogeneous hydrological datasets.

This study bridges the gap between advanced meta-learning techniques and hydrological applications, providing new insights into scalable and adaptable data integration methods for environmental sciences.

Keywords: Model-Agnostic Meta-Learning, hydrological datasets, data integration, heterogeneous data, meta-learning, environmental modelling, machine learning. 

How to cite: Slaimi, A. and Scriney, M.: Model-Agnostic Meta-Learning for Data Integration Across Heterogeneous Hydrological Datasets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19489, https://doi.org/10.5194/egusphere-egu25-19489, 2025.

EGU25-20324 | ECS | Posters virtual | VPS19

Using Remote sensing and geographic information system for delineating suitable sites for artificial groundwater recharge: A multi-criteria decision-making approach. 

Rahma Fri, Andrea Scozzari, Souad Haida, Malika Kili, Lamia Erraoui, Jamal Chaou, Abdelaziz Mridekh, Lahcen Goumghar, and Bouabid El Mansouri

The semi-arid region of Deraa Oued Noun in Morocco faces significant challenges related to water scarcity, which greatly affects the availability of groundwater resources. With recurring droughts and periods of water shortage, it is imperative to address these challenges and implement effective measures for sustainable groundwater resource management. Artificial groundwater recharge has proven to be a viable solution for alleviating water scarcity issues. By capturing and storing excess water during periods of heavy precipitation or surface water availability, artificial recharge can replenish depleted aquifers and provide a reliable water source during drought periods. However, the success of recharge projects depends on identifying suitable sites that meet specific criteria and maximize the efficiency of the recharge process.

The identification of suitable sites for artificial groundwater recharge in Daraa Oued Noun, through the integration of remote sensing, GIS (Geographic Information System), and MCDM (Multi-Criteria Decision Making) techniques, offers a promising solution to address water scarcity challenges in the context of climate change. The proposed research project aims to provide valuable and spatially explicit information for strategic groundwater resource management.

 This study was conducted in the Deraa Oued Noun district, where water shortages have been observed over the years. The research utilized geology, soil, land use, stream data, and Sentinel-2 and DEM images to develop thematic layers, including lithogeology, soil, slope, lineament density, land use, stream density, and water surface. Additionally, data on the vadose zone thickness were incorporated to enhance the analysis.

By integrating GIS and image processing techniques, these thematic layers were utilized to prepare groundwater recharge maps of the area through a weighted overlay method on a GIS platform. The results revealed that artificial recharge potential was high in the northern and western parts of the study area.

By following a systematic and rigorous methodology, including data collection, remote sensing analysis, MCDM evaluation, and site validation, this project aims to contribute to the successful implementation of artificial recharge projects in the region. By maximizing the efficiency of the recharge method, these projects will help ensure sustainable water supply, mitigate the impacts of drought, and promote long-term water security in Derâa Oued Noun and similar semi-arid regions.

How to cite: Fri, R., Scozzari, A., Haida, S., Kili, M., Erraoui, L., Chaou, J., Mridekh, A., Goumghar, L., and El Mansouri, B.: Using Remote sensing and geographic information system for delineating suitable sites for artificial groundwater recharge: A multi-criteria decision-making approach., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20324, https://doi.org/10.5194/egusphere-egu25-20324, 2025.

EGU25-20616 | ECS | Posters virtual | VPS19

LUCIE: A Lightweight Uncoupled ClImate Emulator with long-term stability and physical consistency for O(1000)-member ensembles 

Haiwen Guan, Troy Arcomano, Ashesh Chattopadhyay, and Romit Maulik

We present LUCIE, a data-driven atmospheric emulator that remains stable during autoregressive inference for a thousand of years with minimal drifting climatology. LUCIE was trained using 9.5 years of coarse-resolution ERA5 data, incorporating 5 prognostic variables, 2 forcing variables, and one diagnostic variable (6-hourly total precipitation), all on a single A100 GPU over a two-hour period. LUCIE autoregressively predicts the prognostic variables and outputs the diagnostic variables similar to AllenAI’s ACE climate emulator. Unlike all the other state-of-the-art AI weather models, LUCIE is neither unstable nor does it produce hallucinations that result in unphysical drift of the emulated climate. The low computational requirements of LUCIE allow for rapid experimentation including the development of novel loss functions to reduce spectral bias and improve tails of the distributions. Furthermore, LUCIE does not impose true sea-surface temperature (SST) from a coupled numerical model to enforce the annual cycle in temperature. We demonstrate the long-term climatology obtained from LUCIE as well as subseasonal-to-seasonal scale prediction skills on the prognostic variables. LUCIE is capable of 6000 years of simulation per day on a single GPU, allowing for O(100)-ensemble members for quantifying model uncertainty for climate and ensemble weather prediction.

How to cite: Guan, H., Arcomano, T., Chattopadhyay, A., and Maulik, R.: LUCIE: A Lightweight Uncoupled ClImate Emulator with long-term stability and physical consistency for O(1000)-member ensembles, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20616, https://doi.org/10.5194/egusphere-egu25-20616, 2025.

EGU25-21759 | Posters virtual | VPS19

Assessment of Atmospheric Pollen Presence in Urban Areas of Greece During CALIPSO Overpasses 

Archontoula Karageorgopoulou, Stathopoulos Christos, Georgiou Thanasis, Shang Χiaoxia, Pyrri Ioanna, Amiridis Vassilis, and Giannakaki Elina

Analysis of pollen events was conducted using Hirst-type volumetric samplers in Athens and Thessaloniki in combination with CALIPSO vertical aerosol profiles. While Hirst-type ‎[1] volumetric samplers are used to confirm and characterize pollen at ground level, the understanding of pollen vertical distribution and transport is still limited. The utilization of Light Detection and Ranging (LIDAR) for identifying different pollen types is increasingly prevalent, as the depolarization ratio is related to the shape of the pollen particles while other non-spherical particle types are absent ‎[2].
Samplers are situated on the buildings’ rooftops of the Physics and Biology Departments, in Athens and Thessaloniki, respectively. Following ‎[2], intense pollen events are considered when the pollen concentration exceeds 400 grains m-3 for a minimum of two hours each day.
CALIPSO provides unique vertical profile measurements of the Earth’s atmosphere on a global scale ‎[3], with the ability to distinguish between feature types (i.e., clouds vs. aerosol) and subtypes (i.e., marine, dust, clean continental). Only case studies where CALIPSO aerosol layers were classified as marine, dusty marine, dust, or polluted dust were analyzed.
Model simulations were used to exclude the presence of other depolarizing aerosol types. HYSPLIT (Hybrid Single-Particle Lagrangian Integrated Trajectory) was used to trace the origin of the air masses. The atmospheric model RAMS/ICLAMS (Regional Atmospheric Modeling System/Integrated Community Limited Area Modeling System) was selected to describe dust and sea-salt emissions and transport.
Mean values of lidar-derived optical properties inside the detected pollen layers are provided from CALIPSO data analysis. Specifically, there are three observed aerosol layers, one over Athens (12-3-2021) and two over Thessaloniki (2-3-2020, 10-4-2020). Particulate color ratios of 0.652 ± 0.194, 0.638 ± 0.362, and 0.456 ± 0.284, and depolarization ratios of 8.70 ± 6.26%, 28.30 ± 14.16%, and 8.96±6.87 % for 12-3-2021, 2-3-2020 and 10-4-2020, respectively, were misclassified by CALIPSO as marine-dusty marine, dust and polluted dust. The pollen analysis conducted on the 12th of March 2021 indicated that the dominant pollen types were 69% Pinaceae and 24% Cupressaceae. On the 2nd of March 2020, Cupressaceae accounted for 97% of the total pollen, while on the 10th of April 2020, Carpinus represented 76% and Platanus 15%. Consequently, during periods of intense pollen presence, CALIPSO vertical profiles and aerobiological monitoring techniques may be used synergistically to better characterize the atmospheric pollen layers.

Acknowledgements
The research work was supported by the Hellenic Foundation for Research and Innovation (H.F.R.I.) under the “Basic Research Financing (Horizontal support for all Sciences), National Recovery and Resilience Plan (Greece 2.0)” (Project Number: 015144).

[1] J. M. Hirst, Annals of Applied Biology 39, 157-293 (1952).
[2] X. Shang et al., Atmos. Chem. Phys. 20, 15323–15339 (2020).
[3] D. M. Winker et al, BAMS 91, 1211–1229 (2010).

How to cite: Karageorgopoulou, A., Christos, S., Thanasis, G., Χiaoxia, S., Ioanna, P., Vassilis, A., and Elina, G.: Assessment of Atmospheric Pollen Presence in Urban Areas of Greece During CALIPSO Overpasses, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21759, https://doi.org/10.5194/egusphere-egu25-21759, 2025.

EGU25-2237 | Posters virtual | VPS20

Assessment of GOCI-II satellite remote sensing products in Lake Taihu 

Min Zhao, Huaming Li, Hao Li, Xuan Zhang, Xiaosong Ding, and Fang Gong

The Geostationary Ocean Color Imager-II (GOCI-II), which was launched on February 19, 2020, offers an increased observation times within a day and finer spatial resolution than those of its predecessor, the Geostationary Ocean Color Imager (GOCI), which was launched in 2010. To ensure the reliability of GOCI-II data for practical applications, the accuracy of remote sensing products must be validated. In this study, we employed in situ data from Lake Taihu for validation. We assessed the accuracy of GOCI-II products, including the remote sensing reflectance inverted via two atmospheric correction algorithms (ultraviolet (UV) and near-infrared (NIR) atmospheric correction algorithms), as well as the chlorophyll a (Chl-a) concentration, total suspended matter (TSM) concentration, and phytoplankton absorption coefficient (aph). Our results revealed that the UV atmospheric correction algorithm provided a relatively higher accuracy in Lake Taihu, with average absolute percentage deviations (APDs) of the remote sensing reflectance across different bands of 25.17% (412 nm), 29.69% (443 nm), 22.27% (490 nm), 19.38% (555 nm), 36.83% (660 nm), and 33.0% (680 nm). Compared to the products generated using the NIR atmospheric correction algorithm, the derived Chl-a concentration, TSM concentration, and aph products from the UV algorithm showed improved accuracy, with APD values reduced by 16.92%, 3.32%, and 10.91%, respectively. When using UV correction, the 412 nm band performed better than the 380 nm band, likely due to the lower signal-to-noise ratio of the 380 nm band and smaller extrapolation errors when assuming a zero signal for the 412 nm band. Considering that the NIR algorithm is suitable for open ocean waters while the UV algorithm demonstrates higher accuracy in highly turbid environments, a combined UV-NIR atmospheric correction algorithm may be more suitable for addressing different types of water environments. Additionally, more suitable retrieval algorithms are needed to improve the accuracy of Chl-a concentration and aph in eutrophic waters.

How to cite: Zhao, M., Li, H., Li, H., Zhang, X., Ding, X., and Gong, F.: Assessment of GOCI-II satellite remote sensing products in Lake Taihu, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2237, https://doi.org/10.5194/egusphere-egu25-2237, 2025.

EGU25-4994 | ECS | Posters virtual | VPS20

bibliometric analysis of natural lakes and paleolakes origin of natural events 

Jamal Abbach, Said El Moussaoui, Hajar El Talibi, and Charaf Eddine Bouiss

This study explores studies on lakes and paleolakes originating from natural effects. The main objective is to perform a bibliometric analysis of research on naturally occurring lake environments worldwide, covering the period from 2014 to 2024. Data extracted from 1687 documents in the Scopus database were analyzed using VOSviewer software. The results reveal a strict trend towards a focus on geosciences and the environment, underlined by research. This study particularly highlights the relationships between authors, co-authors, keywords, and publishers of specialized journals in this research field, thus providing essential information to guide future research and to value the role of these geological environments, which are rare in the world, based on essentially multidisciplinary geoscience approaches.

How to cite: Abbach, J., El Moussaoui, S., El Talibi, H., and Bouiss, C. E.: bibliometric analysis of natural lakes and paleolakes origin of natural events, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4994, https://doi.org/10.5194/egusphere-egu25-4994, 2025.

Rock masses are characterized by the complex hierarchical structures involving various scale levels. The deformation of rock masses is primarily controlled in weak structural layers between rocks, whereas the rock block can be regarded as a non-deformable block and can move as a whole. In consequence, a new dynamic phenomenon, namely the pendulum-type wave, has emerged, which is a kind of nonlinear displacement wave caused by the overall movement of relatively intact large-scale rock blocks. Aiming at the complex hierarchical structures of rock masses and low-frequency characteristics of pendulum-type waves, the blocky rock masses composed of granite blocks and rubber interlayers are simplified into the block-spring model and wave motion model. Based on Bloch theorem and d’Alembert’s principle, the dispersion relation and equations of motion of 1D blocky rock masses are determined. Research shows that with the increase of the rock size and geomechanical invariant, the initial frequency of the first attenuation zone gradually decreases, and only the low-frequency waves lower than that frequency can propagate in blocky rock masses, which reveals the mechanism of low-frequency characteristics of pendulum-type waves theoretically. The equivalent substitution for the two models and their errors are given, and the results show that the equivalent substitution of the two models is not universal and unconditional. Finally, the influence of hierarchical structures on the dispersion relation and dynamic response is further studied. The larger the stiffness ratio, or the higher the order of hierarchical structures, the smaller is the effect of ignoring the high-order hierarchical structures.

How to cite: Jiang, K. and Qi, C.: Research on dispersion relation and dynamic properties of pendulum‑type waves in 1D blocky rock masses with complex hierarchical structures, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5033, https://doi.org/10.5194/egusphere-egu25-5033, 2025.

EGU25-5359 | ECS | Posters virtual | VPS20

Innovating Coral Reef Mapping with Drones & NASA Fluid Lensing Technology in the Mariana Islands 

Jonelle Sayama and Keanno Fausto

Coral reefs in the Mariana Islands serve important roles for the islands’ ecology and economy, contributing to the region’s fisheries, tourism, coastal protection, education, and cultural histories. Despite their immense value, the resilience of these marine ecosystems is threatened by an array of climate-change induced stressors, including ocean acidification and coral bleaching. In response, a team from the University of Guam (UOG) launched a large-scale coral reef mapping campaign to monitor priority reef sites throughout the Mariana Islands using drone technology. The UOG team, consisting of researchers and remote pilots funded by the USGS, Pacific Islands Climate Adaptation Science Center, NASA Guam EPSCoR, and NASA Guam Space Grant, has been conducting drone-based missions to capture high-resolution imagery of priority coral reef sites across Guam and Saipan. Their efforts aim to gather aerial data of coral reefs in Micronesia, providing resource managers with essential information regarding response and recovery. Initially, the campaign used of NASA’s fluid lensing technology developed by Chirayath (2019) for coral reef mapping. This technology combines unmanned aerial systems (UAS), off-the-shelf technology, and machine learning algorithms to create detailed coral reef maps by filtering out distortions caused by light and ocean waves, resulting in clear, high-resolution imagery. In 2024, this process was augmented to employ a new methodology that strategically uses RGB sensors and low tides. This system allows the remote pilots to capture the areas and produce orthomosaic maps at much more efficient rates while maintaining high-resolution quality. By providing these datasets within a shorter turn-around time, local natural resource managers are able to get a timely snapshot of the coral reef sites – providing crucial data of the ecosystem’s health that can help inform conservation decisions. This presentation will outline the collaborative efforts between UOG and regional partners, demonstrating how drones and fluid lensing technology are innovating coral reef monitoring efforts. It will explore how the collected data can help local resource managers make informed decisions regarding coral reef management, showcase coral reefs to the general public, ultimately transforming how local communities can contribute to coral reef resilience.

How to cite: Sayama, J. and Fausto, K.: Innovating Coral Reef Mapping with Drones & NASA Fluid Lensing Technology in the Mariana Islands, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5359, https://doi.org/10.5194/egusphere-egu25-5359, 2025.

EGU25-6240 | ECS | Posters virtual | VPS20

Enhancing the use of Geoinformation technologies to assess the socioeconomic impacts of climate change in the Arctic: Insights from the EO-PERSIST Project 

Georgios-Nektarios Tselos, Spyridon E. Detsikas, Beata Kroszka, Patryk Grzybowski, and George P. Petropoulos

In today's changing climate, there is an urgent need to understand the adverse impacts of climate change on natural environments, infrastructures, and industries.Particularly permafrost regions in the Arctic are highly vulnerable to global warming, impacting both the environment and socioeconomic aspects. Thus, systematic monitoring of such environments, is of paramount significance. Advances in Geoinformation technologies, and in particular in Earth Observation (EO), cloud computing, GIS, web cartography create new opportunities and challenges for Arctic research examining the socioeconomic impact of climate change.The rapid advancements in EOin particular have led to an exponential increase in the volume of geospatial data that come from spaceborne EO sensors. This surge, combined with the fast developments in GIS and web cartography present significant challenges for effective management, access, and utilization by researchers, policymakers, and the public. Consequently, there is a growing need for advanced methodologies to organize, process, and deliver geospatial information that comes from EO satellites in an accessible and user-friendly manner.

Recognizing thepromising potential of geoinformation technologies, the European Union (EU) has funded several research projects that leverage advanced technologies such as geospatial databases and WebGIS platforms to streamline EO data handling and dissemination. One such project is EO-PERSIST (http://www.eo-persist.eu), which aims to create a collaborative research and innovation environment focusing on leveraging existing services, datasets, and emerging technologies to achieve a consistently updated ecosystem of EO-based datasets for permafrost applications. To formulate the socioeconomic indicators, the project exploits state of the art cloud processing resources, innovative Remote Sensing (RS) algorithms, Geographic Information Systems (GIS)-based models formulating, exchanging also multidisciplinary knowledge.EO-PERSIST innovative approach is anticipated to contribute to more informed decision-making and broader data accessibility for researchers, policymakers, and other stakeholders.

The present contribution aim is two-fold: at first, to provide an overview of EO-PERSIST Marie Curie Staff Exchanges EU-funded research project; second, to present some of the key project outputs delivered so far relevant to the selected Use Cases of the project and the geospatial database developed for assessing the socioeconomic impacts of climate change in the permafrost Arctic regions.

This study is supported by EO-PERSIST project which has received funding from the European Union's Horizon 2020 research and innovation program under grant agreement No. 101086386.

KEYWORDS:earth observation, cloud platform, Arctic, socioeconomic impact

How to cite: Tselos, G.-N., Detsikas, S. E., Kroszka, B., Grzybowski, P., and Petropoulos, G. P.: Enhancing the use of Geoinformation technologies to assess the socioeconomic impacts of climate change in the Arctic: Insights from the EO-PERSIST Project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6240, https://doi.org/10.5194/egusphere-egu25-6240, 2025.

EGU25-8063 | ECS | Posters virtual | VPS20

Deploying UAV technology to assess typhoon impacts in vulnerable communities in Guam  

Keanno Fausto and Jonelle Sayama

The U.S. territory of Guam is threatened annually by high-intensity storms and typhoons due to its location in the western Pacific Ocean. The island’s infrastructure – buildings, roads, and utilities – bear the brunt of typhoon damage, which in turn affects public health, the economy, and natural resources. Traditionally, these impacts have been observed via satellite, radar, and official weather stations.  Damages are assessed in the aftermath of the typhoon with a manual, on-the-ground approach led by the National Weather Service (NWS). This is often exhaustive and time-consuming for the assessment team. Observations from the ground can inadvertently create data gaps on damage assessments due to inaccessible areas caused by vegetative and construction debris, and flooded roads and pathways. This may not capture many impacts eligible for local or federal assistance. To address these data gaps and augment damage assessments, the University of Guam (UOG) Drone Corps program aims to assist local and federal government agencies (e.g., utility companies, public health, emergency services, and natural resource management) by collecting high-resolution aerial imagery to help prioritize and allocate limited resources. This presentation highlights the results of this novel collaboration of UOG, NWS, Guam Homeland Security (GHS), and the Office of the Governor of Guam in the creation of the damage assessment of Typhoon Mawar, which ravaged Guam on 24-25 May 2023. Following the typhoon, UOG worked with NWS to identify and capture imagery of vulnerable sites that were heavily impacted. This presentation will also share how UOG Drone Corps’ data was disseminated among other agencies as supplemental data for natural disaster recovery efforts. The presentation will conclude with a summary of the UOG Drone Corps program model as a resource for developing resiliency strategies for vulnerable island communities using advanced and emerging technologies. 

How to cite: Fausto, K. and Sayama, J.: Deploying UAV technology to assess typhoon impacts in vulnerable communities in Guam , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8063, https://doi.org/10.5194/egusphere-egu25-8063, 2025.

EGU25-11808 | ECS | Posters virtual | VPS20

The use of InSAR and DInSAR for detecting land subsidence in Albania 

Pietro Belba

INTRODUCTION. InSAR or Interferometric Synthetic Aperture Radar is a technique for mapping ground deformation using radar images of the Earth's surface collected from orbiting satellites. DInSAR or Differential SAR Interferometry is an active remote sensing technique based on the principle that, due to the very high stability of the satellite orbits, it is possible to exploit the informative contribution carried by the phase difference between two SAR images looking at the same scene from comparable geometries.

AIM. In this setting, the main objective of this study is to evaluate the region near the closed rock salt mine in the south of Albania. Our input for this exercise will be two images of the land near the former rock salt mine in Dhrovjan near the Blue Eye (Saranda, Albania).

RESULTS. By combining the phases of 2 images we produce an interferogram where the phase is correlated to the terrain topography and deformation so if the phase shifts related to the topography are removed from the interferogram, the difference between the resulting products will show surface deformation patterns or cure between the two acquisition dates and this methodology is called differential interferometry Processing, Phase Unwrapping, and at the end creating the displacement map. We use in our study the difference in time with the algorithm which consists of working step by step with these operators: Read the two split products, Applying Orbit files, Back-Geocoding, Enhanced Spectral Diversity, Interferogram, TOPSAR Deburst, and Write. The resulting difference of phases is called an interferogram containing all the information on relative geometry. Removing the topographic and orbital contributions may reveal ground movements along the line of sight between the radar and the target.

The next algorithm we worked with these operators: Read the debursted interferogram, TopoPhaseRemoval, Multilook, Goldstain Filtering, and Write. At the same time from Goldstain Filtering, we add the Snaphu Export operator.

Correct phase unwrapping procedures must be performed to retrieve the absolute phase value by adding multiples of 2π phase values to each pixel to extract accurate information from the signal. In this study, we will use SNAPHU, which is a two-dimensional phase unwrapping algorithm consists of working step by step with these operators: read (the wrapped image) and read (2) the unwrapped image, Snaphu Import, PhaseToDisplacement, and Write. We can display it in Google Earth after saving it as .kmz and also make a profile of the displacements.

DISCUSSION AND CONCLUSIONS

One of the SAR Interferometry applications is deformation mapping and change detection. This work demonstrates the capability of interferometric processing for the observation and analysis of instant relative surface deformations in the radar LOS direction. When two observations are made from the same location in space but at different times, the interferometric phase is proportional to any change in the range of a surface feature directly. All three stages of the work are important and require accurate interpretation knowledge, especially when working with the Snaphu program.

KEY-WORDS

InSAR, DInSAR, Interferogram

How to cite: Belba, P.: The use of InSAR and DInSAR for detecting land subsidence in Albania, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11808, https://doi.org/10.5194/egusphere-egu25-11808, 2025.

EGU25-12947 | ECS | Posters virtual | VPS20

Topological fingerprinting of dynamical systems 

Gisela Daniela Charó, Davide Faranda, Michael Ghil, and Denisse Sciamarella

Poincaré established a framework for understanding the dependence of a dynamical system's properties on its topology. Topological properties offer detailed insights into the fundamental mechanisms — stretching, squeezing, tearing, folding, and twisting — that govern the shaping of a dynamical system's flow in state space. These mechanisms serve as a conduit between the system's dynamics and its topology [Ghil & Sciamarella, NPG, 2023]. A topological analysis based on the templex approach [Charó et al., Chaos, 2022] involves finding a topological representation of the underlying structure of the flow by the construction of a cell complex that approximates its branched manifold and a directed graph on this complex. A pivotal feature of the cell complex that facilitates the characterization of the flow dynamics is the joining locus, upon which all the fundamental mechanisms that sculpt the flow leave a pronounced signature.

The local dimension d(x) and the inverse persistence θ(x) of the state x of a dynamical system [Lucarini et al., 2016; Faranda et al., Sci. Rep., 2017] provide information on the rarity and predictability of specific states, respectively. We demonstrate herein that these two measures, d and θ  also provide information about the localization of the joining locus.

The present work proposes a new topological method for fingerprinting a system’s nonlinear behavior using the concept of persistent generatexes. This novel approach integrates the strengths of two topological data analysis methods: the templex and persistent homologies. Rather than employing a single cell complex and a digraph to characterize the flow of the system, our approach emphasizes the localization of the joining locus through the calculation of local dimension and the inverse persistence, leading to the construction of a family of nested digraphs. The dynamical paths, namely the nonequivalent ways of travelling through the flow, are found to be the most persistent cycles; here the concept of persistence is used in the sense of the persistent homology approach [Edelsbrunner & Harer, Contemporary mathematics, 2008]. The dynamical paths give us the ‘topological fingerprinting’ of a system’s dynamics.

How to cite: Charó, G. D., Faranda, D., Ghil, M., and Sciamarella, D.: Topological fingerprinting of dynamical systems, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12947, https://doi.org/10.5194/egusphere-egu25-12947, 2025.

The Er-Rich region is a focal area for understanding the geological evolution of the central-eastern High Atlas, which it covers almost entirely along a north-south transverse line. It is a hinge region between the two major tectonic structures of the High Atlas (the North and South Atlas faults), which reveal a framework of Meso-Cenozoic carbonate, detrital and magmatic rocks.

Previous studies have highlighted the complexity of mapping in this area. To date, no detailed geological map has been produced for this study area, with the exception of the old provisional 1:200,000 map of the Midelt-Rich High Atlas. Remotely-sensed mapping initiatives have also been carried out in the region, except that they do not provide a final interpretation as a geological map, supported by geological maps covering neighboring regions. A detailed geological map of the Er-Rich region, based on the results of remote sensing and field data, is therefore needed in the area. For this purpose, remote sensing geological mapping techniques have been applied to two types of satellite data: 1) Landsat 8 OLI (Optical Land Imager) multispectral optical data, and the Spot 5 panchromatic band acquired by the HRG-2 (High Geometric Resolution) instrument; 2) Sentinel-1 SAR data with dual polarisation (HV-HH).

All the data underwent several pre-processing or correction stages using appropriate software, in particular radiometric and atmospheric correction for Landsat 8 OLI (Optical Land Imager) images using ENVI software. The corrected product of the three Landsat 8 OLI scenes covering the region were then spatially enhanced using the Spot 5 panchromatic band to produce a multispectral image with a high spatial resolution of 5 m using ENVI software. The Sentinel-1 radar data were pre-processed using SNAP toolbox software by applying a series of corrections.

The results obtained by applying the Optimum Index Factor (OIF) method and Principal Component Analysis (PCA), allowing us to select the most significant colored compositions. Moreover, this combination enabled us to delineate with great precision the large outcrops of carbonate rocks (limestones, marl), siliciclastic rocks (conglomerates, sandstones and silts) and magmatic rocks (igneous intrusions).

The lineaments were extracted manually by visual interpretation of Sentinel-1 radar images, after applying directional filtering folowing four general orientations (N0, N45, N90, N135), enabling us to generate a synthetic structural map of the region.

The results obtained were compared with data from geological maps of adjacent areas and approved by field observations, leading to the production of a high-precision geological map, compiled with pre-existing geological literature.

How to cite: Hdoufane, M., Zafaty, O., and Ettaki, M.: Integrated remote sensing data and field investigations for geological mapping and structural analysis in the Er-Rich area (High Atlas, Morocco), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13782, https://doi.org/10.5194/egusphere-egu25-13782, 2025.

EGU25-15569 | Posters virtual | VPS20

Uniform Data Access Layer: Advancing Data FAIRness in FAIR-EASE 

Jorge Mendes and Marc Portier

The Uniform Data Access Layer (UDAL), a central component within the FAIR-EASE project, is designed to revolutionize how researchers access, integrate, and utilize diverse scientific datasets. FAIR-EASE prioritizes FAIR (Findable, Accessible, Interoperable, Reusable) principles to ensure that data becomes a powerful enabler of scientific discovery and informed decision-making. 

The UDAL concept brings a modular and re-usable approach to choosing and using data in data processing workflows. It materializes as a software package that users can use in their pipelines. UDAL serves as a middleware layer, offering a standardized, user-centric framework for data access. By bridging the gap between complex infrastructures and researchers, UDAL simplifies data retrieval, integration, and usage. This solution decouples data usage from technical complexities, ensuring that researchers can focus on analysis without needing detailed knowledge of access protocols or data formats. Its adaptability to a wide range of technologies and protocols enables interoperability across disciplines and geographic regions. UDAL's innovative approach has been validated with data providers such as Argo and Blue-Cloud and various technology stacks and formats like NetCDF, Beacon, SPARQL endpoint, HTTP REST API, demonstrating its capacity to unify diverse datasets into a single, intuitive system. 

A key feature of UDAL is its "named query" mechanism, which standardizes and reuses specific data requests. This enhances reproducibility, shields users from the intricacies of data filtering and retrieval, and promotes efficiency. Additionally, UDAL’s technology-agnostic approach accommodates centralized and distributed data architectures, supporting innovation in data management and usage strategies. 

By addressing critical challenges in data management—such as technical barriers and the diversity of data sources—UDAL aligns with the broader goals of FAIR-EASE. It empowers both researchers and data providers, fostering cross-domain collaboration and innovation. Beyond its technical contributions, UDAL embodies a vision of “data as a commodity,” promoting the sustainability and accessibility necessary for open science. While it does not directly address equitable benefit distribution, its transparent usage measurement capabilities lay a foundation for future policy and governance frameworks. 

In conclusion, UDAL represents a transformative advance in data-driven research, harmonizing access across disciplines and platforms while accelerating discovery and fostering innovation. As a cornerstone of FAIR-EASE, UDAL is set to establish new standards for simplicity, usability, and sustainability in scientific data management. 

How to cite: Mendes, J. and Portier, M.: Uniform Data Access Layer: Advancing Data FAIRness in FAIR-EASE, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15569, https://doi.org/10.5194/egusphere-egu25-15569, 2025.

EGU25-16203 | ECS | Posters virtual | VPS20

Temporal and Spatial Dynamics of Urban Evapotranspiration in Paris: A Multiscale Perspective 

Sitian Zhu, Auguste Gires, Daniel Schertzer, Ioulia Tchiguirinskaia, and Cedo Maksimovic

The impacts of global change, such as extreme heat and water scarcity, are increasingly threatening urban populations. Evapotranspiration (ET) plays a vital role in mitigating urban heat islands and reducing the effects of heat waves. It also serves as a proxy for vegetation water use, making it a critical tool for designing resilient green cities. Despite its importance, high-resolution mapping of urban ET that captures both spatial and temporal dynamics remains limited. This study focuses on the Paris metropolitan area, analyzing ET variability across multiple spatial scales (from 10 m to 10 km) using Sentinel-2 data from the Copernicus system. The Normalized Difference Vegetation Index (NDVI) is calculated with observation scale of 10 m, and then used as a proxy for ET. Universal Multifractal analysis, which have been widely used to characterize and model geophysical fields extremely variable across wide range of space-time scales, are implemented on this new data set. This framework is parsimonious since it basically relies on three parameters only: the mean intermittency codimension C1, the multifractality index a and the non-conservation parameter H.  Specifically, the multifractality index α (1.3–1.5) and the mean intermittency codimension C1 (~0.02) were derived to quantify the spatial and temporal heterogeneity of ET. The analysis, spanning 2019–2023, revealed noticeable temporal and spatial variability in ET. The study focuses on a square region of approximately 60 km × 60 km within the area around Paris. This region was further divided into multiple portions of size ranging from 2 to 10 km to assess potential variability over the studied areas. By incorporating both yearly and monthly data, the analysis captured seasonal trends as well as interannual variability, with higher variability observed during the summer months, driven by increased vegetation activity and water demand. Spatially, yearly data was analyzed and ET variability was most pronounced in densely populated areas, such as central Paris, where anthropogenic influences dominate. In contrast, forested areas and urban parks demonstrated significantly more stable ET patterns, underscoring the moderating effect of vegetation cover. These findings highlight the critical role of urban greening in mitigating extreme variability and stress on urban ecosystems.

How to cite: Zhu, S., Gires, A., Schertzer, D., Tchiguirinskaia, I., and Maksimovic, C.: Temporal and Spatial Dynamics of Urban Evapotranspiration in Paris: A Multiscale Perspective, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16203, https://doi.org/10.5194/egusphere-egu25-16203, 2025.

EGU25-17156 | Posters virtual | VPS20

Leveraging EO for Security and Resilience 

Michela Corvino and the Michela Corvino

The ESA Directorate of Earth Observation Programmes has been actively leveraging satellite-based environmental information to address fragility contexts, focusing on areas such as environmental crimes, crimes against humanity, cross-border crimes, and onset of crises. Over the past decade, ESA has explored digital intelligence crime analysis by employing advanced data mining and machine learning tools to uncover hidden patterns and relationships in historical crime datasets, enabling better detection, prediction, and prevention of criminal activities.

Despite these advancements, the integration of Earth Observation (EO) capabilities into investigative practices remains limited. This is due to several challenges, including low awareness of EO's potential, a lack of illustrative use cases showcasing its benefits, inconsistencies in satellite data collection compared to investigative needs, high costs of very high-resolution imagery, and restricted access to national intelligence sources. To overcome these barriers, ESA has been investigating strategies to systematically incorporate EO-derived information into investigative frameworks also as legal evidence, aiming to enhance situational awareness and support stakeholders in developing procedures to exploit EO and OSINT for addressing international crimes and assessing fragility contexts, in cooperation with international organizations including Interpol, UNODC and ICC.

Recent developments in EO technology and methodologies have created significant opportunities for more impactful applications. ESA has focused on tailoring EO-based services and OSINT to meet the case-sensitive requirements of security and development end-users, enabling better integration of EO-derived insights into intelligence models. These efforts include developing advanced EO information products that go beyond routine offerings, testing and evaluating these products in collaboration with end-users, and demonstrating their value in operational settings.

The GDA Fragility, Conflict, and Security initiative has been a cornerstone of ESA’s work, involving partnerships with International Financial Institutions (IFIs) to co-design tools that provide precise and timely information. These tools have supported initiatives aimed at reducing inequalities, promoting economic development, and enhancing environmental safety in fragile and conflict or post conflict-affected areas. By combining geospatial data with diverse data sources, ESA has delivered customized analyses and reports to improve emerging threats analysis and decision-making processes.

Several ESA initiatives have demonstrated the benefits of EO services for assessing fragility risk exposure, characterizing dynamic needs in fragile contexts, planning post-conflict reconstruction, and managing natural resources. ESA constantly engages with stakeholders, including the OECD, security organizations, and humanitarian actors, and its community of industries and research centres to promote the adoption of EO in international development, humanitarian aid, and peacebuilding. Through these efforts, ESA continues to advance the role of EO in supporting justice, accountability, and sustainable recovery in fragile settings.

How to cite: Corvino, M. and the Michela Corvino: Leveraging EO for Security and Resilience, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17156, https://doi.org/10.5194/egusphere-egu25-17156, 2025.

EGU25-17965 | ECS | Posters virtual | VPS20

Lagrangian Evolution of the Trapping Capacity of Mesoscale Eddies in the Canary Eddy Corridor: A Numerical Modeling Approach 

Daniel Vacca, Borja Aguiar-González, and Tammy Morris

The Canary Eddy Corridor is a dynamic region of mesoscale eddy activity, playing a critical role in the transport of physical properties (heat and salt) and biogeochemical properties (nutrients, larvae, plankton) in the eastern North Atlantic. This study investigates the Lagrangian evolution of the trapping capacity of mesoscale eddies according to their lifecycle phases and vertical structure (surface vs. subsurface eddies).


We combine OceanParcels (an open-source Python toolbox) and an eddy identification and tracking algorithm with the GLORYS12V1 reanalysis product and altimetry data from AVISO to simulate particle release and track trajectories within eddies. Applying the eddy tracking algorithm at surface and subsurface levels in GLORYS12V1 reveals that subsurface eddies with a surface signal exhibit subsurface rotational velocities at the eddy core that occasionally exceed those of surface eddy cores. This highlights the potential misrepresentation of eddy transport capacity when relying solely on altimetry data, without accounting for the vertical structure, which can be better resolved through a combination of model outputs and observational data, such as non-standard Argo float configurations. Furthermore, a detailed analysis of the eddy lifecycle phases shows that mature eddies exhibit substantially greater trapping depths compared to their growth and decay stages. These findings align with earlier modeling analyses of dipoles originating south of Madagascar, which also highlight enhanced trapping depths in mature eddies.


The results provide a comprehensive view of the trapping capacity of mesoscale eddies throughout their lifecycle and vertical structure, emphasizing their critical role in biophysical coupling, ecological connectivity, and the transport of biogeochemical properties, as well as microplastics and other pollutants.

 

Acknowledgments: The first author is grateful for the internship grants ERASMUS +, AMI-MESRI, and TIGER. 

How to cite: Vacca, D., Aguiar-González, B., and Morris, T.: Lagrangian Evolution of the Trapping Capacity of Mesoscale Eddies in the Canary Eddy Corridor: A Numerical Modeling Approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17965, https://doi.org/10.5194/egusphere-egu25-17965, 2025.

EGU25-18606 | ECS | Posters virtual | VPS20

Deep Learning based Paddy Land Abandonment Detection Using Multitemporal Polarimetric SAR Patterns 

Shivam Kasture, Aishwarya Hegde A, and Pruthviraj Umesh

The abandonment of agricultural land in India, especially paddy fields, has emerged as a significant challenge for food security and ecosystem sustainability in the country. Although rice production is vital for national food security, research on paddy land abandonment in India remains limited. Some Indian states have reported an alarming decline in paddy cultivation area over the past two decades. The study employs the Udupi district of Karnataka, India, a high-rainfall coastal region where paddy has traditionally been the dominant crop and where paddy land abandonment has been observed, as the study area. This study addresses crucial research gaps by framing these objectives for the study: (1) developing a deep learning framework that utilizes both intensity and phase information from polarimetric Synthetic Aperture Radar (SAR) data for abandoned paddy land detection, (2) leveraging recurrent neural networks (RNNs) to capture temporal patterns in abandonment, and (3) demonstrating an automated, all-weather monitoring approach that overcomes the limitations of traditional optical remote sensing in tropical regions.

Conventional monitoring approaches struggle with persistent cloud cover in tropical regions which limits effective assessment of abandonment patterns. SAR data provides unique capabilities for continuous monitoring under all weather conditions, making it particularly well-suited for tropical regions. However, previous studies have primarily underutilized SAR's potential by concentrating solely on backscattering intensity from ground range detected (GRD) products, overlooking the valuable phase information that could offer deeper insights into land use changes.  In this study, we employ Sentinel-1 Single Look Complex (SLC) data, which offers both intensity and phase information. Considering the temporal nature of paddy land abandonment, we developed a deep learning framework utilizing RNNs viz. LSTM, BiLSTM and BiGRU to effectively capture time-series patterns in the data. This framework analyzes backscattering coefficients (VV and VH polarizations) and polarimetric parameters (entropy, anisotropy and alpha angle) derived from SLC data collected during the Kharif seasons from 2017 to 2024. We carried out extensive ground truth data collection of active and abandoned paddy lands to train and validate our models. The backscattering coefficients were processed through orbit correction, radiometric calibration, TOPSAR deburst, multi-looking, speckle filtering and terrain correction. For deriving the polarimetric parameters, after basic preprocessing steps, the covariance matrix was generated followed by the polarimetric decomposition of the phase-preserved data. Results indicate that our RNN models show promising performance in detecting temporal patterns of paddy land abandonment. The method exhibits a robust ability to produce reliable abandoned land maps in regions prone to cloudy and rainy conditions. Future research should explore polarimetric features across various vegetation types in abandoned lands, expand the methodology to other agricultural systems, and examine the impact of socio-economic and topographical factors on abandonment patterns to support evidence-based land management policies.

How to cite: Kasture, S., Hegde A, A., and Umesh, P.: Deep Learning based Paddy Land Abandonment Detection Using Multitemporal Polarimetric SAR Patterns, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18606, https://doi.org/10.5194/egusphere-egu25-18606, 2025.

EGU25-19070 | Posters virtual | VPS20

Rainfall Dynamics in Wind Energy Scenarios 

Martin Obligado and Auguste Gires

The presence of rain in wind farms involves several modeling challenges, as the momentum exchanges between turbulent wakes and the particle phase present subtle phenomena. For instance, rain droplets are typically large enough to exhibit inertia relative to the air carrier phase. Under these conditions, it has been found that the gravitational settling of particles in turbulent flows may be either enhanced or hindered compared to stagnant conditions. While this has significant implications for rainfall transport, ash pollutants, and pollen dispersion, very few studies have been conducted in field conditions. Moreover, the scaling laws and non-dimensional parameters governing this phenomenon have not yet been properly identified, and determining which configurations result in the enhancement or hindrance of settling velocity remains an open question.

We propose a hybrid experimental/numerical approach. Field data from a meteorological mast located at a wind farm in Pays d’Othe, 110 km South-East of Paris, France, were used to characterize the background turbulent flow through a set of sonic anemometers. Additionally, disdrometers were employed to characterize the settling velocity of raindrops, discriminating by particle size. Numerical simulations complement this data analysis. Specifically, 3D space and time vector fields that realistically reproduce the observed spatial and temporal variability of wind fields are generated using multifractal tools. Then, 3D trajectories of non-spherical particles are simulated and their settling velocity derived.

Our findings indicate that the presence of turbulence significantly hinders the settling velocity of raindrops in turbulent environments. Our study covers several distinct rainfall events, allowing us to analyze the influence of turbulent flow properties on this phenomenon.

How to cite: Obligado, M. and Gires, A.: Rainfall Dynamics in Wind Energy Scenarios, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19070, https://doi.org/10.5194/egusphere-egu25-19070, 2025.

EGU25-19225 | Posters virtual | VPS20

Customizing Trends.Earth for land degradation assessment in the earth critical zone: a FAIR-EASE approach 

Italia Elisa Mauriello, Giuliano Langella, Fabio Terribile, and Marco Miralto

Land degradation is a critical challenge to sustainable development, impacting ecosystems, economies, and communities globally. As part of the FAIR-EASE Earth Critical Zone (ECZ) pilot, this study develops a tailored Land Degradation Assessment tool based on the Trends.Earth approach. The tool aims to enhance data accessibility, integration, and usability across environmental domains, supporting decision-making and policy frameworks aligned with the United Nations Sustainable Development Goals (SDGs).
Building upon the robust Trends.Earth implementation, we can integrate customized workflows and datasets to reflect regional variability in degradation indicators, including vegetation productivity, soil health, and land cover changes. Our approach prioritizes FAIR (Findable, Accessible, Interoperable, and Reusable) principles to ensure broad usability and collaboration across scientific and policy communities.
Preliminary results demonstrate the tool's capacity to enhace the detail of the analysis and to identify degradation hotspots. Furthermore, the integration of open-source geospatial tools and standards supports a scalable framework applicable to diverse environmental contexts.
The tool is designed to be embedded within the LandSupport platform, a geospatial decision support system, further enhancing its accessibility and integration into decision-making processes for land management.
This work contributes to advancing interdomain digital services and illustrates the potential of FAIR principles in addressing complex environmental challenges. We invite feedback from the community to refine, expand and customise the tool's application, fostering collaboration for sustainable land management.

How to cite: Mauriello, I. E., Langella, G., Terribile, F., and Miralto, M.: Customizing Trends.Earth for land degradation assessment in the earth critical zone: a FAIR-EASE approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19225, https://doi.org/10.5194/egusphere-egu25-19225, 2025.

EGU25-20653 * | Posters virtual | VPS20 | Highlight

The turbulence of solids: a multifractal plate tectonic model with Guttenberg-Richter plate “quakes”  

Shaun Lovejoy, Andrej Spiridonov, and Lauras Balakauskas

Over thirty years ago, Y. Kagan speculated that seismicity could fruitfully be considered as “the turbulence of solids”.  Indeed, fluid turbulence and seismicity have many common features: they are both highly nonlinear with huge numbers of degrees of freedom.  Beyond that, Kagan recognized that they are both riddled with scaling laws in space and in time as well as displaying power law extreme variability and – we could add – multifractal statistics.

Kagan was referring to seismicity as usually conceived, as a sudden rupture process  occurring over very short time periods.  We argue that even at one million year time scales, that the movement of tectonic plates is “quake-like” and is quantitatively close to seismicity, in spite of being caused by relatively smooth mantle convection. 

To demonstrate this, we develop a multifractal model grounded in convection theory and the analysis of the GPlates data-base of 1000 point trajectories over the last 200 Myrs.  We analyzed the statistics of the dynamically important vector velocity differences where Dr is the great circle distance between two points and Dt is the corresponding time lag.  The longitudinal and transverse velocity components were analysed separately.  The longitudinal scaling of the mean longitudinal difference follows the scaling law <Δv(Δr)> ≈ ΔrH with empirical H close to the mantle convection theory value  H = 1.  This high value implies that  mean fluctuations vary smoothly with distance.  Yet at the same time,  the intermittency exponent C1 is extremely high (C1 ≈ 0.55) implying that from time to time there are enormous “jumps” in velocity: “Plate quakes”.  For comparison, laminar (nonturbulent) flow has H = 1 but is not intermittent (C1 = 0), whereas fully developed isotropic fluid turbulence has the (less smooth) value H = 1/3 (Kolmolgorov) but with non-negligible intermittency C1 ≈ 0.07 and seismicity has very large C1 ≈ 1.3.  Our study thus quantitatively shows how smooth fluid-like behaviour for the longitudinal velocity component can co-exist with highly intermittent quake-like behaviour.

Whereas the longitudinal component is well modelled by (highly intermittent) convection, the transverse velocity is well modelled by Brownian motion.  In the temporal domain both components (including their strong correlations) display such diffusion behaviour (i.e. with classical exponent H = ½), but are highly intermittent (C1time = C1space/2 ≈ 0.27).  Finally, the extreme velocity differences (that appear as occasional spikes in the velocities) have power law probability tails; the “Guttenberg-Richter” exponents in the seismology literature.

The advection - diffusion model is based on an underlying multifractal space-time cascade process.  Using mantle convection theory, we show how the driving multifractal flux (ψ) is related to vertical heat fluxes, expansion coefficients, densities, viscosities and specific heats. Taking typical values predict driving fluxes very close to the observed mean <ψ> ≈ 1/(400 Myrs).  Trace moment analysis shows that the outer space-time scales of the cascade process are ≈17000 km in space and ≈ 50Myrs in time.   Whereas the former corresponds to half the Earth’s circumference, the latter is the typical time required for a plate to randomly “walk” the same distance.

How to cite: Lovejoy, S., Spiridonov, A., and Balakauskas, L.: The turbulence of solids: a multifractal plate tectonic model with Guttenberg-Richter plate “quakes” , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20653, https://doi.org/10.5194/egusphere-egu25-20653, 2025.

ESSI1 – Next-Generation Analytics for Scientific Discovery: Data Science, Machine Learning, AI

Gold deposits in New Brunswick, part of the Canadian Appalachians, formed during various stages of the Appalachian orogeny. Significant regional-scale transcurrent faults that are locally controlling cogenetic magmatic, include the Restigouche, Rocky Brook-Millstream, McCormack-Ramsay Brook, McKenzie Gulch, and Moose Lake faults, played a crucial role in shaping the geological framework and enabling the focusing of mineralizing fluifds in northern New Brunswick. The mineral systems approach is applied here to link conceptual models of mineralization processes with available exploration data, aiming to achieve effective mineral prospectivity mapping (MPM). This method is designed to streamline exploration efforts, minimizing both time and cost, which are key priorities in the mineral exploration industry. A machine learning-based data-driven approach was utilized to evaluate 18 predictor maps with a pixel size of 200 meters. These MPM maps integrated diverse features, including geochemical indicators for Au, As, Sb, Zn, Pb, Cu, and Mo in till to define geochemical anomalies, airborne radiometric data for K, eU, and eTh, as well as aeromagnetic and LiDAR datasets, to interpret geological characteristics, structural features, faults, intrusive and extrusive units, and lithological contacts. A series of edge enhancement filters, including Reduced to Pole (RTP), first vertical derivative (FVD), tilt derivative (TDR), and analytic signal (AS), were applied to the dataset, followed by a 3D inversion. Our results show that bimodal felsic to mafic intrusive and extrusive igneous systems exhibit a strong magnetic response, a conclusion validated through correlation with drill core assay data. Moreover, this study utilized principal component analysis (PCA) of till data to determine pathfinder and indicator elements associated with gold mineralization. A MPM model was created for epithermal gold mineralization using a Support Vector Machine (SVM), incorporating the known gold occurrences and deposits of the area. The performance of the resulting MPM maps was evaluated using the area under the receiver operating characteristic curves (AUC-ROC). The study concludes that SVM is a robust tool for mineral exploration, providing a data-driven approach to identifying new mineral deposits with greater accuracy and efficiency.

How to cite: Mami khalifani, F., Lentz, D., and Walker, J.: A Machine Learning-Based Framework for Mineral Prospectivity Modeling: Predicting Epithermal Gold Mineralization in Northern New Brunswick, Canada, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1121, https://doi.org/10.5194/egusphere-egu25-1121, 2025.

Mongolian society and food production depends heavily on livestock farming, which is usually practiced with nomadic systems. Consequently, movement patterns of herders are crucial in respect of finding sufficient forage and sustainable use of pastures. In this study, a combination of InSAR, optical and weather time series data has been explored as a tool for spatio-temporal grazing monitoring. To detect movement patterns, a machine learning (ML) based method to detect breakpoints in vegetation condition has been developed and compared to the widely-used Breaks For Additive Season and Trend (BFAST) algorithm. The results have been validated using test sites spread across the entire eastern Mongolian steppe ecosystem, covering different grassland use intensities. The results indicate that (1) ML method performed superior compared to BFAST, detecting 41.5% of breakpoints. (2) Breakpoints in summer pastures mainly occurred from April to June, while on winter pastures, they emerged in October, November, and the following February and March. (3) Regarding spatial prediction, the model developed in this study predicts breakpoints in areas distinguish between summer and winter camps, However, there is insufficient data to conclusively attribute the occurrence of pasture breakpoints to herder movements.

How to cite: ji, S.: Can Vegetation Breakpoints in Eastern Mongolia grassland be detected using Sentinel-1 coherence time series data?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1321, https://doi.org/10.5194/egusphere-egu25-1321, 2025.

EGU25-4051 | ECS | PICO | ESSI1.2

Score-based Diffusion Models for the Space-Time Interpolation of Sea Surface Turbidity 

Thi-Thuy-Nga Nguyen, Mahima Lakra, Frédéric Jourdin, and Ronan Fablet

This study explores the application of score-based generative diffusion models for mapping sea surface Suspended Particulate Matter (SPM) of the Dutch Wadden Sea using satellite-derived images, focusing on their comparative efficacy against state of the art deterministic methods such as 4DVarNet, UNet, and DInEOF. Although deterministic deep learning approaches provide robust reconstructions, they often struggle with probabilistic uncertainty and extreme values of overly complex real-world scenarios. Our findings indicate that diffusion models, when conditioned with 4DVarNet and DInEOF, offer improved performance over DInEOF and UNet. Although slightly less accurate than 4DVarNet, this discrepancy is not a significant concern, as the primary goal extends beyond merely maintaining accurate reconstructions. Instead, our approach aims to provide a comprehensive view of the distribution through the samples. Our results show that diffusion models are able to generate the tail of the distribution, thereby capturing extreme values more effectively. And they assist in identifying areas of high uncertainty, particularly when the samples show inconsistencies. Furthermore, unlike typical 2D diffusion models, this study employs a 3D approach, incorporating 2D spatial and 1D temporal dimensions, allowing the model to capture dynamic physical changes over time and enhance the accuracy of probabilistic predictions of the image time series.

How to cite: Nguyen, T.-T.-N., Lakra, M., Jourdin, F., and Fablet, R.: Score-based Diffusion Models for the Space-Time Interpolation of Sea Surface Turbidity, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4051, https://doi.org/10.5194/egusphere-egu25-4051, 2025.

Geospatial monitoring of water quality is essential for managing and protecting groundwater resources, particularly in agricultural regions where nitrate contamination poses significant environmental and public health risks. This study presents a novel methodology for generating absence points in geospatial binary classifications applied to nitrate levels in groundwater across Odense, Denmark. We developed machine learning designed to generate absence points using multiple approaches for binary classification: random, buffer-based, similarity-based, and Maxent-based methods. The integration of maximum entropy into the absence generation workflow allowed us to identify low-susceptibility zones, improving the accuracy of binary classification. The dataset comprised geospatial nitrate concentration levels derived from environmental, hydrological, and anthropogenic variables. Spatial data included high-resolution land-use maps and hydrological parameters. Model evaluation was conducted using Random Forest, with results indicating that the Maxent-based approach consistently outperformed other methods across all metrics, including precision (0.96), AUC (0.96), and TSS (0.91). This method proved particularly effective in handling the challenges associated with presence-only data and produced the most reliable predictions for nitrate contamination in groundwater. The findings underscore the importance of leveraging advanced absence generation techniques to enhance model performance in geospatial classification modeling.

How to cite: Naghibi, S. A., Ahmadi, K., and Berndtsson, R.: Bridging Data Gaps in Water Quality Modelling: A Machine Learning Framework for Absence Point Generation in Geospatial Binary Classifications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6262, https://doi.org/10.5194/egusphere-egu25-6262, 2025.

EGU25-6404 | ECS | PICO | ESSI1.2

A new data-efficient, deep learning-based methodology for geological subsurface reconstructions 

Rodrigo Uribe-Ventura, Yoan Barriga-Berrios, Jorge Barriga-Gamarra, Patrice Baby, and Willem Viveen

Deep learning approaches for geological subsurface reconstruction typically require extensive training datasets, limiting their practical application in geosciences where data acquisition is costly and sparse. We present a methodology using sparse convolutional autoencoders that effectively learns from synthetically generated training data while maintaining strong generalization to real-world scenarios. Our model is trained exclusively on synthetic basin boundary configurations and corresponding forward-modeled Vertical Electrical Sounding (VES) responses, thereby eliminating reliance on extensive real-world training datasets. Through transfer learning, the model achieves high reconstruction accuracy with as few as 1000 synthetic training examples. Systematic tests reveal the model preserves strong performance beyond its training distribution, suggesting it learns robust heuristic approximations and remains effective beyond the training range of 3–50 input points.

The trained model was applied to the Huancayo tectonic basin in the Peruvian Andes. There, the 300 to 350-m deep subsurface geometry of the tectonic basin was sucessfully modeled on basis of data input from 41, newly acquired VES logs along two cross sections of 12- and 14-km long.  Surprisingly, the reconstruction also revealed previously unidentified fold and thrust systems, for which the model was not explicitely trained, while also maintaining physical consistency with field measurements.

Our results demonstrate that sparse convolutional autoencoders, when trained on synthetic datasets, can effectively bridge the gap between data-hungry deep learning methods and data-sparse geological applications.

How to cite: Uribe-Ventura, R., Barriga-Berrios, Y., Barriga-Gamarra, J., Baby, P., and Viveen, W.: A new data-efficient, deep learning-based methodology for geological subsurface reconstructions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6404, https://doi.org/10.5194/egusphere-egu25-6404, 2025.

Extreme Learning Machine (ELM) is a fast and efficient learning algorithm designed for single-layer feedforward neural networks. It stands out by randomly initializing input weights and biases, which remain fixed during training, and learning the output weights using a closed-form solution. This approach eliminates the need for iterative optimization, significantly accelerating the training process. ELM is known for its generalization performance and versatility. Kernelized ELM  enhance its capability to model complex nonlinear systems. However, achieving optimal performance requires careful tuning of hyperparameters, such as the number of hidden neurons and the regularization parameter.
 
ELM has been widely applied in environmental risk and natural hazard assessments, climate and meteorological modelling, hydrology, renewable energy analysis  and time series forecasting. Recent advancements have extended the standard ELM model to include multilayer architectures, deep learning methodologies, unsupervised learning, and multiple kernel ELMs, broadening its applicability to more challenging  and diverse problems.
 
This research investigates the application of ELM for intelligent environmental data exploration and modelling. The study focuses on addressing problems in spatial and spatio-temporal data exploration, analysis, and modelling, including feature engineering and selection, multi-scale analysis, data normalization and anisotropy, nonlinearity, multivariate analysis and uncertainty quantification. 
 
The quality of ELM-based modelling is assessed through the examination of unexplained variability in  data and a comprehensive analysis of residuals. Various ELM configurations are applied throughout all phases of the research, enabling a flexible approach. Due to its computational efficiency, ELM facilitates numerous simulations and experiments, providing deeper insights into the data and the resulting models. Both simulated and real-world environmental datasets, including pollution, precipitation, and permafrost data, are utilized. Finally, the performance of ELM is compared with other machine learning algorithms in order to evaluate its effectiveness and reliability.

How to cite: Kanevski, M.: Environmental Data Exploration and Modelling Using Extreme Learning Machine, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7462, https://doi.org/10.5194/egusphere-egu25-7462, 2025.

EGU25-11150 | ECS | PICO | ESSI1.2

Tackling Spatial Multiple Features AI/ML Problems in Geology with Hexagons 

Marie Katrine Traun, Finn Sandø, and Søren Lund Jensen

As Artificial Intelligence (AI) and Machine Learning (ML) methods evolve at an explosive pace, there is an increased need to handle geological data challenges if we wish to (continue to) ride the AI/ML wave. Most geological data is at its core geospatial data in different shapes and formats. A few examples are polygon-based geological maps, geophysical and remote sensing raster grids and a plethora of sample analyses with coordinate data. Applications of geological data are as varied as the Earth is vast. However, these differing geospatial data formats in geology significantly limit the interoperability of datasets in an analytical ML context. Multivariable analyses of geological data often involve extensive spatial interpolation and projection headaches. Consequently, we must first solve geospatial data challenges to fully tackle inter- and intradisciplinary geoscience problems with ML and AI predictions on multivariable cross-disciplinary geological data. At our company, Scandinavian Highlands, we are building a platform and database structure to break down these geospatial format barriers using a hexagonal discrete global grid system called H3. The H3 grid represents all positions on Earth’s surface by hexagon (and 12 pentagon) cells at different levels of coarseness, ie. resolutions, down to 1 m2 cell area. The resolutions are bound together by a systematic parent cell to children cells hierarchy. We process different types of geospatial geological data (raster and vector) to an H3 grid representation at the appropriate resolution for the given dataset. In doing this, we create a database structure where different geological data layers can be seamlessly merged into a single feature “stack” table for AI/ML purposes at either local, regional or global scales and across individual dataset resolutions. In this presentation, we demonstrate the hexagonal multiple feature stack concept in action, from simple grouped/filtered visualisation, regression and descriptive statistics to dimension reduction techniques (e.g. PCA and t-SNE), clustering and other supervised and unsupervised methods. Furthermore, all analysis results can be assessed spatially on the map, grounding them on the Earth’s surface and in real-life decision-making use cases.

How to cite: Traun, M. K., Sandø, F., and Jensen, S. L.: Tackling Spatial Multiple Features AI/ML Problems in Geology with Hexagons, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11150, https://doi.org/10.5194/egusphere-egu25-11150, 2025.

EGU25-12795 | ECS | PICO | ESSI1.2

Optimal Use of Multi-Sensor Data for Precision Agriculture: Sentinel-1 and Sentinel-2 Fusion in Crop Classification 

maryam choukri, ahmed laamrani, and abdelghani chehbouni

Effective land monitoring and land use classification are critical for proper management of resources especially in heterogeneous and climate diverse areas. Consequently, this study seeks to test the hypothesis that the integration of Sentinel-1 radar and Sentinel-2 optical data enhances the degree of discrimination of crops in major farming areas of Morocco from the years 2020 to 2022. A three-dimensional coordinate system was established which included a series of processing stages that started with cloud masking, scaling of reflectance, and radar optical integration. At each year’s end, temporal averages and composites were created using selected Sentinel-2 spectral bands B2, B3, B4, B8, B11, B12 and Sentinel-1 VV & VH dual polarization channels. Ground truth samples from four major crops; Baley, Crop, D. Wheat and S. Wheat were used as the training set in a Random Forest classifier. The results for the three agricultural zones indicated high overall accuracies greater than 80% for each year, with the application of a combination of radar and optical data sets contributing greatly towards the ability to differentiate the crops located in cloud folded and spectral overlapping areas. Many classes had high consumer accuracy (≥70%) levels, yet several crops, like D. Wheat, had poor producer accuracy, possibly due to the uneven distribution of ground truth data sets. The small amount of Kappa coefficients between 0.50 and 0.60 also indicate moderate agreement similar to the validation data and thus more accurate ground truth and class targeted feature detection is needed. This study emphasizes the relevance notes of the multi-sensor data fusion technology for crop monitoring and also landcover classification which contributes to precision farming and resources management. Future work will focus on including temporal characteristics as well as state-of-the-art machine learning techniques to solve class balance issues and improve classification performance.

How to cite: choukri, M., laamrani, A., and chehbouni, A.: Optimal Use of Multi-Sensor Data for Precision Agriculture: Sentinel-1 and Sentinel-2 Fusion in Crop Classification, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12795, https://doi.org/10.5194/egusphere-egu25-12795, 2025.

EGU25-16038 | PICO | ESSI1.2

Spatial autocorrelation in machine learning for modelling soil organic carbon 

Alexander Kmoch, Jeonghwan Choi, Clay Taylor Harrison, and Evelyn Uuemaa

Spatial autocorrelation, the relationship between nearby samples of a spatial random variable, is often overlooked in machine learning models, leading to biased results. We investigated various methods to account for, address, and integrate spatial autocorrelation for modelling and prediction of soil organic carbon (SOC) using random forest models. We created and evaluated five different RF models to incorporate spatial structure through methods like buffer distances, KNN/RFSI coordinates, GWRFR, and kriging/RFRK. These were compared against a baseline models that did not have any added spatial components. Cross-validation showed slight improvements in accuracy for models considering spatial autocorrelation, while Shapley Additive Explanations confirmed the importance of spatial variables. However, no decrease in spatial autocorrelation of residuals was observed. The raster-based models exhibited enhanced prediction detail, but high-resolution validation data availability limited thorough validation. The findings emphasize the value of incorporating spatial autocorrelation for improved SOC prediction in machine learning models. We applied the models to predict SOC for the whole of Estonia in 10m raster resolution. Computational differences provided additional insights into pragmatic choices of models.

How to cite: Kmoch, A., Choi, J., Harrison, C. T., and Uuemaa, E.: Spatial autocorrelation in machine learning for modelling soil organic carbon, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16038, https://doi.org/10.5194/egusphere-egu25-16038, 2025.

EGU25-16905 | ECS | PICO | ESSI1.2

An AI-Copilot for JupyterLab for climate data analyses using FrevaGPT 

Felix Oertel, Etor Lucio Eceiza, Sebastian Willmann, Bianca Wentzel, Martin Bergemann, and Christopher Kadow

JupyterLab is a web-based interactive development platform that is widely used in the Earth science community. Using Jupyter Notebooks, it is possible to perform data analysis tasks, annotate and visualize results in a way that is easy to reproduce, present and share with others. JupyterLab allows the use of “extensions”, which add functionality to the platform. One of these is Jupyter-AI [1], which allows the use of Large Language Models (LLMs), such as ChatGPT, Claude Sonnet and Ollama, within the JupyterLab environment, through  a chat interface or directly within notebooks. By integrating LLMs into JupyterLab, it is possible to leverage their code generation capabilities to assist a user to translate their analysis tasks from an idea to actual executable code in an efficient manner. One drawback of using these LLMs in tasks involving spatio–temporal data is that the models typically do not have access to the data necessary for the analysis task and will often resort to generating fictional data or using placeholders in the code that they create. This requires the user to adapt the provided code to their data, which removes some of the utility provided by the LLM.

In this context we make use of FrevaGPT, an approach for using LLMs in climate data analysis that allows for quick, complex and reproducible analyses of data sets, such as decadal climate model forecasts. Leveraging LLM’s capability to write code and using few-shot prompting (in-context learning) allows the LLM to utilize Freva [2,3] (Free Evaluation Framework), a data search and analysis platform, which provides a standardised interface to spatio-temporal datasets hosted on an HPC cluster [4].  

FrevaGPT integrates seamlessly into Jupyter-AI and, by making use of the Freva library, combines the code-generating capabilities of LLMs with contextual understanding of how to access relevant datasets on the HPC cluster. This in addition with FrevaGPT’s ability to execute generated code in an isolated environment on an HPC node, annotating and explaining any intermediate results, as well as automatically correcting errors encountered along the way, could serve as a starting ramp for researchers to efficiently produce new analysis products based on spatio-temporal climate data. 

This PICO will include examples of using FrevaGPT within JupyterLab to analyse spatio-temporal datasets from the climate of the past, as well seasonal to decadal climate predictions.

 

 

References:

[1] Jupyter-AI GitHub Repository: https://github.com/jupyterlab/jupyter-ai
[2] Kadow, Christopher, Sebastian Illing, Etor E. Lucio-Eceiza, Martin Bergemann, Mahesh Ramadoss, Philipp S. Sommer, Oliver Kunst, et al.. 2021. “Introduction to Freva – A Free Evaluation System Framework for Earth System Modeling”. Journal of Open Research Software 9 (1): 13. https://doi.org/10.5334/jors.253.
[3] Freva GitHub Repository: https://github.com/FREVA-CLINT/freva
[4] Public Freva Instance: https://www.freva.dkrz.de/

How to cite: Oertel, F., Lucio Eceiza, E., Willmann, S., Wentzel, B., Bergemann, M., and Kadow, C.: An AI-Copilot for JupyterLab for climate data analyses using FrevaGPT, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16905, https://doi.org/10.5194/egusphere-egu25-16905, 2025.

EGU25-18720 | ECS | PICO | ESSI1.2

Leveraging AI for Material Identification in Unauthorized Dumps for Circular Economy Applications 

Adi Mager, Vered Blass, Aryeh Gorun, Yoni Tsur, and Moni Shahar

Aerial imagery has emerged as a powerful tool for environmental analysis and decision-making, enabling us to gain valuable insights. We present a comprehensive approach for performing semantic segmentation on aerial images of illegally dumped construction waste. We focus on the detection and analysis of the waste content to utilize it for circular economy. Leveraging the Segment Anything Model (SAM) developed by Meta, we produced highly accurate masks from aerial drone images. We created a dataset of over 46,000 manually labeled masks, which serve as ground truth for training and evaluation. Then we fine-tuned the ResNet-50 classification model together with the deep learning model. Our methodology combines the prediction of the classification model with these detailed masks to produce the final waste stream map. The map offers a comprehensive understanding of the open area allowing for further potential stocks analysis and economic evaluation. Overall, we achieved 86% detection accuracy on our full dataset, where for common classes the accuracy is higher. The waste identification can be used for economic and environmental decisions-making necessity of cleanup operations. The results also allow better planning of potential untapped stocks and treatment of different waste streams, aiding in local circular economy and waste management strategies. Our model development can serve the waste management and recycling sectors as well as municipal and national policy makers. 

How to cite: Mager, A., Blass, V., Gorun, A., Tsur, Y., and Shahar, M.: Leveraging AI for Material Identification in Unauthorized Dumps for Circular Economy Applications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18720, https://doi.org/10.5194/egusphere-egu25-18720, 2025.

EGU25-19037 | ECS | PICO | ESSI1.2

Efficient Large Ensemble Generation of Climate Model Output Using Latent Diffusion and Spatio-Temporal Transformers 

Johannes Meuer, Maximilian Witte, Claudia Timmreck, and Christopher Kadow

Estimating uncertainty in climate scenarios often requires generating large ensembles of high-resolution simulations, a task that is both computationally and memory intensive. To overcome these challenges, we propose a deep learning framework that combines a variational autoencoder for dimensionality reduction with a denoising diffusion probabilistic model built on a spatio-temporal transformer architecture. The model is trained on large ensembles of low-resolution climate model outputs to capture internal variability and a single high-resolution climate model output to generate high-resolution simulations. This innovative approach enables the dynamic generation of large ensembles of high-resolution simulations with minimal computational overhead, eliminating the need for storing extensive precomputed data. By facilitating the efficient quantification of uncertainty, this framework provides a powerful tool for exploring a wide range of high-resolution climate outcomes, supporting the development of informed climate policies and adaptation strategies.


How to cite: Meuer, J., Witte, M., Timmreck, C., and Kadow, C.: Efficient Large Ensemble Generation of Climate Model Output Using Latent Diffusion and Spatio-Temporal Transformers, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19037, https://doi.org/10.5194/egusphere-egu25-19037, 2025.

EGU25-19281 | ECS | PICO | ESSI1.2

From a regional to field scale - transfer learning for Earth observation based crop yield forecasts 

Emanuel Bueechi, Felix Reuß, Miroslav Pikl, Vojtech Lukas, Miroslav Trnka, Lucie Homolova, and Wouter Dorigo

Climate change is threatening food security, necessitating optimized resource management to ensure food availability. Field-scale crop yield forecasts, using machine learning and Earth observation data, have great potential for adaptive farm management, but the development of such models is curbed by the scarcity of field-scale training data. This strongly limits the applicability of traditional machine-learning approaches for field-scale crop yield modeling. However, increasingly popular transfer learning techniques provide a solution to improve this, since they can learn from a different domain than the one they are applied for. Here, we explore transfer learning to forecast crop yields on a field scale by training the model on a regional scale (where we have abundant data in Europe). We use Sentinel-1 and Sentinel-2 data with an artificial neural network to forecast maize, winter wheat, and spring barley yields in southern Czechia. We compared four model setups: two classical machine learning approaches trained and tested on a regional scale and one trained and tested on a field scale as a baseline. We compared these models to two transfer learning models that are trained on a regional scale and tested on a field scale, one with and one without fine-tuning the model using field-scale data. Forecasts were calculated at four lead times (1-4 months) before harvest. We showed that transfer learning with fine-tuning demonstrates superior performance, achieving correlations of approximately 0.75 at a one-month lead time for all crops. It outperformed the field scale-trained model by 0.05-0.12. In addition, transfer learning required significantly less field-level data to achieve a performance comparable to the model trained at the field level: 50% of the data for spring barley and maize, and only 25% for winter wheat. Therefore, this transfer learning approach improves the efficiency of crop yield data utilization and enhances field-level crop yield forecasting. 

The work of this study was conducted in the frame of the project “Yield Prediction and Estimation from Earth Observation” (funded by ESA - Contract No. 4000141154/23/I-EF) 

How to cite: Bueechi, E., Reuß, F., Pikl, M., Lukas, V., Trnka, M., Homolova, L., and Dorigo, W.: From a regional to field scale - transfer learning for Earth observation based crop yield forecasts, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19281, https://doi.org/10.5194/egusphere-egu25-19281, 2025.

EGU25-19971 | PICO | ESSI1.2

Geo Intelligence: A Key to Sustainable Paddy and Maize Production in Hassan District 

Vinay Shivamurthy, Mansoor Palat Ebrahim, and Vinuta M Betegeri

Agriculture, a cornerstone of human civilization, exerts a significant impact on natural resources to fulfill societal food demands. Climate change, exacerbated by anthropogenic activities and environmental consequences, poses a critical threat to agricultural productivity. While modern agronomic practices have enhanced yields, they have also resulted in detrimental consequences such as habitat loss, reduced biodiversity, and resource depletion.

This study investigates crop suitability in Hassan District, India, by integrating Artificial Intelligence (AI) with Geographic Information Systems (GIS). Eight key geo-climatic and pedological factors, relatively stable over time, were considered. Determining optimal land use for targeted crop cultivation is crucial in the face of climate change and global food security concerns.

Geospatial technologies and Sequential AI have demonstrated significant potential in addressing agricultural and environmental challenges through data-driven approaches. This research assesses the suitability of land for paddy, maize, and gram cultivation during the kharif season in Hassan District. A weighted metric approach was employed within a GIS environment, utilizing a Sequential Artificial Neural Network (ANN) model. Initially, an equal-weighted arithmetic mean was used to evaluate seven criteria encompassing soil, climate, and topographic factors. Likewise, criterion weights were derived from a sequential regression model, reflecting their relative importance in crop suitability prediction.Slope, soil depth, and rainfall emerged as the most influential factors, collectively accounting for 76% of the total weight. The results demonstrated an improvement in site suitability assessment compared to conventional methods, highlighting the efficacy of this integrated AI-GIS approach.

How to cite: Shivamurthy, V., Palat Ebrahim, M., and M Betegeri, V.: Geo Intelligence: A Key to Sustainable Paddy and Maize Production in Hassan District, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19971, https://doi.org/10.5194/egusphere-egu25-19971, 2025.

This study presents a comprehensive exploration of the collection and analysis of diverse geological and geophysical datasets from the eastern sector of the Russian Arctic. By leveraging advanced machine learning (ML) techniques, including convolutional neural networks, decision trees, and classical regression models, we provide insights into both data acquisition—encompassing geological, gravimetric, magnetic, and other parameters—and the subsequent analysis and interpretation of these data.

The research is structured around three primary objectives:

  • Data Collection and Structuring: A systematic approach to the acquisition and organization of information on the geological and geophysical conditions in the eastern Russian Arctic.
  • Application of Machine Learning Techniques: Employing cutting-edge ML methods to analyze and interpret the collected datasets.
  • Findings and Practical Implications: Highlighting key results and conclusions, with an emphasis on their practical applications in Arctic geological and geophysical research.

This work aims to introduce conference participants to innovative ML methodologies in geophysical data analysis and emphasizes the significance of employing diverse approaches to enhance understanding and application. The study also underscores the broader potential of these methods for application in other regions and global-scale research.

How to cite: Lisenkov, I. and Soloviev, A.: Analysis of Geological and Geophysical Data in the Eastern Russian Arctic Using Machine Learning Techniques, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20280, https://doi.org/10.5194/egusphere-egu25-20280, 2025.

Temperature plays a critical role in climate systems and resource management. Understanding spatiotemporal evolution of the temperature is vital for effective climate adaptation and resource management. Traditional models often treat spatial and temporal aspects separately, limiting their ability to capture the full correlation between these dimensions. This study evaluates various time series and machine learning models, including Holt-Winters, SARIMA, TSLM, NNAR, and ANN, using a daily dataset from 30 meteorological stations in Apulia region (Italy) from 1982 to 2023. These models are assessed based on RMSE and MAE metrics. The best models are then integrated with spatiotemporal kriging of the residual data, with results showing that the hybrid approach outperforms traditional methods. This generated high-resolution predictive maps provide valuable insights into temperature trends, supporting better decision-making in agriculture, water management, and climate resilience.

Funding information

Financial support from ICSC–National Research Center in High Performance Computing, Big Data and Quantum Computing, funded by European Union–NextGenerationEU”
Project name: PNRR-HPC; Project code: CN00000013; CUP: C83C22000560007. 

How to cite: Iqbal, N., De Iaco, S., and Palma, M.: Integrating Machine Learning and Time Series Models for Spatiotemporal Temperature Prediction: A Case Study from Apulia, Italy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20435, https://doi.org/10.5194/egusphere-egu25-20435, 2025.

EGU25-46 | Posters on site | ESSI1.5

Integrating Unoccupied Aerial Systems and Satellite Data to Map the Patchiness of Bare Ground at the Landscape Scale 

Guillermo E. Ponce-Campos, Philip Heilman, Cynthia L. Norton, Shang Gao, Michael A. Crimmins, and Mitchel P. McClaran

Integrating fine-scale measurements with broad-scale monitoring is a challenge for environmental monitoring, but it is a critical advancement in the face of increasing climate variability. We addressed this challenge by integrating fine-scale measures from Unoccupied Aerial Systems (UAS) to train broad-scale satellite imagery via machine learning algorithms. We applied this integration to detect how the spatial patchiness of bare ground varies over five years across a 100 km² semi-arid landscape in southern Arizona, USA. We used the Largest Patch Index (LPI) as the measure of spatial patchiness of bare ground. Our findings reveal three key advances in monitoring spatial patchiness over time and across a large landscape. First, the UAS-trained satellite estimates of LPI effectively represented the expected bare ground response to extreme climate events, where LPI increased during severe drought (-2.47 Standardized Precipitation-Evapotranspiration Index (SPEI)) and LPI decreased during exceptional wet periods (+1.95 SPEI). Second, the estimates of LPI were consistently 30-60% greater at lower and drier elevations, validating the ability to represent known ecological gradients. Third, and most notably, we confirmed that LPI is a scale-sensitive measure that differs between 3-m and 30-m grids, and that the magnitude of the differences is inversely related to the density of data in the satellite imagery. LPI was greatest using the 30-m grid Landsat 8 data with a density of 0.02 B/m² and LPI was least when using the 3-m grid PlanetScope data with a density of 0.9 B/m². But we found intermediate LPI values when resampling PlanetScope to 30-m grid while maintaining the greater data density. This previously unrecognized role of data density enriches the understanding of scale effects in landscape pattern analysis. In the end, we demonstrated a practical solution for integrating fine-scale UAS and broad-scale satellite observations via machine learning to support broad-scale environmental monitoring.

How to cite: Ponce-Campos, G. E., Heilman, P., Norton, C. L., Gao, S., Crimmins, M. A., and McClaran, M. P.: Integrating Unoccupied Aerial Systems and Satellite Data to Map the Patchiness of Bare Ground at the Landscape Scale, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-46, https://doi.org/10.5194/egusphere-egu25-46, 2025.

EGU25-1195 | Orals | ESSI1.5

Hydrological surfaces classification with Deep Learning using multiple sensors and exogeneous data 

Guillaume Eynard-Bontemps, Stéphane May, Dawa Derksen, Nicolas Dublé, Pierre-Jean Coquard, and Pauline Audenino

Traditional approaches for classifying hydrological surfaces - water, turbid water, salt pan, snow, and ice - usually rely only on one remote sensing dataset (often optical data like Sentinel-2). They face limitations under cloud cover areas and often confuse similar surface types (snow & salt pan, water & shadows). To overcome this, the study explores the use of Convolutional Neural Networks that can integrate spatial context, trained with multiple data sources like SAR (e.g., Sentinel-1), optical imagery, and exogenous inputs (weather, elevation). 

Deep Neural Networks are well-suited for texture extraction in remote sensing imagery and can efficiently handle inputs with multiple spectral bands. However, processing data from various sensor modalities introduces the challenge of aligning these inputs within a shared feature space where correlations can be effectively captured. To address this challenge, we developed a classical encoder-decoder architecture and explored the use of multiple encoders feeding into a single shared decoder. Two types of encoder families – EfficientNet and Swin Transformer – and two types of decoders – UNET and FPN – alongside various fusion methods were tried and showed similar performances.

For this study, a global multimodal database was gathered using open-source data from the Copernicus program. Initial trials with 17 labelled scenes (50GB) showed poor generalisation capabilities, leading to the extension of the dataset to 57 different scenes worldwide. Additional products were integrated, including Sentinel-1 (GRD VV+VH) data, 30m digital elevation models (ASTER GDEM), and meteorological data (from the ECMWF) to build the final 350GB database. Segmentation masks were generated semi-automatically (using a first version of our DL network) and then refined through visual inspection of Sentinel-2 images.

Results showed improved classification performance for all target classes when elevation data was included, and a dedicated dual-encoder-decoder model architecture proved particularly effective. On the other side, the integration of Sentinel-1 SAR data did not improve performance, likely due to the low temporal correlation between Sentinel-1 and Sentinel-2 acquisition (3-days average). Similarly, adding meteorological information did not enhance results, as our experiments showed that the model consistently disregarded scalar inputs regardless of integration approach.

Our model demonstrated notable robustness on the global database and was compared to existing CNES classification chains, including SurfWater (surface water detection) and Let-It-Snow (snow segmentation in mountains). Classification performance was comparable to SurfWater, though snow classification showed limitations in comparison to Let-It-Snow, particularly in the French Pyrenees.

The findings from this study underscore the potential of a multimodal approach in improving hydrological surface classification, particularly by incorporating data such as elevation. Future work could focus on increasing the volume of labelled data used to train the network to further enhance the model’s global applicability and precision across varied geographic and climatic conditions. Additionally, to fully leverage SAR imagery, reworking the database with more precise, directly annotated products would be essential. Finally, other approaches have to be tried in order to take into account meteorological data, for example using seasonality or more complex inputs. Other exogenous data could be added like terrain shadows.

How to cite: Eynard-Bontemps, G., May, S., Derksen, D., Dublé, N., Coquard, P.-J., and Audenino, P.: Hydrological surfaces classification with Deep Learning using multiple sensors and exogeneous data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1195, https://doi.org/10.5194/egusphere-egu25-1195, 2025.

The rapid expansion of Plastic-Mulched Landcover (PML),characterized by its relatively small size and short lifespan, necessitates precisely mapping PML using High-Resolution Remote Sensing Imagery (HRRSI). However, the high costs and limited temporal resolution of acquiring HRRSI pose significant challenges for precise PML identification. Remote Sensing Image Super-Resolution (RSISR) offers a viable solution by reconstructing high-resolution images from lower-resolution inputs, enhancing PML detection capabilities. This study, based on hybrid attention transformer, develops a Multi-Scale Gated Feedforward Attention Network (MSG-FAN) for super-resolution reconstruction of Sentinel-2 data to meter-level resolution. The main contributions include: (1) Construction of a PML RS dataset comprising 5300 pairs of 10-m Sentinel-2 and corresponding 2.5-m Gaofen-2 and Planetscope images from eight globally selected plastic-mulched planting regions. (2) Development of the MSG-FAN model, which enhances multi-channel, multi-scale and global attentions by integrating Gated Multi-Scale Feedforward Layer (GMS-FL), Top-k Token Selective Attention (TTSA) module and Global Context Attention (GCA) module. (3) Demonstration that MSG-FAN outperforms nine state-of-the-art deep learning-based super-resolution networks, achieving an average PSNR 30.81 and SSIM of 0.7287. Our proposed MSG-FAN model advances RSISR techniques and addresses critical challenges in monitoring plastic-mulched planting regions.

How to cite: Lu, L. and Du, Y.: A Multi-Scale Gated Feedforward Attention Network for Super-Resolution Reconstruction of Remote Sensing Images in Plastic-Mulched Planting Regions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2475, https://doi.org/10.5194/egusphere-egu25-2475, 2025.

EGU25-4523 | ECS | Orals | ESSI1.5 | Highlight

AnySat: a Multi-Resolution/Modality/Scale Earth Observation Model 

Guillaume Astruc, Nicolas Gonthier, Clément Mallet, and Loic Landrieu

Learning rich and robust representations of Earth Observation (EO) data is critical for effective and accessible geoanalytics. While the ever-growing volume of EO data suggests high potential for self-supervised learning, most approaches are limited to fixed scales, resolutions, or modalities—thus failing to generalize beyond their original sensor configurations. To address these shortcomings, we introduce AnySat, a novel multimodal framework capable of self-supervised training on multiple, diverse EO datasets simultaneously.

AnySat’s design centers on two key innovations. First, we propose a Joint Embedding Predictive Architecture (JEPA) adapted for multimodal EO. Unlike pixel-level reconstruction methods, JEPA operates in latent space—making it inherently more resilient to cloud cover, time-of-day shifts, and varying acquisition angles. Second, scale-adaptive spatial encoders allow a single network to handle variable spatial and temporal resolutions. Notably, more than 75% of AnySat’s 100M parameters are shared across all supported modalities, scales, and resolutions, enabling the model to fully exploit diverse training corpora—a fundamental requirement for developing a true EO foundation model.

To train AnySat, we compile GeoPlex, a collection of five multimodal datasets (PASTIS-HD, TreeSatAI-TS, PLANTED, FLAIR, and S2NAIP), aiming for diversity: 11 distinct sensors including radar and optical modalities, 0.2–250 m resolution, single-image and time series, and 0.3-2600 ha per input sample. Thanks to its versatility, a single Anysat model can learn powerful representations by training from all five datasets simultaneously. We only use cross-modal alignment as a source of self-supervision, and do not require labels for pretraining.

We fine-tune and evaluate our model on the datasets of GeoPlex, as well as four external datasets to evaluate generalization.  We report state-of-the-art results on seven downstream tasks, including land cover mapping, crop-type classification, tree-species identification, deforestation detection, and disaster mapping. Notably, AnySat yields significant performance gains across multiple benchmarks, such as +2.8 mIoU on PASTIS-HD, +3.6 mIoU on SICKLE, +11.0 accuracy on TimeSen2Crop, and +10.2 IoU on BraDD-S1TS.

A major benefit of AnySat is its high performance when performing linear probing with fixed representations—even for semantic segmentation tasks. This combination of versatility, generalizability, and ease of use positions AnySat as a valuable tool for practitioners facing diverse sensor types, specialized data distributions, and limited annotations.

How to cite: Astruc, G., Gonthier, N., Mallet, C., and Landrieu, L.: AnySat: a Multi-Resolution/Modality/Scale Earth Observation Model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4523, https://doi.org/10.5194/egusphere-egu25-4523, 2025.

EGU25-5955 | ECS | Posters on site | ESSI1.5

Deep learning for detecting landscape features in agricultural lands 

Muhammad Afif Fauzan, Holger Virro, and Evelyn Uuemaa

Agricultural landscape features are small fragments of natural or semi-natural vegetation in agricultural land, which, compared to their relatively small size, are essential in providing various ecosystem services and supporting biodiversity in the agricultural landscape. The Common Agricultural Policy (CAP) includes landscape features in its payment instruments, allowing farmers to receive incentives for preserving landscape features on their land. However, to effectively manage and monitor the status of landscape features requires their mapping, which is often done manually. The potential of deep learning methods has been promising in automatically segmenting particular objects on remote sensing images, but they require large amounts of labelled data to train the model, which is time-consuming to prepare manually. 

The aim of our study was to develop a deep learning methodology to automate the detection of landscape features in agricultural lands. We leveraged the publicly available dataset of landscape features’ polygons that has been created manually by farmers in Estonia to create labelled training data. To ensure that all landscape features in the database still actually exist, we filtered the dataset by applying a threshold of Normalized Difference Vegetation Index (NDVI) value difference between each field island and its surrounding arable land from three Sentinel-2 seasonal composites. Additionally, we checked the digitization quality of field island polygons by comparing them to orthophotos and the digital elevation model. The labelled training data were used to train a U-Net deep learning model to detect landscape features from orthophotos. We also experimented with adding elevation data as input to improve detection accuracy. We used F1-score and Intersection over Union (IoU) to evaluate the model performance. The results showed that the model is reliable for automated landscape feature detection and can be adopted by the relevant stakeholders to automate their workflow in delineating landscape features for incentive schemes to preserve small landscape features. 

How to cite: Fauzan, M. A., Virro, H., and Uuemaa, E.: Deep learning for detecting landscape features in agricultural lands, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5955, https://doi.org/10.5194/egusphere-egu25-5955, 2025.

Groundwater quality is a critical concern in agricultural regions, where nitrate contamination poses environmental and health risks, and ammonium levels play a pivotal role in nitrogen cycling processes. This study introduces a multi-task learning (MTL) framework designed to jointly predict nitrate and ammonium levels in groundwater, addressing the interdependencies between these variables. Conducted in Odense, Denmark, the study leverages spatial and temporal data, including hydrological, environmental, and anthropogenic variables, alongside land-use maps. The MTL approach outperforms traditional single-task models by capturing shared environmental and hydrological variables. By sharing information across tasks, the model identifies overlapping spatial, enabling robust predictions even in data-scarce scenarios. Additionally, the shared layers of the MTL model reduce overfitting, improving generalizability and providing deeper insights into the drivers of groundwater quality. The dataset used in this study includes geospatial nitrate and ammonium measurements, which were modeled alongside predictor variables such as land use, soil characteristics, and topographical variables. Model evaluation metrics demonstrated the superiority of the MTL approach, with increased accuracy, R², and reduced root mean squared error (RMSE) compared to separate models. The results highlight the potential of MTL to improve predictions and foster integrated groundwater management strategies. This study underscores the importance of advanced machine learning techniques in environmental modeling, showcasing a novel approach to jointly predict interrelated water quality variables.

How to cite: Ahmadi, K. and Naghibi, A.: Integrated Multi-Task Learning Framework for Groundwater Nitrate and Ammonium Prediction in Odense, Denmark, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6248, https://doi.org/10.5194/egusphere-egu25-6248, 2025.

Detailed and up-to-date information on urban land use plays a key role in understanding urban environment, evaluating urban planning and promoting the development of sustainable cities and communities. Recent years have witnessed many efforts dedicated to developing effective land use classification methods and generating products at different scales. Nevertheless, an accurate and fine-grained delineation of parcel-level urban land use for the entire China is still lacking. In this study, we developed a novel urban land use mapping framework to identify accurate land use categories by integrating multimodal deep learning model and multisource geospatial data. With complete and precise land parcels generated by road networks from two public source as minimum classification units, we produced a nationwide Urban Essential Land Use Categories (EULUC) map covering all cities in China for 2022, named as EULUC 2.0. The mapping results show that residential, industrial and park and greenspaces are the dominant land use categories across China, collectively accounting for nearly 80% of the urban area. The spatially explicit information provided by EULUC 2.0 can reveal distinct spatial patterns of the heterogeneous land use landscape in each city. The evaluation results found the overall accuracies of Level-I and Level-II classification could be as high as 72% and 79%, with substantial improvements across all categories over previous product. The advancements can be mainly attributed to the effectiveness of deep learning for multi-modal input, especially the graph modeling of Point-of-interest (POI) data. The free-access product and insights in this study can potentially help researcher and practitioners to investigate and address the pressing urban challenges in the process of urbanization.

How to cite: Li, Z. and Chen, B.:  Mapping Nationwide Essential Urban Land Use Categories by Integrating Multimodal Deep Learning and Multi-source Geospatial Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6757, https://doi.org/10.5194/egusphere-egu25-6757, 2025.

EGU25-8714 | ECS | Orals | ESSI1.5

Leveraging multi-modal classification of historical aerial images and topographic maps to derive past land cover 

Mareike Dorozynski, Franz Rottensteiner, Thorsten Dahms, and Michael Hovenbitzer

For the analysis of the evolution of landscapes, it is required to determine not only current states of the Earth’s surface, but to gain knowledge about past states, too. Sources of information for historic land cover are historic remote sensing imagery, and scanned historic topographic maps. To make the contained information explicitly available for subsequent computer-aided spatio-temporal analysis, classification techniques can be exploited. Against this background, multi-modal land cover classification from maps and aerial orthoimages is developed in the context of the Gauss Center (Gauss Centre, 2025), aiming to benefit from both the textural and geometrical details contained in aerial images, as well as the small intra-class variability in topographic maps.

The proposed deep learning-based classifier is a variant of a UPerNet (Xiao et al., 2018) with four down-sampling stages and takes aerial orthoimagery and topographic maps of the same epoch as an input. Each input modality is processed by an individual encoder, either based on convolutions, e.g. a ResNet (He et al., 2016), or on attention mechanisms, e.g. a Swin Transformer (Liu et al., 2022). This results in uni-modal map features and uni-modal aerial image features at four levels of detail. As the aerial images provide finer details about the texture and boundaries of the land cover objects, the aerial features of the first three stages are directly presented to the decoder, while the highest level aerial image features are fused with those of the topographic maps in a mid-level fusion. To focus on the most relevant features of the two modalities in spatial and feature dimension both, locally and globally, features are weighted by attention weights that are learned following the strategy in (Song et al., 2022). The lower-level aerial features and the high-level multi-modal features are presented to the decoder to predict to multi-modal land cover.

Experiments are conducted on two multi-modal datasets; one for binary building classification and one for multi-class vegetation classification. Both datasets consist of pixel-aligned aerial orthoimages, topographic maps and reference data at a ground sampling distance of 1 m. For all experiments, weights obtained in a pre-training on ImageNet (Russakovsky et al., 2015) are selected for the two encoder branches, while all remaining network weights are randomly initialized based on variance scaling (He et al., 2015). Training is proceeded utilizing the ADAM optimizer (Kingma & Ba, 2015) with standard parameters and a learning rate of 10-2 until the validation F1-score does not improve for 30 epochs. For both datasets, multi-modal predictions are compared to uni-modal predictions. Furthermore, attention-based feature extraction is compared to the one based on convolutions. The achieved mean F1-socres are the highest for the multi-modal variants of the classifier, where a higher score of 90.1% can be achieved utilizing convolutions on the building dataset (multi-modal, attention: 86.9%; aerial, convolution: 89.2%; map, convolution: 84.6%) and attentions are to be preferred for vegetation classification, resulting in a mean F1-score of 83.0% (multi-modal, convolution: 82.2%; aerial, attention: 82.1%; map, attention: 54.0%).

How to cite: Dorozynski, M., Rottensteiner, F., Dahms, T., and Hovenbitzer, M.: Leveraging multi-modal classification of historical aerial images and topographic maps to derive past land cover, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8714, https://doi.org/10.5194/egusphere-egu25-8714, 2025.

EGU25-9376 | ECS | Orals | ESSI1.5

Advancing environmental monitoring through deep learning: wildfire segmentation using time-series of images from the Sentinel constellation 

Gioacchino Alex Anastasi, Giuseppe Piparo, and Alessia Rita Tricomi

The integration of remote sensing and deep learning has revolutionized environmental monitoring, leveraging cutting-edge technologies to assist the decision-making processes in resource management and offering advanced tools for rapid disaster response. Our work employs satellite imagery to address pressing challenges in Earth observation, integrating multi-sensor, multi-resolution, and multi-temporal data for studying the aftermath of disastrous events by means of deep learning models, capable of handling such diverse data modalities.

We focused on the segmentation of wildfire-affected areas, using multispectral images from the Sentinel-2 satellites combined with the information from the Copernicus Emergency Management Service, in particular the geolocation and impact assessments, for more that 100 events occurred mostly in the European Mediterranean region. This dataset is further enriched with the observations from the Sentinel-1 and Sentinel-3 satellites, ensuring a comprehensive representation of the effects of each wildfire event by integrating measurements from multiple sensors with varying resolutions and revisit time. To streamline the workflow, a custom library based on the SentinelHub API has been developed, facilitating the download, preprocessing, and combination of data from different sources.

The study is performed on time-series of images, incorporating pre-event and post-event data, processed with a deep learning approach that combines Convolutional Long Short-Term Memory (ConvLSTM) layers in a UNet-like architecture. The results demonstrate the effectiveness of our model in accurately segmenting the affected areas, thus providing actionable insights for emergency management and recovery. Furthermore, the varied dataset, which comprises wildfire events occurring in diverse geographical conditions, enhances the robustness and generalizability of the described methodology.

This work is supported by ICSC – Centro Nazionale di Ricerca in High Performance Computing, Big Data and Quantum Computing, funded by European Union – NextGenerationEU, and it has been carried out within the Spoke 2 (“Fundamental Research and Space Economy”) as part of the activities in the Working Group 6 (“Cross-Domain Initiatives and Space Economy”) under the flagship use-case “AI algorithms for (satellite) imaging reconstruction”.

How to cite: Anastasi, G. A., Piparo, G., and Tricomi, A. R.: Advancing environmental monitoring through deep learning: wildfire segmentation using time-series of images from the Sentinel constellation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9376, https://doi.org/10.5194/egusphere-egu25-9376, 2025.

EGU25-10406 | ECS | Posters on site | ESSI1.5

PATCH-FILL: Multiscale and Univariate Gap-Filling in Remote Sensing Data 

Charly Zimmer, Anja Neumann, Miguel Mahecha, and Josefine Umlauft

Many applications in Earth system sciences require continuous, gap-free data sets. However, remote sensing data in particular are plagued by gaps due to clouds, incomplete coverage, or low-quality flags. Gap-Filling in remote sensing data often requires model architectures that are tailored specifically to underlying dataset characteristics such as scale, resolution or range of values. This limits the transferability to other gap-filling scenarios. Training these models is further hindered by the lack of adequate training samples, as they must be gathered from gap-afflicted data themselves. In this work, we present a spatiotemporal, univariate and multiscale gap-filling method that is independent of any specific dataset. A modular implementation allows for the customization of system parameters, so that the method can be adjusted and applied to various datasets, even outside the Earth Science domain. By employing a patch-wise gap-filling approach, introducing masked loss functions, and providing effective methods for synthetic gap generation, we are able to leverage gap-afflicted datasets and gather large amounts of training samples from them. To demonstrate the flexibility of the system, we perform gap-filling on multiple climatic variables from Earth System Data Cubes (ESDC) (Mahecha et al. 2020) using a 3D CNN architecture, making this the first global-scale gap-filling solution on ESDC. By capturing both spatial and temporal relations, the model is able to generate predictions that are coherent on large scale and across patches, thus demonstrating the potential of the patch-wise gap-filling framework and the use of 3D CNN architectures for spatiotemporal gap-filling tasks.

How to cite: Zimmer, C., Neumann, A., Mahecha, M., and Umlauft, J.: PATCH-FILL: Multiscale and Univariate Gap-Filling in Remote Sensing Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10406, https://doi.org/10.5194/egusphere-egu25-10406, 2025.

EGU25-11619 | ECS | Orals | ESSI1.5

Graph Neural Networks for Crop Cover Mapping 

Elif Donmez Altindal, Johannes Leonhardt, Ribana Roscher, Thomas Heckelei, and Hugo Storm

Crop maps provide valuable insights for a range of applications, including water resources management, crop yield prediction, and the planning of domestic and foreign policies. Information on seasonal or yearly agricultural land cover can help governments and organizations make informed decisions to address agricultural challenges and promote environmental sustainability. However, large-scale land cover mapping remains a significant challenge due to the high computational demands of processing remote sensing data, especially when using high-resolution imagery for large-scale applications such as country-wide mapping.

This high computational requirement can be partially mitigated by performing object-based classification, where data is summarized into segments or fields. A challenge in object-level mapping is selecting an appropriate method for analyzing and interpreting the data. For instance, convolutional neural networks (CNNs), a commonly used deep learning algorithm, are not directly applicable in their basic form because they require a gridded structure. Graph Neural Networks (GNNs) present a novel approach, effectively analyzing relationships between objects represented as nodes and edges in a graph. The ability of GNNs to capture complex relationships between segments in a non-grid structure offers distinct advantages, such as handling irregular or non-Euclidean data and exploiting spatial and temporal dependencies within a region. This makes GNNs particularly well-suited for high-resolution remote sensing tasks where traditional grid-based methods may struggle with spatial context and object interactions.

This study applies GNNs to multitemporal Sentinel-1 Interferometric Wide Swath data (VV and VH polarizations), leveraging ten-day composites from May to September to capture seasonal crop growth dynamics. Training, testing, and validation datasets cover 40×40 km², 20×20 km², and 20×20 km² areas, respectively, within North Rhine-Westphalia, Germany. Sentinel-1 images are segmented using the Felzenszwalb-Huttenlocher algorithm, grouping pixels into objects. Each segment’s average backscatter values are calculated, and crop class labels are assigned using InVeKos ground truth data, which includes field boundaries and crop information. This data is transformed into a graph, where nodes represent segments, and edges define adjacency. The GraphSAGE framework is employed to train the GNN model.

Performance comparisons include segment-level and pixel-level neural networks (NNs). Preliminary results show that GNNs achieved the highest accuracy (88.01%), outperforming segment-level NN (86.02%) and pixel-level NN (78.89%). GNNs also demonstrated efficient computational performance, with shorter inference times (0.19 seconds) compared to pixel-based methods (10.7 seconds), and generated more homogeneous maps, minimizing the salt-and-pepper effect.

These results highlight the potential of GNNs for scalable, object-based mapping at high resolution. The approach will be expanded to classify cropland across Germany, generating a 10-meter spatial resolution crop map. By leveraging the temporal dynamics of Sentinel-1 data and incorporating 2022 data, this method offers an efficient and robust framework for large-scale applications in crop management, land-use monitoring, and resource planning.

How to cite: Donmez Altindal, E., Leonhardt, J., Roscher, R., Heckelei, T., and Storm, H.: Graph Neural Networks for Crop Cover Mapping, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11619, https://doi.org/10.5194/egusphere-egu25-11619, 2025.

EGU25-12622 | ECS | Posters on site | ESSI1.5

Instrument-to-Instrument translation: An AI tool to intercalibrate and homogenize observations from Earth-observing satellites 

Anna Jungbluth, Lilli Freischem, J. Emmanuel Johnson, Robert Jarolim, Christoph Schirninger, and Anne Spalding

Climate change is fundamentally altering Earth's natural systems, from shifting weather patterns and sea level rise to increasingly frequent extreme events. Understanding and responding to these changes demands continuous, reliable observations of our planet. While Earth-observing satellites have collected terabytes of data in recent decades with ever-increasing temporal, spatial, and spectral resolution, synthesizing these diverse data sources into homogeneous, long-term records remains a significant challenge for climate monitoring and situational awareness. 

We address this challenge with Instrument-to-Instrument Translation (ITI), an artificial intelligence framework that learns to translate between different satellite imaging domains. Building on unpaired image-to-image translation techniques, ITI overcomes a fundamental limitation in satellite data integration - it does not require the instruments to observe the same location at the same time. This flexibility enables ITI to perform instrument intercalibration, enhance image quality, mitigate sensor degradation, and achieve super-resolution asynchronously across multiple wavelength bands to enable multi-vantage point observations

Building on ITI's proven success in harmonizing solar observations, we extend the framework to address the unique challenges of Earth observation and atmospheric monitoring. More specifically, we demonstrate ITI’s capability by harmonizing observations from two geostationary weather satellites with complementary coverage: the Meteosat Second Generation (MSG) monitoring Europe and Africa with 11 spectral bands, and the Geostationary Operational Environmental Satellite (GOES-16) observing the Americas with 16 spectral bands. For this, we developed rs_tools, a comprehensive software package that streamlines the creation of machine learning-ready datasets, and adapted the ITI pipeline to handle the specific complexities of Earth observation data, e.g. missing observations of visible bands at night. 

Our results reveal good agreement between the ITI-translated imagery and actual high-quality observations, especially for infrared spectral channels. We conduct a multi-faceted performance analysis using image quality metrics (PSNR, histogram distributions, power spectra) across varying spatial scales, spectral bands, and geographic features (land/ocean). The unique overlap in MSG and GOES-16 coverage over the Atlantic Ocean enables additional validation through paired metrics (MSE, Pearson correlation, SSIM) after projecting both observing systems into a common reference frame.

The ITI tool is available as open-source software for the research community, and can easily be adapted to novel datasets and research applications. This research outcome is supported by NASA award 22-MDRAIT22-0018 (No. 80NSSC23K1045) and managed by Trillium Technologies, Inc.

How to cite: Jungbluth, A., Freischem, L., Johnson, J. E., Jarolim, R., Schirninger, C., and Spalding, A.: Instrument-to-Instrument translation: An AI tool to intercalibrate and homogenize observations from Earth-observing satellites, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12622, https://doi.org/10.5194/egusphere-egu25-12622, 2025.

EGU25-12681 | ECS | Orals | ESSI1.5

PlantTraitNet: A Multi-Modal, Multi-Task Approach to Learning Global Plant Trait Patterns Using Citizen Science Data and Noisy Labels 

Ayushi Sharma, Daniel Lusk, Johanna Trost, and Teja Kattenborn

As primary producers in Earth's system, plants drive global matter and energy fluxes. Understanding the global distribution of plant functional traits and their biodiversity is, therefore, critical for understanding ecosystem behavior and Earth system dynamics in the face of climate and global change. However, we lack observations for various plant functional traits, such as plant height, leaf size, and nitrogen content, at a global scale.

These data gaps could be addressed through citizen science projects, where thousands of individuals have already recorded millions of plant photographs for species identification purposes. While these photographs do not include direct information about plant traits, trait data for thousands of plant species can be accessed from scientific databases. By linking these two data sources—crowd-sourced plant photographs and trait information from scientific databases—through plant species, we can supervise computer vision models to infer plant traits from plant images. The principle of "form follows function" suggests that a plant's appearance can provide valuable insights into its functional properties.

To assess the potential of citizen science data for plant trait estimation, we propose testing the feasibility of using weak and noisy labels for effective trait prediction. Considering that different plant traits are not independent of each other, we leverage multi-trait learning. Additionally, our approach incorporates plant images along with ancillary environmental data, such as soil conditions and Earth observation satellite data, to provide crucial context on factors like climate or land surface properties.

To fairly evaluate model performance, we curate a clean dataset spanning diverse geographic regions, as well as taxonomic and phylogenetic groups. We conduct a comprehensive study on the resilience of trained models across these distribution shifts. Furthermore, we assess which traits can be effectively learned from noisy labels and explore the extent of trait transferability under different conditions.

Our findings indicate that models trained on noisy data can, to a notable extent, predict a series of plant traits, including plant height, leaf area, and specific leaf area. This approach provides an efficient, scalable, and non-destructive method for estimating important plant functional traits. It could lay the groundwork for large-scale biodiversity monitoring and ecosystem assessment, with the potential to revolutionize how we track the functional properties of ecosystems at a global scale.

How to cite: Sharma, A., Lusk, D., Trost, J., and Kattenborn, T.: PlantTraitNet: A Multi-Modal, Multi-Task Approach to Learning Global Plant Trait Patterns Using Citizen Science Data and Noisy Labels, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12681, https://doi.org/10.5194/egusphere-egu25-12681, 2025.

EGU25-14877 | ECS | Posters on site | ESSI1.5

Crowd-sourced images geo-localization method based on multi-modal deep learning 

Qianbao Hou, Ce Hou, Fan Zhang, and Qihao Weng

Urban dwellers today frequently document their daily experiences through digital photography, sharing these moments across popular social networking platforms. While these platforms host vast collections of images, the geographical data associated with many photos is often imprecise or missing entirely. Accurately determining the geographic coordinates of user-submitted photographs adds substantial value to these visual records, offering practical applications in city development planning, architectural studies, and public safety monitoring. Despite its potential benefits, the process of precise image geo-localization remains technically complex and challenging. This study presents an innovative approach to geo-localize crowd-sourced images in urban settings, addressing the limitations of traditional methods. By combining street-view panoramas and satellite imagery through a novel contrastive learning framework, we significantly improve localization accuracy. Using Hong Kong as a case study, we demonstrate substantial improvements over existing approaches, reducing median and average errors by 77.4% and 63.6%, respectively. Surprisingly, our findings reveal that satellite imagery alone outperforms street-view data in geo-localization tasks, challenging previous assumptions. This research not only advances the field of urban image geo-localization but also provides a valuable multi-source benchmark, paving the way for future innovations in urban sensing, mapping, and analysis across various disciplines.

How to cite: Hou, Q., Hou, C., Zhang, F., and Weng, Q.: Crowd-sourced images geo-localization method based on multi-modal deep learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14877, https://doi.org/10.5194/egusphere-egu25-14877, 2025.

EGU25-16027 | Orals | ESSI1.5

Harmonizing Multisatellite Geostationary Observations Using Zero-Shot Video Frame Interpolation 

Yeji Choi, Hyun Gon Ryu, Minseok Seo, and Doyi Kim

Geostationary satellites such as GOES, GK2A, Himawari, and MTG/MSG have been providing a wealth of observational data over the past few decades, capturing the evolution and movement of clouds and precipitation systems with unprecedented detail. This extensive record has been instrumental in advancing our understanding of atmospheric dynamics and supporting the continuous monitoring of extreme weather events and natural disasters. Despite their capability to observe wide areas and the advantage of covering the entire globe with data from just three satellites, utilizing these geostationary satellites to model the Earth system on a global scale has proven challenging due to differences in observation intervals, spectral channels, and coverage footprints. To address these challenges, we propose a zero-shot Video Frame Interpolation (VFI) framework designed to harmonize imagery from multiple geostationary satellites. This method adapts the Many-to-Many Splatting VFI model, originally developed for RGB video processing, to work with single-channel infrared satellite imagery. By generating intermediate frames at high temporal resolutions (e.g., 2–5-minute intervals), our approach enables near-synchronous global coverage with improved temporal consistency. This method offers two key benefits. First, it enhances uniform sampling across satellites, creating a cohesive global view that is particularly valuable in data-sparse regions. Second, the interpolated frames improve the ability to capture and track critical meteorological phenomena. In addition, we address practical considerations such as computational efficiency, consistency with radiative or brightness temperature fields, and the robustness of zero-shot generalization. Our findings suggest that this zero-shot VFI framework can significantly advance global nowcasting providing a pathway to more accurate and timely Earth system modeling.

How to cite: Choi, Y., Ryu, H. G., Seo, M., and Kim, D.: Harmonizing Multisatellite Geostationary Observations Using Zero-Shot Video Frame Interpolation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16027, https://doi.org/10.5194/egusphere-egu25-16027, 2025.

EGU25-16891 | ECS | Orals | ESSI1.5

Reconstructing 3D cloud fields from multispectral satellite images using deep learning 

Stella Girtsou, Lilli Freischem, Kyriaki-Margarita Bintsi, Guiseppe Castiglione, Emiliano Diaz Salas-Porras, Michael Eisinger, Emmanuel Johnson, William Jones, Anna Jungbluth, and Joppe Massant

Clouds affect Earth’s radiation balance by reflecting incoming sunlight (cooling effect) and trapping outgoing infrared radiation (warming effect). Their vertical distribution in the atmosphere significantly influences their radiative properties and overall climate impacts. However, how clouds will respond to climate change remains unknown: cloud feedbacks are the largest source of uncertainty in climate projections. Global 3D cloud data can help reduce these uncertainties, improve climate predictions, and support better decision-making.

Clouds are observed globally from space using satellites, which provide insights into their distribution, structure, and evolution. Observations from the Cloud Profiling Radar (CPR) aboard NASA’s CloudSat mission have provided valuable information on the vertical distribution of clouds. However, its long revisit times (~ 16 days), narrow swath (1.4 km) and observations limited to the same local time each day hinder our ability to study the temporal evolution of clouds or their diurnal cycle. In contrast, imaging instruments observe larger regions with higher temporal resolution but only offer a top-down view with limited vertical information.

In this work, we apply deep learning to images observed by geostationary satellites paired with vertical cloud profiles to extrapolate the vertical profiles beyond the observed tracks. Specifically, we use 11-channel imagery from the MSG/SEVIRI instrument, colocated with CPR vertical profiles. First, we pre-train models using self-supervised learning methods, specifically (geospatially-aware) Masked Autoencoders, applied to MSG/SEVIRI data from 2010. The pre-trained models are then fine-tuned for the 3D cloud reconstruction task using paired image-profile data. As only a small fraction of images overlap with CloudSat observations, the pre-training step enables us to exploit the full information contained in the MSG/SEVIRI images. We find that pre-training consistently improves reconstruction performance, particularly in complex regions such as the inter-tropical convergence zone. Notably, geospatially-aware pre-trained models incorporating time and coordinate encodings outperform both randomly initialized networks and simpler U-Net architectures, leading to improved reconstruction results compared to previous work.

In the future, we plan to extend this method to longer time periods and apply it to ESA’s EarthCARE data, once available, to further improve 3D reconstructions and enable the development of long-term 3D cloud products.

How to cite: Girtsou, S., Freischem, L., Bintsi, K.-M., Castiglione, G., Diaz Salas-Porras, E., Eisinger, M., Johnson, E., Jones, W., Jungbluth, A., and Massant, J.: Reconstructing 3D cloud fields from multispectral satellite images using deep learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16891, https://doi.org/10.5194/egusphere-egu25-16891, 2025.

EGU25-18922 | Orals | ESSI1.5

Land-Surface Temperature Super Resolution from Geostationary to Low Earth Orbit Satellite Products with a Masked Autoencoder 

Manuel Traub, Matthias Karlbauer, Florian M. Hellwig, Thomas Jagdhuber, and Martin V. Butz

Remote sensing data from satellites offer real-world observations on large spatial scales without incorporating model biases and model simplifications such as contained in reanalysis datasets. Numerical weather prediction models benefit largely from data with high temporal and spatial resolution, as provided by Earth observation remote sensing missions. Yet, while geostationary (GEO) satellites provide data at high temporal, e.g., 15 minutes, but low spatial resolution, e.g., 5 km, low earth orbit (LEO) satellites deliver data at low temporal, e.g., 16 days, but high spatial resolution, e.g., 90 m. In this research study, we therefore train a combination of a masked autoencoder and a ResNet model to learn a mapping from GEO to LEO Land-Surface Temperature (LST) products. The model receives the coarse-resolution 5 km LST from the Copernicus Global Land Service (apart from other static inputs) to approximate the fine-grained 70 m LST product from NASA’s ECOSTRESS mission. We use the spatial domain extent over Europe defined by the Land Atmosphere Feedback Initiative (LAFI). In theory, our algorithm allows the generation of 70 m LST estimates at a temporal resolution of 15 minutes. However, missing or corrupted input patches, when covered by clouds or in the event of missing sensor coverage or outages, challenges this optimal resolution. Therefore, we aim at 70 m daily LST estimates across continental Europe. We will present examples of the super resolution results from different biome regions across Europe, highlighting the potential and limitations of our approach.

How to cite: Traub, M., Karlbauer, M., Hellwig, F. M., Jagdhuber, T., and Butz, M. V.: Land-Surface Temperature Super Resolution from Geostationary to Low Earth Orbit Satellite Products with a Masked Autoencoder, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18922, https://doi.org/10.5194/egusphere-egu25-18922, 2025.

EGU25-19171 | ECS | Posters on site | ESSI1.5

Towards a Scalable Deep Learning Framework for Forest Monitoring under Challenging Conditions with Multimodal Data 

Lorenzo Beltrame, Jules Salzinger, Jasmin Lampert, and Phillipp Fanta-Jende

Frequent cloud cover and terrain-induced shadows pose significant challenges for reliable forest monitoring. Traditional monitoring methods, such as ground-based observations and aerial surveys, often suffer from low temporal resolution, making it difficult to track seasonal changes or detect sudden forest anomalies, such as windthrow damage. Earth Observation (EO), particularly Sentinel-2 imagery, offers the potential for high revisit rates and global coverage, but these advantages are diminished by the persistent presence of clouds and shadows, particularly during winter months in mountainous areas. The tasks of forest anomaly detection and windthrow damage assessment particularly benefit from the increased temporal resolution provided by cloud and shadow free Sentinel-2 imagery. 

The SAFIR project aims to develop a scalable and robust framework for comprehensive forest monitoring, with a focus on resilience in complex terrains, including mountainous regions. Within the project and to fully leverage the advantages of EO, it is crucial to implement effective preprocessing techniques to address cloud and shadow disturbances. These challenges can be overcome by employing a method that predicts missing image information by reconstructing the albedo. This process involves integrating spatial, spectral, temporal, and physical priors into the image restoration, allowing for the extraction of meaningful information from partially obscured satellite measurements. 

This contribution introduces a concept for a modular deep learning framework designed to process cloudy or shadowed satellite images and predicting the corresponding albedo values. The framework consists of two core modules: a shadow remover and a cloud remover. Both modules undergo pretraining on large cloud-free satellite datasets to build robust spatiotemporal embeddings. They are subsequently fine-tuned using physics-based methods to improve accuracy in restoring obscured and clouded image areas. Unlike traditional approaches that prioritize visual clarity, this framework is optimized for machine learning. The objective is to create enhanced data products for downstream forest monitoring applications. The effectiveness of this approach is validated by comparing the results with non-enhanced Sentinel-2 data, making the downstream tasks a methodological validation step.  

Validation is also conducted using multimodal data, integrating satellite imagery with high-resolution Unmanned Aerial Vehicle (UAV) data. The planned UAV campaigns, conducted in Portugal, Germany and Austria, capture low-altitude imagery at 120 m. Hence, they provide ground-truth validation by revealing surface conditions beneath cloud cover. This validation step supports the fine-tuning of the image restoration models and ensures that restored satellite images align closely with real-world conditions.  

By leveraging heterogeneous data sources, including high-quality in situ UAV data, this contribution introduces a scalable concept for high-frequency satellite monitoring. The framework aims to go beyond experimental setups and achieve operational deployment in the GTIF initiative by ESA, making EO more efficient. 

How to cite: Beltrame, L., Salzinger, J., Lampert, J., and Fanta-Jende, P.: Towards a Scalable Deep Learning Framework for Forest Monitoring under Challenging Conditions with Multimodal Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19171, https://doi.org/10.5194/egusphere-egu25-19171, 2025.

Urban heat islands (UHI) exacerbate health and environmental challenges, disproportionately affecting vulnerable populations. This study identifies high-risk areas for UHI effects in the Americas, including their metropolitan regions, using a suitability analysis model. It highlights the interplay between urban expansion, social vulnerability, and climate stress, emphasizing the urgency of addressing these issues in rapidly urbanizing contexts.

High-resolution satellite imagery and geospatial data were used to build the model. Key criteria included population density (dasymetric layers from WorldPop), relative wealth index, land surface temperature (LST) from MODIS, land cover from MODIS, PM2.5 and NO2 concentrations (Sentinel-5P), and road network layers derived for the analysis. Each criterion was reclassified, transformed to a common scale, and weighted equally to ensure consistency and comparability.

The suitability index was generated using raster algebra (weighted sum), producing a continuous map where higher values indicate greater susceptibility to heat stress and lower socioeconomic status. The analysis revealed spatial patterns that highlight areas with high potential impacts due to UHI characteristics.

The suitability index serves as a tool for identifying priority areas for targeted interventions and climate mitigation actions. This integrative approach highlights the need for sustainable urban development policies that reduce socio-environmental disparities and promote resilience in vulnerable communities

How to cite: Rocha, T.: Identifying Urban Heat Island Risk Areas with Vulnerable Populations: A Suitability Analysis Approach to support Health Interventions in the Americas, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-742, https://doi.org/10.5194/egusphere-egu25-742, 2025.

EGU25-1299 | Posters on site | ESSI1.6

Building a framework for the design and deployment of digital twins: the Digital Twin Factory project 

Thibault Xavier, Dawa Derksen, Vincent Martin, and Pierre-Marie Brunet

The digital twin is a useful tool for scientists and decision makers to understand the present (what now), explore future trajectories (what next) to to investigate the future impacts of current risk mitigation actions (what if), or of a system. Working at the local scale allows detailed physics to be implemented in an approach that better captures the complexity of the study site (city, watershed, etc.) in an approach that complements the global scale. The availability of very high-precision spatial products (optical, 3D, thermal, etc.) enables this high-precision local analysis anywhere on the Earth.This growing interest is leading a number of actors to build digital twins at the local scale. However, building this type of representation requires a dedicated effort from the user, usually a scientist, which prevents him from focusing on the scientific added value he could bring with his thematic expertise.
The Digital Twin Factory (DTF, 2024-2026) project, coordinated by the French National Centre of Space Studies (CNES), aims to provide users with a framework capable of building, deploying and operating a digital twin at the scale of the site. It is designed as a Digital Twin as a Service API (PaaS) to abstract the underlying infrastructure, with possibility of accessing both the HPC resources and usual Cloud providers. The DTF also provides users with methodological building blocks to access (catalogue harvester), manipulate (ingester, data processing pipeline), visualize and analyze (plot, dashboarding) the data. In this way, the instantiators of the digital twin can focus on their thematic expertise and deploy their physical solvers with access to multi-source data.
While high performance computing resources can be made available to run these physical models, parametric studies or climate trajectories may require high cost and long simulation times. Partial or full data based surrogate model is an approach that can overcome this barrier and provide results in a reactive manner. Part of the DTF's work is therefore aimed at providing users with methodological building blocks for surrogate modelling, based on the expertise of the scientific community.
This contribution presents the multi-layered architecture of the DTF project, its different components and the services offered to users. We illustrate this work with the construction of the digital Twin of Nokoué Lake in Benin that integrates flood forecasting, pollution control, salinity management, long-term risk evolution, risk governance, and adaptation measures. Satellite data are used as input for a hydrodynamic code, on which first developments of surrogate models are presented.

How to cite: Xavier, T., Derksen, D., Martin, V., and Brunet, P.-M.: Building a framework for the design and deployment of digital twins: the Digital Twin Factory project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1299, https://doi.org/10.5194/egusphere-egu25-1299, 2025.

An Earth System Digital Twin (ESDT) is a dynamic, interactive, digital replica of the state and temporal evolution of Earth systems. It integrates multiple models along with observations, and connects them with analysis, AI, and visualization tools. Together, these enable users to explore the current state of the Earth system, predict future conditions, and run hypothetical scenarios to understand how the system would evolve under various assumptions. To establish a sustainable, extensible, and expandable ESDT solution, a formal software architecture is needed. Through the collaboration across multiple NASA centers, the teams partnered with the Apache Software Foundation to establish the professional open-source Integrated Digital Exploratory Analysis System (IDEAS) framework for digital twins. IDEAS is designed with three basic concepts. 

  • Design to Exploit– ESDT architect needs to be aware of existing freely available provisioned data, model, and visualization services that the ESDT could benefit from. In computer science, we instructed our students the simple Don’t Repeat Yourself (DRY) principle. The is especially true for ESDT, because ESDT is aiming for operational big data solutions, not just a quick demonstration. If an organization is already actively making their curated data, analysis, or model outputs available, the ESDT should find ways to exploit them. Software reuse is another form of exploitation. Why reinvent? Exploiting existing services and software reuse would significantly reduce future operation costs. An ESDT should also be prepared to be exploited by other ESDTs. It will be discussed in the Design to Expand principle. 
  • Design to Extend– An ESDT needs to be extensible to support new measurements, models (both numeric and AI-based), and interfaces. As we are connecting our digital assets, we will encounter gaps and limitations, both in data and technology. As we identify new climate phenomena or scientific needs, the ESDT needs to be prepared for these changes in technology and requirements. 
  • Design to Expand– It would be unrealistic to expect a single ESDT is capable to replicate the complex Earth System that is equipped with all the past and present observations, to drive all possible global and regional numerical models, and process all AI capabilities concerning instruments, data, and predictions. We could try to acquire all publicly available data, but an ESDT should also be able to integrate private, commercial data. The Babel-like ESDT would require significant computing and store, and well as staff (science and technical) to operate. This is the motivation behind the open-source IDEAS framework, to encourage collaboration to set up common public-facing architecture. Imagine being able to orchestrate a federation of ESDTs using exploration tools like a Jupyter notebook without having to setting up a local ESDT instance. We think federated ESTDs is the answer to making actionable science available to Disadvantaged Communities. 

The presentation will discuss IDEAS architecture, its latest progress, and its growing portfolio of digital twins for the Earth including air quality, wildfire, hydrology, and coastal zone.

How to cite: Huang, T.:  Integrated Digital Exploratory Analysis System (IDEAS) – An Open-Source Software Framework for Digital Twins, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2622, https://doi.org/10.5194/egusphere-egu25-2622, 2025.

EGU25-3702 | ECS | Posters on site | ESSI1.6

What if Data story telling was the corner stone for environmental digital twins? 

Faten EL outa and Guillaume Dechambenoit

Environmental digital twins face significant interdisciplinary challenges in their development and operation, particularly in managing complex environmental data and facilitating effective communication among diverse stakeholders. While these virtual representations of environmental systems offer powerful capabilities for monitoring and decision-making, they often struggle to bridge the communication gap between technical experts, decision-makers, and end-users.

Data storytelling is the practice of narrating messages derived from data to address specific needs and visually communicating these messages to an audience in an ordered manner that is easily understandable. Interestingly, digital twins share a similar objective: both aim to simplify and communicate complex data through intuitive and meaningful narratives.

Building on this shared characteristic, we propose an approach that adapts data storytelling techniques to the creation of digital twins. This abstract focuses on how data storytelling can enhance the creation and communication of digital twin data through visual formats tailored to specific audiences, addressing their unique needs to support monitoring, decision-making, and actionable insights.

This innovative integration of data storytelling and environmental digital twins establishes a comprehensive approach to address three key challenges:

  • Documenting and structuring the development process  from data to communication to incorporate stakeholder needs and communication requirements from the outset.
  • Facilitating collaboration among interdisciplinary teams through shared narrative frameworks.
  • Ensuring environmental insights are effectively translated into actionable knowledge.

We present a methodology that leverages data storytelling techniques to enhance the accessibility and impact of environmental digital twins, ultimately improving their effectiveness in environmental monitoring, decision-making, and stakeholder engagement.

 

How to cite: EL outa, F. and Dechambenoit, G.: What if Data story telling was the corner stone for environmental digital twins?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3702, https://doi.org/10.5194/egusphere-egu25-3702, 2025.

EGU25-4499 | ECS | Orals | ESSI1.6

Coupling approaches for data-driven Earth system models 

Lorenzo Zampieri, Harrison Cook, Rachel Furner, Sara Hahner, Florian Pinault, Baudouin Raoult, Nina Raoult, Mario Santa Cruz, and Matthew Chantry

Machine learning models have emerged as powerful tools for simulating Earth system processes. Following their successful application in capturing atmospheric evolution for medium-range weather forecasts, attention has increasingly shifted towards other components of the Earth system, such as the marine and land environments. This interest is further driven by the potential to enhance forecasting capabilities beyond the medium range. Machine learning frameworks offer remarkable flexibility in integrating these model components to achieve a coherent Earth system representation. At one end of the spectrum, model components can be trained jointly within a unified framework optimised using a shared loss function. At the other end, components may be developed independently and coupled by exchanging physically relevant information at multiple interfaces, mirroring the traditional coupling strategies employed in numerical models. In this presentation, we will examine the advantages and challenges of these approaches, with a particular emphasis on coupling the atmospheric, land, and marine components within the deterministic AIFS model, the machine learning-based forecast system developed at ECMWF. Furthermore, we will compare the coupling strategies of data-driven models with those of traditional numerical models, highlighting their strengths and limitations.

How to cite: Zampieri, L., Cook, H., Furner, R., Hahner, S., Pinault, F., Raoult, B., Raoult, N., Santa Cruz, M., and Chantry, M.: Coupling approaches for data-driven Earth system models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4499, https://doi.org/10.5194/egusphere-egu25-4499, 2025.

EGU25-5110 | Orals | ESSI1.6

Digital twin for weather-induced extremes 

Roger Randriamampianina and the DestinE Extremes Digital Twin Team

Our presentation aims to describe the development and operationalisation of the Destination Earth (DestinE) Extremes Digital Twin (DT), including the On-Demand component, a system designed to improve the prediction and management of extreme weather events in Europe. The system leverages high-resolution weather models using information from Extreme Detection (EDF) and Triggering (DTF) Frameworks, as well as ECMWF ensemble, incorporating impact-specific models for hydrology, air quality, renewable energy, and more. A key component is a configuration lookup table prioritising end-user needs and available resources. The system incorporates various masking techniques (ACCORD models configurations, geographical, capacity, event type) to refine forecasts. The presentation describes the system's architecture, data sources, and workflow, emphasising the integration of multiple models and data sources, and the use of cutting-edge technologies such as GPUs and machine learning for enhanced forecasting and efficient resource utilisation. Pilot regions are used for testing and operationalisation, with a phased approach planned for broader deployment. The project addresses challenges in forecasting accuracy, communication of uncertainty, and the integration of forecasts into decision-making processes across various sectors.

How to cite: Randriamampianina, R. and the DestinE Extremes Digital Twin Team: Digital twin for weather-induced extremes, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5110, https://doi.org/10.5194/egusphere-egu25-5110, 2025.

EGU25-5909 | ECS | Orals | ESSI1.6

Enhancing the Skill of Medium Range Forecasts with a Machine Learning Based Multi-Model Super-Ensemble (MMSE) 

Karan Purohit, Mitali Sinha, Aniruddha Panda, Subhasis Banerjee, and Ravi S Nanjundiah

In recent years, medium-range AI weather forecasting models have improved significantly, now offering forecasting accuracy comparable to classical numerical weather prediction (NWP) models, while also being faster and (once trained) less computationally demanding.

Due to inherent assumptions and limitations, all weather prediction models exhibit some degree of persistent systematic errors, also called biases, in their forecast output, with certain models performing better than others for specific variables and regions.

To address these persistent biases, we introduce a machine learning-based multi-model super-ensemble (MMSE), which collectively reduces model biases by combining the complementary strengths of each model. The MMSE assigns optimized weights to each model's forecast based on its historical performance to leverage each model’s strengths under different conditions (both spatial and temporal) rather than equally weighting models as in a simple ensemble mean.

In this work, we developed two regional MMSE models tailored to specific regions, seasons, and variables of interest. One model targets 2-meter air temperature and 10-meter wind components in Germany’s winter season, while the other targets Indian summer monsoon rainfall.

We trained the MMSE using an Extreme Gradient Boosting framework (XGBoost) to capture spatiotemporal features more effectively. The training data consisted of past forecasts from multiple AI models (FourCastNet, Pangu-Weather, GraphCast) and relevant climatology and topology data. ERA5 reanalysis served as the ground truth. The details of MMSE development will be presented.

Our MMSE developed for 2-meter temperature over Germany showed approximately a one-day improvement in forecast gain time compared to the best-performing individual model. In other words, the MMSE’s 11th-day forecast matched the accuracy of the 10th-day forecast from the best-performing model, effectively adding an extra day of reliable lead time. These findings suggest that the proposed MMSE offers a promising, computationally efficient alternative to traditional ensembles for real-time weather forecasting, with potential applications in domains requiring high-precision predictions. With a view to make these results interpretable and to identify the relative strengths of participating models, we will also present the analysis of SHAP values for various variables and regions.

How to cite: Purohit, K., Sinha, M., Panda, A., Banerjee, S., and S Nanjundiah, R.: Enhancing the Skill of Medium Range Forecasts with a Machine Learning Based Multi-Model Super-Ensemble (MMSE), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5909, https://doi.org/10.5194/egusphere-egu25-5909, 2025.

EGU25-6965 | ECS | Posters on site | ESSI1.6

Global post-processing of ERA5 precipitation product via graph-based neural networks  

Patrick Ebel, Linus Magnusson, and Rochelle Schneider

Total precipitation is a key variable of the weather state, accumulated over a given period. Beyond their direct relevance, high-quality precipitation data are of importance for driving downstream applications in hydrology, e.g. river streamflow and runoff forecasting. However, common measurements of precipitation are either precise but sparse (as for in-situ recordings) or global but uncertain (as for spaceborne observations). Though reanalysis products such as ECMWF’s ERA5 provide a best estimate of the state of the atmosphere, the quality of their total precipitation reconstruction is imperfect. Following reports that ERA5 is prone to overestimating the occurrence of drizzle at the cost of underestimating extreme precipitation, prior work explored data-driven models for local post-processing to address the latter. However, the local models employed in preceding work do not easily extend to a global post-processing setup and an exclusive emphasis on outliers limits the ability to represent the full distribution of precipitation intensity, which limits their relevance.

 

In this work, we propose a novel approach for precipitation post-processing which models the entire globe in a single forward pass and models dryness, light rain and heavy rain alike. The post-processer is based on a graph neural network architecture, trained on decades of gauge-calibrated multi-source weighted estimates of precipitation. We demonstrate that our model learns to bias-correct ERA5 total precipitation information and consistently improves upon the baseline while maintaining its global applicability. Further experiments will detail the nature of its improvements and may explore its benefits for downstream applications.  

How to cite: Ebel, P., Magnusson, L., and Schneider, R.: Global post-processing of ERA5 precipitation product via graph-based neural networks , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6965, https://doi.org/10.5194/egusphere-egu25-6965, 2025.

Developing region-specific radar quantitative precipitation estimation (QPE) products for South China (SC) is crucial due to its unique climate and complex terrain over there. Deep learning (DL) has emerged as a promising avenue for radar QPE, especially graph neural networks (GNNs). Many studies have tested the DL models in radar QPE, but virtually no studies have evaluated the performance of DL models in different precipitation intensities, types, or organizations. Moreover, limited attention has been given to whether DL-based methods can mitigate radar QPE errors caused by orographic influences in complex terrains, such as those in SC.

This study investigates the advantages of DL methods for QPE tasks in South China, utilizing nearly three years of hourly gauge data as labels and ground-based radar reflectivity as inputs. Firstly, multi-layer perceptron (MLP), Convolutional Neural Networks (CNNs), and GNNs with similar architectures are constructed and compared to traditional Z-R relationships considering precipitation types. DL methods outperform traditional Z-R relationships and GNNs perform the best. More importantly, this study conducts a systematic evaluation of the proposed GNN. For extreme precipitation (>30 mm/h), GNN achieves the smallest MAE, highlighting its potential for hazardous event estimation. It also demonstrates stable performance for stratiform and organized precipitation, with minimal bias and standard deviation. However, GNN is less effective for isolated precipitation, whereas CNNs are a better choice due to their ability to estimate scattered rainfall accurately. Last but not least, the Z-R relationship shows systematic spatial biases, overestimating precipitation in coastal plains and underestimating it in inland high-altitude regions. DL methods alleviate these terrain-induced biases by incorporating spatial information. Overall, this study highlights the advantages of DL methods across different precipitation scenarios and demonstrates their ability to mitigate systematic biases from complex terrain.

How to cite: Zhu, K. and Xu, W.: Deep Learning for Radar Quantitative Precipitation Estimation over Complex Terrain in Southern China, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8155, https://doi.org/10.5194/egusphere-egu25-8155, 2025.

EGU25-8169 | ECS | Orals | ESSI1.6 | Highlight

Exploring AI-Driven Event-based Storylines 

Amal John, Thomas Rackow, Nikolay Koldunov, Sebastian Beyer, Antonio Sanchez Benitez, Helge Gößling, Marylou Athanase, and Thomas Jung

Artificial Intelligence-based Numerical Weather Prediction (AI-NWP) models have recently emerged as powerful tools for weather forecasting, offering computational efficiency and high accuracy. This study explores the extreme weather events simulated by the Artificial Intelligence Forecasting System (AIFS), initialised with conditions derived from kilometer-scale storyline experiments using the IFS-FESOM model where the atmospheric circulation is constrained to observations. We present two case studies: the 2023 South Asian humid heatwave and the 2024 Storm Boris. These two events are reproduced in the present climate, but also simulated if they were to unfold in pre-industrial and +2K future climates, effectively creating AI-driven storylines. The methodology we employ offers a complementary framework, where the use of AI-driven ensembles provides a scalable and rapid way to assess the potential uncertainty and variability associated with such events, by enabling us to explore a broader range of plausible outcomes at very low computational costs. By combining the strengths of physics-based modelling with the efficiency and flexibility of AI-driven simulations, this dual approach offers a pathway to operationalise ensemble-based extreme weather storylines.

How to cite: John, A., Rackow, T., Koldunov, N., Beyer, S., Sanchez Benitez, A., Gößling, H., Athanase, M., and Jung, T.: Exploring AI-Driven Event-based Storylines, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8169, https://doi.org/10.5194/egusphere-egu25-8169, 2025.

EGU25-9956 | ECS | Orals | ESSI1.6

 Xaurora: Advancing subseasonal-to-seasonal forecasting by fine-tuning foundation weather models with spectral consistency  

Eliot Walt, Wessel Bruinsma, Maurice Schmeits, Efstratios Gavves, and Dim Coumou

Sub-seasonal to seasonal (S2S) timescales range from two weeks to three months and are crucial to make informed climate change-related decisions, including renewable energy resources allocation, extreme events’ risks mitigation, and the development of effective early warning systems. Unfortunately, traditional physics-based forecasting systems achieve poor skill on these lead times. Recently, deep learning (DL) has shown promising results in weather forecasting on timescales up to 10 days, reaching performance competitive with that of physical models. However, these DL approaches currently struggle on S2S timescales.  

Following previous studies on neural solvers for partial differential equations and weather forecasting, we propose a fine-tuning framework aimed at improving the S2S prediction skill of foundation weather models. Our approach has two core components. First, we implicitly condition the latent space embeddings to retain the predictable signals at a given lead time using an additional regression head. Second, we design a novel frequency-domain decoder and loss function to ensure spectral consistency. These steps should ensure that the model focuses on the most predictable frequencies. We apply this methodology to the recently published Aurora foundation model and propose Xaurora, standing for “extended Aurora”. Our fine-tuning approach represents an important milestone in data-driven S2S forecasting, addressing key challenges in the field while remaining broadly applicable with minimal assumptions on the underlying model’s architecture. 

The relevance of our framework is evaluated through ablation studies, comparing our spectral consistency fine-tuning to the original Aurora model. Furthermore, we provide standard deterministic and probabilistic skill scores on S2S timescales, as well as relevant teleconnection indexes. We present preliminary outputs of this analysis. 

How to cite: Walt, E., Bruinsma, W., Schmeits, M., Gavves, E., and Coumou, D.:  Xaurora: Advancing subseasonal-to-seasonal forecasting by fine-tuning foundation weather models with spectral consistency , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9956, https://doi.org/10.5194/egusphere-egu25-9956, 2025.

EGU25-10817 | ECS | Orals | ESSI1.6

MLOps on DestinE Data Lake – Towards Reproducible AI on Edge Services 

Sina Montazeri, Miruna Stoicescu, Oriol Hinojo Comellas, Danaële Puechmaille, and Michael Schick

Destination Earth (DestinE) is European Commission’s initiative to gradually develop highly accurate Digital Twins (DT)s of the Earth with unprecedented accuracy and resolution. DestinE will initially provide DTs for adapting to climate change, forecasting extreme events and interactive use of high-resolution climate data. Insights from these models support scientists and policymakers to study and plan for future weather- and climate-induced events. 

Stakeholders implementing what-if scenarios and/or ready to use applications on DestinE require the optimum storage and the seamless provision of access to a sheer volume of heterogeneous data often available from different data origins. EUMETSAT has implemented the DestinE Data Lake (DEDL) to address the above challenges. The DEDL offers the Harmonised Data Access (HDA) service that enables access to diverse data from the DEDL data portfolio via a unified STAC API. Furthermore, it offers, for power users, DEDL edge services on request, which are a dynamic suite of distributed big data processing components that operate close to DestinE’s massive data repositories. The edge services offered  are:  STACK (DEDL-managed software applications such as JupyterHub, DASK and Open Data Cube), ISLET (project-managed compute and storage services such as configurable virtual machines and S3 object storage) and HOOK (schedule and run pre-defined or user-defined high-level workflows, such as setting up a data processing pipeline). 

To efficiently exploit the wealth of data available on DestinE, DEDL edge services will extend their abilities to accommodate the necessary infrastructure and software to enable Artificial Intelligence/Machine Learning (AI/ML) activities. DEDL will offer an ML Operations (MLOps) service tailored to Earth Observation (EO) data, which allows users to engage in various steps of AI/ML such as data preprocessing, model training and evaluation, experiment tracking, model deployment, model inference and monitoring. The modularized DEDL MLOps architecture will allow the users to use components as required without the need to be bound to pre-defined workflows and pipelines. The users, furthermore, can develop their AI/ML algorithms according to CI/CD best practices and have multiple environments for development, staging and production. 

A specific focus of DEDL will be to define and work with highly flexible data pipelines. The framework will allow to convert DestinE data portfolio datasets to AI-ready formats, which can readily be used as inputs for various AI/ML models. The framework will have the capability to combine and harmonise data from various sources and formats and provides typical EO-based pre-processing steps such as data collocation, re-projection, and re-gridding among other operations. 

This presentation will highlight the AI/ML and MLOps capabilities of the DEDL, demonstrating how they empower users to efficiently analyse data and derive valuable insights. By seamlessly integrating with DestinE’s data ecosystem, these advancements enable users to focus on innovation and address critical challenges such as climate adaptation and extreme event forecasting, rather than on managing complex workflows or infrastructure. 

How to cite: Montazeri, S., Stoicescu, M., Hinojo Comellas, O., Puechmaille, D., and Schick, M.: MLOps on DestinE Data Lake – Towards Reproducible AI on Edge Services, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10817, https://doi.org/10.5194/egusphere-egu25-10817, 2025.

EGU25-11883 | ECS | Orals | ESSI1.6

Developing a data-driven global ocean model at ECMWF  

Rachel Furner, Rilwan Adewoyin, Mario Santa Cruz, Sara Hahner, Sarah Keeley, Kristian Mogensen, and Lorenzo Zampieri

Machine learning (ML) techniques have emerged as a powerful tool for predicting weather and climate systems, particularly in predicting the short-term evolution of the atmosphere. Here, we look at the potential for ML to predict the evolution of the 3d-ocean. 

We present a data-driven global ocean model, developed within the Destination Earth project, to form the ocean component of a fully data-driven earth system model. Following the skill shown by the AIFS (Lang et al, 2024), we use a graph-based encoder-decoder design, with a transformer backbone. Our model is trained on the ECMWF ORAS6 reanalysis dataset (Zuo et al, 2024). Work focuses on short-term predictions, up to a 2-week forecast period. The model predicts temperature, salinity, zonal and meridional current components throughout the full ocean depth, along with sea-surface height and sea-ice. 

In this presentation we will discuss the design choices of our network architecture, including comparisons between networks trained to predict future fields, and those trained to predict increments to fields. We will show results from our data-driven model and put these into the context of other similar models. 

Simon Lang, Mihai Alexe, Matthew Chantry, Jesper Dramsch, Florian Pinault, Baudouin Raoult, Mariana C. A. Clare, Christian Lessig, Michael Maier-Gerber, Linus Magnusson, Zied Ben Bouallègue, Ana Prieto Nemesio, Peter D. Dueben, Andrew Brown, Florian Pappenberger, and Florence Rabier (2024). AIFS – ECMWF’s data-driven forecasting system. arXiv preprint https://arxiv.org/abs/2406.01465.  

Hao Zuo, Magdalena Alonso-Balmaseda, Eric de Boisseson, Philip Browne, Marcin Chrust, Sarah Keeley, Kristian Mogensen, Charles Pelletier, Patricia de Rosnay, Toshinari Takakura (2024). ECMWF’s next ensemble reanalysis system for ocean and sea ice: ORAS6. ECMWF newsletter. https://doi.org/10.21957/hzd5y821lk  

How to cite: Furner, R., Adewoyin, R., Santa Cruz, M., Hahner, S., Keeley, S., Mogensen, K., and Zampieri, L.: Developing a data-driven global ocean model at ECMWF , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11883, https://doi.org/10.5194/egusphere-egu25-11883, 2025.

EGU25-11988 | Orals | ESSI1.6

Development of an offline and online hybrid model for the Integrated Forecasting System 

Marcin Chrust, Alban Farchi, Massimo Bonavita, Marc Bocquet, and Patrick Laloyaux

Systematic model errors significantly limit the predictability horizon and practical utility of the current state-of-the-art forecasting systems. Even though accounting for these systematic model errors is increasingly viewed as a fundamental challenge in the field of numerical weather prediction, estimation and correction of the predictable component of the model error has received relatively little attention. Modern implementations of weak-constraint 4D-Var are an exception here and a promising avenue within the variational data assimilation framework, showing encouraging results. Weak-constraint 4D-Var can be viewed as an online hybrid data assimilation and machine learning approach which gradually learns about model errors from partial and imperfect observations, allowing to improve the state estimation. We propose a natural extension of this approach by applying deep learning techniques to further develop the concept of online model error estimation and correction.

In this talk, we will present recent progress in developing a hybrid model for the ECMWF Integrated Forecasting System (IFS). This system augments the state-of-the-art physics-based model with a statistical model implemented via a neural network, providing flow dependent model error corrections. While the statistical model can be pre-trained offline, we demonstrate that by extending the 4D-Var control vector to include the parameters of the neural network, i.e. the model of model error, we can further improve its predictive capability. We will discuss the impact of applying the flow dependent model error corrections in the medium range forecasts on the forecast quality.

How to cite: Chrust, M., Farchi, A., Bonavita, M., Bocquet, M., and Laloyaux, P.: Development of an offline and online hybrid model for the Integrated Forecasting System, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11988, https://doi.org/10.5194/egusphere-egu25-11988, 2025.

EGU25-12042 | Posters on site | ESSI1.6

Implementing FAIR Agrobiodiversity Workflows for the Destination Earth Data Lake 

Claus Weiland, Daniel Bauer, Desalegn Chala, Dag Endresen, Jonas Grieb, Marcella Orwick Rydmark, and Gabriela Zuquim

The Horizon Europe project “Biodiversity Digital Twin for Advanced Modelling, Simulation and Prediction Capabilities” (biodt.eu) combines state-of-the-art supercomputing resources with FAIR data practices to address biodiversity grand challenges such as the loss of diversity at genetic, species and ecosystem levels driven by factors such as anthropogenic climate change, intensified land use and the spreading of alien invasive species.

Key tools for the integration and assessment of such information in BioDT are so-called prototype Digital Twins (pDTs) which provide digital replicas of particular phenomena of the biosphere. The data produced by pDTs and the computational workflows themselves are annotated with rich machine-interpretable metadata, notably schema.org and its Bioschemas and Croissant (high-level metadata format for machine learning datasets) extensions to enable discovery and interaction with the data for humans and machines. 

The crop wild relatives digital twin developed in BioDT for the identification of novel genetic resources in crop wild relatives (CWR) aims to support the development of mitigation strategies against food-related crises associated with climate change (doi:10.3897/rio.10.e125192). Underutilised crops and CWR are a valuable source of untapped genetic diversity for implementing such strategies because their greater genetic diversity in relation to crops makes adaptation to droughts, cold waves, heavy precipitation or other weather-induced extreme events possible. They are therefore of particular importance to guarantee food security.  

We developed a pilot study to migrate and functionally integrate the pDT CWR in the Destination Earth Data Lake (DEDL) to increase its availability and, more importantly, use both data provided in the data space and the near data processing features of DEDL to improve models and prediction. The framework makes use of web-based technologies such as RO-Crates and FAIR Signposting to facilitate reuse and repurposing of the data within and across different data spaces as well as orchestrated interplay of multiple digital twins (involving the Climate Change Adaptation Digital Twin). 

Building on this blueprint, we will showcase in this presentation the deployment of FAIR agrobiodiversity workflows in DEDL’s compute service Islet and the subsequent publication of results in a lightweight service framework (figure).

How to cite: Weiland, C., Bauer, D., Chala, D., Endresen, D., Grieb, J., Orwick Rydmark, M., and Zuquim, G.: Implementing FAIR Agrobiodiversity Workflows for the Destination Earth Data Lake, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12042, https://doi.org/10.5194/egusphere-egu25-12042, 2025.

EGU25-12349 | Orals | ESSI1.6

How Destination Earth Data Lake support Destination Earth users 

Patryk Grzybowski, Marcin Ziółkowski, Aubin Lambare, Christoph Reimer, and Michael Schick

Destination Earth (DestinE) is a flagship initiative led by the European Commission, implemented by EUMETSAT, ESA and ECMWF. It aims to create highly detailed Digital Twins (DTs) of the Earth, enabling precise simulations for a variety of uses. Currently, the initiative focuses on two primary Digital Twins:  the Weather Extremes Digital Twin (ExtremeDT) and the Climate Change Adaptation Digital Twin (ClimateDT). Over the coming years, the scope of DTs is set to expand, necessitating improved access to data and streamlined methods for working with it. This is where the Destination Earth Data Lake (DEDL) plays a pivotal role, offering comprehensive data discovery, access, and processing services tailored to the needs of DestinE users.

The DEDL operates on two key levels: ‘Data Discovery and Access’ and ‘Edge Services’. DEDL Discovery and Data Access services is provided by Harmonized Data Access (HDA) tool which provides a single, federated entry point to the services and data, including resources from existing datasets and complementary sources such as in-situ and socio-economic data. Notably, it also provides access to the unique datasets generated by DestinE’s DT’s. The services rely on use of the SpatioTemporal Asset Catalogs (STAC) standard which means:

  • The search in the dataset is done according to the STAC protocol;
  • The Federated Catalog search proxy component converts STAC queries into queries adapted to the underlying catalog and returns the results to the user in STAC format.

The cloud computing service is powered by the ISLET infrastructure, a distributed Infrastructure as a Service (IaaS) built on OpenStack. It allows users to manage virtual machines, s3 storage, and run advanced computations via a graphical user interface or command-line interface. A standout feature of ISLET is its proximity to data sources, operating near High-Performance Computing (HPC) facilities. This is achieved through data bridges, enabling efficient processing of large datasets, including those from Digital Twins, in conjunction with HPC systems.

The STACK environment supports application development using JupyterHub and DASK, with Python, and R languages. Users can create DASK clusters on selected infrastructure (sites) to process data directly where it resides, removing the need for extensive local setup and optimization.

Hook Services is a set of pre-defined workflows which could be used by users as a ready-to-use processors like: Sentinel-2: MAJA Atmospheric Correction; Sentinel-1: Terrain-corrected backscatter. It also enables workflow functions to generate on-demand higher-level products, such as temporal composites.

DEDL is a transformative initiative that revolutionizes how Earth Observation data is managed and utilized. By integrating innovative infrastructure (ISLET), data services (HDA), reliable processors (Hook Services), and user-friendly development tools (STACK), DEDL enables unprecedented levels of data harmonization, federation, and processing. Moreover, the DEDL plays a crucial role in empowering DestinE users by providing them with seamless access to vast datasets and advanced computational tools. It simplifies the process of data exploration, integration, and analysis, enabling researchers, policymakers, and developers to focus on innovation and decision-making rather than technical barriers. This cutting-edge system enhances climate research capabilities and supports sustainable development efforts on a scale previously unattainable.

How to cite: Grzybowski, P., Ziółkowski, M., Lambare, A., Reimer, C., and Schick, M.: How Destination Earth Data Lake support Destination Earth users, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12349, https://doi.org/10.5194/egusphere-egu25-12349, 2025.

EGU25-13817 | Posters on site | ESSI1.6

Real-time Prediction of Global Tropical Deciduous Ecosystem Phenology with Deep Learning 

Minchao Wu, Torbern Tagesson, Zhanzhang Cai, and Zheng Duan

Tropical deciduous ecosystems play a critical role in terrestrial ecological processes and the global carbon cycle, influencing seasonal climates through phenology-induced biophysical and biogeochemical feedbacks. Phenological processes for tropical deciduous ecosystems are complex with multiple intertwining climatic and physiological factors that co-shape the underlying dynamics. Here, we present a deep learning framework based on Temporal Fusion Transformer for predicting tropical deciduous phenology globally in real-time with high accuracy. The framework integrates long-term AVHRR-derived vegetation greenness data, high-resolution climate data from ERA-Land, and land surface features including physical and chemical properties to account for terrestrial spatial heterogeneities that affect phenological processes. Our preliminary results demonstrate the ability of the developed framework to accurately predict historical phenological dynamics across 35 growing seasons in the pan-tropical regions. Key phenological metrics, including the start, peak, and end of the growing season, are identified with high accuracy. We believe the framework provides a powerful tool for real-time predictions and reconstructions of phenological states for tropical deciduous ecosystems, especially in regions where human activities like deforestation and agriculture heavily influence the estimates of tropical carbon cycle potential. With insight into the potential phenological states, this framework may help inform sustainable land management practices in pan-tropical regions.

How to cite: Wu, M., Tagesson, T., Cai, Z., and Duan, Z.: Real-time Prediction of Global Tropical Deciduous Ecosystem Phenology with Deep Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13817, https://doi.org/10.5194/egusphere-egu25-13817, 2025.

EGU25-14658 | Posters on site | ESSI1.6

Comparison of the Minimum Bounding Rectangle and Minimum Circumscribed Ellipse of Rain Cells from TRMM 

HongKe Cai, YaQin Mao, XuanHao Zhu, YunFei Fu, and RenJun Zhou

Based on the TRMM dataset, this paper compares the applicability of the improved MCE (minimum circumscribed ellipse), MBR (minimum bounding rectangle), and DIA (direct indexing area) methods for rain cell fitting. These three methods can reflect the geometric characteristics of clouds and apply geometric parameters to estimate the real dimensions of rain cells. The MCE method shows a major advantage in identifying the circumference of rain cells. The circumference of rain cells identified by MCE in most samples is smaller than that identified by DIA and MBR, and more similar to the observed rain cells. The area of rain cells identified by MBR is relatively robust. For rain cells composed of many pixels (N > 20), the overall performance is better than that of MCE, but the contribution of MBR to the best identification results, which have the shortest circumference and the smallest area, is less than that of MCE. The DIA method is best suited to small rain cells with a circumference of less than 100 km and an area of less than 120 km2, but the overall performance is mediocre. The MCE method tends to achieve the highest success at any angle, whereas there are fewer “best identification” results from DIA or MBR and more of the worst ones in the along-track direction and cross-track direction. Through this comprehensive comparison, we conclude that MCE can obtain the best fitting results with the shortest circumference and the smallest area on behalf of the high filling effect for all sizes of rain cells.

How to cite: Cai, H., Mao, Y., Zhu, X., Fu, Y., and Zhou, R.: Comparison of the Minimum Bounding Rectangle and Minimum Circumscribed Ellipse of Rain Cells from TRMM, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14658, https://doi.org/10.5194/egusphere-egu25-14658, 2025.

EGU25-14783 | ECS | Posters on site | ESSI1.6

A Unified Model of Forecasting Ozone by Deep Learning 

Zhenze Liu

We propose a simple yet effective framework for real-time surface ozone forecasting using deep learning. The framework highlights three key modules: independent channel encoders, frequency information extraction, and fine-tuning, all of which consistently enhance model performance. This unified model is built well to autonomously capture different spatial and temporal patterns of ozone concentrations, with an averaged RMSE of 8 ppb for day 1 forecasting. The performance of day 4 forecasting is slightly lower. We find that chemistry becomes less important than meteorology over time, indicating their different roles in short-term and long-term forecasting. Most high ozone episodes can be simulated, though capturing extremely high ozone values remains a challenge. Observations from China are trained and tested to demonstrate our model.

How to cite: Liu, Z.: A Unified Model of Forecasting Ozone by Deep Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14783, https://doi.org/10.5194/egusphere-egu25-14783, 2025.

EGU25-15162 | ECS | Posters on site | ESSI1.6

Uncertainty Evaluation of Deep Learning Models Using an Artificial Rainfall 

Younghun Kim and Giha Lee

Accurate rainfall prediction is essential not only for water resource management but also for forecasting and mitigating the impacts of climate change-driven weather events such as floods and droughts. Due to the high spatiotemporal variability of complex meteorological phenomena like rainfall, effective prediction necessitates in high-quality data collection, model application, and uncertainty analysis. Unlike existing studies that focus primarily on developing deep learning models to improve rainfall prediction accuracy, this study evaluates the uncertainty of rainfall predictions using pre-existing deep learning models, U-Net and ConvLSTM, with artificially generated elliptical rainfall data. Artificial rainfall data were designed with four temporal patterns: constant, gradually increasing, gradually decreasing, and time-varying. These patterns were applied in horizontal, vertical, and diagonal movements to evaluate the models' ability to handle spatiotemporal complexity. The results indicate that both deep learning models exhibited spatial smoothing issues on rainfall predictions over time. However, the U-Net model demonstrated superior spatiotemporal performance compared to ConvLSTM. While this study focuses solely on deep learning models for rainfall prediction, future research will consider factors such as data complexity and loss functions to conduct a comprehensive evaluation of prediction uncertainty. This work is expected to contribute to the development of methodologies for rainfall modeling using deep learning approaches.

 

Funding: This research was supported by Disaster-Safety Platform Technology Development Program of the National Research Foundation of Korea(NRF) funded by the Ministry of Science and ICT. (No. 2022M3D7A1090338)

How to cite: Kim, Y. and Lee, G.: Uncertainty Evaluation of Deep Learning Models Using an Artificial Rainfall, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15162, https://doi.org/10.5194/egusphere-egu25-15162, 2025.

EGU25-15669 | ECS | Orals | ESSI1.6

Post-Processing Neural Weather Model Outputs for Tropical Cyclone Intensity Forecasts 

Milton Gomez, Tom Beucler, Alexis Berne, and Louis Poulain--Auzéau

Numerical Weather Prediction (NWP) models, which integrate physical equations forward in time, are the traditional tools for simulating atmospheric processes and forecasting weather in modern meteorology. With recent advancements in deep learning, Neural Weather Models (NeWMs) have emerged as competent medium-range NWP emulators with reported performances that compare favorably to state-of-the-art NWP models. However, they are commonly trained on reanalysis with limited spatial resolution (e.g., 0.25° horizontal grid spacing) and thus smooth out key features associated with a number of weather phenomena. For example, tropical cyclones—among the most impactful weather events due to their devastating effects on human activities—are challenging to forecast, as extrema like wind gusts, which serve as proxies for tropical cyclone intensity, are smoothed in deterministic forecasts at 0.25° resolution. To address this, we use our best global observational estimate of wind gusts and minimum sea level pressure to train models that post-process NeWM outputs and enable accurate and reliable forecasts of TC intensity. We present a tracking-independent post-processing algorithm and show that even naïve, linear models extract useful information from NeWM model outputs beyond what is present in the initial conditions used to roll out NeWM predictions. We explore how the NeWM’s spatial context may further improve the forecast through masking and convolutional architectures. Our post-processing framework thus presents a step towards democratization of tropical cyclone intensity forecasting, given the reduction in computational requirements for producing global weather forecasts with NeWMs compared to traditional NWP approaches and the algorithmic simplicity of the tracking-independent approach.

How to cite: Gomez, M., Beucler, T., Berne, A., and Poulain--Auzéau, L.: Post-Processing Neural Weather Model Outputs for Tropical Cyclone Intensity Forecasts, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15669, https://doi.org/10.5194/egusphere-egu25-15669, 2025.

EGU25-15811 | Orals | ESSI1.6

Climate Adaptation Digital Twin – building an operational climate information system to support decision-making 

Jenni Kontkanen, Mario Acosta, Pierre-Antoine Bretonnière, Miguel Castrillo, Paolo Davini, Francisco Doblas-Reyes, Barbara Früh, Jost von Hardenberg, Thomas Jung, Heikki Järvinen, Daniel Klocke, Nikolay Koldunov, Pekka Manninen, Sebastian Milinski, Jarmo Mäkelä, Devaraju Naraynappa, Suraj Polade, Irina Sandu, Outi Sievi-Korte, and Stephan Thober

The Climate Change Adaptation Digital Twin (Climate DT) is part of the Destination Earth (DestinE) initiative, developing Digital Twins of Earth to increase resilience against environmental changes. More specifically, Climate DT provides capabilities supporting climate change adaptation at regional and national levels at multi-decadal time scales. We present here an overview of Climate DT, highlighting the added value for the users and discussing the transition of the system towards the operations.  

The development of Climate DT has started in Phase 1 of DestinE, during which the first prototype of the new climate information system has been developed. A key innovation of Phase 1 was the introduction of a generic state vector (GSV), which is evolved by the Earth system models (ESMs) and streamed to applications from climate adaptation impact sectors.  This has created a basis for a pioneering climate information system that enables (i) provision of global climate information at an unprecedented granularity, (ii) scaling the system across a number applications that have access to all the data they need, (iii) user-centric approach with new ways of co-design and opportunities for enhancing interactivity. In Phase 2, which started in May 2024, our focus is on operationalizing Climate DT to deliver high-quality climate and impact-sector information regularly while incorporating new interactive features.

The operational model of the Climate DT is built around three storm- and eddy resolving ESMs; ICON, IFS-NEMO and IFS-FESOM. The operational framework utilizes a DevOps-like cycle, including three set-ups: d-suite for development, e-suite for testing the operational set-up and o-suite for operating the system. The o-suite simulations will provide data covering both past (1990-2020) and future periods (2020-2050) with a 5 km global grid. Additionally, capabilities for special simulations are developed, including story-line simulations for future periods of extremes as well as what-if scenario simulations enabling a new level of interactivity.

The added value of Climate DT to users is demonstrated through four impact sector applications. These applications operate on the streamed GSV as part of the operational framework, and they are improved in co-design with key users. The impact sector applications cover societally relevant climate change adaptation domains, including wind energy management, disaster risk management (with regards to wildfires and floods), as well as agriculture and water management. Climate DT output, including high-resolution climate simulations, storyline simulations, user-relevant indicators and impact assessments are made available to users via DestinE Service Portal.

How to cite: Kontkanen, J., Acosta, M., Bretonnière, P.-A., Castrillo, M., Davini, P., Doblas-Reyes, F., Früh, B., von Hardenberg, J., Jung, T., Järvinen, H., Klocke, D., Koldunov, N., Manninen, P., Milinski, S., Mäkelä, J., Naraynappa, D., Polade, S., Sandu, I., Sievi-Korte, O., and Thober, S.: Climate Adaptation Digital Twin – building an operational climate information system to support decision-making, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15811, https://doi.org/10.5194/egusphere-egu25-15811, 2025.

EGU25-15912 | Orals | ESSI1.6

ClimateBenchPress: A Benchmark for Compression of Climate Data 

Tim Reichelt, Juniper Tyree, Milan Kloewer, Peter Dueben, Bryan Lawrence, Dorit Hammerling, Alisson Baker, Sara Faghih-Naini, and Philip Stier

The rapid growth of weather and climate datasets is increasing the pressure on data centres and hinders scientific analysis and data distribution. For example, kilometre-scale weather and climate models can generate 20 gigabytes of data per second when run operationally, making it generally infeasible to store all output unless advanced compression is applied. 

To address this challenge, novel lossy compression techniques, including recently so-called neural compressors which learn smaller representations of climate data, have been proposed with compression factors beyond 100x. However, if applied without care, lossy compression can remove valuable information from a dataset for often unknown downstream applications. It is therefore important to validate that the compression process does not alter scientific conclusions drawn from the data. Whether the compression error is tolerable is often easier to assess for domain experts and rarely well defined. 

Here, we address this challenge by presenting a benchmark suite for lossy compression of climate data (atmosphere, ocean, and land). We are defining data sets that can be used to train neural compressors as well as corresponding evaluation methods. Compressors have to pass a set of tests for each data set while compressing into the smallest file size possible at a reasonable (de)compression speed. To ensure evaluation on a diverse set of inputs, the benchmark covers climate variables following various statistical distributions at medium to very high resolution in time (hourly to yearly) and space (~1 km to 150 km). Evaluation tests are for single and multi-variable compression of gridded data with stable or changing statistics, random data access or large archives, in medium to very large datasets.

To provide references towards what compression levels can be achieved with current state of the art lossy compressors, we also evaluate a set of baseline compressors (SZ3, ZFP, Real Information) on our benchmark tasks. The benchmark is a quality check for new compressors towards a standardization of climate data compression, aiming to make compressors with high compression factors safe to use and widely supported.

How to cite: Reichelt, T., Tyree, J., Kloewer, M., Dueben, P., Lawrence, B., Hammerling, D., Baker, A., Faghih-Naini, S., and Stier, P.: ClimateBenchPress: A Benchmark for Compression of Climate Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15912, https://doi.org/10.5194/egusphere-egu25-15912, 2025.

EGU25-16981 | ECS | Posters on site | ESSI1.6

Flow matching for in situ, spatially consistent weather forecast downscaling 

David Landry, Anastase Charantonis, and Claire Monteleoni

Weather forecast downscaling, the problem of recovering accurate local predictions given a lower resolution forecast,  is commonly used in operational NWP pipelines. Its purpose is to recover some of the sub-grid processes that could not be represented by the underlying numerical model due to a limited resolution. This misrepresentation provokes statistical mismatches between the observation data gathered from stations and the nearest grid point in the numerical simulation.

Using a downscaling model typically requires making a compromise between spatial consistency and statistical calibration. Traditionally, these models are trained to target a traditional verification metric. Consequently, they suffer from the double penalty issue and fail to correctly model spatial correlation structures by becoming overly smooth. This is detrimental to downstream modeling tasks such as power grid management, which require a good assessment of spatially-correlated phenomena. 

Recently, the finer details of the atmospheric state have successfully been recovered using generative models such as denoising diffusion [2-4]. We propose a similar strategy for in situ downscaling by introducing a flow matching [1] model for that task. A cross-attention transformer [5] backbone allows us to build an internal representation for the gridded numerical forecast as well as the in situ downscaled forecast. 

Our model avoids the numerical instability and mode collapse issues related to Generative Adversarial Networks. It produces well-calibrated forecasts that better represent the spatial correlations between the stations when compared to non-generative alternatives. Our model makes no assumptions about the underlying forecast, and thus can be thought of in two ways. It can be considered a hybrid NWP/AI model, where we first run a numerical simulation and then downscale it. It can also be considered a supplementary forecasting product in a full machine learning pipeline.

Using our flow matching weather forecast downscaling model, we run experiments on the EUPPBench post-processing dataset to predict surface temperature and wind speed. Particular care is given to evaluating the model, where we assess both the marginal performance (via the CRPS, reliability histogram, and spread-error ratio) and the joint performance (via the Energy Score, local Variogram Score and forecast spatial frequency content). The accurate representation of extreme events is evaluated using Brier scores. Further experiments discuss the pitfalls of fitting the Energy Score directly without a generative model.

 

[1] Lipman, Y. et al. (2023) ‘Flow Matching for Generative Modeling’. arXiv. Available at: https://doi.org/10.48550/arXiv.2210.02747.

[2] Couairon, G. et al. (2024) ‘ArchesWeather & ArchesWeatherGen: a deterministic and generative model for efficient ML weather forecasting’. arXiv. Available at: https://doi.org/10.48550/arXiv.2412.12971.

[3] Price, I. et al. (2023) ‘GenCast: Diffusion-based ensemble forecasting for medium-range weather’. arXiv. Available at: https://doi.org/10.48550/arXiv.2312.15796.

[4] Lang, S. and Chantry, M. (2024) ‘Enter the ensembles’, AIFS Blog, 21 June. Available at: https://www.ecmwf.int/en/about/media-centre/aifs-blog/2024/enter-ensembles (Accessed: 15 January 2025).

[5] Vaswani, A. et al. (2017) ‘Attention is All you Need’, in Advances in Neural Information Processing Systems. Curran Associates, Inc. Available at: https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html (Accessed: 17 May 2022).

How to cite: Landry, D., Charantonis, A., and Monteleoni, C.: Flow matching for in situ, spatially consistent weather forecast downscaling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16981, https://doi.org/10.5194/egusphere-egu25-16981, 2025.

EGU25-17080 | ECS | Orals | ESSI1.6

Spatial Gap Filling in a Geostationary Land-Surface Temperature Product with a Masked Autoencoder 

Matthias Karlbauer, Florian M. Hellwig, Thomas Jagdhuber, and Martin V. Butz

With the increasing availability and demand of remote sensing data from Earth observation satellites, the accuracy of weather prediction models can be improved substantially. Satellite products, such as Land-Surface Temperature (LST), however, suffer from missing data, either caused by clouds that cover the ground, by missing spatial coverage of the mission, or by outages of the sensors. Such spatial data gaps in LST products impose strict limitations when aiming to process the data further with, e.g., numerical weather prediction models assuming spatial continuity with gapless input data. We therefore propose a gap-filling algorithm based on a masked autoencoder that only receives a small percentage from a 32x32 LST snapshot and learns to reconstruct the missing patches. We use the spatial domain defined by the Land Atmosphere Feedback Initiative (LAFI) over central Europe and operate on geostationary LST data from the Copernicus Global Land Service in June 2023 at 5 km resolution. Our approach indicates considerable potential when filling spatial gaps in LST products, however, we emphasize one critical aspect. The LST estimates below clouds cannot be expected to be realistic and would require a sophisticated atmospheric correction. To mitigate this limitation, we aim to incorporate microwave data in future that penetrates clouds and therefore could help to estimate LST below clouds. In its current formulation, our algorithm can be used to fill gaps in LST products as if there were no clouds. We will show the potential and limitations of the autoencoder-based gap-filling algorithm for several showcases across Europe. 

How to cite: Karlbauer, M., Hellwig, F. M., Jagdhuber, T., and Butz, M. V.: Spatial Gap Filling in a Geostationary Land-Surface Temperature Product with a Masked Autoencoder, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17080, https://doi.org/10.5194/egusphere-egu25-17080, 2025.

The Children’s Climate Risk Index (CCRI) was first released in 2021, providing a comprehensive, global view of children’s exposure and vulnerability to the impacts of climate change. The CCRI is a composite index that aims to rank countries where children are exposed to climate and environmental hazards. The CCRI 2.0 builds on the previous index by integrating two pillars; Pillar 1 focusing on climate hazards and Pillar 2 on inherent vulnerabilities to WASH, health, education and other relevant dimensions. 

 

We highlight our contributions to CCRI 2.0, using a cluster methodology for quantifying children’s exposure to climate risks including riverine and coastal flooding, tropical storms, heatwaves, and drought. Using unsupervised learning, we allow for a data-driven approach to provide an interpretation of the ranking of children’s exposure to climate risks on a global scale between countries, as well as at the sub-national and local levels. It complements the previous method of constructing the synthesized index, which involved calculating the simple average of multiple indicators. We further discuss our techniques in tackling the challenges of multisource data processing, analysis, and visualization of geospatial data for user insight.

How to cite: Kim, D. and Doerksen, K.: Quantifying Children’s Exposure to Climate Risks using Unsupervised Learning with Multi-Source Geospatial Datasets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17875, https://doi.org/10.5194/egusphere-egu25-17875, 2025.

EGU25-18447 | ECS | Orals | ESSI1.6

ArchesClimate: Ensemble Generation for Decadal Prediction using Flow Matching 

Graham Clyne, Guillaume Couairon, Juliette Mignot, Guillaume Gastineau, Anastase Charantonis, and Claire Monteleoni

Sampling from climate models to generate ensembles of predictions is computationally expensive (Hawkins et al., 2015). Climate model ensembles are used to understand probabilities of climatic events and identify internal variability in climate models. In the short term, model uncertainty and inter-annual variability dominate uncertainty in climate predictions (Smith et al., 2019). A typical approach to address these uncertainties is to use large ensembles of non-learned, physical numerical global circulation models (GCM) (Eade et al., 2014). These ensembles allow for statistical analysis of distributions and determination of internal variability in the climate model.

Our approach demonstrates that we can efficiently learn to emulate a GCM. We use ensembles generated by the IPSL submission to the Decadal Climate Prediction Project (DCPP). The dataset ranges from 1960-2016 and produces 10-member, 10-year forecast ensembles for each year. On this dataset, we train a modified version of ArchesWeatherGen, a Swin Transformer based on PanguWeather that can be used in a generative way using flow matching (Couairon et al. 2024). The model was modified to predict additional climatic variables (e.g. air temperature, specific humidity, ocean potential temperature at depth, sea surface temperature, sea level pressure) at a monthly temporal resolution. Once trained, the model probabilistically generates ensemble members rapidly which can be auto-regressively rolled out. We show that they are physically reliable via evaluation methods that assess physical processes derived from the variables represented in the machine learning model, such as by evaluating it on El Niño/La Niña events. This model demonstrates that machine learning can enhance climate models by expanding ensemble sizes to improve our understanding of climatic processes. We aim to output physically realizable month-to-month trajectories to estimate future climate and its uncertainties across various domains, including land, ocean, and atmospheric processes.



Couairon, G., Singh, R., Charantonis, A., Lessig, C., & Monteleoni, C. (2024). ArchesWeather & ArchesWeatherGen: a deterministic and generative model for efficient ML weather forecasting. arXiv preprint arXiv:2412.12971.

Eade, Rosie, Doug Smith, Adam Scaife, Emily Wallace, Nick Dunstone, Leon Hermanson, et Niall Robinson. « Do Seasonal-to-Decadal Climate Predictions Underestimate the Predictability of the Real World? » Geophysical Research Letters 41, no 15 (2014): 5620‑28. https://doi.org/10.1002/2014GL061146.

Hawkins, Ed, Robin S. Smith, Jonathan M. Gregory, et David A. Stainforth. « Irreducible Uncertainty in Near-Term Climate Projections ». Climate Dynamics 46, no 11 (1 juin 2016): 3807‑19. https://doi.org/10.1007/s00382-015-2806-8.

Smith, D. M., R. Eade, A. A. Scaife, L.-P. Caron, G. Danabasoglu, T. M. DelSole, T. Delworth, et al. « Robust Skill of Decadal Climate Predictions ». Npj Climate and Atmospheric Science 2, no 1 (17 mai 2019): 1‑10. https://doi.org/10.1038/s41612-019-0071-y.

How to cite: Clyne, G., Couairon, G., Mignot, J., Gastineau, G., Charantonis, A., and Monteleoni, C.: ArchesClimate: Ensemble Generation for Decadal Prediction using Flow Matching, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18447, https://doi.org/10.5194/egusphere-egu25-18447, 2025.

EGU25-19021 | ECS | Posters on site | ESSI1.6

RAINA - High-resolution nowcasting of precipitation and wind extremes with a foundation model for the atmosphere 

Erik Pavel, Michael Langguth, Martin G. Schultz, Christian Lessig, Stefanie Hollborn, Jan Keller, Roland Potthast, Britta Seegebrecht, Sabrina Wahl, Juergen Gall, Anas Allaham, Mohamad Hakam Shams Eddin, and Ilaria Luise

Data-driven weather prediction models based on deep learning have been on the rise for several years and have outperformed traditional physics-based numerical models in various benchmark forecasting scores. However, a significant challenge remains: accurately predicting extreme events on a local scale, such as thunderstorms and wind gusts. Previous models struggle in this area, as they were primarily developed for medium-range forecasting and operate at relatively coarse spatio-temporal resolutions. However, the capability of weather models to predict extreme events at a local level is essential for preventing severe consequences for communities, ecosystems, and the financial and material losses they entail. Recently, task-agnostic foundation models, trained on extensive and diverse datasets using self-supervised methods, have demonstrated remarkable skill and robustness, especially in their ability to generalize to rare extreme events. 

The RAINA project aims to develop a foundation model for the atmosphere, with an emphasis on delivering reliable, high-resolution forecasts of extreme wind and precipitation events. In partnership with the EU Horizon-funded WeatherGenerator project, which aims to create advanced digital twins for Destination Earth, RAINA will extend the pioneering AtmoRep model (Lessig et al., 2023) by employing a multi-modal learning approach.
The foundation model seeks to develop a comprehensive, statistically robust, and multi-scale understanding of atmospheric dynamics by incorporating a wide range of meteorological datasets from both models and observations. Innovative deep learning methods, including diffusion models and test-time adaptation, will be investigated to facilitate short-range forecasts of temperature, wind, and precipitation at kilometer-scale resolution over Germany.

In a first demonstrator, short-range forecasts are generated using the AtmoRep model and subsequently refined with the CorrDiff downscaling approach (Mardani et al., 2024) that combines a generative diffusion model with a residual approach. This two-step strategy delivers high-resolution forecasts with a maximum lead time of six hours while disentangling uncertainties inherent in the forecasting and downscaling processes, a separation that can enhance training quality when properly applied. By using ERA5 and COSMO REA2 reanalysis data, the approach enhances the precision of high-resolution forecasts over Germany. 
Initial results from the first demonstrator will be presented in a poster, along with the overall timeline and key milestones of the RAINA project.

How to cite: Pavel, E., Langguth, M., Schultz, M. G., Lessig, C., Hollborn, S., Keller, J., Potthast, R., Seegebrecht, B., Wahl, S., Gall, J., Allaham, A., Shams Eddin, M. H., and Luise, I.: RAINA - High-resolution nowcasting of precipitation and wind extremes with a foundation model for the atmosphere, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19021, https://doi.org/10.5194/egusphere-egu25-19021, 2025.

EGU25-19238 | Orals | ESSI1.6

Destination Earth Data Lake user story on Danube Delta water reservoir  

Charlotte Delmas, Aubin Lambaré, and Arnaud Le Carvennec

The Danube Delta, the second-largest delta in Europe and a critical economic waterway, represents a dynamic yet fragile ecosystem requiring effective preservation strategies. Monitoring water reservoirs is crucial for both ecological sustainability and socio-economic management. The increasing availability of diverse datasets from multiple sources offers new opportunities to enhance real-time observation and forecasting efforts. 

Implemented by EUMETSAT, the Destination Earth Data Lake (DEDL) provides seamless access to these datasets and integrates high-performance computing for complex scientific modeling. Its edge services provide efficient, scalable data processing, empowering researchers to analyze environmental phenomena with speed and precision. 

Leveraging DEDL services enables to consolidate key hydrological datasets offering important features to monitor the ecosystem’s health state: 

  • Daily live in situ data: Real-time measurements of water level, temperature, and discharge from DanubeHIS ground stations along the river and its delta, has been retrieved via the DEDL Harmonized Data Access (HDA). 
  • Outputs from existing scientific algorithms: The integration and evolution of the Surfwater algorithm within the DEDL environment allows leveraging Earth Observation data (Landsat 8/9) to detect water bodies in the area. This makes it possible to generate time series of surface areas, volumes, and fill rates of water bodies within the region. 
  • Hourly radar data: Rainfall rates are computed using OPERA radar observations on the European Weather Cloud instances. 
  • Precipitation forecasts: Predictive data from ECMWF (Destination Earth Digital Twin Outputs), accessed via HDA, are leveraged to provide valuable forecasting insights. 

The outcome of those algorithms and analysis are provided live through a dashboard. By enabling cross-referencing of diverse data streams, it allows stakeholders to obtain a complete view of the Danube Delta’s environmental conditions, supporting informed decision-making for ecosystem preservation. 

Leveraging advanced geoscience tools, this integrated approach highlights the transformative power of modern data platforms in tackling global environmental challenges. 

How to cite: Delmas, C., Lambaré, A., and Le Carvennec, A.: Destination Earth Data Lake user story on Danube Delta water reservoir , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19238, https://doi.org/10.5194/egusphere-egu25-19238, 2025.

EGU25-19942 | Orals | ESSI1.6

A Spatial Multi-Grid Neural Operator-Transformer Mdoel for High-Resolution Climate Modeling 

Maximilian Witte, Johannes Meuer, and Kadow Christopher

High-resolution machine learning faces the challenge of balancing local computation with large physical context windows. GPU memory limitations and the slow training process when distributing the model across multiple GPUs further complicate this task. 

We present a transformer model for high-resolution climate-related tasks that uses neural operators within a multi-grid architecture. This approach allows resolution independence, large physical context windows, and the handling of discontinuities such as coastlines.

The model is spatially flexible, supporting both regional and global training schemes. It is also independent of the number of input variables, allowing training to be scaled to large numbers of input variables.

We demonstrate the ability of the model to scale, both spatially and in terms of variables. The model forms the foundation of an approach to learn from the rich and diverse climate data available, enabling high-resolution downscaling, infilling, and predictions.

How to cite: Witte, M., Meuer, J., and Christopher, K.: A Spatial Multi-Grid Neural Operator-Transformer Mdoel for High-Resolution Climate Modeling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19942, https://doi.org/10.5194/egusphere-egu25-19942, 2025.

EGU25-20187 | Orals | ESSI1.6

TWINE: TWInning capability for the Natural Environment 

Joana Mendes, Edward Pope, Zorica Jones, Andrew Cottrell, Michael Eastman, Joshua Wiggs, Hannah Findley, Angela Heard, Paul Hallett, Emilie Vanvyve, Remy Vandaele, Hywell Williams, Milto Miltiadou, Finley Gibson, Kirstine Dale, Anna Angus-Smyth, Simon Gardner, and Sam Tailby

Digital twins are an exciting and rapidly developing research area, with the potential to provide a step change in the way we understand our evolving environment and its impact on sensitive systems.

The TWInning Capability for the Natural Environment (TWINE) programme is being co-delivered by the Met Office and Natural Environment Research Council (NERC) to explore the potential of this technology for transforming environmental science and across priority areas including climate change adaptation and mitigation, biodiversity and ecosystems, and natural hazards and their mitigation.

Through the TWINE programme, NERC and the Met Office have funded six digital twin pilot projects across a range of applications, including harmful algal blooms, flooding and coastal overtopping, optimising data collected by ocean gliders and aircraft, and multi-objective land-use decisions.

We will introduce the TWINE programme, giving a brief overview of the projects which are advancing our understanding of how we can harness the potential of digital twinning technology. These include cross-sector challenges such as: risk to natural resources, Net Zero targets, and addressing the science-to-policy lag.

How to cite: Mendes, J., Pope, E., Jones, Z., Cottrell, A., Eastman, M., Wiggs, J., Findley, H., Heard, A., Hallett, P., Vanvyve, E., Vandaele, R., Williams, H., Miltiadou, M., Gibson, F., Dale, K., Angus-Smyth, A., Gardner, S., and Tailby, S.: TWINE: TWInning capability for the Natural Environment, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20187, https://doi.org/10.5194/egusphere-egu25-20187, 2025.

EGU25-1534 | ECS | Posters on site | ESSI1.9

Features from Multispectral Drone Data: Curating, training and distributing Transformers for all 

Jurrian Doornbos and Önder Babur

Uncrewed aerial vehicles (UAVs) have been identified as an important tool supporting more detailed remote sensing applications compared to satellite-based platforms, from agriculture to forest monitoring and sensing mountains. The insights the UAV can offer, due to its flexibility, high precision and sensor variety, are far beyond the previous approaches to measure the health of forests, yield in field crops, and even rockfall risk. The flexibility however also poses a problem, flight conditions, sensor types, flight height and angles all affect the generalization of developed approaches using UAVs. These supervised approaches also rely on large amounts of human-labelled datasets. A pathway to reduce high-label requirements is to utilize unsupervised training with Vision Transformers (ViTs). Pretrained Vision Transformers on large datasets generalize well to unseen data, with only supervised few samples required to specify the application. However, these models are often trained on massive web-scraped RGB datasets. Furthermore, RGB-ViTs miss the infrared domain to handle crucial vegetation information. Finally, UAV imagery is exclusively from the aerial perspective, this is missing in existing pretraining datasets.

We present an openly available, pre-trained Vision Transformer specifically for UAV multispectral imagery across various domains. Furthermore, various downstream applications such as canopy height modelling and semantic segmentation are evaluated and compared against RGB baselines. The main contribution is the openly available training dataset, and the pre-trained models, with recipes for finetuning a task-specific head.

The dataset is built around multispectral image contributions from the ICAERUS Drone Data Analytics Library and an additional database search on Zenodo and Data in Brief (Table 1.). This is followed by a quality check after, including radiometric calibration, and spectral alignment. Furthermore, all data is quantized into 16-bit float and sliced into smaller 224x224 chips with four channels (Green, Red, Red Edge and NIR).  A summary of included datasets is presented in Table 1. DINOv2-s and DINOv2-b were chosen for the architecture as there is much available documentation and provide a state-of-the-art vision foundation model. The training was done in minibatches of size 32, for 6 days on two V100 GPUs.

Early experiments suggest that the pre-trained models outperform existing DINOv2-s and DINOv2-b pre-trained foundation models in both the clarity of the features, as well as tuned on UAV-specific tasks (canopy height modelling, and semantic segmentation).

Table 1. Included datasets for pretraining, total size on disk is 399GB

How to cite: Doornbos, J. and Babur, Ö.: Features from Multispectral Drone Data: Curating, training and distributing Transformers for all, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1534, https://doi.org/10.5194/egusphere-egu25-1534, 2025.

EGU25-3302 | Orals | ESSI1.9

An Explainable AI (XAI) Benchmark for Geospatial Foundation Models 

Hamed Alemohammad, Sam Khallaghi, Denys Godwin, Rufai Balogun, Sujit Roy, and Rahul Ramachandran

There is a significant growth in development and utilization of foundation models for geospatial applications. These models are trained on large scale unlabeled data and commonly evaluated on downstream tasks using labeled datasets. While this approach provides a platform to assess the performance of the model for specific downstream tasks, there has been limited effort to quantify the characteristics of the foundation model after pre-training. 

Explainable AI (XAI) approaches aim to increase the accuracy and transparency of AI models and to make their results interpretable. In the case of geospatial foundation models, it is essential to assess if the model learns the spectral, spatial and temporal properties of geospatial data, and how this learning impacts the accuracy of model predictions. 

To this end, we introduce a new global XAI benchmark for geospatial foundation models using multispectral remote sensing imagery. This benchmark contains separate tasks that allows the user to test a foundation model’s properties in the embedding space, and demonstrate whether the model has learned spectral, spatial and temporal features. The spectral task consists of a set of chips with homogeneous spatial patterns from all major land cover classes. The spatial task consists of the same data used for spectral taks but regular spatial patterns are replaced with heterogeneous features representative of their true distribution. Finally, the temporal task includes a set of chips with time series imagery of pre- and post-event for disturbances such as wildfire and flood. 

In this presentation, we will demonstrate the results of using this benchmark to evaluate the properties of multiple geospatial foundation models.

How to cite: Alemohammad, H., Khallaghi, S., Godwin, D., Balogun, R., Roy, S., and Ramachandran, R.: An Explainable AI (XAI) Benchmark for Geospatial Foundation Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3302, https://doi.org/10.5194/egusphere-egu25-3302, 2025.

EGU25-3328 | Orals | ESSI1.9

AI Foundation Models for Science: Current Initiatives, Workflow, and Future Roadmap 

Rahul Ramachandran, Tsengdar Lee, and Kevin Murphy

NASA has collected—and continues to amass—petabytes of scientific data, ranging from the vastness of galaxies to the intricacies of cellular biology. These ever-expanding datasets provide unparalleled opportunities for discovery but pose significant challenges for managing data and extracting meaningful insights. Artificial intelligence (AI) and machine learning (ML) are emerging as transformative tools for addressing these issues. However, state-of-the-art deep neural networks often require large volumes of labeled training data, which are costly and time-intensive to generate. AI foundation models (FMs) offer a promising alternative by leveraging self-supervised learning to identify patterns within data. These FMs enable diverse applications with reduced dependence on compute resources and labeled datasets.

NASA’s Office of the Chief Science Data Officer has formulated a "5+1" strategy to develop AI foundation models for science. This strategy emphasizes creating foundation models (FMs) pre-trained using flagship datasets from each of NASA’s science divisions while building a science-specific language model to support cross-divisional applications. Key achievements include the release of INDUS, an encoder language model trained on scientific publications and technical documents; two versions of the Prithvi Geospatial model for environmental monitoring applications; and the Prithvi Weather and Climate model, designed to reconstruct atmospheric states from incomplete data and forecast future states. Additionally, a heliophysics foundation model for space weather applications is under development and is scheduled for release by mid-2025.

To encourage NASA’s research and application communities to use these FMs in their work and to support NASA’s new Earth Science to Action Strategy, the Earth Science Division has developed additional research and application solicitations to further enhance these FMs and to build applications and tools leveraging these FMs. These announcements are available in NASA’s Research Opportunities in Space and Earth Science (ROSES 2025).

NASA has forged strategic partnerships with private sector organizations, academia, and other entities grounded in open science principles to build these models. Each model is designed around a specific set of scientific use cases to ensure relevance and practical impact. All models and associated use case notebooks are shared openly. 

This presentation will provide an overview of the foundation models released to date, the workflows used in their design and development, and the roadmap for future models. It will also highlight upcoming workshops aimed at equipping the broader scientific community to effectively integrate these models into their research.

How to cite: Ramachandran, R., Lee, T., and Murphy, K.: AI Foundation Models for Science: Current Initiatives, Workflow, and Future Roadmap, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3328, https://doi.org/10.5194/egusphere-egu25-3328, 2025.

EGU25-7220 | ECS | Orals | ESSI1.9

Ground-level land surface classification and thermal analysis using foundation models 

Konlavach Mengsuwan and Masahiro Ryo

Foundation models have shown substantial potential in enhancing Earth observation by providing high accuracy while minimizing the need for manual annotation. However, the combined application of multiple foundation models for processing ground-truthing data remains largely underexplored. This study introduces a new approach for ground-level land use classification and ground-truthing by integrating a vision foundation model, the Segment Anything Model (SAM), with a general-purpose large language model (GPT-4o). Using high-resolution thermal and RGB imagery captured from human-eye height with a handheld camera, the proposed method generates object-level land use classifications and surface temperature profiles. Data collection was conducted in the Lusatia region of Germany, covering diverse land use types. SAM was utilized to segment complex landscape structures into meaningful elements such as roads, water bodies, and trees, followed by GPT-4o, which classified these segments into custom-defined land use categories. At a broad level (7 classification types), the workflow achieved approximately 80% accuracy, with high F1 scores for categories such as Road (0.89), Vegetation (0.82), and Built Structure (0.81). At a finer level (28 classification types), the method attained around 64% accuracy, effectively classifying detailed sub-classes such as Asphalt-Concrete Road (F1 = 0.85), Brick Road (F1 = 0.86), Tree (F1 = 0.74), and Arable Land (F1 = 0.68). By overlaying thermal imagery with classified segments, the method revealed distinct microclimatic patterns across land use types, with agricultural land showing the lowest surface temperatures (p < 0.001). The proposed workflow underscores the potential of combining SAM and GPT-4o to deliver robust ground-truthing data using portable cameras, advancing AI-enabled environmental monitoring.

How to cite: Mengsuwan, K. and Ryo, M.: Ground-level land surface classification and thermal analysis using foundation models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7220, https://doi.org/10.5194/egusphere-egu25-7220, 2025.

EGU25-8272 | Posters on site | ESSI1.9

Unsupervised Change Detection Using Sentinel-1 and Sentinel-2 Imagery with the Clay Foundation Model: A Case Study of Flood-Affected Areas in Valencia Spain 

Mohanad Albughdadi, Marica Antonacci, Vasileios Baousis, Federico Fornari, Tolga Kaprol, and Claudio Pisa

The detection of environmental changes caused by natural disasters is critical for rapid response and effective management. In this study, we present a methodology for unsupervised change detection that leverges optical Sentinel-2 [1] and Synthetic Aperture Radar (SAR) Sentinel-1 [2] accessed through public SpatioTemporal Asset Catalogs (STAC) [3] along with Earth Observation (EO) foundation model, namely, Clay [4]. The analysis was conducted independently for each dataset to capitalize on the unique properties of these satellite sensors. Sentinel-1 offers robust surface texture sensitivity with its all-weather, day-and-night imaging capability, while Sentinel-2 provides detailed spectral and spatial information critical for vegetation and land-use analysis.

Clay foundation model, a large-scale pretrained Vision Transformer trained on EO data from various missions (Sentinel-1, Sentinel-2, Landsat, Planet, NAIP, LINZ, and MODIS), was used to extract spatially and spectrally rich embedding from Sentinel-1 and Sentinel-2 images. The model takes as an input the satellite imagery along with information about location and time and outputs mathematical representations of a given area at a certain time on Earth’s surface. The images were fed to the model as patches of size 256×256 along with the timestap of the scene, the spatial location and other metadata of the input image to estimate the embeddings that can be rearranged to be of size (1024×32×32). These embeddings were then analyzed using pixel-wise distance metrics to quantify changes between pre- and post-even imagery and the resulting distance image was then spatially interpolated to the size of the input image.

The approach was validated on satellite imagery of the Valencia region in Spain, an area significantly impacted by recent flooding on the 29th October 2024. For Sentinel-1, the method effectively highlighted surface water changes and structure affected by the floods in two scenes acquired on the 7th October and the 12th November 2024, while Sentinel-2 data captured variations in vegetation areas that was impacted by the floods using two scenes acquired on the 1st October and the 10 November 2024. By analyzing the datasets independently, this framework demonstrates the complementary insights offered by radar and optical imagery in assessing disaster impacts.

This study highlights the potential of leveraging open satellite data available via STAC catalogs and EO foundation models for unsupervised change detection in disaster monitoring, enabling rapid response without relying on specialized models tailored to specific regions. Unlike traditional approaches that require retraining for new areas due to geographical variability, this methodology is both scalable and adaptable, providing a generalizable framework for environmental monitoring, disaster response, and resilience planning. The results emphasize the value of integrating multi-sensor satellite imagery to enhance understanding of disaster impacts, facilitating more informed and timely decision-making.

References:

[1] https://earth-search.aws.element84.com/v1

[2] https://planetarycomputer.microsoft.com/api/stac/v1

[3] https://stacspec.org/en

[4] https://clay-foundation.github.io/model/index.html

How to cite: Albughdadi, M., Antonacci, M., Baousis, V., Fornari, F., Kaprol, T., and Pisa, C.: Unsupervised Change Detection Using Sentinel-1 and Sentinel-2 Imagery with the Clay Foundation Model: A Case Study of Flood-Affected Areas in Valencia Spain, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8272, https://doi.org/10.5194/egusphere-egu25-8272, 2025.

EGU25-8460 | ECS | Orals | ESSI1.9

SeCo-Eco: Global multiband seasonal pre-training dataset and self-supervised model for ecological applications 

Elena Plekhanova, Damien Robert, Johannes Dollinger, Philipp Brun, Jan Dirk Wegner, and Niklaus E. Zimmermann

With the biodiversity crisis and land use intensification, macroecological questions related to biodiversity assessment and conservation are becoming increasingly pressing. Such questions require global datasets such as satellite imagery. Traditional methods using satellite data rely heavily on supervised learning and annotated datasets, which are limited and difficult to generalize across geographical scales. In recent years, self-supervised learning (SSL) has opened the doors to learning expressive representations of massive datasets without annotations,  thus revolutionizing the analysis of remote sensing imagery. However, currently available datasets for pre-training such models have a skewed geographical distribution, focusing on cities and agricultural areas while failing to adequately represent regions of high ecological interest, such as rainforests or polar latitudes.

We propose a new Sentinel 2A (10m resolution) multiband dataset, globally distributed on a regular grid across the landmass(250k locations). At each location, the dataset captures four different seasons determined based on the local EVI-curve and includes NDVI index, which is widely used in ecological applications. Our temporal sampling is specifically designed to align with plant phenology rather than ad-hoc calendar dates. We use this data to pre-train Momentum Contrast and Seasonal Contrast SSL models that have shown similar performance on commonly-used benchmarks and advanced performance on macroecological downstream tasks, such as species distribution modelling. We anticipate that the dataset and model will be valuable for macroecological applications, such as deep species distribution modeling or large-scale biodiversity assessments.

How to cite: Plekhanova, E., Robert, D., Dollinger, J., Brun, P., Wegner, J. D., and Zimmermann, N. E.: SeCo-Eco: Global multiband seasonal pre-training dataset and self-supervised model for ecological applications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8460, https://doi.org/10.5194/egusphere-egu25-8460, 2025.

EGU25-9632 | Orals | ESSI1.9

Developing Global Embeddings from Sentinel-1 and Sentinel-2 Data to Enhance Earth Observation Analysis 

Marcin Kluczek, Mikolaj Czerkawski, and Jędrzej S. Bojanowski

The rapid growth of Earth Observation (EO) data from the Copernicus programme presents new opportunities for applying artificial intelligence (AI) and machine learning (ML) techniques. This work introduces a global embedding framework designed to improve the analysis of large EO datasets from Sentinel-1 and Sentinel-2 imagery. Following the Major TOM standard, we process over 8 million images, encompassing 9.368 trillion pixels of raw data, to generate more than 170 million embeddings from 62 terabytes of satellite data.

To enable that, a set of commonly used vision models (from both general and remote sensing domain, such as SigLIP, DINOv2, SSL4EO, DeCUR and MMEarth) are employed to derive efficient embedding representations of the input data. These embeddings support various applications, including text-to-image and image-to-image retrieval, as well as zero-shot classification, allowing for more effective integration of EO data into AI pipelines and providing valuable insights into global phenomena.

The current approach efficiently processes large-scale data, built on the CloudFerro cloud platform, with experiments demonstrating its usefulness in Earth Observation analysis. The results highlight the system’s reliability across different applications, emphasizing its potential to support data-driven decision-making on a global scale. This study also discusses key strategies for scalable cloud computing, GPU optimization, and multithreaded CPU processing to handle large volumes of EO data efficiently. 

How to cite: Kluczek, M., Czerkawski, M., and Bojanowski, J. S.: Developing Global Embeddings from Sentinel-1 and Sentinel-2 Data to Enhance Earth Observation Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9632, https://doi.org/10.5194/egusphere-egu25-9632, 2025.

EGU25-12338 | Posters on site | ESSI1.9

Aurora Evolution Model During the Substorm Expansion Phase using Machine Learning based method  

Yang Lu, Jianan Jiang, Jia Zhong, and Ziming Zou

 Aurora is an important manifestation of solar-terrestrial physical processes. The aurora activities have rapid changes in spatial and intensity distribution during a substorm, especially the expansion phase. In this work, a newly developed aurora evolution model is built including the aurora images prediction and the substorm expansion duration prediction. The aurora images prediction model is trained based on the Convolutional Long Short-Term Memory network, using the aurora images captured by the ultraviolet imager on the Polar satellite during the substorm expansion phases. Given the images after the onset, the model can predict the following aurora images sequences during the substorm expansion phase. However, the images prediction model works well only for 30-45 minutes, which is close to the substorm expansion phase duration. Considering this, the expansion phase duration prediction model is trained using the solar wind and interplanetary magnetic field data. Using the traditional machine learning method, the duration is predicted by inputting these physics parameters.

How to cite: Lu, Y., Jiang, J., Zhong, J., and Zou, Z.: Aurora Evolution Model During the Substorm Expansion Phase using Machine Learning based method , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12338, https://doi.org/10.5194/egusphere-egu25-12338, 2025.

EGU25-14068 | Posters on site | ESSI1.9

 Data efficiency: The master key for unlocking energy efficiency 

Valentine Anantharaj, Takuya Kurihana, Gabriele Padovani, and Sandro Fiore

AI foundation models have already demonstrated their usefulness in harnessing their potential in a wide range of science application domains. They derive their power from the large volumes of data, along with the computational methods used to exploit them using unprecedented amounts of compute power. We are inundated with data but have managed to exploit only a small fraction of the available data.  The Earth System Grid Federation (ESGF) is hosting nearly 16 PB of data collection from the Coupled Model Intercomparison Project (CMIP6), expected to grow to 5 - 10 more in the CMIP7 era. The NASA Earth Observation Data and Information System (EOSDIS) archive is expected to exceed 600 PB by 2030. 

AI-enabled solutions will require integrating multimodal data while being cognizant of the energy footprint introduced by the data and the computational methods. Currently, the energy consumption of transformer-based foundation models scale with the amount of data and corresponding model sizes. This impediment needs to be mitigated by developing data-efficient methods that lead to energy efficiency as well across all scales. There is little guidance in the research community on developing a computational plan for the optimal use of the resources for developing foundation models using multimodal scientific data. The benchmarks based on LLM scaling are still insufficient for vision transformers (ViTs), commonly adopted for geoscientific applications. We need a suite of community benchmarks based on ViT backbones and other methods at different scales to understand energy efficient methods for different classes of science problems.

Relatively few studies have focused on the issue of data efficiency for training science foundation models. We have adopted a smart sampling approach to extract the most informative samples is an effective means of significantly reducing the training data. We trained two ViT models, one with all available MODIS data over the ocean and another using an intelligently-sampled subset. We applied the models to classify clouds over the ocean.  Our preliminary results indicate that reasonably accurate models can be trained with only a fraction of total training data. Improvements in reduction of data translate directly into improvements for energy efficiency. 

How to cite: Anantharaj, V., Kurihana, T., Padovani, G., and Fiore, S.:  Data efficiency: The master key for unlocking energy efficiency, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14068, https://doi.org/10.5194/egusphere-egu25-14068, 2025.

EGU25-14422 | ECS | Posters on site | ESSI1.9 | Highlight

Comprehensive Validation of the Prithvi WxC FM Through Atmospheric Process Analysis 

Ankur Kumar, Sujit Roy, Udaysankar Nair, Manil Maskey, and Rahul Ramachandran

Weather Foundation Models in general represent a significant advancement in computational weather prediction by leveraging data-driven techniques to improve accuracy and speed. This study evaluates the Prithvi WxC foundation model for weather and climate by systematically assessing its adherence to fundamental physical constraints governing atmospheric processes. While traditional model validation primarily focuses on error statistics of modeled fields, this work takes a more comprehensive approach, incorporating a series of process-based tests to ensure the model's consistency with key atmospheric principles.

We first test the model's compliance with conservation of mass, ensuring that it respects the fundamental principle that mass is neither created nor destroyed within the atmosphere. Next, we examine the model's representation of geostrophic balance, critical for large-scale flow, by evaluating the relationship between the pressure gradient and the Coriolis force. The hypsometric equation is also applied to assess the vertical consistency of the model’s simulations, verifying that changes in pressure are appropriately related to temperature and height. To further evaluate large-scale flow dynamics, we analyze the model’s consistency with thermal wind relations, ensuring that temperature gradients are correctly reflected in the vertical wind profile. Finally, we tested our model on radiative and convective parameterization by comparing its performance of convection to established methods in conventional weather models, testing its ability to properly simulate convective processes. 

The results of these tests highlight the Prithvi WxC model's strengths and areas for improvement in terms of physical consistency. By adhering to these atmospheric principles, the findings of this study offer valuable insights into how the model can be refined, enhancing its potential applications in both weather forecasting and climate research.

 

How to cite: Kumar, A., Roy, S., Nair, U., Maskey, M., and Ramachandran, R.: Comprehensive Validation of the Prithvi WxC FM Through Atmospheric Process Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14422, https://doi.org/10.5194/egusphere-egu25-14422, 2025.

EGU25-14848 | ECS | Posters on site | ESSI1.9

Fine-tuning Foundation Models for Benchmarking Prediction Skills for Sub-seasonal Forecasting  

Takuya Kurihana, Valentine Anantharaj, and Moetasim Ashfaq

Few hundred million to billion parameters of autoregressive transformer-based weather foundation models (FMs) have demonstrated generalizabilities for various downstream applications such as regional forecasting to downscaling. They occasionally outperform traditional physics-based models for medium-range forecasting skills as well as enable significantly faster execution speeds. These ML approaches are designed for timeframes ranging from hourly to up to 10 days, and sub-seasonal forecasting, defined as a range spanning two weeks to two months, often receives less attention for their downstream tasks due to the inherent challenges in predicting the chaotic nature of atmospheric systems. However, the sub-seasonal to seasonal forecast has socio-economic impacts influencing actions from seasonal extreme weather events and economic activities. While community standards for benchmarking studies have been conducted for the medium-range forecasts, the benchmarking of sub-seasonal forecasts still needs further efforts. In this study, we are aiming to fine-tune foundation models to predict sub-seasonal forecasts for various variables to conduct comprehensive benchmarking for weather foundation models. Particularly, to reduce the complexity of tasks, our fine-tune task forecasts two-week averaged atmospheric variables with a forecasting lead-time of two weeks. For this task, we resample the community standard dataset, WeatherBench, for the two-week averaged dataset. We primarily work with the Oak Ridge Base Foundation Model for Earth System Predictability (ORBIT), and extend the benchmarking to other FMs across Aurora, ClimaX, and Prithvi WxC models. Our initial fine-tuning task uses a 100 million parameters ORBIT model to predict geopotential height at 200 hPa with two-week lead time, a key indicator for extreme precipitation in Central Southeast Asia. The preliminary results demonstrate that the fine-tuned ORBIT predicts realistic spatial distributions achieving an MSE of 24.32 m when evaluated against the 2018 data. The comprehensive sub-seasonal forecasting benchmarking can highlight the potential of weather FMs whether they capture underlying principles of atmospheric dynamics, thereby enabling their performance to be extended to longer forecast lead-times. 

How to cite: Kurihana, T., Anantharaj, V., and Ashfaq, M.: Fine-tuning Foundation Models for Benchmarking Prediction Skills for Sub-seasonal Forecasting , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14848, https://doi.org/10.5194/egusphere-egu25-14848, 2025.

EGU25-16635 | ECS | Posters on site | ESSI1.9

Generalist Geospatial Foundation Model PhilEO on Satellite Sentinel-2 MajorTOM Multi-Spectral Data 

Nikolaos Dionelis, Riccardo Musto, Giancarlo Paoletti, Jente Bosmans, Peter Naylor, Simone Sarti, Fabio Di Matteo, Giacomo Cascarano, Casper Fibaek, and Nicolas Longepe

Recent advancements in AI and Self-Supervised Learning (SSL) have revolutionized large-scale computer vision models, enabling exceptional performance on downstream tasks in remote sensing with minimal labelled data. These models are pre-trained on large amounts of unlabelled data and, then, fine-tuned on specific downstream Earth Observation (EO) applications [1-3]. The scarcity of large-scale labelled datasets and the technical challenges of annotating the vast volumes of data collected by satellites pose significant barriers to achieving high accuracy in many important downstream tasks, which require extensive labeling at a large scale to be effective. Furthermore, the dynamic nature of Earth adds complexity, as labels at a particular moment in time are not enough. 

Consequently, SSL and Foundation Models offer a powerful solution to these challenges. By pre-training a Foundation Model on extensive unlabelled data, only a small amount of labelled data is required for supervised fine-tuning on downstream tasks. This approach reduces the need for labelled EO data. This enables the development of general-purpose EO Foundation Models capable of solving a diverse range of problems. 

In this work, we train the EO Foundation Model PhilEO Version 1.0 [2] on the dataset MajorTOM [6]. The model is trained on Sentinel-2 data, i.e. 23TB, Core-S2L2A. We scale up the pre-training data from less than 1TB in the dataset PhilEO-Globe [7] to 23TB. The model is trained using a combination of reconstruction and auxiliary task losses, including the Mean Squared Error (MSE) for geo-location longitude and latitude prediction. The architecture is a modified U-Net. The training is conducted on the Leonardo Davinci-1 supercomputer. 

In this work, we also extend the capabilities of the evaluation framework for EO Foundation Models we recently introduced in [2]. We develop PhilEO-Bench++. To take advantage of multi-level features and of a U-Net-like architecture, for fine-tuning on downstream tasks, we use the decoder UPerNet [4]. Furthermore, to strengthen the evaluation of EO Foundation Models, we also perform confidence quantification and assessment [5] on both classification and regression tasks, including on land cover semantic segmentation and building density pixel-wise regression. 

Experiments on the PhilEO-Bench downstream tasks of building density estimation, road segmentation and land cover mapping demonstrate the effectiveness of our model. For building density regression, for n-shots n=50 and n=100, the PhilEO model trained on MajorTOM achieves the MSEs 0.0191 and 0.0058, respectively.

 

References:  

[1] D. Szwarcman, et al., “Prithvi-EO-2.0: A Versatile Multi-Temporal Foundation Model for Earth Observation Applications,” arXiv:2412.02732, 2024. 

[2] C. Fibaek, et al., “PhilEO Bench: Evaluating Geo-Spatial Foundation Models,” in Proceedings IGARSS, 2024. 

[3] N. Dionelis, et al., “Evaluating and Benchmarking Foundation Models for Earth Observation and Geospatial AI,” arXiv:2406.18295, 2024. 

[4] T. Xiao, et al., “Unified Perceptual Parsing for Scene Understanding,” 2018.

[5] N. Dionelis and N. Longepe, “Fine-Tuning Foundation Models with Confidence Assessment for enhanced Semantic segmentation,” 2024. 

[6] A. Francis and M. Czerkawski, “MajorTOM: Expandable Datasets for Earth Observation,” IGARSS, 2024. 

[7] B. Le Saux, et al., “The PhilEO Geospatial Foundation Model Suite,” EGU, 2024.

How to cite: Dionelis, N., Musto, R., Paoletti, G., Bosmans, J., Naylor, P., Sarti, S., Di Matteo, F., Cascarano, G., Fibaek, C., and Longepe, N.: Generalist Geospatial Foundation Model PhilEO on Satellite Sentinel-2 MajorTOM Multi-Spectral Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16635, https://doi.org/10.5194/egusphere-egu25-16635, 2025.

EGU25-16756 | Posters on site | ESSI1.9

Earth Observation embeddings at the test: A novel benchmark to evaluate (neural) compression for satellite imagery 

Rikard Vinge, Michael L Marszalek, Jannik Schneider, and Conrad M Albrecht

With the rapidly growing production and utilization of Earth Observation (EO) data, the past decade sparked interest in the efficient compression of EO data into low-dimensional embeddings. In a parallel development, EO Foundation Models (FM), trained on large amounts of unlabeled data to be used in a wide range of applications, also utilize low-dimensional embeddings to distill representations of EO data [1, 2, 3]. In one aspect, EO FMs may serve as (lossy) neural compressors to improve data transfer and lower storage needs – effectively reducing the carbon footprint of EO data [4].

While the development in EO FMs rapidly advances, there is need for a novel benchmark scheme to evaluate the quality of (compressed) embeddings. The statement “foundational” or “general purpose representation” needs a test.

As part of the Horizon Europe project “Embed2Scale” [5], co-funded by the European Union (Horizon Europe contract No. 101131841), the Swiss State Secretariat for Education (SERI), and UK Research and Innovation (UKRI), we present a novel approach to benchmark learnt compression of multimodal Copernicus Sentinel data for various relevant application domains. In the form of a competition, contestants provide embeddings that are evaluated on a diverse set of problems based on real-life use cases relevant for the research community, governments, and corporate businesses. The problems are hidden from the contestants to evaluate the applicability of the embeddings to unknown problems. The benchmark statistically evaluates the performance of downstream tasks through fine-tuning of neural networks that fit into commodity hardware. We underline a practically relevant scenario where end users rarely have access to costly and energy-intensive acceleration hardware. The overall performance, i.e. the evaluation across all the benchmark’s problems, is crucial and ensures a diverse and fair evaluation of the embeddings. After the competition, the datasets in the benchmark are published and made available to the community.

[1] X. Sun et al., “RingMo: A remote sensing foundation model with masked image modeling,” IEEE Transactions on Geoscience and Remote Sensing, 2022.

[2] D. Wang et al., “Advancing plain vision transformer toward remote sensing foundation model,” IEEE Transactions on Geoscience and Remote Sensing, 2022.

[3] C. Bodnar et al., “Aurora: A foundation model of the atmosphere,” Tech. Rep., 2024.

[4] R. Wilkinson, M.M. Mleczko, R.J.W. Brewin, K.J. Gaston, M. Mueller, J.D. Shutler, X. Yan, K. Anderson, Environmental impacts of earth observation data in the constellation and cloud computing era,Science of The Total Environment, Volume 909,2024,168584,ISSN 0048-9697, https://doi.org/10.1016/j.scitotenv.2023.168584

[5] https://embed2scale.eu/

How to cite: Vinge, R., Marszalek, M. L., Schneider, J., and Albrecht, C. M.: Earth Observation embeddings at the test: A novel benchmark to evaluate (neural) compression for satellite imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16756, https://doi.org/10.5194/egusphere-egu25-16756, 2025.

EGU25-17133 | Posters on site | ESSI1.9

Hierarchical Causal Graph-Based Methods for Imputing Food Insecurity 

Esther Rodrigo-Bonet, Jordi Cerda, Michele Ronco, and Gustau Camps-Valls

Food insecurity is typically modelled using inter-regional data comprising economical, geophysical and social variables. Such datasets are often of varying granularity, with each variable corresponding to a certain granularity level (e.g., GDP is a national variable, while disaster displacement can be local or regional). Additionally, each level shows a specific causal relation of its variables.  Since countries affected by food insecurity are usually underdeveloped, collecting such variables is a challenging task, leading to highly-incomplete datasets. To deal with the multi-level complexity and incomplete nature of the data, we propose to build a hierarchical causal graph (HCG) structure of the variables, that can then be injected in different imputation methods. Specifically, we propose to classify the variables at different granularity levels, and use causal graph discovery to learn a causal graph at each level. We test the proposed approach for imputing food insecurity using a dataset of 300+ economical, geophysical and social variables for more than 70 countries.

How to cite: Rodrigo-Bonet, E., Cerda, J., Ronco, M., and Camps-Valls, G.: Hierarchical Causal Graph-Based Methods for Imputing Food Insecurity, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17133, https://doi.org/10.5194/egusphere-egu25-17133, 2025.

EGU25-17360 | ECS | Posters on site | ESSI1.9

PrithviWxC Foundation Model Validation on Weather Downscaling for Cross Domain Learning 

Gabriele Padovani, Ankur Kumar, Takuya Kurihana, Sandro Fiore, and Valentine Anantharaj

AI foundation models hold considerable promise for leveraging the vast and diverse datasets available in atmospheric and geoscientific research. These models have the potential to advance scientific discovery by capturing complex spatial and temporal relationships inherent in earth system processes. However, the development and deployment of such models is often hindered by limited computational resources.

Accurate reconstruction of fine-scale atmospheric features from coarse-resolution data is a critical challenge in geoscientific modeling, as well as a benchmark for understanding the performance of climate-related models. High-resolution atmospheric data are essential for capturing localized phenomena, such as convective systems, topographic effects, and land-atmosphere interactions, that influence weather patterns and climate processes. However, the generation and storage of high-resolution datasets are computationally expensive, necessitating methods that can infer fine-scale structures from lower-resolution observations.

The primary objective of this study is to validate PrithviWxC [4], a ViT-based foundation model [1], for the task of downscaled image reconstruction. While Prithvi WxC was trained on 160 atmospheric variables from the Modern-Era Retrospective Analysis for Research and Applications, Version 2 (MERRA-2) dataset [2], we implement a smaller version of the initial model, with 21 million parameters, and pretrained on the same dataset with a set of six variables. 

The evaluation process involves the assessment of the model's capacity to reconstruct fine-scale features through the process of downscaling atmospheric data. In this procedure, inputs at a high spatial resolution of 1 km from [5] are first coarsened to 25 km resolution, the same as European Centre of Medium-range Weather Forecasts Reanalysis v.5 (ERA5) [3], and subsequently upscaled to recover the original fine-grained structure. This process serves as a benchmark for assessing the model's capacity to learn and preserve spatial details during resolution transformations, which is an essential requirement for geoscientific modeling tasks.

We fine-tune the model for downscaled image reconstruction on a set of 30784 128x128 patches, and validate its output, produced after learning on a limited temporal period, on tiles coming from the ERA5 dataset, which encompass all seasonality. In particular, we aim at highlighting the model's ability to generalize to data domains beyond its pretraining distribution, demonstrating its adaptability and the transferability of knowledge embedded within ViT architectures. By applying PrithviWxC to a knowledge domain that is distinct from its original training context, we demonstrate the potential for cross-domain learning in geoscientific applications.

 

REFERENCES

[1] Dosovitskiy, Alexey. "An image is worth 16x16 words: Transformers for image recognition at scale." arXiv preprint arXiv:2010.11929 (2020).

[2] Gelaro, Ronald, et al. "The modern-era retrospective analysis for research and applications, version 2 (MERRA-2)." Journal of climate 30.14 (2017): 5419-5454.

[3] Hersbach, Hans, et al. "The ERA5 global reanalysis." Quarterly Journal of the Royal Meteorological Society 146.730 (2020): 1999-2049.

[4] Schmude, Johannes, et al. "Prithvi wxc: Foundation model for weather and climate." arXiv preprint arXiv:2409.13598 (2024).

[5] Wedi, Nils P., et al. "A baseline for global weather and climate simulations at 1 km resolution." Journal of Advances in Modeling Earth Systems 12.11 (2020): e2020MS002192.

How to cite: Padovani, G., Kumar, A., Kurihana, T., Fiore, S., and Anantharaj, V.: PrithviWxC Foundation Model Validation on Weather Downscaling for Cross Domain Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17360, https://doi.org/10.5194/egusphere-egu25-17360, 2025.

EGU25-18029 | ECS | Orals | ESSI1.9

GeoDINO: A Vision Foundation Model for Earth Observation Leveraging DINO Architecture and Sentinel-2 Multi-Spectral Data 

Riccardo Musto, Giancarlo Paoletti, Nikolaos Dionelis, Simone Sarti, Fabio Di Matteo, Jente Bosmans, Peter Naylor, Giacomo Donato Cascarano, Casper Fibaek, and Nicolas Longépé

Foundation Models are emerging as a transformative paradigm in Earth observation, offering powerful solutions to the challenges of processing and understanding satellite imagery at scale. The scarcity of large-scale labeled datasets and the technical challenges of annotating the vast volumes of data collected by satellites pose significant barriers to achieving high accuracy in many important downstream tasks. Furthermore, the dynamic nature of Earth adds complexity, as labels tied to a specific geographical region at a particular moment in time are insufficient to capture the evolving characteristics of the environment. Self-supervised learning techniques have emerged as a promising solution, enabling models to learn rich representations from unlabeled data while requiring minimal supervised fine-tuning for specific applications.
In this work, we present GeoDINO, a novel foundation model that adapts the DINO self-supervised learning architecture to handle multi-spectral Sentinel-2 data. While the original DINO framework has shown remarkable success in computer vision tasks through its teacher-student architecture and self-distillation approach, we extend it significantly for Earth observation applications. Our key innovation lies in the addition of multiple supervised auxiliary tasks: after the encoder generates representations, we attach specialized MLPs designed to predict various geospatial attributes including climate zones, permanent water bodies and geographical coordinates. Both the teacher and student networks are trained to predict these auxiliary labels, with the teacher network being updated through Exponential Moving Average (EMA) of the student's weights. This modification enables our model to learn not only from the self-supervised distillation process but also from the rich spatial and temporal information inherent in satellite imagery.
We are currently training GeoDINO on MajorTOM, a comprehensive Sentinel-2 dataset comprising 23TB of Core-S2L2A data, exploiting the Leonardo Davinci-1 Supercomputer. Furthermore, to validate our approach, we are also training the model on FastTOM and TinyTOM, two subsets of MajorTOM. Finally, the model will be evaluated within the PhilEO Bench framework to assess its performance on different tasks, including land cover classification, change detection, and building density estimation. Looking ahead, we plan to transition to the DINOv2 architecture to further enhance our model's capabilities. Through this research, we aim to demonstrate how self-supervised learning techniques, when properly adapted for Earth observation data, can address the fundamental challenges of data scarcity and temporal dynamics in remote sensing applications. The development of GeoDINO represents a step toward more efficient and adaptable Earth observation systems that can leverage the vast amounts of available satellite data while minimizing the need for extensive labeled datasets.
References:  
[1] M. Caron, et al., “Emerging Properties in Self-Supervised Vision Transformers”, arXiv:2104.14294, 2021
[2] C. Fibaek, et al., “PhilEO Bench: Evaluating Geo-Spatial Foundation Models,” in Proceedings IGARSS, 2024. 
[3] N. Dionelis, et al., “Evaluating and Benchmarking Foundation Models for Earth Observation and Geospatial AI,” arXiv:2406.18295, 2024. 
[4] N. Dionelis and N. Longepe, “Fine-Tuning Foundation Models with Confidence Assessment for enhanced Semantic segmentation,” 2024. 
[5] A. Francis and M. Czerkawski, “MajorTOM: Expandable Datasets for Earth Observation,” IGARSS, 2024. 
[6] B. Le Saux, et al., “The PhilEO Geospatial Foundation Model Suite,” EGU, 2024.

How to cite: Musto, R., Paoletti, G., Dionelis, N., Sarti, S., Di Matteo, F., Bosmans, J., Naylor, P., Donato Cascarano, G., Fibaek, C., and Longépé, N.: GeoDINO: A Vision Foundation Model for Earth Observation Leveraging DINO Architecture and Sentinel-2 Multi-Spectral Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18029, https://doi.org/10.5194/egusphere-egu25-18029, 2025.

EGU25-19016 | ECS | Posters on site | ESSI1.9

Scalable Efficient Compression in Large-Scale Earth Observation 

Erik Scheurer, Jiangtao Wang, Rocco Sedona, Stefano Maurogiovanni, Benedikt Blumenstiel, Johannes Jakubik, Paolo Fraccaro, Thomas Brunschwiler, Stefan Kesselheim, and Gabriele Cavallaro

Earth observation (EO) yields large-scale, multimodal datasets collected from various satellite missions, DEMs, land-use data, and textual metadata. Foundation models like 4M (Massively Multimodal Masked Modeling) can learn a joint embedding space that bridges modality gaps, mitigates missing data issues, and facilitates partial spatio-temporal alignment [1]. However, directly training such foundation models on the vast, high-dimensional original EO datasets is not only computationally intensive but also imposes substantial demands on storage resources.

To address this, one can leverage VQ-VAE (Vector Quantized-Variational AutoEncoder) as neural compressors to transform high-dimension multimodal inputs into a few discrete indices, significantly reducing data volume while preserving critical information. By inverting the tokenization process, we can reconstruct the original high-dimensional data with minimal quality loss, aided by adversarial and perceptual losses that enhance reconstruction fidelity.

Traditional VQ-based approaches, however, face challenges such as inefficient codebook utilization and limited latent space representation. To overcome these, we propose scaling strategies that complement 4M’s tokenizer-based architecture. By expanding the codebook size, latent dimensions, and network depth, our method captures the complexity of EO modalities more effectively. Specifically, we employ spherical quantization techniques like Grouped Spherical Quantization (GSQ) to address limitations in traditional approaches [2]. GSQ constrains codebook vectors to a spherical surface, stabilizing training, preventing code collapse, and promoting uniform codebook usage. Unlike standard VQ, GSQ uses spherical initialization and normalization to maintain consistent distances among codebook entries, ensuring robust latent space coverage even under extreme compression or large codebooks. From our empirical and ablation studies, alternative methods like LFQ (Lookup-Free Quantization), FSQ (Finite Scalar Quantization), and RVQ (Residual Vector Quantizer) often exhibit limitations, such as tightly coupling the latent dimension to codebook size or relying on specialized training losses. In contrast, spherical-based techniques effectively decouple latent dimensions from codebook vocabulary, providing greater flexibility and scalability as data demands increase.

Our approach enables neural compressors to adapt to varying scales of compression and complexity without compromising performance. Comprehensive scalability experiments—examining large codebooks, deeper networks, and diverse compression ratios—assessed the generalizability of the proposed compression strategies and demonstrated their effectiveness on high-dimensional, large-scale EO data with minimal information loss. By integrating advanced compression techniques with scalable architectures, this framework establishes a robust foundation for addressing multimodal challenges in EO research that significantly reduces the difficulty of training foundation models on multimodal high-dimensional EO data.

References

[1] Mizrahi, D., Bachmann, R., Kar, O. F., Yeo, T., Gao, M., Dehghan, A., & Zamir, A. (2023). 4M: Massively Multimodal Masked Modeling (Version 1). arXiv. https://doi.org/10.48550/ARXIV.2312.06647

[2] Wang, J., Qin, Z., Zhang, Y., Hu, V. T., Ommer, B., Briq, R., & Kesselheim, S. (2024). Scaling Image Tokenizers with Grouped Spherical Quantization (Version 2). arXiv. https://doi.org/10.48550/ARXIV.2412.02632

Acknowledgments

This work is performed in the Embed2Scale (Earth Observation & Weather Data Federation With AI Embeddings) project, funded by the EU’s Horizon Europe program under Grant Agreement number 101131841.

How to cite: Scheurer, E., Wang, J., Sedona, R., Maurogiovanni, S., Blumenstiel, B., Jakubik, J., Fraccaro, P., Brunschwiler, T., Kesselheim, S., and Cavallaro, G.: Scalable Efficient Compression in Large-Scale Earth Observation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19016, https://doi.org/10.5194/egusphere-egu25-19016, 2025.

EGU25-19131 | Posters on site | ESSI1.9

A Practical Guide to Hyperspectral Foundation Models 

Conrad Albrecht, Ruben Gonzalez, Nassim Ait Ali Braham, Ranjini Bangalore, and Thomas Brunschwiler

Hyperspectral imagery (HSI) provides rich spectral information that is the basis for applications such as mineral mapping, trace gas identification, and precision agriculture. Yet, the development of HSI Foundation Models (FMs) is less advanced compared to multi-spectral remote sensing modalities.

In this study, we leverage the SpectralEarth dataset [1] to explore practical aspects of training robust HSI FMs. In particular, we shed light on the role of:

  • the impact of model architecture (transformers vs. convolutional networks),
  • self-supervised learning methods (contrastive vs. masked autoencoders),
  • model size & training data volume,
  • and the resulting computational requirements.

Through extensive experiments, this study aims to provide concrete guidelines for the development and effective application of FMs in the HSI domain. Moreover, we report on findings to identify downstream applications where hyperspectral imagery has an edge over multi-spectral photos [2], and where such an advantage is less likely to expect.

 

References

[1] Braham, Nassim Ait Ali, et al. "SpectralEarth: Training Hyperspectral Foundation Models at Scale." arXiv preprint arXiv:2408.08447 (2024)

[2] Bangalore, Ranjini, et al. "Hyperspectral foundation model trained by spectral reconstruction for greenhouse gas emission estimation", annual meeting of the American Geophysical Union (2024)

How to cite: Albrecht, C., Gonzalez, R., Braham, N. A. A., Bangalore, R., and Brunschwiler, T.: A Practical Guide to Hyperspectral Foundation Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19131, https://doi.org/10.5194/egusphere-egu25-19131, 2025.

EGU25-19378 | Orals | ESSI1.9

Towards a Foundation Model for Global Terrestrial 3D Above and Below Ground Carbon Stock Mapping (3D-ABC) 

Guido Grosse, Pedram Ghamisi, Gabriele Cavallaro, Martin Herold, Andreas Huth, and Irena Hajnsek and the 3D-ABC Team

Understanding the global carbon budget with its carbon sources and sinks is scientifically important and economically relevant. In particular, vegetation and soils are major and highly dynamic carbon pools in the Earth System and a substantial part of the terrestrial carbon budget is influenced by land use changes, vegetation dynamics, and soil processes.

Recent advances in Foundation Models (FMs) are transforming AI, enabling remarkable generalization and zero-shot learning capabilities. Within the Helmholtz Foundation Model Initiative, we are developing the 3D-ABC FM, a tool targeting the accurate mapping of above- and below-ground carbon stocks in vegetation and soils at high spatial resolution. 3D-ABC aims to provide a seamless understanding of terrestrial carbon distribution by integrating multimodal remote sensing, climate, and elevation datasets, and addressing complex challenges such as multi-dimensionality and multi-resolution in FMs. Our unique 3D-ABC partnership brings together key capacities from the domains remote sensing, carbon monitoring, AI, and high-performance computing to take on such FM development.

The 3D-ABC FM integrates large-scale remote sensing data, including multispectral satellite imagery from the Harmonized Landsat-Sentinel-2 (HLS) dataset, TanDEM-X InSAR coherence data, and 3D lidar data from space (GEDI, ICESat 1&2), aircraft, and ground-based platforms. We also aim to incorporate ERA-5 Land climate reanalysis information, GLO-30 digital elevation data, as well as local lidar and field data on vegetation, soils, and carbon flux parameters. High-resolution forest models will be used to benchmark carbon fluxes.

To accommodate the diverse data modalities assembled for 3D-ABC and to address eight selected downstream tasks, the AI model employs an adaptive architecture, integrating a multi-modal input processor, an FM encoder, an adaptive fusion neck, and task-specific prediction heads. The multi-modal input processor handles data with varying spectral dimensions, automatically mapping inputs to a unified feature space. The FM encoder extracts generalized deep features from the normalized inputs, which are then integrated into universal feature representations through the adaptive fusion neck. This fusion enhances interactions across modalities. Finally, the universal features are decoded into various outputs tailored to the specific needs of downstream tasks. In the first FM training phase, a pretraining strategy leverages a masked autoencoder to train the multi-modal input processor, the encoder, and the fusion neck in an unsupervised manner, enabling the model to develop robust representation capabilities. In the second phase, by leveraging the principles of transfer learning, the pretrained model is fine-tuned using labeled data from various downstream tasks.

3D-ABC targets use of the JUWELS Booster and JUPITER high-performance computing (HPC) systems located at the Jülich Supercomputing Centre (JSC). The JUWELS Booster comprises 936 compute nodes, each equipped with four NVIDIA A100 GPUs. JUPITER, the first European exascale supercomputer, is currently being installed at JSC. Its Booster module will consist of ~6,000 compute nodes, each featuring four NVIDIA GH200 GPUs. To maximize efficient JUPITER utilization, 3D-ABC is leveraging the JUPITER Research and Early Access Program, which provides early access for code optimization and preparation to ensure FM applications are optimized and ready for deployment when the system becomes operational in 2025.

How to cite: Grosse, G., Ghamisi, P., Cavallaro, G., Herold, M., Huth, A., and Hajnsek, I. and the 3D-ABC Team: Towards a Foundation Model for Global Terrestrial 3D Above and Below Ground Carbon Stock Mapping (3D-ABC), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19378, https://doi.org/10.5194/egusphere-egu25-19378, 2025.

EGU25-19821 | Posters on site | ESSI1.9

From architecture to atmospheric sensitivity: studying forecast uncertainty with Prithvi-WxC 

Eloisa Bentivegna, Valentine Anantharaj, Johannes Schmude, Sujit Roy, Ankur Kumar, Amy Lin, Sharana Shivanand, Theodore Papamarkou, Richard Allmendinger, Manil Maskey, and Rahul Ramachandran

AI-based weather emulators have begun to rival the accuracy of traditional numerical solvers, for a fraction of the computational cost. The question of whether they can be reliably deployed in all use cases (e.g., for the forecast of extreme scenarios), however, is still open. We outline an ensembling strategy based on architectural variations of the Prithvi WxC foundation model (FM), highlighting the impact of each of these variations on physical accuracy and ability to capture the distributional extremes. A simple of ensemble of 100 models is sufficient to observe the complex mapping between configuration parameters and the forecast sensitivity of different atmospheric variables. We characterize some features of this mapping and connect them to the task of predicting various weather extremes.

How to cite: Bentivegna, E., Anantharaj, V., Schmude, J., Roy, S., Kumar, A., Lin, A., Shivanand, S., Papamarkou, T., Allmendinger, R., Maskey, M., and Ramachandran, R.: From architecture to atmospheric sensitivity: studying forecast uncertainty with Prithvi-WxC, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19821, https://doi.org/10.5194/egusphere-egu25-19821, 2025.

EGU25-19892 | Orals | ESSI1.9

Enabling Seamless Provenance Collection in Large-Scale Machine Learning Tasks 

Sandro Fiore, Gabriele Padovani, Takuya Kurihana, Massimiliano Fronza, and Valentine Anantharaj

The growing interest in deep learning and large language models (LLMs) in recent years highlights their remarkable adaptability and ability to generalize, drawing researchers from a wide array of disciplines. Despite their promise, in many instances, these advancements have exposed a lack of transparency and rigor during development processes. Although this rapid pace of research undoubtedly offers numerous benefits, it has also led to an increasing prevalence of works conducted without rigor and in a superficial way. Code that is not accompanied by documentation and results that are not reproducible inevitably lead to confusion among researchers and an environment in which trust is not a fundamental aspect of the proposed work. The complexity of data manipulation, characterized by ad hoc transformations, exacerbates these issues by hindering the traceability of processes, and hyperparameter tuning introduces additional difficulties, requiring repeated experimentation that consumes excessive computational resources, especially for large models. 

To address these challenges, we introduce yProv4ML, a python library which provides an accessible option for tracking dataset and model statistics, hyperparameters, and energy metrics. It allows for the comparison of sets of experiments, and introduces a suite of directives to easily track the flow of information through provenance metadata. 

yProv4ML is a component of the yProv framework, a research project on multi-level provenance management which provides scientists with a rich software ecosystem consisting of a web service to manage track and analyze provenance documents. Leveraging the PROV-JSON standard for provenance artifact recording, yProv4ML ensures comprehensive documentation and reproducibility while facilitating a seamless integration process similar to well-known libraries such as MLFlow.

During the last year, yProv4ML was integrated in a variety of use cases in different domains (i.e., Climate Science, High Energy Physics and Earth Observation) in the context of the interTwin (https://www.intertwin.eu/) and ClimateEurope2 (https://climateurope2.eu/) EU projects, as well as the ICSC Italian National Project (https://www.supercomputing-icsc.it/en/icsc-home/). The collection of provenance data in these use cases not only helped facilitate the reproducibility of experiments, but also helped diagnose performance bottlenecks and ensure the reliability and integrity of results, all of which are critical to advancing the field of large-scale ML in a trustworthy manner.

How to cite: Fiore, S., Padovani, G., Kurihana, T., Fronza, M., and Anantharaj, V.: Enabling Seamless Provenance Collection in Large-Scale Machine Learning Tasks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19892, https://doi.org/10.5194/egusphere-egu25-19892, 2025.

EGU25-20521 | ECS | Posters on site | ESSI1.9

Integrating Prithvi WxC with a Hurricane Intensity Estimation Model for Accurate Hurricane Forecasting 

Sujit Roy, Ankur Kumar, Rohit Lal, Udaysankar Nair, Manil Maskey, and Rahul Ramachandran

Accurate hurricane intensity estimation is critical for disaster preparedness, yet remains challenging for weather models trained on coarse-resolution datasets. This study proposes a hybrid approach that integrates NASA-IBM's Prithvi WxC model with a deep learning-based Hurricane Intensity Estimation (HIE) model. While the Prithvi WxC model excels in global atmospheric predictions, its coarse-grained outputs can struggle with precise hurricane intensity estimation. To address this, the HIE model is triggered when it identifies a hurricane in the Prithvi model output, providing corrected intensity predictions based on high-resolution data.

A dataset was created for training and evaluation, consisting of 6,000 unique initial conditions from 1980 to 2024 that resulted in hurricanes across all major basins. Ground truth hurricane tracks and intensity data were obtained from the HURDAT database The training phase focused on hurricane cases from 1980 to 2000, building a foundational understanding of global hurricane characteristics. Subsequently, the model was fine-tuned with 2000–2020 data to account for basin-specific variations and improve regional accuracy. The remaining cases (2020–2024) are reserved for validation and assessment. The HIE model employs advanced deep learning techniques to refine key intensity metrics, such as maximum sustained wind speeds and central pressure. By addressing the limitations of Prithvi WxC's coarse-resolution training data, the HIE model achieves greater precision, leveraging fine-grained atmospheric and oceanographic features. This two-step framework, hurricane detection by Prithvi WxC followed by intensity refinement by the HIE model, capitalizes on the strengths of both models to deliver improved predictions.

This highlights the potential of combining foundation models like Prithvi WxC with specialized deep-learning frameworks to overcome existing limitations in hurricane intensity estimation. By incorporating diverse data sources and leveraging modern machine-learning techniques, this hybrid approach bridges the gap between coarse-grained global models and the need for precise regional forecasting.

How to cite: Roy, S., Kumar, A., Lal, R., Nair, U., Maskey, M., and Ramachandran, R.: Integrating Prithvi WxC with a Hurricane Intensity Estimation Model for Accurate Hurricane Forecasting, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20521, https://doi.org/10.5194/egusphere-egu25-20521, 2025.

EGU25-5082 | Posters on site | ESSI1.11

Region identification in spacecraft data using supervised machine learning 

Maryam Aghabozorgi Nafchi, Gilbert Pi, Frantisek Nemec, Tsung-Che Tsai, and Kun-Han Lee

The classification of near-Earth plasma regions, i.e., distinguishing the region in which a spacecraft is located at any given time, is beneficial for both understanding the dynamics of the interaction between the Earth’s magnetosphere and the solar wind, and for modeling the characteristic boundaries separating these regions. We use measurements from the THEMIS B spacecraft between 2008 and 2010 (340 days in total) with a time resolution of one minute. The data include solar wind velocity and density, magnetic field magnitude, and standard deviation of magnetic field magnitude calculated over one-minute intervals. These data are used for manual labeling of four distinct plasma regions: solar wind, foreshock, magnetosheath, and magnetosphere. Ion energy flux data are used to classify the foreshock, if necessary. An automated classification of the respective regions based on measured plasma and magnetic field parameters is then achieved using either neural network or random forest classifiers. The performance of these classifiers is evaluated and compared. Generally, very high accuracy is achieved, but distinguishing between solar wind and foreshock remains an issue.

How to cite: Aghabozorgi Nafchi, M., Pi, G., Nemec, F., Tsai, T.-C., and Lee, K.-H.: Region identification in spacecraft data using supervised machine learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5082, https://doi.org/10.5194/egusphere-egu25-5082, 2025.

EGU25-6747 | Posters on site | ESSI1.11

 Automatic detection of the electron density from de WHISPER instrument onboard CLUSTER II 

Emmanuel De Leon, Maxime Vandevoorde, Xavier Vallieres, and Pierre Henri

The Waves of HIgh frequency and Sounder for Probing Electron density by Relaxation
(WHISPER) instrument, is part of the Wave Experiment Consortium (WEC) of the ESA
CLUSTER II mission. WHISPER is designed to measure the electric field fluctuation and derive the electron density, i.e. the plasma density, a key parameter of scientific interest for
magnetospheric and near-Earth solar wind studies. The electron density is the WHISPER highest level product and is provided, among other products, to the scientific community through the CLUSTER Science Archive (CSA).
The instrument consists of a receiver, a transmitter, and a wave spectrum analyzer. It delivers both ambient (in natural mode) and active (in sounding mode) electric field spectra. The characteristic signatures of ambient plasma waves or active plasma resonances, combined with the spacecraft position, reveal the different magnetosphere regions. These spectral signatures are used to derive the electron density. Until recently, ad-hoc algorithms have been used to derive the electron density from WHISPER measurements, but at the cost of time-consuming manual steps. These algorithms are dependent on measurements provided by other instruments onboard CLUSTER, thus introducing dependencies and potential delays in the data production.

In this context, the goal of this work is to significantly reduce human intervention by fully
automating the WHISPER electron density derivation, exclusively using WHISPER data.
For this purpose, we develop a two-step derivation process, based on neural networks: first, the plasma region is identified with a Multi-Layer Perceptron classification algorithm; second, the electron density is derived using a Recurrent Neural Network, adapted to each plasma region. These networks have been trained with WHISPER spectra and electron density previously derived from ad-hoc algorithms. The resulting accuracy is up to 98% in some plasma regions. This derivation process has been implemented in a production pipeline, now routinely used to deliver WHISPER electron density to the CSA and dividing by 10 the human intervention. The pipeline has already delivered 3+ years of data and will be used to reprocess some of the archive focusing on the most complex plasma regions with recent improvements. This work will present the implemented methods and models for each region focusing on results and performance. 

How to cite: De Leon, E., Vandevoorde, M., Vallieres, X., and Henri, P.:  Automatic detection of the electron density from de WHISPER instrument onboard CLUSTER II, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6747, https://doi.org/10.5194/egusphere-egu25-6747, 2025.

EGU25-7486 | ECS | Orals | ESSI1.11

Missing Interplanetary Data Estimation for Space Missions via Symbolic Rule Induction 

Federico Sabbatini and Catia Grimani

The necessity to limit budget, size, weight and power consumption of the instruments placed on board space mission satellites results in several drawbacks, including the exclusion of dedicated instrumentation for the monitoring of the spacecraft environment. Understanding the environmental conditions of space missions is essential to correctly analyse their observations. Seldom the necessary interplanetary parameters, not measured in situ, can be gathered from nearby dedicated missions, however this is not always feasible. Other solutions envisage the application of machine learning models to estimate the missing parameters on the basis of those that are available on board the satellites. Despite the high performance of machine learning predictors, they come along with issues related to the model selection and training, the data pre-processing and the opaqueness of the outcomes returned to end-users. The application of tools developed in the explainable artificial intelligence (XAI) field can be considered to encode through symbolic knowledge the functional relationship between parameters observed in situ and correlated parameters for which measurements are lacking but useful. In this context, XAI methods in general, and symbolic knowledge extraction in particular, constitute a promising alternative to traditional machine learning models, enabling users to avoid the model selection and training phases and to obtain completely interpretable results. This presentation provides an overview on the application of symbolic knowledge-extraction techniques to perform rule induction from available in-situ data, aimed at carrying out a human-interpretable estimation and forecasting of missing platform parameters. Potentialities, drawbacks and challenges of this approach are discussed to highlight the direction from current results to future applications.

How to cite: Sabbatini, F. and Grimani, C.: Missing Interplanetary Data Estimation for Space Missions via Symbolic Rule Induction, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7486, https://doi.org/10.5194/egusphere-egu25-7486, 2025.

EGU25-8698 | ECS | Orals | ESSI1.11

Evaluating Solar Imaging Feature Extraction Techniques for Enhancing Space Weather Prediction with Deep Learning Models 

Maria Tahtouh, Guillerme Bernoux, and Antoine Brunet

Many machine learning models have provided significant results in predicting the geomagnetic activity quantified by Earth-measured geomagnetic indices. For instance, one such model is the SERENADE model that provides probabilistic forecasts of the Kp index, days ahead solely from solar imaging. It consists of three modules combining convolutional, recurrent, and linear neural network layers that first extract the important information contained in the input solar imagery and transform them into an intelligible forecast. To improve the performance of this model, we evaluate solar-imaging-adapted dimensionality reduction techniques that extract the features from the images and can therefore be used as the first layer of the forecast model. We use a solar imagery dataset formatted specifically for machine-learning research (SDOML). We applied the Principal Component Analysis method and trained AutoEncoders and Variational AutoEncoders (VAE) targeting several reduced dimensions. We consider the convolutional GoogLeNet method, which was pre-trained on the ImageNet dataset, as a baseline for our comparison. We analyze the information retained by the extracted features in terms of solar activity physical parameters and find high correlations between the latter and the the reduced representations of the images, with the VAE results standing out. In addition, we re-train the SERENADE model to predict the daily maximum of the Kp index two days in advance using the extracted features by the new dimensionality reduction methods as input to the model. We first use the same hyperparameters that were optimized for the GoogLeNet model and obtain more stable predictions using the dedicated solar imaging feature extractors than when using the baseline model, specifically in the VAE case. Furthermore, when fine-tuning SERENADE's hyperparameters to the VAE model, the predictive performance of the model was enhanced, notably during geomagnetic storms, which indicates that the use of adapted feature extractors could improve the geomagnetic activity forecasting.

How to cite: Tahtouh, M., Bernoux, G., and Brunet, A.: Evaluating Solar Imaging Feature Extraction Techniques for Enhancing Space Weather Prediction with Deep Learning Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8698, https://doi.org/10.5194/egusphere-egu25-8698, 2025.

EGU25-9116 | ECS | Orals | ESSI1.11

Machine Learning for Space Weather: Solar Flare Forecasting Using SDO/HMI Magnetogram Time Series 

Elizabeth Doria Rosales, Prof. Vincenzo Carbone, Prof. Mariarosaria Falanga, Prof. Angelo Ciaramella, and PhD. Emanuel Di Nardo

Solar flares, sudden bursts of electromagnetic energy originating from magnetically active regions on the solar surface, pose significant risks to satellite infrastructure, communication systems, and power grids. Accurate forecasting of these events is crucial for advancing space weather prediction and safeguarding technological infrastructure. The interconnected nature of the Sun's atmospheric layers—from the corona to the lower photosphere—highlights the need for comprehensive data analysis techniques that leverage modern advancements in machine learning (ML) and physically informed models.

Traditional approaches have relied on features extracted from line-of-sight (LoS) magnetograms of solar active regions, historically linked to increased flare activity. However, recent studies employing LoS magnetogram time series have shown limited improvements, prompting the need for novel methodologies that integrate learning-based and physics-based insights.

To address this challenge, we present a deep learning-based framework for solar flare forecasting, leveraging the Solar Dynamics Observatory’s Helioseismic and Magnetic Imager (SDO/HMI) LoS magnetograms. Our model frames flare forecasting as a binary time series classification problem, aiming to distinguish active regions likely to produce M- or X-class flares within a 24-hour window. The approach integrates a Convolutional Neural Network (CNN) autoencoder for feature extraction and a Long Short-Term Memory (LSTM) binary classifier for flare activity prediction, achieving a 90% test accuracy.

By leveraging advanced ML techniques, this methodology demonstrates the potential of data-driven models in heliophysics. Our results highlight the transformative role of AI-powered science in advancing solar flare prediction and contributing to the development of reliable early warning systems for space weather forecasting.

How to cite: Doria Rosales, E., Carbone, P. V., Falanga, P. M., Ciaramella, P. A., and Di Nardo, PhD. E.: Machine Learning for Space Weather: Solar Flare Forecasting Using SDO/HMI Magnetogram Time Series, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9116, https://doi.org/10.5194/egusphere-egu25-9116, 2025.

EGU25-9149 | ECS | Orals | ESSI1.11

Space Weather Forecasts of Ground Level Space Weather with Machine Learning: Performance, Limitations and Challenges 

Andy Smith, Jonathan Rae, Colin Forsyth, John Coxon, Maria-Theresia Walach, Christian Lao, Shaun Bloomfield, Sachin Reddy, Mike Coughlan, Amy Keesee, and Sarah Bentley

Space weather describes the dynamic conditions in near-Earth space, mostly driven by the variable interaction between the continuous flow of the solar wind and the Earth’s magnetic field.  Extreme space weather has the potential to disrupt or damage key infrastructure on which we rely, for example through the generation of large, anomalous Geomagnetically Induced Currents (GICs) in power networks and transformers.  Accurately forecasting a risk of large GICs would enable key actions to be taken to mitigate their impact.

Given the sparsity of direct GIC measurements, and their inherent specificity to the contemporaneous network properties and configuration, we turn to forecasting the driving factor: the changing ground magnetic field (R).  In this talk we discuss a recent model developed to forecast whether the rate of change of the ground magnetic field (R) will exceed specific, high thresholds in the United Kingdom.  The model uses a common space weather forecasting framework: an interval of data from the upstream solar wind is used to make a prediction as to future conditions at the Earth.  We will use this model as an example to discuss forecasting performance, particularly with respect to different magnetospheric driving and processes.  We demonstrate the use of techniques such as SHAP (Shapley Additive exPlanations) to investigate how and why the model is making the predictions that it does.  What physical processes can this model set up capture?  Where do we need to go in the future?

How to cite: Smith, A., Rae, J., Forsyth, C., Coxon, J., Walach, M.-T., Lao, C., Bloomfield, S., Reddy, S., Coughlan, M., Keesee, A., and Bentley, S.: Space Weather Forecasts of Ground Level Space Weather with Machine Learning: Performance, Limitations and Challenges, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9149, https://doi.org/10.5194/egusphere-egu25-9149, 2025.

EGU25-9587 | ECS | Posters on site | ESSI1.11

Integrating Machine Learning and Solar Physics for Enhanced Prediction of CME Arrival Times and Near-Sun Solar Wind Conditions 

Yucong Li, Yi Yang, Fang Shen, Rongpei Lin, Haopeng Wang, and Stefaan Poedts

The timely and precise prediction of coronal mass ejection (CME) arrival times and the characterization of near-Sun solar wind conditions are essential for space weather forecasting and planetary sciences. We develop a novel deep-learning framework that integrates imaging observations and physical parameters to predict CME arrival times with improved accuracy. Using time-series data from synchronized solar white-light and EUV observations of 156 geoeffective CME events (2000–2020), we train two models: Model A, a convolutional neural network (CNN) regression model, and Model B, an enhanced version incorporating 11 key physical parameters of CMEs and background solar wind. Model B achieves a minimum mean absolute error (MAE) of 5.12 hours, a 33% improvement over Model A. This demonstrates the value of combining observational and physical data in forecasting CME arrival times.

In addition, we explore the use of GONG/ADAPT magnetograms with a U-Net-based architecture to model solar wind conditions at 0.1 AU. The training labels are derived from the COCONUT coronal model, which offers a potential acceleration in generating initial driving conditions for heliophysical models like ICARUS. While preliminary, this approach highlights a pathway to streamline the modeling of near-Sun solar wind environments, further supporting interplanetary CME propagation studies.

Our results underscore the potential of machine learning when synergized with solar physics to advance predictions critical to heliophysics and planetary sciences.

How to cite: Li, Y., Yang, Y., Shen, F., Lin, R., Wang, H., and Poedts, S.: Integrating Machine Learning and Solar Physics for Enhanced Prediction of CME Arrival Times and Near-Sun Solar Wind Conditions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9587, https://doi.org/10.5194/egusphere-egu25-9587, 2025.

EGU25-9849 | ECS | Orals | ESSI1.11

Efficient Segmentation and Clustering of Solar Coronal Structures: A Comparison of U-Net and Classical Computer Vision Techniques Using SDO Data 

Panagiotis Gonidakis, Francesco Carella, George Miloshevich, and Stefaan Poedts

Segmentation and characterization of solar coronal structures are essential for advancing our understanding of solar atmosphere and accurately identifying key regions, such as active regions and coronal holes, which are precursors to phenomena like solar flares and coronal mass ejections (CMEs). In this study, we investigate two complementary approaches to automate this process. First, we employ a previously presented deep learning-based U-Net architecture tailored for segmenting and characterizing solar coronal structures. Second, we develop a lightweight algorithm aimed at optimizing resource efficiency, consisting of classical computer vision techniques, which include thresholding and morphological filtering. The approach that best balances segmentation performance and computational efficiency will be selected for integration into a prototype designed to support future space exploration missions.

To characterize the segmented regions, we propose a set of carefully designed hand-crafted features to represent and characterize the resulting segmentations. These representations are analyzed using unsupervised clustering techniques, such as K-means and t-SNE, to distinguish solar coronal structures, including active regions, coronal holes and bright points.

Our dataset spans multiple layers of the solar atmosphere, incorporating HMI magnetograms (photosphere) and AIA wavelengths—94 Å (flaring regions), 171 Å (quiet Sun), 193 Å (coronal structures), and 304 Å (chromosphere). The performance of both segmentation approaches is thoroughly evaluated using metrics such as Dice score and Intersection over Union (IoU), with comparisons made against state-of-the-art methods.

Future work will focus on developing feature encoding techniques to better understand and predict solar phenomena, such as solar flare emissions, while investigating the impact of different feature extraction strategies on model performance.

References:

  • Galvez, Richard, et al. "A machine-learning data set prepared from the NASA solar dynamics observatory mission." The Astrophysical Journal Supplement Series 242.1 (2019): 7.
  • Šimon Mackovjak et al. “SCSS-Net: solar corona structures segmentation by deep learning”, Monthly Notices of the Royal Astronomical Society, Volume 508, Issue 3, December 2021, Pages 3111–3124, https://doi.org/10.1093/mnras/stab2536
  • Gonidakis, Panagiotis & Sóñora-Mengana, Alexander & Jansen, Bart & Vandemeulebroucke, Jef. (2023). Handcrafted Features Can Boost Performance and Data-Efficiency for Deep Detection of Lung Nodules From CT Imaging. IEEE Access. PP. 1-1. 10.1109/ACCESS.2023.3331315. 

 

How to cite: Gonidakis, P., Carella, F., Miloshevich, G., and Poedts, S.: Efficient Segmentation and Clustering of Solar Coronal Structures: A Comparison of U-Net and Classical Computer Vision Techniques Using SDO Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9849, https://doi.org/10.5194/egusphere-egu25-9849, 2025.

EGU25-10383 | Posters on site | ESSI1.11

Predicting characteristics of bursty bulk flows in Earth’s plasma sheet using machine learning techniques 

Xuedong Feng, Jian Yang, Jacob Bortnik, Chih-Ping Wang, and Jiang Liu

Bursty bulk flows (BBFs) play a crucial role in transporting energy, mass, and magnetic flux in the Earth's magnetotail, particularly in the earthward direction. However, their impulsive nature and small spatial scale present significant challenges for in-situ observation, as only a limited number of spacecraft operate within the vast expanse of the magnetotail. Consequently, studying their statistical characteristics is a highly demanding task, and accurately predicting their behavior remains a distant goal. In this study, we analyze key characteristics of BBFs and apply regression-based models to predict their parameter behaviorUsing observational data from the THEMIS mission collected between 2007 and 2023, we conducted a feature analysis on parameters associated with BBFs evolution, including velocity, magnetic field, electric field, temperature, density, pressure, and specific entropy indices. Through statistical techniques, we identified parameters exhibiting predictable patterns during BBF events, distinguishing them from background conditions. Furthermore, we used XGBoost regression model, optimized for different parameter combinations, to forecast BBF duration, physical parameters’ average minimum, and peak intensity. This study also tested combinations of parameter predictions across instruments. When using observed background value in parameter combination, our models achieved Mean Absolute Percentage Errors of under 35% for critical variables, including Bz, Btotal, plasma pressure, and ion temperatures, and ion specific entropy and so on. Additionally, we observed BBF duration’s spatial distribution trends: it peaked at approximately X=-13Re, while decreasing with increasing Z distance from the plasma sheet, showing dawn-dusk asymmetry consistent with prior observations. This work highlights the potential of regression methods in forecasting BBFs characteristics and offers insights into their spatial behavior, supporting enhanced prediction capabilities in magnetospheric studies. Future research will aim to improve accuracy with enriched datasets.

How to cite: Feng, X., Yang, J., Bortnik, J., Wang, C.-P., and Liu, J.: Predicting characteristics of bursty bulk flows in Earth’s plasma sheet using machine learning techniques, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10383, https://doi.org/10.5194/egusphere-egu25-10383, 2025.

EGU25-11565 | Orals | ESSI1.11

Parametrization of SHARP Vector Magnetic Field Using Disentangled Representation Learning 

Ekatarina Dineva, George Miloshevich, Giovanni Lapenta, Jasmina Magdalenic Zhukov, and Stefaan Poedts

The rapid growth of high-dimensional data in solar physics presents significant challenges for analysis and interpretation, making it an excellent domain for the application of machine learning (ML) algorithms. Synoptic full-disk observations with the Solar Dynamics Observatory (SDO)  provide continuous observations of the solar magnetic activity over more than one solar cycle, facilitating the study of solar variability and space weather impacts. The Space-weather HMI Active Region Patches (SHARP) vector magnetic field (VMF) maps and parameters, based on Helioseismic and Magnetic Imager (HMI) full-disk observations, are developed to study the magnetic evolution of individual active regions and flare triggering mechanisms. We present a method for active region parametrization by combining empirical parameters and ML-extracted features. Time series of SHARP VMF maps are used as input for the Disentangled Variational Autoencoder (VAE), a Disentangled Representation Learning (DRL) algorithm that facilitates the extraction of a low-dimensional feature representation. The VAE model is used to encode generalized information about nonlinear dynamical systems, i.e., a solar active region, aiming to isolate distinct factors of variation in the data, allowing a clearer interpretation of physical processes. We demonstrate how the ML features can be used to identify and study the stages of the magnetic patches evolution. These are benchmarked with SHARP parameters, relating empirical and learned features. Furthermore, the empirical dataset enhanced with ML features can be used to analyze the development of individual active regions and searching for eruption precursors.

How to cite: Dineva, E., Miloshevich, G., Lapenta, G., Magdalenic Zhukov, J., and Poedts, S.: Parametrization of SHARP Vector Magnetic Field Using Disentangled Representation Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11565, https://doi.org/10.5194/egusphere-egu25-11565, 2025.

EGU25-11587 | ECS | Orals | ESSI1.11

Solar Wind Speed Forecasting From Solar Images Using Distributional Regression  

Daniel Collin, Yuri Shprits, Stefan Hofmeister, Stefano Bianco, Nadja Klein, and Guillermo Gallego

The solar wind, a stream of charged particles originating from the Sun, poses significant risks to technology and astronauts. It is driven by large structures on the solar surface like coronal holes and active regions, which can be identified in extreme ultra-violet (EUV) solar images several days before they become geoeffective. In this work, we propose to use a distributional regression algorithm to forecast the solar wind speed at the Lagrange 1 point from solar images. Instead of predicting a single value, this method models the entire conditional distribution as a function of input features. It allows computing the uncertainty of predictions and specifying the probability of the solar wind speed exceeding certain thresholds, which is especially useful for extreme event predictions like coronal mass ejections and high-speed solar wind streams. We employ a convolutional neural network to encode solar images from multiple wavelength channels into unstructured low-dimensional representations. Using a semi-structured distributional regression approach, we couple the deep learning encoder with structured physical input parameters, such as past solar wind properties and solar cycle information. Thereby, we incorporate physical knowledge into the model and enhance explainability. We predict the solar wind speed distributions with a one-hour cadence four days in advance. We train and evaluate our method using cross-validation on 15 years of data and compare it to current state-of-the-art models. We find that it provides an accurate forecast and especially models the heavy-tailed solar wind speed distribution well. We further show the advantages over standard regression approaches and how to use the predicted conditional quantiles to improve extreme event predictions, highlighting the potential for operational space weather forecasts.

How to cite: Collin, D., Shprits, Y., Hofmeister, S., Bianco, S., Klein, N., and Gallego, G.: Solar Wind Speed Forecasting From Solar Images Using Distributional Regression , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11587, https://doi.org/10.5194/egusphere-egu25-11587, 2025.

EGU25-11680 | ECS | Orals | ESSI1.11

On Covariance Estimation in Physics Informed Neural Networks for Orbit Determination 

Fabian Dallinger, Benedikt Aigner, Thomas Andert, Benjamin Haser, Martin Pätzold, and Matthias Hahn

Artificial intelligence (AI), particularly machine learning (ML), is widely applied in fields such as medicine, autonomous driving, and manufacturing. Over time, ML has also seen increasing use in space and geosciences, where its algorithms hold the potential to enhance orbit prediction and orbit determination (OD) by utilizing measurement data. However, ML models like Artificial Neural Networks (ANNs) are limited to problems with abundant data and are often considered "black boxes", as their predictions lack interpretability in a scientifically meaningful way. To address these challenges, Raissi et al. 2018 introduced Physics Informed Neural Networks (PINNs), a specialized type of ANN. PINNs integrate the governing differential equations of a system into the learning process, imposing a physical constraint on the network's training and predictions. This approach allows effective training with small datasets, removing the reliance on large amounts of measurements. Additionally, PINNs can estimate unknown or poorly defined parameters within the differential equations, making them conceptually similar to classical OD algorithms like the Weighted Least Squares method. Building on this, Scorsoglio et al. 2023 successfully applied a variant of PINNs, called Physics Informed Extreme Learning Machines (PIELMs), for OD. In this study, a similar approach is employed for OD within the AI4POD (Artificial Intelligence for Precise Orbit Determination) software tool, focusing on resident space objects (RSOs) in low Earth orbit. Following this, we explore various methods, such as output perturbation, to determine the covariance matrix for the PINN-based OD approach. The covariance matrix provides an assessment of uncertainty in the predicted orbit and therefore being an essential tool in real space missions and collision avoidance. These methods are compared for their realism and effectiveness, both against each other and against the covariance matrix results from classical approaches. This study aims to evaluate whether the proposed methods can replicate and potentially improve upon traditional covariance estimation techniques.

How to cite: Dallinger, F., Aigner, B., Andert, T., Haser, B., Pätzold, M., and Hahn, M.: On Covariance Estimation in Physics Informed Neural Networks for Orbit Determination, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11680, https://doi.org/10.5194/egusphere-egu25-11680, 2025.

EGU25-11790 | ECS | Orals | ESSI1.11

AI-Enhanced Orbit Determination: The AI4POD Framework 

Benedikt Aigner, Fabian Dallinger, Thomas Andert, Benjamin Haser, Martin Pätzold, and Matthias Hahn

In recent years, the field of space situational awareness (SSA) has gained increasing attention, driven by the rapid rise in both active satellites and orbital debris. Therefore, being able to predict the orbit of a resident space object (RSO) as accurately as possible is more critical than ever in order to reduce collision risks and to preserve the orbital environment. However, incomplete knowledge of debris geometry, uncertain object characteristics, or simplified force models can cause prediction errors which exceed orders of several kilometers within just a few days, making it useless for reliable collision avoidance operations. Using modern Machine Learning (ML) algorithms can enhance prediction accuracy by addressing these challenges as recent studies have shown. In this context we present Artificial Intelligence for Precise Orbit Determination (AI4POD), a Python package that is designed to simplify the integration of ML-algorithms within the orbit prediction and determination process. AI4POD is structured as a comprehensive toolbox that includes a high-fidelity force model, various measurement functions, and classical orbit determination (OD) algorithms such as the batch least-squares estimation method. This integrated approach allows users to combine traditional orbit simulations with data-driven approaches to improve accuracy and to extend the predictability horizon. Based on this catalog, several approaches from artificial intelligence (AI) shall be tested in the future. Inspired by already proposed methodologies we are generating a training set of historical tracking data along with their corresponding orbit determinations using the AI4POD toolbox. Several machine learning algorithms will be explored to learn the nonlinear prediction errors, aiming to compensate for unmodeled or uncertain factors such as incomplete knowledge of satellite geometry or environmental conditions.

How to cite: Aigner, B., Dallinger, F., Andert, T., Haser, B., Pätzold, M., and Hahn, M.: AI-Enhanced Orbit Determination: The AI4POD Framework, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11790, https://doi.org/10.5194/egusphere-egu25-11790, 2025.

EGU25-12327 | Posters on site | ESSI1.11

Automated Identification of Auroral Luminosity Boundaries using pyIntensityFeatures 

Angeline Burrell, Gareth Chisham, Nicola Longden, and Kate Zawdie

Imagers that observe emissions from the atmosphere are commonly used to study various ionospheric phenomena.  These phenomena include the auroral oval, equatorial plasma bubbles, and travelling ionospheric disturbances.  A difficulty in using imager observations is accurately and automatically retrieving the locations of interest from these images.  We present an automated method designed to identify the auroral luminosity boundaries from space-based imager data.   These boundaries are important for high-latitude studies that use statistical or machine learning approaches, as geographic and magnetic coordinate systems that do not account for changes in the polar cap or equatorward auroral oval boundaries will mix together data from regions experiencing different types of coupling with the magnetosphere.

The boundary identification method was originally developed for the Imager for Magnetopause-to-Aurora Global Exploration (IMAGE) observations, and has been further adapted for use in a wider variety of situations.  We will discuss the updated detection method and demonstrate the process on two different satellite data sets.  The updated detection method will be made publicly accessible through a new Python package, pyIntensityFeatures.

How to cite: Burrell, A., Chisham, G., Longden, N., and Zawdie, K.: Automated Identification of Auroral Luminosity Boundaries using pyIntensityFeatures, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12327, https://doi.org/10.5194/egusphere-egu25-12327, 2025.

EGU25-12654 | Orals | ESSI1.11

Hybrid AI Approaches for Solar Feature Recognition Using Ground-Based Instrument Data 

Oleg Stepanyuk, Werner Pötzi, Kamen Kozarev, Momchil Dechev, and Rositsa Miteva
The dynamic behavior of solar prominences and filaments is a preursor to coronal mass ejections (CMEs), which can disrupt Earth's magnetosphere and affect satellite communications. Systematic ground-based solar observations, conducted with high temporal resolution, are instrumental in monitoring these structures. Analysis of the morphological changes and destabilization processes of filaments and prominences captured in datasets can help to identify early warning signs of potential eruptions. This capability is vital for developing reliable space weather forecasting systems, thereby mitigating the adverse effects of solar disturbances on Earth's technological infrastructure. Previously we introduced Wavetrack, a wavelet-based feature recognition software, which allowed, to a certain extent, to automate feature recognition for multiple events. We have since developed a convolutional neural network (CNN) model set which uses Wavetrack outputs as ground truth. Our initial model performance was shown on a set of SDO AIA instrument data performing segmentation of EUV and shock waves. In this work, we extend this hybrid approach for algorithmic and data-driven segmentation of on-disk solar features (prominences and filaments) using data from ground based-instruments, primarily focusing on Kanzelhöhe Observatory data. We discuss our approach to engineering training sets on real and synthetic data and the development of a CNN architecture generated within a general hyperparameter search routine. We showcase its performance on a set of filament/prominence events.

How to cite: Stepanyuk, O., Pötzi, W., Kozarev, K., Dechev, M., and Miteva, R.: Hybrid AI Approaches for Solar Feature Recognition Using Ground-Based Instrument Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12654, https://doi.org/10.5194/egusphere-egu25-12654, 2025.

EGU25-13602 | ECS | Orals | ESSI1.11

Identification of fast solar wind flows and CMEs in the in situ data using Self-Organizing Maps and clustering techniques 

Francesco Carella, Jasmina Magdalenić, and Alessandro Bemporad

The identification and characterization of the coronal mass ejections (CMEs) and fast solar wind flows in the in situ data are important for understanding dynamics of these phenomena and consequently for space weather forecasting. In this study, we apply Self-Organizing Maps (SOMs) and clustering techniques to analyze in situ solar wind observations. SOMs (Kohonen, T, 1982) [1] an unsupervised learning technique, is employed to project high-dimensional interplanetary plasma parameters such as velocity, density, temperature, and magnetic field onto a lower-dimensional representation, preserving the topological structure of the data. Clustering algorithms, such as k-means, are then applied to the SOM output to distinguish between ICME events, fast and slow solar wind flows.
Our approach is validated using a few months long interval of the ACE and Wind in situ observations, with labeled CME intervals from Richardson and Cane [2] as a benchmark. This combination of SOMs and clustering provides a framework for automated identification of interplanetary plasma structures, important for space weather studies but also for operational services. 

[1] T. Kohonen, ‘Self-organized formation of topologically correct feature maps’, Biol. Cybern., vol. 43, no. 1, pp. 59–69, Jan. 1982, doi: 10.1007/BF00337288
[2] Richardson, Ian; Cane, Hilary, 2024, "Near-Earth Interplanetary Coronal Mass Ejections Since January 1996"https://doi.org/10.7910/DVN/C2MHTH

How to cite: Carella, F., Magdalenić, J., and Bemporad, A.: Identification of fast solar wind flows and CMEs in the in situ data using Self-Organizing Maps and clustering techniques, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13602, https://doi.org/10.5194/egusphere-egu25-13602, 2025.

EGU25-14036 | ECS | Orals | ESSI1.11

Using Transformers to Integrate Irregular Data for Improved Ionospheric Modeling 

Liam Smith and Morris Cohen

The ionosphere has important impacts on many different systems, such as communications, thus modeling it is a crucial task. The influence of the ionosphere is closely linked to its electron density, but this is difficult to measure adequately. Because of this, modeling requires the use of additional correlated values, such as solar activity metrics. These measures do not capture enough to reproduce small-scale changes in electron density, so we have developed a technique to expand our input space to include sparse measurements of Total Electron Content (TEC), or the integral of electron density.

TEC data is measured more densely than electron density, although it is still not consistent spatially, with many gaps in measurement coverage. Despite this, it is collected very consistently throughout time so it presents itself as a good candidate for an input to an ionospheric model. Even so, TEC has not been used as an input to such models, especially Machine Learning (ML) models, as the irregular coverage of the measurements makes it difficult to deal with.

We have developed a technique to use transformer-like architectures to move from an irregular domain to a fixed size embedded domain to facilitate further usage of the TEC data. This approach has enabled us to use TEC as a direct input to electron density models, noticeably improving performance. Our technique also enables the use of a variety of irregular inputs all at once, enabling a wider range of possible model inputs. Lastly, as a byproduct of the process, we can use the inverse of our embedding technique (which is also how we train the model) to perform TEC map completion, where we can predict TEC values even where no measurements have been taken.

How to cite: Smith, L. and Cohen, M.: Using Transformers to Integrate Irregular Data for Improved Ionospheric Modeling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14036, https://doi.org/10.5194/egusphere-egu25-14036, 2025.

EGU25-14724 | Posters on site | ESSI1.11

Subgrid-scale modeling of MHD turbulence 

Dmitri Kondrashov and Anthony Sciola

Numerical magnetohydrodynamic models (MHD) are often used to simulate the global interaction between the solar wind and the magnetosphere system. Increasingly, such MHD models require very computationally expensive, high numerical resolutions for realistic global magnetosphere simulations of multiscale turbulent plasma flows. To address this problem, we investigate and compare several ML-based approaches for subgrid-scale (SGS) parameterizations in the coarse-scale Grid Agnostic MHD for Extended Research Applications (GAMERA) model, starting with the Large-Eddy Simulation (LES) formalism. We use a 2D simulation of MHD turbulence in the Orszag-Tang vortex as a testbed to diagnose from benchmark high-resolution GAMERA  solutions the distributions of subgrid-scale (SGS) and large-scale (LS) fields, and model subgrid-scale (SGS) forcing that encapsulates induced feedbacks on the LS fields. 

How to cite: Kondrashov, D. and Sciola, A.: Subgrid-scale modeling of MHD turbulence, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14724, https://doi.org/10.5194/egusphere-egu25-14724, 2025.

EGU25-14910 | Orals | ESSI1.11

Ambient Solar Wind Speed Forecast with Physics-Informed Machine Learning  

Enrico Camporeale and Andong Hu

We present a novel physics-informed machine learning (ML) model designed to forecast the background (ambient) solar wind up to five days in advance. Solar wind speed is a critical driver of geomagnetic activity, and inaccuracies in its prediction significantly contribute to large errors in forecasting the arrival times of coronal mass ejections (CMEs), which are typically off by at least 10 hours.

Predicting solar wind speed has historically been a challenging task, with even state-of-the-art models often failing to consistently outperform a simple 27-day persistence model. Operational physics-based (3D MHD) models, in particular, struggle to accurately forecast high-speed streams associated with co-rotating interaction regions. These regions arise from fast solar wind generated by coronal holes, which are not clearly captured in the magnetogram maps routinely used as inputs. While recent empirical and data-driven methods have shown relatively better performance, significant challenges remain.

Our approach integrates lessons from prior models into what we believe represents the current state-of-the-art. Specifically, we use GONG synoptic maps (magnetograms) and full-disk SDO EUV images as inputs to a neural network. This network estimates the optimal inner boundary condition for the radial solar wind velocity profile at 10 solar radii, which is then propagated to 1 AU using a simplified 1D hydrostatic model.

The key innovation lies in seamlessly integrating the physics-based model within the neural network, creating a true physics-informed ML framework.

We will present validation metrics to assess the model’s performance and discuss plans to make the forecast outputs available to the community 24/7 via the swx-trec.com portal.

How to cite: Camporeale, E. and Hu, A.: Ambient Solar Wind Speed Forecast with Physics-Informed Machine Learning , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14910, https://doi.org/10.5194/egusphere-egu25-14910, 2025.

EGU25-15985 | ECS | Posters on site | ESSI1.11 | Highlight

TRANSCENDENCE - A TRANSit Capture ENgine for DEtection and Neural network Characterization of Exoplanets 

Hendrik Schmerling, Rok Hribar, Sascha Grziwa, and Martin Pätzold

Although the search for exoplanets currently incorporates various computational methods, it still heavily relies on manual analysis of light curves, a process that is both time-intensive and demanding. Our research in the EXOWORLD project addresses these challenges by integrating advanced machine learning techniques, including convolutional, into the transit search process, combining them with recurrent networks to create a fully integrated machine learning-based transit detection and characterization pipeline. This approach reimagines transit search as a pattern recognition task, employing self-learning algorithms to efficiently process vast amounts of astronomical data. We aim to explore and apply a range of machine learning methods, establishing a foundation for comparison not only among these methods but also against traditional transit search techniques. This comparison is expected to focus on potential improvements in efficiency, accuracy, and computational demands. Although still in the early stages, our research aims to significantly enhance exoplanet detection methods, streamlining the process and building a framework for making new discoveries through light curve analysis.

In this context, we present TRANSCENDENCE, our machine learning-based pipeline, which has demonstrated the ability to identify exoplanets larger than 2 Earth radii consitently. Moreover, the pipeline is capable of detecting smaller planets, albeit with lower detection probabilities. One of TRANSCENDENCE's key strengths lies in its remarkably low false positive rate, which ranges between 5% and 10% of all identified transits. By significantly reducing the need for manual intervention and minimizing false positives, this pipeline has the potential to strongly immprove the efficiency of exoplanet detection and characterization.

 

How to cite: Schmerling, H., Hribar, R., Grziwa, S., and Pätzold, M.: TRANSCENDENCE - A TRANSit Capture ENgine for DEtection and Neural network Characterization of Exoplanets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15985, https://doi.org/10.5194/egusphere-egu25-15985, 2025.

EGU25-16713 | ECS | Posters on site | ESSI1.11

Machine Learning Algorithms for Autonomous Space Mission Operations 

Tommaso Torda, Tommaso Alberti, Giuseppe Consolini, Rossana De Marco, Ekaterina Dineva, Jonah Ekelund, Panagiotis Gonidakis, Monica Laurenza, Maria Federica Marcucci, Stefano Markidis, George Miloshevich, Stefaan Poedts, Begnamino Sanò, and Nicolina Chrysaphi

The Automatics in SpAce exploration (ASAP) project has as a goal the design and development of Machine Learning algorithms for the automation of operations to be implemented on the on-board processors of space missions. In the framework of ASAP a set of ML algorithms for on-board science operations of space missions have been developed/optimized on consumer-grade computing systems to be further selected for orting of existent ML models directly on an FPGA prototype. In more detail, algorithms pertaining to four main use cases have been considered: the autonomous triggering of special measurement modes and the selective downlink of plasma environment parameters; the advanced on-board data analysis of three-dimensional particle distribution functions; the on-board analysis of solar images; the on-board prediction capability of SEP related hazards. Here we describe the algorithms, their performances and requirements for the on-board implementation. ASAP has received funding from the EU’s HORIZON Research and Innovation Action (GA no.101082633)

How to cite: Torda, T., Alberti, T., Consolini, G., De Marco, R., Dineva, E., Ekelund, J., Gonidakis, P., Laurenza, M., Marcucci, M. F., Markidis, S., Miloshevich, G., Poedts, S., Sanò, B., and Chrysaphi, N.: Machine Learning Algorithms for Autonomous Space Mission Operations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16713, https://doi.org/10.5194/egusphere-egu25-16713, 2025.

Remote sensing observations, whether astronomical or within the solar system, are constrained by instrumental limitations, such as the point spread function in imaging. Ensuring the reliability of scientific analysis from such data requires robust deconvolution techniques. We present a spatio-temporal deconvolution method, to minimise the effect of an extended or complex-shaped point spread function, applicable to dynamic systems with various timescales. This approach enhances observational data by improving image contrast and resolving small-scale dynamic features.

Our method employs a deep neural network trained on state-of-the-art numerical simulations, enabling it to identify dynamic patterns in both spatial and temporal dimensions and to estimate and correct the degradation of intensity contrast. The resulting improvements in intensity representation and resolution facilitate more accurate analyses of small-scale features.

We apply this methodology to solar observations in the millimeter wavelength regime, recovering fine-scale structures critical for understanding the complex behaviour of the solar atmosphere, predict the generation of potentially harmful events, solar flares and the solar wind. By incorporating the temporal domain, our approach surpasses traditional 2D deconvolution techniques.

While initially developed for solar imaging, the method is versatile and can be adapted to various observational contexts across different wavelength regimes. This makes it a valuable tool for advancing future observational studies and expanding research capabilities.

How to cite: Eklund, H.: Spatio-temporal deconvolution method for enhanced image analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17531, https://doi.org/10.5194/egusphere-egu25-17531, 2025.

EGU25-18050 | ECS | Posters on site | ESSI1.11

Modeling Magnetic Storms' Dynamics with Physics-Informed Neural Networks 

Manuel Lacal, Enrico Camporeale, Giuseppe Consolini, and Mirko Piersanti

Solar activity significantly influences the near-Earth environment, leading to magnetic storms and magnetospheric substorms that can impact both technological and human systems. Understanding the physical processes that govern the Sun-Earth relationship and developing models to forecast magnetic disturbances on Earth are therefore of critical importance. In this context, we present a preliminary work to model and forecast the dynamics of magnetic storms, as measured by the SYM-H geomagnetic index, using Physics-Informed Neural Networks (PINNs). This approach is applied to models based on deterministic ordinary differential equations (ODEs), such as those described by Burton et al. (1975) and others, which were proposed to describe the evolution of geomagnetic indices during magnetic storms. The findings and significance of this approach are discussed in the context of Earth's magnetospheric dynamics and the relevance of PINN techniques in space weather research.

How to cite: Lacal, M., Camporeale, E., Consolini, G., and Piersanti, M.: Modeling Magnetic Storms' Dynamics with Physics-Informed Neural Networks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18050, https://doi.org/10.5194/egusphere-egu25-18050, 2025.

EGU25-18475 | ECS | Orals | ESSI1.11

Prediction of Solar Surface Magnetic Fields Using an AI-based Surface Flux Transport Model 

Hyun-Jin Jeong, Mingyu Jeon, Daeil Kim, Youngjae Kim, Ji-Hye Baek, Yong-Jae Moon, and Seonghwan choi

In this study, we develop an artificial intelligence (AI)-based solar surface flux transport (SFT) model. We predict global magnetic field distributions on the solar surface up to the next solar rotation (27.3 days) using deep-learning. Here we train and evaluate our deep-learning model, based on the Pix2PixCC architecture, using data sets of SDO/HMI, SOHO/MDI, and NSO/GONG synoptic maps with a resolution of 360 by 180 (longitude and sine-latitude) from 1996 to 2023. We present results of our model and compare them with those from the persistence model and the conventional SFT model, including the effects of differential rotation, meridional flow, and diffusion on the solar surface. Our AI-based SFT model generates magnetic field distributions for the next solar rotation, better than the conventional SFT model and the persistence model in the quantitative metrics such as RMSE, FSIM, and pixel-to-pixel CC. Our model successfully generates magnetic features, such as the diffusion of solar active regions and the motions of supergranules. Our model also generates small-scale magnetic features better than the conventional SFT models. Using synthetic input data with bipolar structures, we confirm that our model successfully reproduces differential rotation and meridional flow. Finally, we discuss the advantages and limitations of our model in view of magnetic field evolution and its potential applications.

How to cite: Jeong, H.-J., Jeon, M., Kim, D., Kim, Y., Baek, J.-H., Moon, Y.-J., and choi, S.: Prediction of Solar Surface Magnetic Fields Using an AI-based Surface Flux Transport Model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18475, https://doi.org/10.5194/egusphere-egu25-18475, 2025.

EGU25-20652 | ECS | Orals | ESSI1.11

A Deep-learning-based Model of the Three-dimensional Ion Flux in the Earth’s Northern Cusp 

Gonzalo Cucho-Padin, David Sibeck, Daniel Da Silva, and Xueyi Wang

Magnetic reconnection on the dayside magnetopause is considered the primary mechanism for transporting mass, momentum, and energy from the solar wind into the terrestrial magnetosphere. Several studies have demonstrated that the spatiotemporal dynamics of the dayside magnetic reconnection can be inferred remotely from the analysis of the time-energy dispersion of ions in the Earth’s cusps. Despite the immense number of in-situ cusp measurements acquired by numerous space-based instruments, it is still challenging to determine the overall cusp behavior owing to the intermittency of the measurement acquisition. To overcome this issue, this work implements a regression model of the three-dimensional (3-D) ion flux in the Earth’s Northern cusp based on deep learning techniques and numerous measurements of the cusp under varying solar wind conditions. For the training process, we have used solar wind parameters obtained from NASA's OMNI database as input and in-situ ion flux measurements acquired by the CIS/HIA instruments on board ESA’s multi-spacecraft Cluster mission during the period from 2001 to 2010 for supervised output. The model allows the reconstruction of the time-dependent, 3-D ion flux distribution within the cusp region, which serves to determine the boundaries of the high-altitude cusp, analyze its structural response to time-dependent solar wind conditions, and investigate the relationship between the cusp and dayside magnetic reconnection. The experiments under controlled input parameters show that our model is capable of reproducing  expected ion dispersion signatures as a response to variable solar wind conditions.

How to cite: Cucho-Padin, G., Sibeck, D., Da Silva, D., and Wang, X.: A Deep-learning-based Model of the Three-dimensional Ion Flux in the Earth’s Northern Cusp, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20652, https://doi.org/10.5194/egusphere-egu25-20652, 2025.

EGU25-872 | ECS | Posters on site | CL4.8

A new orographic drag parameterization package for the GLOBO model: implementation and evaluation  

Guido Davoli, Daniele Mastrangelo, Annalisa Cherchi, and Andrea Alessandri

Orography plays a fundamental role in shaping the atmospheric circulation and affects key atmospheric processes. Therefore, weather and climate models must adequately represent its effects to obtain accurate predictions. Since all orographic scales are found to influence the atmospheric flow, the parameterization of unresolved orographic drag has been recognized as crucial to simulate a realistic mid-latitude circulation. Moreover, in the last few years, it has become clear that orographic gravity wave drag (OGWD) and turbulent orographic form drag (TOFD) parameterization schemes play a crucial role in reducing some of the long-standing circulation biases affecting climate models. However, they are still considered a potential source of errors, due to the uncertainties which affect some poorly constrained physical parameters. Furthermore, these schemes need boundary conditions suitable to characterize the physical features of sub-grid orography. The strategies for the generation of such boundary conditions can vary a lot between different modelling centres, and it has been shown to be an important source of uncertainty. 

GLOBO is a global atmospheric general circulation model developed at the Institute for Atmospheric Science and Climate of the Italian National Research Council (ISAC-CNR). It is currently in use within many operational frameworks, including a global monthly probabilistic forecast system that contributes to the Subseasonal-to-seasonal (S2S) project database. In an effort to improve and modernize the model, we implemented a novel orographic drag parameterization package, based on state-of-the-art OGWD and TOFD schemes. Simultaneously with the development of the orographic drag parameterizations, we developed a novel software package, OROGLOBO (OROGraphic ancillary files generator for GLOBal atmospheric mOdels) designed for the generation of the orographic boundary conditions. This unique open-source tool is designed to exploit a state-of-the-art, high resolution global Digital Elevation Model to generate boundary conditions for OGWD and TOFD schemes, gathering the main algorithms and techniques available in the literature in a single software. 

Here, we present the results of this model update. A new set of retrospective forecasts was performed, consisting of an 8-members ensemble, initialized every 5 days and integrated for 35 days, during the period 2001-2020, including the developments in orographic physical parameterization and boundary conditions. This set of simulations is compared to the corresponding hindcasts set performed with the standard model configuration and used to calibrate the operational ensemble of global sub-seasonal probabilistic forecasts. We evaluate the impact of the improved representation of unresolved orographic drag on the simulation and prediction of the Northern Hemisphere mid-latitudes circulation. We assess the change in prediction skill for atmospheric blocking events and associated extreme temperature and wind conditions. 

How to cite: Davoli, G., Mastrangelo, D., Cherchi, A., and Alessandri, A.: A new orographic drag parameterization package for the GLOBO model: implementation and evaluation , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-872, https://doi.org/10.5194/egusphere-egu25-872, 2025.

EGU25-1040 | ECS | Orals | CL4.8

Understanding Soil Modulation of Drought Persistence in CMIP6 Models 

Marco Possega, Emanuele Di Carlo, Vincenzo Senigalliesi, and Andrea Alessandri

 Drought persistence is a critical factor in assessing water availability and its impacts on agriculture, ecosystems, and society. In this respect, poorly constrained soil properties in climate models such as field capacity – i.e. the maximum water a soil can retain after drainage of excess moisture – may strongly affect severity and persistence of simulated soil drought conditions. This study examines for the first time the regulating role of soil properties, particularly of field capacity, in shaping drought memory and its broader impacts. Using the CMIP6 multi-model ensemble and observations, we analyze drought dynamics across various phases of the hydrological cycle applying non-parametric standardized indices: Standardized Precipitation Index (precipitation deficits), Standardized Precipitation-Evapotranspiration Index (precipitation-evapotranspiration balance), Standardized Soil Moisture Index (soil moisture deficits), and Standardized Runoff Index (reduced runoff). Our analysis investigates the persistence between hydrological drought indicators, showing that soils with greater field capacity sustain drought conditions longer, emphasizing the importance of accurately modeling soil properties to capture drought persistence effectively. The historical CMIP6 simulations are compared with observational datasets, including GLEAM and CRU, to assess the deviation between model outputs and observed climate conditions. The future scenarios (SSP126, SSP245, SSP370, SSP585) are also examined, revealing significant regional differences in projected drought behavior depending on the degree of radiative-forcing increase during 21st century. High-emission scenarios project prolonged drought conditions due to increased temperatures and evapotranspiration feedback, while low-emission pathways are effective in preserving more stable hydrological dynamics. Our results show that, in water limited and transition areas such as the Euro-Mediterranean region, the persistence of droughts and its projected change considerably depend on the modeled field capacity. This study highlights the essential role of field capacity and other soil characteristics in regulating the variability and the persistence of drought events. By bridging historical validation with future projections, it provides a comprehensive understanding of drought dynamics and trends, also identifying observational constraints for the Earth System Models. These findings are crucial for refining predictions of agricultural and hydrological drought impacts and for guiding adaptation strategies in water-limited regions that are vulnerable to drought exacerbation under climate change.

How to cite: Possega, M., Di Carlo, E., Senigalliesi, V., and Alessandri, A.: Understanding Soil Modulation of Drought Persistence in CMIP6 Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1040, https://doi.org/10.5194/egusphere-egu25-1040, 2025.

EGU25-1545 | Posters on site | CL4.8

Sub-seasonal to Seasonal Arctic Summer Sea Ice Forecasts Using Dynamical Downscaling with the Regional Arctic System Model 

Younjoo Lee, Wieslaw Maslowski, Anthony Craig, Jaclyn Clement Kinney, and Robert Osinski

The Arctic region has been warming at a rate significantly faster than the global average, leading to an accelerated decline in sea ice. This trend is expected to continue, potentially resulting in a "low-ice regime," which could make sea ice conditions more unpredictable. Anticipating changes in Arctic sea ice and climate states is therefore crucial for guiding various human activities, from natural resource management to risk assessment decisions. While global climate and Earth system models project continuous sea ice decline over decadal time scales, achieving reliable seasonal forecasts remains challenging. To address this, we apply dynamical downscaling with the state-of-the-art Regional Arctic System Model (RASM), which enables us to forecast Arctic sea ice on time scales ranging from weeks to six months. RASM is a fully coupled regional climate model that integrates components for the atmosphere, ocean, sea ice, and land, interconnected through the flux coupler of the Community Earth System Model. In our study, we simulate RASM at a horizontal resolution of 1/12 degree (approximately 9 km) for both the ocean and sea ice, with 45 vertical levels in the ocean and five thickness categories for sea ice. The atmosphere is configured on a 50-km grid with 40 vertical levels, dynamically downscaled from the NOAA/NCEP Climate Forecasting System version 2 (CFSv2) at 72-hour intervals for the upper half of the atmosphere. Monthly ensemble forecasts extending up to six months are generated using initial conditions derived from a fully-coupled RASM hindcast simulation without bias correction and assimilation. This presentation highlights results for September sea ice predictions initialized on April 1, May 1, June 1, July 1, August 1, and September 1, covering pan-Arctic and regional sea ice spatio-temporal conditions from 2012 to 2021. Specifically, we examine how lead time and initial conditions affect the quantitative skill of seasonal predictability for Arctic sea ice and demonstrate skillful predictions of September sea ice up to six months in advance. Overall, our study underscores that enhancing model physics and obtaining more realistic initial conditions are crucial for achieving skillful sub-seasonal to seasonal predictions.

How to cite: Lee, Y., Maslowski, W., Craig, A., Clement Kinney, J., and Osinski, R.: Sub-seasonal to Seasonal Arctic Summer Sea Ice Forecasts Using Dynamical Downscaling with the Regional Arctic System Model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1545, https://doi.org/10.5194/egusphere-egu25-1545, 2025.

EGU25-1847 | Orals | CL4.8

AI deep learning for climate forecasts 

Jing-Jia Luo

AI deep learning for climate science has attracted increasing attentions in recent years with rapidly expanded applications to many areas. In this talk, I will briefly present our recent progresses on using various deep learning methods for seasonal-to-multi-seasonal predictions of ENSO, the Indian Ocean Dipole (IOD), summer precipitation in China and East Africa, Arctic sea ice cover, ocean waves, as well as the bias correction and downscaling of dynamical model’s forecasts. The results suggest that many popular deep learning methods, such as convolutional neural networks, residual neural network, long-short term memory, ConvLSTM, multi-task learning, cycle-consistent generative adversarial networks and vision transformer, can be well applied to improve our understanding and predictions of climate. In addition, a brief introduction of AI large models for ensemble weather-subseasonal-seasonal-decadal forecasts, together with the perspective on the future development of AI methods, will also be presented.

How to cite: Luo, J.-J.: AI deep learning for climate forecasts, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1847, https://doi.org/10.5194/egusphere-egu25-1847, 2025.

EGU25-2210 | ECS | Posters on site | CL4.8

The role of Pacific Tropical Instability Wave in Sub-Seasonal SST predictability  

Li Tianyan, Yu Yongqiang, and Zhen Weipeng

Tropical Instability Waves (TIWs) play a crucial role in modulating Sea Surface Temperature (SST) variability in tropical oceans, yet their representation in current forecast systems remains challenging. This study investigates the relationship between TIWs and sub-seasonal SST predictability while evaluating the performance limitations of the Licoms Forecast System. Through comprehensive analysis of observational data and model outputs, we demonstrate that TIWs provide significant potential for enhancing sub-seasonal SST forecast skill through their regular wave patterns and predictable evolution characteristics. However, our findings reveal that the current Licoms forecast systems systematically underestimate both TIW intensity and wavelength. Critical examination of error sources indicates that these deficiencies primarily originate from initialization fields rather than model physics or dynamics. 

How to cite: Tianyan, L., Yongqiang, Y., and Weipeng, Z.: The role of Pacific Tropical Instability Wave in Sub-Seasonal SST predictability , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2210, https://doi.org/10.5194/egusphere-egu25-2210, 2025.

EGU25-3974 | Orals | CL4.8

Increased multi-year ENSO predictability under greenhouse gas warming accounted by large ensemble simulations and deep learning 

Young-Min Yang, Jae-Heung Park, June-Yi Lee, Soon-Il An, Sang-Wook Yeh, Jong-Seoung Kug, and Yoo-Geun Ham

The El Niño/Southern Oscillation (ENSO) is the primary internal climatic driver shaping extreme events worldwide1,2,3. Its intensity and frequency in response to greenhouse gas (GHG) warming has puzzled scientists for years, despite consensus among models about changes in average conditions4-16. Recent research has shed light on changes not only in ENSO variability5,7,8,10,13, but also in the occurrence of extreme5,6,11,12,13,14 and multi-year El Niño4,15, and La Niña9,11,16 events under GHG warming. Here, we investigate potential changes in ENSO predictability associated with changes in ENSO dynamics in the future by using long-range deep-learning forecasts trained on extensive large ensemble simulations of Earth System Models under historical forcings and the future high GHG emissions scenario. Our results show a remarkable increase in the predictability of ENSO events, ranging from 35% to 65% under the high GHG emissions scenario due to reduced ENSO irregularity, supported by a broad consensus among multi-models. Under GHG warming, an El Nino-like warming flattens the thermocline depth with upper ocean stratification. This flattening of the thermocline depth leads to an increased transition frequency between El Niño and La Niña events, driven by strengthened recharge-discharge oscillation with enhanced thermocline feedback and SST responses to zonal wind stress. As a result, ENSO complexity would reduce with increased regularity and reduced skewness, increasing ENSO predictability. These results imply that the future social and economic impacts of ENSO events may be more manageable, despite an expected increase in the frequency of extreme ENSO events.

How to cite: Yang, Y.-M., Park, J.-H., Lee, J.-Y., An, S.-I., Yeh, S.-W., Kug, J.-S., and Ham, Y.-G.: Increased multi-year ENSO predictability under greenhouse gas warming accounted by large ensemble simulations and deep learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3974, https://doi.org/10.5194/egusphere-egu25-3974, 2025.

EGU25-5233 | Orals | CL4.8

Standardisation of equitable climate services by supporting a community of practice 

Francisco Doblas-Reyes, Asun Lera St Clair, Marina Baldissera Pacchetti, Paula Checchia, Joerg Cortekar, Judith E.M. Klostermann, Werner Krauß, Angel Muñoz, Jaroslav Mysiak, Jorge Paz, Marta Terrado, Andreas Villwock, Mirjana Volarev, and Saioa Zorita

Climate services are essential to support climate-sensitive decision making, enabling adaptation to climate change and variability, and mitigate the sources of anthropogenic climate change, while considering the values and contexts of those involved. The unregulated nature of climate services can lead to low market performance and lack of quality assurance. Best practices, guidance, and standards serve as a form of governance, ensuring quality, legitimacy, and relevance of climate services. The Climateurope2 project (www.climateurope2.eu) addresses this gap by engaging and supporting an equitable and diverse community of climate services to provide recommendations for their standardisation. Four components of climate services are identified (the decision context, the ecosystem of actors and co-production processes, the multiple knowledge systems involved, and the delivery and evaluation of these services) to facilitate analysis. This has resulted in the identification of nine key messages summarising the susceptibility for the climate services standardisation. The recommendations are shared with relevant standardisation bodies and actors as well as with climate services stakeholders and providers.

How to cite: Doblas-Reyes, F., Lera St Clair, A., Baldissera Pacchetti, M., Checchia, P., Cortekar, J., Klostermann, J. E. M., Krauß, W., Muñoz, A., Mysiak, J., Paz, J., Terrado, M., Villwock, A., Volarev, M., and Zorita, S.: Standardisation of equitable climate services by supporting a community of practice, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5233, https://doi.org/10.5194/egusphere-egu25-5233, 2025.

EGU25-7271 | ECS | Posters on site | CL4.8

Impact-Based Forecasting Model for Flood Hazard Mitigation in Java, Indonesia 

Dendi Rona Purnama, Simon F. B. Tett, Ruth Doherty, and Ida Pramuwardani

Flooding is the most frequent and damaging hydrometeorological disaster in Indonesia, with Java being especially vulnerable due to its dense population and rapid urbanization. This study aims to refine the Impact-Based Forecasting (IBF) model to improve flood hazard predictions and mitigation efforts. Using Global Precipitation Measurement (GPM-IMERG) rainfall data as the hazard component combined with vulnerability and capacity datasets from InaRISK, this research focuses on enhancing the precision and reliability of impact assessments.

Initial analyses highlight the potential of impact-based rainfall thresholds and assessment probabilistic impacts to refine the IBF model and reduce subjectivity in impact assessments. By linking calculated impact values and disaster magnitudes for the 2014 – 2023 period, this study shows a promising skill for significant and severe flood events, although improvements are needed for minor and minimal disaster classifications.

This research lays the groundwork for a more robust and scalable IBF model tailored to Java’s unique challenges. The findings aim to support BMKG’s operational needs, enabling the delivery of more actionable early warnings and targeted disaster preparedness measures. By addressing critical gaps in existing IBF systems, this study contributes to bridging the divide between hazard-impact forecasts and societal resilience, ultimately mitigating the impacts of floods in Indonesia.

How to cite: Purnama, D. R., Tett, S. F. B., Doherty, R., and Pramuwardani, I.: Impact-Based Forecasting Model for Flood Hazard Mitigation in Java, Indonesia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7271, https://doi.org/10.5194/egusphere-egu25-7271, 2025.

EGU25-9933 | Orals | CL4.8

Hydroclimate services are more than just providing data 

Jean-Philippe Vidal, Eric Sauquet, Louis Héraut, Sonia Siauve, Guillaume Evin, Jean-Michel Soubeyroux, Flore Tocquer, Audrey Bornançin-Plantier, Claire Magand, and Maud Berel

The concept of hydroclimate services is predominantly recognised as web portals dedicated to the dissemination of data to potential users. However, the scope of climate services extends beyond the sole provision of data. This communication presents a comprehensive ecosystem of tools and resources associated with the development of an updated national hydrological projection dataset in France. The ecosystem was brought to life through a close collaboration between scientists and water managers in two joint projects: Explore2 and LIFE Eau&Climat. Tools and resources were thus developped with and for water resource managers, and designed to enhance the comprehension of both the conceptual framework and the data itself, facilitating utilisation in accordance with best practices for climate change adaptation.

The project websites serve as gateways to the ecosystem and the tools: the Explore2 website contains interviews with the scientific contributors, and the LIFE Eau&Climat website is hosted by the national website dedicated to water managers. A summary of the joint final public event accompanies the replay of the one-day conference and debates on a dedicated website. A compendium of antecedent research projects on climate change impacts on hydrology has been collated to summarise the state of the art prior to the two projects. A MOOC has been developed in conjunction with scientists to facilitate the comprehension of the Explore2 project, its design, and its application in adaptation studies.

Moreover, the Explore2 dataverse (https://entrepot.recherche.data.gouv.fr/dataverse/explore2) brings together a variety of products in an organised and searchable way, including thematic scientific reports, GIS layers, and other key metadata. It also contains three types of station datasheets aimed at locally contextualising outputs: hydrological model performance datasheets, projection results datasheets, and uncertainty quantification datasheets. The MEANDRE interactive data visualisation tool (https://meandre.explore2.inrae.fr/) offers a guided tour of the salient take-home messages and a comprehensive exploration of the Explore2 hydrological projection dataset. This multi-model dataset (GCMs/RCMs/bias correction methods/hydrological models) is made available through the DRIAS-Eau portal (https://drias-eau.fr/), which functions as a water mirror of the established DRIAS-Climat portal. The utilisation of this dataset for local climate change impact studies is facilitated by a methodological guide written as an adventure gamebook (https://livreec.inrae.fr/) and based on real-life studies carried out by water managers during the LIFE Eau&Climat project. Furthermore, experiments of sonification of hydrological projections offer a novel approach to apprehending future changes (https://explore2enmusique.github.io/).

This ecosystem has been met with great anticipation and acclaim by local to national-scale water managers, paving the way for ongoing local prospective studies. These will be able to confront future resources with the ecological needs of aquatic environments and human water usage.

This work is funded by the EU LIFE Eau&Climat project (LIFE19 GIC/FR/001259).

How to cite: Vidal, J.-P., Sauquet, E., Héraut, L., Siauve, S., Evin, G., Soubeyroux, J.-M., Tocquer, F., Bornançin-Plantier, A., Magand, C., and Berel, M.: Hydroclimate services are more than just providing data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9933, https://doi.org/10.5194/egusphere-egu25-9933, 2025.

EGU25-11600 | ECS | Orals | CL4.8

Windows of Opportunity for Seasonal Prediction of droughts: the case of the Middle East 

Thomas Dal Monte, Andrea Alessandri, Annalisa Cherchi, Markus Donat, and Marco Gaetani

Drought warnings are vital to sectors like agriculture and water management, especially at the seasonal time scale. Identifying the sources of drought predictability in regions where a prediction system demonstrated potential for useful applications of the forecasts, represents an important step toward building confidence in the predictions and refining the seasonal predictions. To better identify higher forecast skill in this context, one possible approach is to focus on specific “windows of opportunity”. The approach aims to identify periods when persistent anomalies occurring in the ocean, the atmosphere or the land surface may positively precondition the predictive ability of the seasonal forecast. In the case of SPI3, a high potential for preconditioned predictive skill is identified in the Middle East region, as suggested by a robust relationship with large-scale climate modes. Building on these results, this study explores the contributions of individual years to the skill for the region during the autumn season and in the hindcast period 1993-2016. We used a Multi Model Ensemble (MME) of eight seasonal prediction systems (SPSs) provided by the Copernicus Climate Data Store (CDS) and observations from the Climate Research Unit (CRU) to calculate the SPI3 time series and the values of the Pacific and Indian teleconnection indices, the Oceanic Nino Index (ONI) and the Dipole Mode Index (DMI), respectively. A novel methodology is implemented to cluster the year-by-year MME contributions to the Pearson correlation coefficient (PCC) that are preconditioned by the large-scale teleconnections. 

Results indicate that years with extreme high or low values of ONI and DMI are the main contributors to the forecasting skill of the MME drought predictions over the Middle East. In particular, a window of opportunity is identified in four (out of 24) years that show significantly high contribution to overall skill. These years are robustly preconditioned by El Niño or La Niña events. Among the years with higher contributions, 1994 stands out as being more influenced by the DMI, thus driven primarily by SST anomalies in the Indian Ocean rather than the Pacific Ocean.  The methodological approach developed in this study successfully highlighted the potential windows of opportunity for seasonal prediction in the Middle East region, and could be applied extensively to develop early warnings for the coming seasons to serve agriculture and water management operations.

How to cite: Dal Monte, T., Alessandri, A., Cherchi, A., Donat, M., and Gaetani, M.: Windows of Opportunity for Seasonal Prediction of droughts: the case of the Middle East, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11600, https://doi.org/10.5194/egusphere-egu25-11600, 2025.

EGU25-11821 | Orals | CL4.8

Seamless seasonal to multi-annual climate predictions by constraining transient (CMIP6) climate model simulations 

Juan C. Acosta Navarro, Alvise Aranyossy, Paolo De Luca, Markus G. Donat, Arthur Hrast Essenfelder, Rashed Mahmood, Andrea Toreti, and Danila Volpi

Seamless climate predictions combine information across different timescales to deliver information potentially useful for sectors like agriculture, energy, and public health. Seamless operational forecasts for periods spanning from sub-annual to multi-annual timescales are currently not available throughout the year. We show that this gap can be closed by using a well-established climate model analog method. The method consists in sampling model states from the CMIP6 transient simulation catalog based on their similarity with the observed sea surface temperature as a means of model initialization. 

Here we present the methodology and basic skill evaluation of the analog-based temperature and standardized precipitation index retrospective predictions with forecast times ranging from 3 months up to 4 years. We additionally compare these predictions with the non-initialized CMIP6 ensemble and with two operational benchmarks produced with state-of-the-art dynamical forecasts systems: one on seasonal timescales and the other on annual to multi-annual timescales.

The analog method provides skillful climate predictions across the different timescales, from seasons to several years, offering temperature and precipitation forecasts comparable to those from state-of-the-art initialized climate prediction systems, particularly at the annual to multi-annual timescales. However, unlike operational decadal prediction systems that provide only one or two initializations per year, the analog-based system can generate seamless predictions with monthly initializations, offering year-round climate information. Additionally, analog predictions are computationally inexpensive once the multi-model transient climate simulations have been completed. We argue that these predictions are a valuable complement to existing operational prediction systems and may improve regional climate adaptation and mitigation strategies. 

 

How to cite: Acosta Navarro, J. C., Aranyossy, A., De Luca, P., Donat, M. G., Hrast Essenfelder, A., Mahmood, R., Toreti, A., and Volpi, D.: Seamless seasonal to multi-annual climate predictions by constraining transient (CMIP6) climate model simulations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11821, https://doi.org/10.5194/egusphere-egu25-11821, 2025.

EGU25-13385 | Posters on site | CL4.8

Representation of Temporal Variations of Vegetation in Reanalysis and Climate Predictions: Diverging Soil-Moisture Response in Land Surface Models 

Andrea Alessandri, Marco Possega, Emanuele Di Carlo, Annalisa Cherchi, Souhail Boussetta, Gianpaolo Balsamo, Constantin Ardilouze, Gildas Dayon, and Fransje van Oorschot

Vegetation plays a crucial role in the land surface water and energy balance modulating the interactions and feedback with climate at the regional to global scale. The availability of unprecedented Earth observation products covering recent decades (and extended up to real-time) are therefore of paramount importance to better represent the vegetation and its time evolution in the land surface models (LSMs) used for offline analysis/initialization and for the seasonal-to-decadal predictions. 

Here, we integrate realistic vegetation Leaf Area Index (LAI) variability from latest generation satellite campaigns, available through Copernicus Land Monitoring Service (CLMS), in three different LSMs that conducted the same coordinated set of offline land-only simulations forced by hourly atmospheric fields derived from the ERA5 atmospheric reanalysis. The experiment implementing realistic interannually-varying LAI (SENS) is compared with simulations utilizing a climatological LAI (CTRL) to quantify the vegetation feedback and the effects on the simulation of near-surface soil moisture.

The results show that the inter-annually varying LAI considerably affects the simulation of near-surface soil moisture anomalies in all three models and over the same water-limited regions, but surprisingly the effects diverge among models: compared with ESA-CCI observations, the near-surface soil moisture anomalies significantly improve in  one of the three LSMs (HTESSEL-LPJGuess) while the other two (ECLand and ISBA-CTRIP) display opposite effects with significant worsening of the anomaly correlation coefficients. It is found that the enhanced simulation of near-surface soil moisture is enabled by the positive feedback that is activated by the effective vegetation cover (EVC) parameterization, implemented only in HTESSEL-LPJGuess. The EVC parameterization works such that the effective fraction of the bare soil being covered by vegetation does vary with LAI following an exponential function constrained by available satellite observations. The increased (reduced) soil-moisture limitation during dry (wet) periods produces negative (positive) LAI and therefore EVC anomalies, which in turn generate a dominating positive feedback on the near-surface soil moisture of HTESSEL-LPJGuess by exposing more (less) bare soil to direct evaporation from the sub-surface layer. On the other hand, in the EC-Land and ISBA-CTRIP models, EVC is fixed in time as it cannot vary with LAI and so the positive feedback described cannot be activated. The only feedback on near-surface soil moisture anomalies that operates  in these two models is negative and comes from the reduced (increased) transpiration related to the negative (positive) LAI anomalies.

Simply prescribing observed vegetation data into LSMs does not guarantee the introduction of the correct coupling and feedback on climate. In this respect, this multi-model comparison experiment demonstrates the fundamental role of the inclusion of the underlying vegetation processes in LSMs. Ignoring the proper representation of the vegetation processes could lead to unrealistic (and even the opposite effects compared with observations) behaviour in reanalysis and climate predictions.

How to cite: Alessandri, A., Possega, M., Di Carlo, E., Cherchi, A., Boussetta, S., Balsamo, G., Ardilouze, C., Dayon, G., and van Oorschot, F.: Representation of Temporal Variations of Vegetation in Reanalysis and Climate Predictions: Diverging Soil-Moisture Response in Land Surface Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13385, https://doi.org/10.5194/egusphere-egu25-13385, 2025.

EGU25-14900 | ECS | Posters on site | CL4.8

Psychological Drivers of Climate Silence: A Challenge to Indonesia's Climate Action 

Anggi Dewita and Balgis Inayah

Despite growing awareness of climate change, many Indonesians remain climate-silent, posing a significant challenge to the country's efforts to mitigate its impacts. This study aims to analyze the factors contributing to climate silence in Indonesia, using psychological theories related to climate science denial. A rapid systematic review was conducted to gather evidence, revealing five key drivers of climate denial: limited cognitive abilities, ideological beliefs, sunk costs, perceived risks, and discredence. These barriers are further shaped by factors such as government policies, economic conditions, religious influences, and insufficient environmental education.
This skepticism towards climate change undermines adaptation and mitigation efforts by disrupting community engagement and participation. The findings highlight the importance of government support in addressing the root causes of climate skepticism. Employing the concept of inoculation through a misconception-based learning approach—integrated into religion and education—can help reshape mindsets. Enhancing public understanding of climate change is essential to fostering community involvement and support for effective climate mitigation initiatives.

Keywords: climate silence, climate denial, psychological drivers, Indonesia.

How to cite: Dewita, A. and Inayah, B.: Psychological Drivers of Climate Silence: A Challenge to Indonesia's Climate Action, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14900, https://doi.org/10.5194/egusphere-egu25-14900, 2025.

EGU25-15251 | Posters on site | CL4.8

Assessment of the skill of seasonal probabilistic hydrological forecasts with ParFlow/CLM over central Europe 

Alexandre Belleflamme, Suad Hammoudeh, Klaus Goergen, and Stefan Kollet

In recent years, alternating drought and extreme precipitation events have highlighted the need for subseasonal to seasonal forecasts of the terrestrial water cycle. In particular, predictions of the impacts of dry and wet extremes on subsurface water resources are crucial to provide stakeholders in agriculture, forestry, the water sector, and other fields with information supporting the sustainable use of these resources.

In this context, we release an experimental Water Resources Bulletin (https://adapter-projekt.de/bulletin/index.html) four times per year, offering probabilistic forecasts of the total subsurface water storage (TSS) anomaly at a 0.6 km resolution, from the surface down to 60 m depth, for the upcoming seven months across Germany. These seasonal forecasts are generated using the integrated, physics-based hydrological model ParFlow/CLM, forced by 50 ensemble members of the SEAS5 seasonal forecast from the European Centre for Medium-Range Weather Forecasts (ECMWF).

To evaluate our forecasts, we evaluated six 7-months probabilistic forecasts covering the vegetation period (March to September) for the years 2018 to 2023 with a reference long-term historical time series based on the same ParFlow/CLM setup. The forecast skill was assessed by comparing these seasonal forecasts to a climatology-based 10-member pseudo-forecast over the 2013–2023 period (using the leave-one-out method), extracted from the reference time series.

The monthly Continuous Ranked Probability Skill Score (CRPSS), which evaluates the ensemble distribution based on daily TSS data, indicates that the probabilistic forecast outperforms the climatology-based pseudo-forecast in most regions, except in 2018 and, to a lesser extent, in 2020 and 2022. This can be attributed to an under-representation of extremely dry members in the ensemble, combined with the memory effect of the initial conditions at increasing soil depths. For example, while March 2018 started with a slightly above-average TSS and experienced a strong meteorological drought leading to an agricultural drought, the initial TSS anomaly in March 2019 was already negative, with a less pronounced precipitation deficit during the vegetation period. This resulted in a much higher forecast skill, because of the memory effect accurately simulated with the physics-based model. Notably, the forecast skill only slightly decreases with increasing lead time, both for precipitation and TSS.

The analysis of the Relative Operating Characteristic Skill Score (ROCSS) for the lower quintile of the TSS distribution assesses whether negative TSS anomalies (i.e., droughts) are adequately represented within the probabilistic forecast ensemble. The results are consistent with those of the CRPSS, showing lower skill in 2018. Nevertheless, the ROCSS analysis overall indicates moderate to high skill for the probabilistic forecast, while the climatology-based pseudo-forecast demonstrates no skill. This confirms that the dry conditions experienced in central Europe in recent years were captured within the probabilistic forecast, underlining the added value of these forecasts and their usefulness in the experimental Water Resources Bulletin.

How to cite: Belleflamme, A., Hammoudeh, S., Goergen, K., and Kollet, S.: Assessment of the skill of seasonal probabilistic hydrological forecasts with ParFlow/CLM over central Europe, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15251, https://doi.org/10.5194/egusphere-egu25-15251, 2025.

EGU25-15484 | ECS | Posters on site | CL4.8

From Policy to Action: Empowering Women to Lead Climate Resilience in Indonesia 

Asri Rachmawati and Anggi Dewita

Women face disproportionate impacts from climate change due to significant barriers to accessing education and protection. In Indonesia, women often lack access to essential resources and opportunities, particularly in urban informal settlements. However, women also hold a pivotal position in the community in advancing climate literacy. Despite progressive regulations supporting women’s rights, gaps in implementation persist, highlighting the need for targeted initiatives to enhance women’s understanding of climate issues and their capacity to lead resilience efforts. Indonesia has established strong policies for gender equality and climate action, such as Presidential Regulation No. 59/2017 and the National Action Plan for Climate Change Adaptation (RAN-API), which emphasize gender-responsive strategies. However, translating these policies into real-world actions remains a challenge, highlighting the need to better connect scientific research and community insights to effective governance and implementation. This study identifies a critical gap in urban climate literacy and proposes empowering women as a solution. By leveraging women’s social network in Indonesia, the project disseminates climate knowledge and fosters collective action. Key initiatives include training women in climate literacy, introducing sustainable practices such as urban gardening, and developing accessible educational tools like songs, games, and visual materials. These programs are designed to position women as trusted leaders within their communities. Structured monitoring and evaluation methods, including annual surveys and peer-led literacy programs, ensure continuous improvement and scalability. Preliminary findings demonstrate that women-led climate literacy initiatives significantly enhance community resilience and resource allocation. Empowered women influence their families and peers, creating a ripple effect that strengthens societal adaptability. This scalable model integrates women-centered initiatives into governance frameworks, building pathways for sustainable, inclusive development. By empowering women, we transform vulnerability into strength, paving the way for a resilient future.

How to cite: Rachmawati, A. and Dewita, A.: From Policy to Action: Empowering Women to Lead Climate Resilience in Indonesia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15484, https://doi.org/10.5194/egusphere-egu25-15484, 2025.

EGU25-15694 | Orals | CL4.8 | Highlight

The Use of Social Media on Weather and Climate Information Dissemination To Support Effective Climate Action 

Radjab Achmad Fachri and Achmad Ezra Reynara

Timely and fast dissemination are some of the key factors for the effective climate information services in order to support effective climate action. Various mean of communication channel has been used by an authoritative agency to disseminate their climate information, including social media. Currently, social media become one of the most effective chanel to disseminate of weather and climate information. Social media is not only a powerfull tools to ensure the timely, fast, massive dissemination of weather and climate information, but it is also easy to use and access by the general public. Social media also can be optimized public outreach and public education in order to raising awareness and mobilizing an effective climate action. Through it’s real time response tools, social media also can be used to strengthen the engagement between meteorological and hydrological services with their users. Our research will describe the effectiveness of social media to disseminate weather and climate information in order to support climate action in Indonesia.

How to cite: Achmad Fachri, R. and Ezra Reynara, A.: The Use of Social Media on Weather and Climate Information Dissemination To Support Effective Climate Action, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15694, https://doi.org/10.5194/egusphere-egu25-15694, 2025.

EGU25-19049 | Posters on site | CL4.8

The Role of Afforestation in Modulating Arid Climate 

Thang M. Luong, Matteo Zampieri, and Ibrahim Hoteit

Afforestation and greening initiatives are increasingly considered viable strategies for mitigating climate change, particularly in arid regions. In this study, we assess the climate impacts of large-scale afforestation in the Arabian Peninsula (AP). The afforestation is represented by replacing sandy bare soil with woody savanna vegetation, assumed to be naturally sustained by rainfall, in the absence of overgrazing. Using a 30-year regional climate model simulation, we prescribe afforestation within a circular area of 4.5° radius (approximately 71.9 million hectares) centered at 24.2°N, 44.3°E. The afforestation modifies surface characteristics, including darker albedo (0.25 vs. 0.38 for bare soil), a green fraction of 0.3, and a leaf area index (LAI) of 0.1.

Our results show that the afforestation slows down near-surface winds and due to darker surface, increases sensible heat flux, leading to enhanced warming of the atmosphere over vegetated areas. Despite these warming effects, the additional vegetation promotes higher rainfall due to increased moisture availability and reduction of subsidence. This study underscores the dual role of afforestation in modulating regional climate, serving as both a climate mitigation measure and a potential warming source, depending on regional conditions. These findings highlight the importance of considering water availability and local climate factors when designing greening policies for arid regions.

How to cite: Luong, T. M., Zampieri, M., and Hoteit, I.: The Role of Afforestation in Modulating Arid Climate, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19049, https://doi.org/10.5194/egusphere-egu25-19049, 2025.

EGU25-19889 | Orals | CL4.8

Progresses and Challenges for Subseasonal to Interdecadal Prediction 

Ángel G. Muñoz, William J. Merryfield, and Debra Hudson

Subseasonal to decadal predictions provide essential information that bridges the gap in timescales between weather forecasts and long-term climate projections. The science and practice of making such predictions using global climate models initialized with observational data has advanced considerably in recent years, and as a result operational subseasonal, seasonal and decadal prediction services are now a reality. Nonetheless, important remaining challenges must be overcome if these predictions are to more fully realize their potential value for society. This talk highlights five key challenges recommended as targets for focused international research; these are set against a backdrop of wider challenges encompassing climate modelling and services across time scales.

How to cite: Muñoz, Á. G., Merryfield, W. J., and Hudson, D.: Progresses and Challenges for Subseasonal to Interdecadal Prediction, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19889, https://doi.org/10.5194/egusphere-egu25-19889, 2025.

EGU25-21425 | Orals | CL4.8

A large ensemble of decadal predictions using MIROC6 

Takahito Kataoka, Hiroaki Tatebe, Hiroshi Koyama, and Masato: Mori

The climate fluctuates on various timescales and in various patterns, giving rise to extreme events over the globe. Skillful predictions of such climate variations would therefore benefit society, and there have been substantial efforts. For the CMIP6 Decadal Climate Prediction Project (DCPP), we performed decadal predictions with ten ensemble members using the Model for Interdisciplinary Research on Climate version 6 (MIROC6). However, since models tend to underestimate signal-to-noise ratio in some sectors, such as the Atlantic, a large ensemble size appears to be required for skillful predictions of those variations. To better understand the predictability on timescales out to a season to a decade, we have prepared a set of initialized predictions using MIROC6 that consists of 10-year-long hindcasts starting every November between 1960-2021, with 50 ensemble members. Compared to the original 10-member ensemble hindcast, both seasonal and decadal prediction skills are broadly improved (e.g., SAT and SLP over southeast China and Scandinavia for the first winter, North and South Pacific SSTs for decadal prediction). Regarding the decadal prediction skill, the impact of initialization is seen up to lead year 7-10 for the North and eastern tropical Pacific Oceans.
Also, building on our experience with decadal climate predictions, we have been working on decadal carbon predictions in recent years. Our efforts on earth system predictions will be introduced as well.

How to cite: Kataoka, T., Tatebe, H., Koyama, H., and Mori, M.: A large ensemble of decadal predictions using MIROC6, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21425, https://doi.org/10.5194/egusphere-egu25-21425, 2025.

EGU25-275 | ECS | Orals | GI2.4

Application of Machine Learning Algorithms to Predict Rock Types Using Geochemical Data: A Case Study from the Obuasi Gold District, Ghana 

Abdullah Bello Muhammed, Emmanuel Daanoba Sunkari, and Abdul Wahab Basit

Artificial Intelligence and Machine Learning (AI/ML) are gaining increasing interest due to their capacity to increase precision and productivity in the current big data era. Machine learning has indicated its robustness in geosciences, particularly rock-type classification. Lithological classification in the traditional way has raised critical concerns and the need to curb the limitations it breeds, such as time consumption and subjective results. The gold mineralisation occurrence is structurally controlled in the Obuasi Gold district of Ghana. It exhibits complex patterns and relationships that may not be readily discernible through traditional methods, leading to missing out on discovering new resources or potential exploration targets. Consequently, this work attempts to create a predictive model by exploring the best machine-learning algorithms to predict rock types in the Obuasi Gold District using X-ray fluorescence (XRF) geochemical data. Here we established comparative predictive modelling using four supervised classification algorithms: Gradient Boosting (GBoost), Adaptive Boosting (AdaBoost), Support Vector Machine (SVM) and Random Forest (RF). The acquired XRF data was integrated with the model using the Google Collaboratory cloud-based platform. Results show that the performance evaluation of the models indicated SVM as the best algorithm for deployment with a Classification Accuracy (CA) of 0.902. Therefore, ML algorithms have been a great tool in rock-type classification, whereby SVM emerged as the best in the case of the Obuasi Gold District. However, it is encouraged to understand the geology of a particular area before employing the tool and the datasets must be balanced to yield good results and avoid model overfitting.

Keywords: Artificial intelligence; Machine learning algorithm; Support vector machine; Lithogeochemistry; Rock-type classification; Obuasi Gold District

How to cite: Muhammed, A. B., Sunkari, E. D., and Basit, A. W.: Application of Machine Learning Algorithms to Predict Rock Types Using Geochemical Data: A Case Study from the Obuasi Gold District, Ghana, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-275, https://doi.org/10.5194/egusphere-egu25-275, 2025.

Preserving clean water resources and efficiently treating wastewater is critical for ensuring human survival on Earth and in extraterrestrial environments. Major pollutants, including ammonium, heavy metals, industrial dyes, and chemicals, threaten limited clean water supplies and soil. Among the renowned absorbent materials, natural zeolite minerals have demonstrated their effectiveness for pollutant removal compared to clay minerals and synthetic equivalents like biochar, activated carbon, and MOFs, owing to their relatively extensive reserves and eco-friendly nature. Recently, investigating and optimizing pollutant removal rates from water without conducting laboratory experiments is getting more crucial, considering the time-consuming, expensive, and error-prone nature of laboratory testing due to human factors and potential calibration issues among the chosen analytical techniques.

This study aims to forecast the ammonium removal efficiency (% adsorption) and capacity (mg/g) of natural and modified zeolites from aqueous solutions using the regression ensemble LSBoost (MATLAB R2024b) machine learning (ML) algorithm, which is equivalent to XGBoost open-source library. A total of 527 experiments on 15 different zeolite compositions were gathered from a combination of 14 suitable moderately recent (≥ 2005) and highly referred studies to assess the performance of zeolite minerals on ammonium removal rates from aqueous solutions. The LSBoost algorithm achieved over 0.99 R2 fitting for training and overall, 0.95 R2 for prediction on the quarterly partitioned testing data for both efficiency and capacity. Throughout the improvement of the ML models using different random forest ML approaches, the number of predictors was successfully reduced to 8 based on importance rates among 31 different features in the initial dataset, with a negligible accuracy loss (<0.1 R2) on both training and testing. This research provides a valuable contribution to optimizing applicable experimental parameters in water treatment processes by effectively identifying the significance of predictors within a comprehensive data set. In addition to this, our model not only provides a robust predictive tool for optimizing zeolite performance in water treatment but also represents the first open-sourced web application in the literature to estimate the water treatment performance of zeolites.

How to cite: Akkaş, E., Ünal, B. C., and Ersoy, O.: AI-Based Prediction and Optimization of Ammonium Removal Efficiency and Capacity of Natural Zeolites Using LSBoost (XGBoost) for Sustainable Ecosystems, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-809, https://doi.org/10.5194/egusphere-egu25-809, 2025.

EGU25-882 | ECS | Orals | GI2.4

Assessment of a deep learning framework for time-lapse seismic monitoring 

Giovanni Pantaleo and Michele Pipan

In the context of CO₂ storage, cost-effective monitor methods are essential to ensure safe and long-term storage. This work explores the use of seismic time-lapse monitoring, combined with deep learning (DL) techniques, to assess potential leakage and migration pathways. The goal is to develop a cost-effective monitoring method while guaranteeing the safety of storage operations. To this end, we propose a Siamese Neural Network (SNN)-based framework to analyse shot gathers, designed to detect and localize changes within the storage complex. We aim to address the challenges of working with large seismic datasets, enabling the identification of significant events with high confidence, while avoiding the need for event-by-event processing. This framework can allow experts to rely on semi-automatic detections while ensuring human evaluation for interpreting and validating the results.

The proposed SNN architecture processes pairs of shot gather from baseline and monitor surveys in a cross-well configuration. It uses two identical neural networks with shared weights to encode the shot gathers into latent feature embeddings, which are then compared to identify similarities and detect changes. By transforming the data into a shared latent space, the model focuses on capturing relevant patterns while filtering out irrelevant variations, ensuring robust and accurate comparisons. When the SNN detects changes between the baseline and the monitor surveys, it highlights the regions where these changes occur. This approach is particularly effective for identifying subtle but important changes in seismic data, such as those caused by CO₂ migration, which alters the velocity and density of the subsurface. Even in noisy data, the SNN can detect these variations, thanks to its ability to learn features that are highly sensitive to small but meaningful changes. The SNN architecture is scalable and can be adaptable to various seismic monitoring tasks, requiring minimal preprocessing. The proposed framework harnesses the power of deep learning to provide insights into the dynamics of the storage complex, with a focus on identifying changes in time-lapse seismic data related to localized variations. The proposed migration detection tool offers a cost-effective and reliable solution to the modern challenges of gas storage monitoring. This study aims to enable operators to identify and address problems promptly, thereby minimising the impact of potential leakages.

How to cite: Pantaleo, G. and Pipan, M.: Assessment of a deep learning framework for time-lapse seismic monitoring, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-882, https://doi.org/10.5194/egusphere-egu25-882, 2025.

Joint inversion is an essential technique in potential field data processing. The current methodology largely relies on the geology model of anomalous bodies, especially for deep, complex structures. Inspired by the excellent nonlinear mapping capability of the image semantic segmentation model and the advantages of supervised learning, a regressive, end-to-end, encoder-decoder structural, convolutional neural network with a double-branch structure called PFInvNet(Potential Field Inversion Neural Network) is proposed for joint 3D  inversion of physical properties from gravity and magnetic data. Its input is a four-channel dataset consisting of gravity and magnetic anomalies and their vertical gradients, and its output is a 3D matrix representing the spatial distribution of the remnant density and the magnetic susceptibility, which are predicted independently through the double-branch structure of the decoders and then concatenated in the final layer. For network training, a large amount of precisely labeled sample is exceedingly demanding; thus, forward modeling becomes a prerequisite approach. Two discretized forward modeling algorithms for gravity and magnetic anomalies of 3D homogeneous arbitrary-shaped bodies based on surface integrals are deduced and verified with analytic solutions of the sphere model. Furthermore, the neural network needs to learn from the anomalies generated by various forms of abnormal bodies with different physical properties. Therefore, different sizes and quantities of cuboids are randomly distributed in the model space to simulate different forms of abnormal bodies. The label represents the combined spatial distribution of remanent density and magnetic susceptibility for the cuboids, encompassing both spatial location information and physical properties information. With the help of the Marching Cubes(MC) algorithm, the surface of the cuboids can be easily extracted and divided into a triangular surface mesh. The surface mesh is then used to calculate the gravity and magnetic anomalies synchronously through the forward modeling algorithms. The anomalies are concatenated in the channel direction as a sample. A set of optimal network parameters has been determined, including the weight initialization method, the gradient calculation methods, the loss function, the training hyperparameters, the regularization method, and the normalization method. The PFInvNet is trained with 500 and 10000 pairs of samples and labels, respectively. The analysis and comparison of training results prove that PFInvNet has two crucial features: one is that the branch structure enables independent prediction of magnetic susceptibility and remanent density; the other is efficient anti-overfitting ability and efficient solution-finding ability .The prediction error of small samples is very close to that of large samples and is also not obviously enhanced by the noise-contaminated data , demonstrating the strong generalization and robustness of the network. Finally, the network is tested with magnetic and gravity anomalies of the Victoria Land Basin in the western Ross Sea through transfer learning and retraining, and definite 3D distributions of apparent remnant density and apparent magnetic susceptibility have been obtained and can be checked with geological evidences.

How to cite: Jiang, W., Zhao, Y., Gao, J., Ge, S., and Xie, Z.: 3D property inversion of gravity and magnetic data based on a double branch regressive CNN trained by synchronous forward modeling :a case study of the Western Ross Sea, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1141, https://doi.org/10.5194/egusphere-egu25-1141, 2025.

EGU25-2156 | ECS | Orals | GI2.4

TOC Intelligent Prediction Model in Shale Reservoir: Integrating Data Enhancement with Empirically Driven Algorithm 

Yuzhen Hong, Shaogui Deng, Zhijun Li, and Zhoutuo Wei

Shale oil and shale gas are important unconventional resources. Organic matter serves as the primary source of shale oil and gas generation, and high TOC values typically indicate better oil and gas reservoir conditions and production. Therefore, an accurate TOC prediction model is conducive to low-cost evaluation of reservoir hydrocarbon potential and improvement of development efficiency. However, geochemical experimental measurements are costly, and the data obtained is discrete. It is unable to meet the requirements for fine-scale assessment of shale reservoirs. The multiple regression method and ΔlogR method, when directly applied to shale reservoirs, often result in significant errors. In this study, we propose a composite model for accurate TOC prediction in shale reservoirs based on data enhancement and empirically driven. We first address the issue of poorly characterized logging responses and discrete experimental data. The features and quantities of the dataset are enhanced by introducing reconstruction curves and generative adversarial networks (GAN). The validity of the synthesized data is then verified by plotting the data density. In the empirically-driven module, we optimize a density-gamma modified method on traditional ΔlogR method according to the characteristics of shale reservoirs. The modified ΔlogR method will be integrated into the GWO-SVR model as an empirically driven subject in the form of a fitness function. Above, a composite model with both empirical and data-driven components is constructed. We use the Dongying Depression in China as an example for model experiments. The composite model was generalized to wells X and Y. The R² (coefficient of determination) was 0.95 and 0.97, the RMSE (Root Mean Square Error) was 0.31 and 0.29, and the MAE (Mean Absolute Error) was less than 0.3, which indicated a high degree of consistency between the model predictions and the experimental values. Further controlled experiments revealed that the composite model predicted better than the ΔlogR method and the GWO-SVR model alone. Finally, we also performed SHAP interpretability analysis on the model. By revealing the decision-making mechanism inside the model, we verified the rationality of the empirical drive and enhanced the credibility of the model. This provides strong technical support and decision-making basis for the subsequent oil and gas exploration and development work.

How to cite: Hong, Y., Deng, S., Li, Z., and Wei, Z.: TOC Intelligent Prediction Model in Shale Reservoir: Integrating Data Enhancement with Empirically Driven Algorithm, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2156, https://doi.org/10.5194/egusphere-egu25-2156, 2025.

EGU25-2242 | Posters on site | GI2.4

A Conditional Neural Operator Approach for Resolution-Flexible and Parameter-Controlled Gravity Forward Modelling 

Ruiyuan Kang, Meixia Geng, Qingjie Yang, and Felix Vega

We present a novel approach to gravity forward modeling using conditional neural operators that establishes a forward generative model from the basin models and hyperparameters (reference basement depth, etc.) to gravity anomaly. Our methodology introduces an innovative adaptive embedding mechanism where scalar hyperparameters are first embedded into a 32-dimensional space and then adaptively expanded to match the dimensions of the basin depth model, enabling effective fusion with basin depth model data. Subsequently, Fourier Convolution Layers are employed to transform the fused data into gravity anomalies. The model demonstrates superior performance compared to existing convolutional neural networks on the test dataset, showcasing improved accuracy in capturing complex geological structures and their gravity responses. A key advantage of our architectural design is that it not only preserves the super-resolution capability of conventional neural operators but also enables controlled generation through different hyperparameters. This dual capability allows for both resolution-flexible modeling and parameter-controlled generation, while training on low-resolution data and producing high-resolution outputs, significantly reducing training data requirements and computational costs. The model's adaptive architecture effectively bridges the resolution gap between training and application scenarios, offering a practical solution for real-world geological surveys. Our results suggest that this approach could substantially improve the accessibility and applicability of gravity forward modeling in various geological settings, particularly in regions with limited high-resolution training data.

How to cite: Kang, R., Geng, M., Yang, Q., and Vega, F.: A Conditional Neural Operator Approach for Resolution-Flexible and Parameter-Controlled Gravity Forward Modelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2242, https://doi.org/10.5194/egusphere-egu25-2242, 2025.

EGU25-2478 | Orals | GI2.4

International standards for responsible AI in disaster management 

Monique Kuglitsch and Elena Xoplaki

The combination of big data and AI/ML technologies shows tremendous promise within the domain of disaster management. So that benefits of AI/ML can be realized, and risks can be adverted, internationally agreed upon standards are an important mechanism. These can provide guidance on how to apply (and develop policy around) data collection and preprocessing, model training and evaluation, and operational implementation. In conjunction, they can cultivate interoperability and harmonization of AI-based systems. At the United Nations, the Global Initiative on Resilience to Natural Hazards through AI Solutions brings together experts from different disciplines (geosciences, disaster risk management, computer sciences) and sectors (government, research, NGO) to analyze use cases and lay the groundwork for such standards. Through proof-of-concept projects (e.g., HEU-funded MedEWSa), these standards can be further refined. Finally, through education and capacity building activities, the Global Initiative can help to democratize the responsible use of AI for this domain.

How to cite: Kuglitsch, M. and Xoplaki, E.: International standards for responsible AI in disaster management, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2478, https://doi.org/10.5194/egusphere-egu25-2478, 2025.

EGU25-4298 | ECS | Posters on site | GI2.4

Advances in the identification of geological discontinuities in boreholes with deep learning 

Rushan Wang, Martin Ziegler, Michele Volpi, and Andrea Manconi

Geological discontinuities significantly influence rock mass behaviour. Understanding the origin, setting, and properties of discontinuities is of major relevance, especially in boreholes. Traditionally, manual interpretation of borehole logs is done by geologists, a process that is time-consuming, costly, and subject to variability based on the interpreter's expertise. Recent advancements in artificial intelligence have made it feasible to use machine learning models and automatically detect and differentiate various features in digital images. In this study, we employ a state-of-the-art semantic segmentation model to tackle domain-specific challenges, enabling the identification of discontinuity types (e.g., natural faults, fault zones) and rock mass behaviour features (e.g., breakouts, induced cracks). We applied the SegFormer semantic segmentation model, which integrates a hierarchically structured transformer encoder with a multilayer perceptron (MLP). The borehole data used in this study was collected from the Mont Terri underground rock laboratory. Specifically, we labelled several high-resolution optical logs from one borehole and divided the dataset into training and testing subsets. The borehole considered is an experimental borehole designed to investigate the spatial and temporal evolution of damage around an underground opening in faulted clay shale. Our strategy achieved robust and accurate segmentation results on borehole images. Following segmentation, post-processing techniques were employed to extract critical information such as the total length of induced cracks and the total area of breakouts, as well as their locations and frequencies. The experimental results demonstrate high performance, with the pixel accuracy of 96 % in under three minutes for a 10-meter borehole. Our study lays the groundwork for future research by introducing a powerful tool for extracting geological structures and demonstrating the potential of AI models in geological analysis. By reducing processing time and increasing consistency in the identification, mapping, and classification of geological features, our approach can reveal spatial and temporal patterns associated with the evolution of rock masses.

How to cite: Wang, R., Ziegler, M., Volpi, M., and Manconi, A.: Advances in the identification of geological discontinuities in boreholes with deep learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4298, https://doi.org/10.5194/egusphere-egu25-4298, 2025.

EGU25-4404 | ECS | Orals | GI2.4

AI-Enhanced Meteorological Data Retrieval Systems for Improved Forecast Operations 

Byeongyeon Kim, Hayan Shin, Areum Cho, Junsang Park, Hyesook Lee, ChaeHun Park, Jinkyung Joe, Jaegul Choo, and Minjoon Seo

Meteorological data are vast and complex, and their rapid and accurate retrieval is essential for forecasting operations. However, traditional systems have struggled with limited search accuracy and inefficient processing speeds, hindering effective forecast support. To address these challenges, this study developed an AI-based system capable of performing speech recognition, URL search, extreme value detection, and local forecast error analysis. In speech recognition, the Whisper-large model achieved a character error rate (CER) of 3.19%, with GPU memory usage reduced by 15.7% and inference time by 38.18%, enabling real-time processing and scalability in multi-GPU environments. The URL search systems translated natural language inputs into SQL queries and URLs, achieving a Mean Reciprocal Rank (MRR) of 0.92, thereby enhancing data retrieval precision. The extreme value detection systems utilized GPT-4-based template augmentation to expand training data by approximately 111%, significantly improving detection performance and search accuracy. For local forecast error analysis, a prototype chatbot was implemented using prompt engineering and a Text-to-SQL model, allowing for the automated identification of inconsistencies in local forecasts and streamlining the analysis process. These systems have substantially enhanced operational workflows across meteorological tasks, facilitating rapid data retrieval through voice commands, precise responses to complex queries, and real-time analytical support. Future research will focus on further refining these technologies to tackle a wider range of meteorological challenges and integrate them into global forecasting systems for enhanced accuracy and reliability.

How to cite: Kim, B., Shin, H., Cho, A., Park, J., Lee, H., Park, C., Joe, J., Choo, J., and Seo, M.: AI-Enhanced Meteorological Data Retrieval Systems for Improved Forecast Operations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4404, https://doi.org/10.5194/egusphere-egu25-4404, 2025.

EGU25-4699 | ECS | Orals | GI2.4

S-wave velocity prediction of shale reservoirs based on explainable physically-data driven model 

Zhijun Li, Shaogui Deng, and Yuzhen Hong

The shear wave (S-wave) velocity is a key basis for shale reservoir development, particularly for fracability evaluation. Additionally, S-wave velocity also plays a significant role in prestack seismic inversion and amplitude versus offset (AVO) analysis. However, the actual logging data often lack S-wave velocity data, so it is of significant importance for S-wave velocity prediction. We propose a rapid and precise prediction method for the S-wave velocity in shale reservoirs based on class activation maps (CAM) model combined with physically constrained two-dimensional Convolutional Neural Network (2D-CNN). High sensitivity curves related to S-wave velocity are selected as the foundation. Meanwhile, based on the petrophysical theory of pore medium, the petrophysical model of complex multi-mineral components is established. The dispersion effect is reduced to a certain extent and the results are used to constrain the model. The Adam optimization algorithm is used to construct a 2D-CNN model under the constraint of petrophysical model. The CAM is obtained by replacing the global average pooling (GAP) layer with a fully connected layer, which in turn leads to interpretable results. Then, the model is applied to wells A, B1, and B2 in the southern Songliao Basin. Afterwards, comparisons are made with unconstrained model and petrophysical model. The results show that the correlation coefficients and relative errors in the three test wells are 0.96 and 2.14%, 0.97 and 2.35%, and 0.97 and 2.9%, respectively. The higher prediction accuracy and generalization ability of the new method is confirmed. Finally, we present the defined C-factor as a means of evaluating the extent of concern regarding CAMs in regression problems. The C-factor confirms that the focus of 2D-CNN can be significantly enhanced by incorporating the petrophysical model, thereby imposing physical constraints on the 2D-CNN. In addition, we establish the SHAP model to assist in proving the importance of constraints.

Keywords: S-wave velocity prediction; Physically constrained 2D-CNN; Petrophysical model; Class activation mapping technique; Explainable results

How to cite: Li, Z., Deng, S., and Hong, Y.: S-wave velocity prediction of shale reservoirs based on explainable physically-data driven model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4699, https://doi.org/10.5194/egusphere-egu25-4699, 2025.

EGU25-7000 | ECS | Posters on site | GI2.4

Multivariate generative modelling of subsurface properties with diffusion models  

Roberto Miele and Niklas Linde

Accurate multivariate parametrization of subsurface properties is essential for subsurface characterization and inversion tasks. Deep generative models, such as variational autoencoders (VAEs) and generative adversarial networks (GANs), are known to efficiently parametrize complex facies patterns. Nonetheless, the inherent complexity of multivariate modeling poses significant limitations to their applicability when considering multiple subsurface properties simultaneously. Presently, diffusion models (DM) offer state-of-the-art performance and outperform GANs and VAEs in several tasks of image generation. In addition, training is much more stable compared to the training of GANs. In this work, we consider score-based DMs in multivariate geological modeling, specifically for the parametrization of categorical (facies) and continuous (acoustic impedance – I­P) distributions, focusing on a synthetic scenario of sand channel bodies in a shale background. We benchmark modeling performance against results obtained by GAN and VAE networks previously proposed in literature for multivariate modeling. As for the GAN and VAE models, the DM was trained with a training dataset of 3000 samples, consisting of facies realizations and co-located I­P geostatistical realizations. Overall, the trained DM shows significant improvements in modeling accuracy, for all evaluation metrics considered in this study, except for the sand-to-shale ratio, where the values are comparable to those of the GAN and VAE. In particular, the DM is 26% more accurate at reproducing the average (nonstationary) facies distribution and up to 90% more accurate at reproducing the IP marginal distributions for both sand and shale classes. Higher accuracy is also found in the reproduction of the facies-to-I­P joint distribution, whereas the spatial I­P distributions generated by the DM honour the two-point statistics of the training samples. The iterative generative process in DMs generally makes these networks more computationally demanding than VAEs and GANs. However, we demonstrate that with appropriate network design and training parametrization, the DM can generate realizations with significantly fewer sampling iterations while maintaining accuracy comparable to these benchmarking networks. Finally, since the proposed DM parametrizes the joint prior probability density function with a Gaussian latent space, it is straightforward to perform inversion. In addition to improved modeling accuracy, the mapping between the latent and image representations preserves a better topology than that of GANs, overcoming the well-known limitation of the latter for inference tasks, particularly for gradient-based inversion.

How to cite: Miele, R. and Linde, N.: Multivariate generative modelling of subsurface properties with diffusion models , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7000, https://doi.org/10.5194/egusphere-egu25-7000, 2025.

EGU25-9590 | Posters on site | GI2.4

Dynamic Time Warping algorithm: A geoscience aware AI for automatic interpretation in lithostratigraphy? Insights from an application to the Gulf of Corinth (Greece) 

Alina-Berenice Christ, Zahraa Hamieh, John Armitage, Renaud Divies, Sébastien Rohais, Luca Mattioni, and Antoine Bouziat

Stratigraphic correlation of well log data is a fundamental step in geosciences. It involves correlating stratigraphic units across multiple wells to build a comprehensive understanding of subsurface geology. Currently, stratigraphic correlation is predominantly performed “manually” by geoscientists. The process is labor-intensive and time-consuming, and interpretations may vary among interpreters due to differences in expertise, experience, and perspective.

Recent advancements in the application of the Dynamic Time Warping (DTW) algorithm have demonstrated its potential to automate and enhance the stratigraphic correlation of well logs. DTW can generate multiple correlation scenarios highlighting different interpretations of subsurface continuity. Thus, the aim of this work is to explore the potential of DTW as a supporting tool in the standard workflows of geoscientists and test it on well log data from IODP Expedition 381 from the Gulf of Corinth. We automatically correlate lithostratigraphic subunits within a 700 m thick stratigraphic unit across two wells, using Natural Gamma Ray (NGR) and Magnetic Susceptibility (MAGS) logs. We selected this dataset because it illustrates the evolution of geological interpretations over time. Between the first version of the IODP data interpretations and a second version published a few years later, significant differences in interpretation were proposed. These differences highlight the critical role of geological expertise in refining subsurface data interpretations and correlations.

The automatic correlations interpreted by DTW showed a minimal average absolute difference with the most recent and updated published correlation, making the human and the machine correlation almost identical. By applying DTW to this dataset, we demonstrate it would have been possible to identify discrepancies and challenges in the interpretations of subunits at the initial stages after data acquisition. This approach could have flagged potential issues even before the IODP data were made available on the public site. Such early identification highlights the potential of DTW as a valuable tool for providing immediate feedback and guiding more accurate stratigraphic interpretations faster.

While DTW significantly reduces the time required for the correlation phase, the time investment needed for data formatting upstream should not be underestimated. Future work on larger datasets will be crucial to better quantify and validate the overall time savings provided by DTW, as well as to optimize the preparatory steps to ensure efficiency in broader applications.

In conclusion, we show that DTW can offer innovative approaches to enhance geological investigations and speed up interpretations. More generally, we consider this work illustrates how data science methods can be leveraged to assist geologists in routine tasks, with our Corinth case study highlighting both the promises and current limitations of digital transformation in well correlations.  

How to cite: Christ, A.-B., Hamieh, Z., Armitage, J., Divies, R., Rohais, S., Mattioni, L., and Bouziat, A.: Dynamic Time Warping algorithm: A geoscience aware AI for automatic interpretation in lithostratigraphy? Insights from an application to the Gulf of Corinth (Greece), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9590, https://doi.org/10.5194/egusphere-egu25-9590, 2025.

EGU25-9671 | ECS | Orals | GI2.4

Kalpa: Empowering Artificial Intelligence-Driven Geospatial Analysis for Multidisciplinary Applications 

Sudhir Sukhbir, Satyam Pratap Singh, Utpal Singh, Mohit Kumar, and Tushar Goyal

In the age of big data, artificial intelligence (AI) is transforming Earth sciences by enabling efficient analysis and visualisation of complex datasets and fostering innovative approaches to solve age-old geoscientific challenges. Kalpa, a Python-based free and cross platform software, represents a pioneering step in this direction. Built with versatility at its core, Kalpa seamlessly integrates AI and machine learning workflows into geoscience applications, offering  customization through its Python plugin architecture. What sets Kalpa apart is its ease of use, even for non-experts. Its intuitive interface lowers the learning curve, enabling a broader audience—including researchers, professionals, and enthusiasts—to leverage advanced geospatial and AI tools without requiring extensive technical expertise.

Kalpa's capabilities span advanced 3D visualization, geospatial data processing, and machine learning model development. It supports global and regional raster and vector dataset visualisation and processing, allowing for interactive analysis in both geographic and cartesian coordinates. With tools to process satellite imagery, geological and geophysical data, uncrewed aerial vehicle (UAV) data, and digital geological maps, Kalpa caters to a wide range of applications, from mineral exploration to natural hazard forecasting. Its machine learning integration supports supervised and unsupervised algorithms for applications such as lithological mapping, mineral prospectivity mapping, land cover and land usage studies, agricultural productivity mapping and natural disaster management. In this study, we demonstrate Kalpa’s transformative potential through three case studies:

  • Lithological Mapping in Ladakh, India: Utilizing LANDSAT and SRTM data, we produced accurate lithological maps for this geologically complex region.
  • Copper Prospectivity Mapping in Northwest India: Combining remote sensing, geophysical, and geological data, Kalpa predicted copper mineralization zones, with all known deposits falling within areas of predicted probabilities exceeding 0.70.
  • Landslide Susceptibility Mapping in Uttarakhand, India: Using remote sensing datasets, Kalpa identified high-risk landslide zones, supporting disaster management efforts.

Kalpa’s user-friendly interface, robust machine learning integration, and publication-ready export capabilities position it as a powerful tool for advancing geoscience research and practical applications. By bridging the gap between domain expertise and cutting-edge AI methodologies, Kalpa empowers Earth scientists, environmental researchers, and GIS professionals to analyze, model, and predict with unprecedented efficiency and precision. This software marks a new frontier in the application of AI to Earth sciences, enabling multidisciplinary research and fostering innovative solutions to pressing geoscientific challenges.

How to cite: Sukhbir, S., Singh, S. P., Singh, U., Kumar, M., and Goyal, T.: Kalpa: Empowering Artificial Intelligence-Driven Geospatial Analysis for Multidisciplinary Applications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9671, https://doi.org/10.5194/egusphere-egu25-9671, 2025.

EGU25-9968 | ECS | Posters on site | GI2.4

Advancing Landscape Archaeology with AI-driven insights from Airborne Laser Scanning data 

Nejc Coz, Žiga Kokalj, Susan Curran, Anthony Corns, Dragi Kocev, Ana Kostovska, Stephen Davis, and John O'Keeffe

Artificial intelligence (AI) is transforming landscape archaeology by enabling the automated analysis of high-resolution datasets, such as airborne laser scanning (ALS). The Automatic Detection of Archaeological Features (ADAF) tool is an example of the potential of AI to streamline the identification of subtle surface features and demonstrate their value in uncovering and understanding archaeological landscapes. By improving the detection of archaeological sites, the ADAF plays a crucial role in the research, management and preservation of cultural heritage.

ADAF uses advanced AI models, including convolutional neural networks (CNNs) for semantic segmentation and object detection, to detect features in ALS datasets. The tool has been trained on a large archive of ALS data from Ireland and processes visualised inputs to detect patterns indicative of archaeological structures. The workflow integrates pre-processing with the Relief Visualisation Toolbox, inference with trained AI models and post-processing to refine the results to ensure reliable outputs with minimal false positives.

Designed with accessibility in mind, ADAF features an intuitive user interface that removes the barriers traditionally associated with AI-driven analyses. Users can process ALS data and export GIS-compatible results without the need for specialised knowledge, making the tool suitable for a wide audience. This approach democratises the use of AI in landscape archaeology and extends its utility to professionals and researchers in the field.

Tests with Irish ALS datasets have shown that ADAF is able to detect both known and previously unrecognised archaeological features in the landscape, while enhancing the spatial accuracy of identified sites. By automating complex data analysis, ADAF underlines the efficiency, precision and scalability of AI in landscape archaeology. In addition, the tool contributes to the preservation of cultural heritage by identifying sites that would otherwise remain undiscovered and enabling their preservation and integration into cultural heritage management strategies.

ADAF represents a significant advance in the application of AI in landscape archaeology, providing a powerful and accessible solution for surface feature recognition. Its development underlines the transformative potential of AI to revolutionise the study and interpretation of archaeological landscapes.

How to cite: Coz, N., Kokalj, Ž., Curran, S., Corns, A., Kocev, D., Kostovska, A., Davis, S., and O'Keeffe, J.: Advancing Landscape Archaeology with AI-driven insights from Airborne Laser Scanning data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9968, https://doi.org/10.5194/egusphere-egu25-9968, 2025.

Borehole images (BHI) are crucial for resource exploration, providing detailed fracture analysis at millimeter-scale resolution. However, their interpretation is typically carried out manually, a process that is time-consuming, costly, and subject to significant uncertainty due to interpreter bias and variability. Current state-of-the-art AI methods for automated or semi-automated fracture analysis of BHI often rely on field data for training, using manual interpretations as labels. This approach inherently embeds both aleatoric (data-related) and epistemic (manual) uncertainties, which may undermine the reliability and adaptability of these methods. This study proposes an alternative, synthetic data-driven approach to train a set of two deep neural networks (DNNs) connected in sequence. These DNNs are designed to replicate the primary cognitive tasks involved in manual interpretation: the segmentation of the BHI to identify potential edge zones and the tracing of sinusoids over these edges to approximate their best-fitting 2D representation. By utilizing synthetic data, we are able to systematically assess the sensitivity of both networks and explore various training strategies, including curriculum learning (CL) and self-attention mechanisms. Our proposed solution is designed for post-hoc human-machine collaboration, where the model supports but does not replace human expertise. This framework enables the possibility of a multi-level uncertainty assessment—at the human, machine, and human-machine interface levels—opening the door to new ways of understanding and quantifying the sources of uncertainty in BHI analysis. Additionally, the synthetic data-driven approach ensures the generalizability and scalability of the method, as demonstrated by its successful application to low-resolution logging-while-drilling (LWD) and high-resolution fullbore formation microimager (FMI) datasets from multiple global locations. By combining advanced AI techniques with geoscientific knowledge, this study outlines a potential pathway toward more robust, ethical, and sustainable fracture analysis workflows. Beyond the traditional benefits of reduced cost and time, the approach may provide a scientifically grounded framework for exploring the benefits of human-machine collaboration and uncertainty quantification in geoscience practices. If adopted, this framework could significantly advance the field of BHI analysis, offering new tools for resource exploration in hydrocarbon and geothermal applications.

How to cite: Molossi, A., Roncoroni, G., and Pipan, M.: NeuroFit: a robust and scalable synthetic data-driven deep learning solution for automated borehole image analysis at LWD and wireline resolution, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10472, https://doi.org/10.5194/egusphere-egu25-10472, 2025.

EGU25-11754 | ECS | Orals | GI2.4

Exploring Pretraining Possibilities of Crop Classifiaction Models Using Large-Scale Sentinel-1 Datasets 

Mátyás Richter-Cserey, Máté Simon, Gabriel Magyar-Santen, Vivien Pacskó, and Dániel Kristóf

Since 2014, ESA Sentinel missions have been producing an ever-growing amount of data for Earth Observation. This creates the opportunity to monitor changes in high temporal and spatial resolution, however, interpreting this huge quantity of data is challenging. In recent years, the rapid advancement and widespread adoption of applied Artificial Intelligence (AI) methods made it feasible to create deep learning models for specific Earth Observation applications. Combining Sentinel datasets with the appropriate amount of ground truth, robust pre-trained models can be created and applied to produce good-quality thematic maps for different years and large areas. Due to responsibilities related to the European Union’s Common Agriculture Policy (CAP) and the motivation for regional yield estimation, crop classification is one of the most frequently studied remote sensing problems these days. Several papers investigate the possible methods to construct robust and generally functioning models to map the spatial distribution of crops as accurately as possible.

In this study, we present the development of a modular, pre-trained deep learning model designed specifically for crop type mapping. The model is tailored to classify the most prevalent crops in Hungary, including winter and autumn cereals, corn, sunflower, alfalfa, rapeseed, grasslands, and other significant types. For pre-training, we leverage country-wide Sentinel-1 Synthetic Aperture Radar (SAR) data such as Sigma Naught or polarimetric descriptors from H-A-alpha decomposition, collected during the 2021–2024 time period. This dataset comprises annual time series of Sentinel-1 pixels at a spatial resolution of 20 meters. Our approach builds upon prior findings that Sentinel-1-based crop type classification performs comparably to methods using Sentinel-2 optical data. However, Sentinel-1 has the added advantage of producing consistent and regular time series, as it is unaffected by atmospheric conditions such as cloud cover.

The proposed model employs a hybrid methodology integrating self-supervised and fully supervised learning paradigms. This modular architecture allows for seamless integration of task-specific classifiers, which can be fine-tuned using supervised learning to address both current and future classification requirements effectively. The fully supervised component is supported by an extensive ground truth dataset covering over 60% of Hungary’s total land area. This proprietary dataset is derived from the national agricultural subsidy database, providing detailed and accurate annotations. The abundance and quality of labeled data enable the construction of robust, highly generalizable models, ensuring reliable performance across diverse classification tasks. This methodology offers great potential to advance national-scale operational tasks, such as early land cover prediction and annual crop type mapping.

How to cite: Richter-Cserey, M., Simon, M., Magyar-Santen, G., Pacskó, V., and Kristóf, D.: Exploring Pretraining Possibilities of Crop Classifiaction Models Using Large-Scale Sentinel-1 Datasets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11754, https://doi.org/10.5194/egusphere-egu25-11754, 2025.

EGU25-12065 | ECS | Orals | GI2.4

Calibration of the Hypoplastic Clay model with a deep neutral network 

Phuong Do and Tomas Kadlicek

Advanced constitutive models, here represented by the hypoplastic clay model, are powerful tools which provide engineers with reliable responses in various practical applications. However, the model calibration is not an easy task. Calibration of these models can be addressed with several approaches, which are generally distinguished as stochastic or deterministic approaches. In general, these approaches extract information from the experimental data and the subsequent optimisation process finds the best combination of parameters to fit the desired constraints. . The deterministic approach was integrated and combined in development of the online automated calibration tool ExCalibre. This paper presents a Machine Learning approach for automated calibration of the Hypoplastic Clay model. By using pairs of input experimental data and calibrated results performed by ExCalibre as training data, a Deep Neural Networks (DNNs) model is constructed to recognise how the experimental data can be used to derive the asymptotic state parameters such as the slope and the interception of the Normal Compression Line (NCL), or the critical friction angle, and the optimised stiffness parameters. The training and testing data comprise of In-house protocols and User-upload data over 3 years of launching the ExCalibre, and synthetic data with small distortion to prevent overfitting. Finally, investigations on how the DNNs model recognises the asymptotic patterns, as well as its calibration results will be presented.

How to cite: Do, P. and Kadlicek, T.: Calibration of the Hypoplastic Clay model with a deep neutral network, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12065, https://doi.org/10.5194/egusphere-egu25-12065, 2025.

EGU25-12242 | Posters on site | GI2.4

Automated Particle Size Distribution Estimation of Rock Avalanches using Deep Learning 

Ruoshen Lin, Michel Jaboyedoff, Marc-Henri Derron, and Tianxin Lu

Accurate estimation of Particle Size Distribution (PSD) in rock avalanche deposits is essential for understanding the fragmentation processes and spatial distribution characteristics during mass movement. However, traditional methods, such as physical sieving or visual field estimation, are time-consuming, labor-intensive, and impractical for large-scale field measurements. To address these limitations, this study presents an automated PSD estimation framework that combines UAV imagery and deep learning-based segmentation. A synthetic dataset was used to train the segmentation model, improving its robustness across different scenarios. Image resolution adjustments were applied to improve detection accuracy for small and overlapping particles. Additionally, Fourier analysis was utilized to reconstruct smooth and continuous particle contours, to effectively handle overlapping particles. The reconstructed 2D outlines were further used to estimate 3D particle volumes through the shape-volume model based on laboratory and literature data. Projection correction was applied to mitigate image distortions to ensure precise volume predictions. The proposed approach overcomes the limitations of traditional methods dealing with complex particle distributions in real field environments. The results demonstrate the effectiveness of the proposed method for large-scale particle detection and volume estimation, providing new insights into rock avalanche fragmentation dynamics.

Keywords: Rock avalanche; Particle size distribution (PSD); deep learning; UAV Imagery

How to cite: Lin, R., Jaboyedoff, M., Derron, M.-H., and Lu, T.: Automated Particle Size Distribution Estimation of Rock Avalanches using Deep Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12242, https://doi.org/10.5194/egusphere-egu25-12242, 2025.

EGU25-12675 | ECS | Orals | GI2.4

Automatic Detection of Cimiciato Defect in Hazelnuts Using Deep Learning and X-ray Radiography 

Antonio Gaetano Napolitano, Giacomo Mele, Laura Gargiulo, Matteo Giaccone, and Andrea Vitale

Hazelnuts are a significant crop, with global production exceeding 1.25 million tons by 2023 (INC, 2023). Quality is threatened by biotic agents, including insects causing the cimiciato defect. This defect, from insect bites during fruit growth, results in off-flavor, tissue alterations, and lipid oxidation (De Benedetta et al., 2023). Damage can be external or internal (hidden cimiciato). Industrial quality standards often exceed official regulations, making effective selection crucial. Traditional visual inspection is time-consuming and subjective. Non-destructive methods like NIR and NMR have potential but limited applicability. Deep Learning (DL) has revolutionized image classification, proving effective in agriculture, including disease and pest management (Mohanty et al., 2016; Dhaka et al., 2021; Meena et al., 2023). This study explores Deep Learning (DL) for automated detection of the cimiciato defect in hazelnuts using X-ray radiographs. Cimiciato, caused by insect feeding, degrades hazelnut quality, requiring product selection. Traditional methods are time-consuming and subjective. We propose a Convolutional Neural Network (CNN) model trained on X-ray images to classify hazelnuts as healthy or infected. Results demonstrate the model's effectiveness, offering a non-destructive, automated quality control solution.
Radiographs were acquired using a cone-beam micro-tomograph. Each hazelnut was positioned on a rotating stage. A CNN model was used for classification. CNNs effectively extract features from images. Convolutional layers apply filters to identify features; pooling layers reduce data dimensionality; fully connected layers combine features for classification. The Inception-ResNet-V2 architecture was chosen, combining Inception modules and residual connections (Szegedy et al., 2017). The model was trained (128 image batch size, 0.001 learning rate, 30 epochs), comparing SGD, ADAM, and RMSP optimizers. Images were pre-processed: resizing, pixel normalization, and data augmentation. 
The test dataset evaluated the trained network. SGD, ADAM, and RMSP yielded similar results. Confusion matrices visualize performance. ADAM performed best, but all achieved good results, especially for cimiciato detection. 

Keywords: Cimiciato defect, Hazelnut, Deep Learning, X-ray radiography, CNN

How to cite: Napolitano, A. G., Mele, G., Gargiulo, L., Giaccone, M., and Vitale, A.: Automatic Detection of Cimiciato Defect in Hazelnuts Using Deep Learning and X-ray Radiography, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12675, https://doi.org/10.5194/egusphere-egu25-12675, 2025.

EGU25-13315 | Posters on site | GI2.4

Harnessing AI and Decentralized Networks for Next-Generation Geophysical Forecasting 

Gabriel Moraga, Steven Hristopoulos, and Noah Pearson Kramer

Artificial Intelligence (AI) is revolutionizing geosciences, aligning perfectly with the themes of session GI2.4 by enabling the analysis of complex, multidimensional datasets and delivering actionable insights at unprecedented scales. This study presents two innovative AI-driven frameworks addressing critical challenges in soil moisture prediction and geomagnetic disturbance forecasting. Both approaches leverage decentralized networks to achieve scalability, foster collaboration, and enhance model performance through continuous refinement by a distributed community of contributors.

For soil moisture prediction, our multi-stream base model integrates data from Sentinel-2 (high-resolution spectral imagery), SMAP L4 (volumetric water content), ERA5 (meteorological variables), and SRTM (elevation data). The model predicts surface and rootzone soil moisture with six-hour lead times, achieving RMSE values of 0.1087 m³/m³ and 0.1183 m³/m³, respectively, across diverse Köppen-Geiger climate zones. By utilizing a decentralized network, contributors perform inference on 100 km² global regions, generating predictions evaluated against SMAP data using Root Mean Square Error (RMSE) and R² metrics. This system ensures robust model performance while addressing the spatial and temporal gaps inherent in traditional observational networks. These advances have significant implications for agriculture, hydrology, and climate modeling, enabling better water resource management, crop planning, and drought mitigation strategies. 

In geomagnetic disturbance forecasting, our GeoMagModel leverages Prophet, a time-series forecasting library, to predict the Disturbance Storm Time (Dst) index, a key indicator of geomagnetic activity. The model achieves an RMSE of 6.37 for December 2024 datasets, effectively capturing both trend shifts and weekly seasonality. The decentralized community enhances predictive accuracy by dynamically integrating historical and real-time Dst data, which is validated by benchmark predictions of the Kyoto World Data Center’s hourly records. This approach provides near-real-time forecasts critical for safeguarding power grids, satellite systems, and other infrastructure vulnerable to space weather events.

By integrating machine learning with decentralized computing and state-of-the-art data sources, these frameworks offer scalable solutions to longstanding challenges in geophysical monitoring. The decentralized network not only improves scalability but also incentivizes the geoscience community to refine baseline models, fostering innovation and enabling systems to outpace state-of-the-art benchmarks. The implications of this work extend beyond immediate applications, paving the way for hybrid models that combine AI-driven predictions with physical process-based simulations. This fusion has the potential to improve understanding and resilience in critical domains such as water resource planning, disaster mitigation, and space weather forecasting. By addressing the limitations of traditional observation systems and delivering actionable insights at scale, these AI-driven frameworks represent a paradigm shift in how we approach and solve complex geoscientific problems.

How to cite: Moraga, G., Hristopoulos, S., and Pearson Kramer, N.: Harnessing AI and Decentralized Networks for Next-Generation Geophysical Forecasting, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13315, https://doi.org/10.5194/egusphere-egu25-13315, 2025.

EGU25-15949 | ECS | Posters on site | GI2.4

Advanced Super-Resolution Techniques for Optical Payloads in Earth Observation: Combining Traditional and Deep Learning Methods 

Camilla De Martino, Vincenzo Della Corte, Laura Inno, Fabio Cozzolino, Giacomo Ruggiero, Vania Da Deppo, Paola Zuppella, Lama Moualla, and Sara Venafra

Small satellite platforms are increasingly used for Earth observation due to their cost-effectiveness and flexibility. However, their limited payload size often results in reduced spatial resolution of captured images. In our work, we address this challenge by proposing an advanced multi-image super-resolution (MISR) approach tailored for small satellite applications.

It integrates:

  • Sub-pixel image registration and on curvelet transform-based interpolation to preserve high-frequency details while reducing artifacts;
  • A novel hybrid method called SP-MISR (Subpixel Multi-Image Super-Resolution), which leverages Convolutional Neural Networks (CNNs) for local detail analysis and Transformers for global spatial relationships.

Our experimental results demonstrate that this combined approach  significantly improves image sharpness, preserves fine details, and reduces artifacts, outperforming traditional super-resolution techniques. Moreover, SP-MISR exhibits robustness in processing noisy and distorted images, making it particularly suitable for the constrained imaging systems of small satellites.

Future developments will focus on improving computational efficiency, reducing interpolation errors, and extending the method to multi-spectral imaging and interplanetary missions, by exploring explore pure deep learning techniques.

This work highlights the potential of integrating traditional and deep learning methodologies to enhance image quality, thus expanding the scientific and operational capabilities of small satellite missions.

How to cite: De Martino, C., Della Corte, V., Inno, L., Cozzolino, F., Ruggiero, G., Da Deppo, V., Zuppella, P., Moualla, L., and Venafra, S.: Advanced Super-Resolution Techniques for Optical Payloads in Earth Observation: Combining Traditional and Deep Learning Methods, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15949, https://doi.org/10.5194/egusphere-egu25-15949, 2025.

Soil carbonates are critical players in the global carbon cycle and have a profound influence on soil health and agricultural productivity. Their quantification is also central to carbon sequestration efforts, where accurate measurement of soil carbonates can inform strategies for reducing atmospheric CO2. However, conventional methods of carbonate analysis in soil---while effective---are often slow, costly, and labor-intensive (1).

In this study, we introduce Shifted Excitation Raman Difference Spectroscopy (SERDS) as a rapid, non-destructive alternative, further enhanced by advanced preprocessing techniques and machine learning algorithms. Specifically, we employ Asymmetric Least Squares (ALS) for background correction, Standard Normal Variate (SNV) for normalization, and Savitzky–Golay filtering for smoothing. Unlike conventional Raman spectroscopy, SERDS effectively eliminates background fluorescence and reduces overlapping peaks, resulting in clearer spectral signatures (2)

We employed Partial Least Squares Regression (PLSR) and eXtreme Gradient Boosting (XGBoost) to predict the inorganic carbon content from the carbonate vibrational modes in conventional Raman and SERDS spectra, benchmarked against total inorganic carbon (TIC) measurements from coulometric titration. Our results show that switching to dual-laser SERDS substantially boosted model performance. For PLSR, the coefficient of determination (R2) improved from 0.8 to 0.88 (an increase of about 10.5%), and the root-mean-square error (RMSE) declined from 0.29 to 0.22 (26% decrease). The XGBoost model exhibited an even greater increase, with R2 increasing from 0.63 to 0.93 (approximately 49% improvement) and RMSE dropping from 0.39 to 0.16 (59% reduction).

Figure 1: Left: All SERDS data of soil samples showing the main carbonate peak; Right: XGBoost model prediction of soil inorganic carbon using SERDS data.

These findings underscore the potential of SERDS to replace conventional methods for carbonate quantification, offering reduced cost, faster analysis, and essentially no sample preparation. Furthermore, by providing highly accurate carbonate measurements, this methodology can be pivotal for carbon sequestration assessments and large-scale soil management practices, helping to advance both environmental sustainability and agricultural productivity.

References:

1) Barra I, Haefele SM, Sakrabani R, Kebede F. Soil spectroscopy with the use of chemometrics, machine learning and pre-processing techniques in soil diagnosis: Recent advances–A review. TrAC Trends in Analytical Chemistry. 2021 Feb 1;135:116166.

2) Orlando A, Franceschini F, Muscas C, Pidkova S, Bartoli M, Rovere M, Tagliaferro A. A comprehensive review on Raman spectroscopy applications. Chemosensors. 2021 Sep 13;9(9):262.

How to cite: Poursorkh, Z., Solomatova, N., and Grant, E.: Quantifying Soil Inorganic Matter: Integrating Shifted Excitation Raman Difference Spectroscopy (SERDS) with Machine Learning for Enhanced Analysis of Carbonates, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16705, https://doi.org/10.5194/egusphere-egu25-16705, 2025.

Seismic exploration heavily relies on the accurate processing of seismic data, as high-quality reconstructed data is essential for reliable imaging and interpretation. In recent years, data-driven approaches have shown great promise in seismic data processing. However, supervised learning methods require large amounts of labeled data, while generative models, such as GANs, often encounter issues like mode collapse and instability. On the other hand, generative diffusion models, leveraging principles from nonequilibrium thermodynamics and Markov processes, have emerged as powerful tools for capturing complex data distributions.

Despite these advantages, Denoising Diffusion Probabilistic Models (DDPM) purely generate data distributions from the latent space with the reliance on random noise, making it inadequate for seismic data reconstruction where the goal is to accurately recover missing traces. Thus, DDPM is  lacks interpretability in seismic data restoration, and may disrupt the structured patterns crucial for interpolating seismic signals. Furthermore, we view the reverse process that starts from noise as unnecessary and inefficient for reconstruction task.

To address these challenges, we propose a novel Conditional Residual Diffusion Model (CRDM) that enhances both certainty and interpretability by incorporating residual diffusion and conditional constraints derived from observed seismic data (Fig.1). This approach better aligns with the inherent structure of seismic signals, enabling more accurate and interpretable reconstruction. The model is grounded in DDPM, with mathematical derivations for loss functions, conditional probability distributions, and reverse inference steps, ensuring both theoretical rigor and practical applicability.

Additionally, Our CRDM utilizes a shallow U-Net architecture featuring one down-sampling and one up-sampling layer integrated with Multi-Head Self-Attention (MHSA), which significantly enhances the model's efficiency and effectiveness. Experimental results (Fig.2) demonstrate that CRDM outperforms DDPM, denoising convolutional neural network (DnCNN), and fast projection onto convex sets (FPOCS), achieving a 15.1% improvement in reconstruction SNR and reducing computation time by 139 times compared to DDPM. Notably, CRDM achieves optimal results in a few diffusion steps, whereas DDPM typically requires thousands of steps.

The innovative approach generates data through residuals for determinism, while guiding the processing with noise for diversity. This not only enhances the interpretability and efficiency of seismic data reconstruction, but also positions the model as a promising tool for advancing data-driven seismic processing through flexible coefficient adjustment. Therefore, we believe this model has great potential for broader applications in geophysical data analysis, offering significant value in accurately depicting complex geological structures and providing more effective guidance for petroleum exploration.

Fig.1 The framework of CRDM. The model consists of two stage: (a) the training stage with forward diffusion process; (b) the sampling stage for seismic data reconstruction.

Fig.2 Reconstruction results and residuals of the 1994 BP seismic data with 50% irregular missing traces. (a) Complete data, reconstruction using (b) FPOCS (SNR=10.5dB), (c) DnCNN (SNR=15.6dB), (d) DDPM(SNR=18.4dB), (e) CRDM(SNR=21.2dB), and (f) observed seismic data with 50% missing traces. (g-j) display the residuals corresponding to reconstructions (b-e), respectively. The red box highlights a zoomed-in region, which is shown in detail in (k-t).

 

How to cite: Gong, X. and chen, S.: Enhancing Seismic Data Reconstruction with a Conditionally Constrained Residual Diffusion Model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16835, https://doi.org/10.5194/egusphere-egu25-16835, 2025.

Recent trends in climate change and global warming have amplified the occurrence of severe convective initiation (CI) and other extreme weather events, underscoring the importance of high-frequency remote sensing. Modern satellite constellations such as EUMETSAT, GOES, Himawari, and GK2A now offer global monitoring intervals of 10 minutes and regional updates as frequent as 1–2 minutes. However, these capabilities are not universally accessible—many developing countries lack in-orbit assets or historical high-frequency data archives.

This study presents a zero-shot video frame interpolation (VFI) approach to generate high-frequency (1–10 minute) satellite imagery from legacy or sparsely sampled observations. Leveraging a flexible Many-to-Many Splatting VFI model, our framework avoids domain-specific retraining while delivering reliable intermediate frames. We validate the method using overlapping data from the KMA GK2A (10-minute full-disk, 2-minute Asia-Pacific) and KMA COMS (3-hour full-disk, 15-minute Asia-Pacific) satellites over the period July 25, 2019, to March 31, 2020.

Our results indicate significant improvements in both PSNR and SSIM metrics, confirming the model’s efficacy in three critical applications:

  • Up-sampling Archived Geostationary Data

    • Enhancing the temporal resolution of older satellite imagery (e.g., COMS 30-minute or 3-hour intervals) to match or approximate modern satellite capabilities (e.g., GK2A 10-minute intervals). This harmonization facilitates unified climatology analyses spanning multiple generations of instruments.
  • Sensor-Error Correction and Gap Filling

    • Recovering missing or corrupted frames resulting from attitude adjustments, sensor calibrations, or malfunctions on geostationary satellites. Ensuring a continuous record provides more robust inputs to operational forecasting and climate assessments.
  • Delayed High-Frequency Observation Services

    • Enabling resource-constrained meteorological agencies to retrospectively produce and disseminate high-frequency satellite products (e.g., near 1-minute intervals) to improve nowcasting, risk assessment, and disaster preparedness.

Our preliminary findings show minimal computational overhead per inference step, making this cost-effective method feasible for near-real-time deployment and post-event analyses alike. By bridging temporal gaps in global satellite datasets, this technique supports advanced level-2 products such as cloud-tracking and convective-initiation alerts, thereby driving broader socioeconomic and scientific benefits in both developed and developing regions.

How to cite: Ryu, H. and Choi, Y.: Toward High-Frequency Satellite Observations in Data-Sparse Regions: A Zero-Shot Interpolation Framework for Missing and Historical Imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17456, https://doi.org/10.5194/egusphere-egu25-17456, 2025.

EGU25-17957 | Posters on site | GI2.4

Fuzzy clustering of electrical and seismic data for the detection of geophysical targets 

Giorgio De Donno, Michele Cercato, Davide Melegari, Valeria Paoletti, Guido Penta de Peppo, and Ester Piegari

Tomographic methods, such as electrical resistivity tomography (ERT), induced polarization (IP) and seismic refraction tomography (SRT) are often effective for detecting geophysical targets in disparate real-world scenarios. However, a final reconstruction expressed only in terms of individual geophysical parameters (resistivity, chargeability, P-wave velocity) leaves room for ambiguity in complex sites exhibiting several transitions between layers or zones having different geophysical properties. In such cases, the sensitivity of the geophysical parameters for the various methods can differ significantly, so that a univocal interpretation based only on a visual comparison of the different models is often ineffective. To overcome these limits, in this work we present a machine learning-based quantitative approach for the detection of geophysical targets associated with both geological and anthropogenic scenarios. We integrate two-dimensional ERT, IP and SRT tomographic data with a soft clustering analysis by the Fuzzy C-Means (FCM) to obtain a final combined section, where each pixel is characterized by a cluster index and an associated membership value. The membership function of the Fuzzy C-Means is a good estimator of the accuracy of the subsurface reconstruction, as it ranges from 0 to 1, with 1 reflecting a high reliability of the clustering analysis. We apply this method to two case studies, related to the detection of leachate accumulation areas in a municipal solid waste landfill and to the bedrock characterization in a site prone to instability. In both cases, we detect the cluster associated with the geophysical targets of interest and our final sections are validated by a good agreement with the available direct information (boreholes and wells). The accuracy of the reconstruction is consistently high across most areas (membership values > 0.75), even though it is reduced in areas where the resolution of geophysical data is lower. Therefore, this approach may be a valuable automatic tool for optimizing the cost-effectiveness of projects where new constructions or remediation interventions have to be planned.

How to cite: De Donno, G., Cercato, M., Melegari, D., Paoletti, V., Penta de Peppo, G., and Piegari, E.: Fuzzy clustering of electrical and seismic data for the detection of geophysical targets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17957, https://doi.org/10.5194/egusphere-egu25-17957, 2025.

EGU25-18082 | ECS | Orals | GI2.4

Integrating FDEM data with K-Means clustering for improved archaeological site identification 

Angelica Capozzoli, Valeria Paoletti, Federico Cella, Mauro La Manna, and Ester Piegari

The Frequency Domain Electromagnetic (FDEM) method is a cost-effective geophysical technique that simultaneously studies the electrical and magnetic properties of a medium, providing data as in-phase and out-of-phase components of the electromagnetic field. Although FDEM yields valuable insights, its results can be complex to interpret, and the two EM field components are normally only visually inspected to support findings from other techniques. This study aims to enhance FDEM data interpretation using an unsupervised learning technique. The proposed approach seeks to automate and expedite the interpretative phase. By applying the K-Means clustering algorithm, we divided the FDEM data into several clusters based on specific intervals of the in-phase and quadrature components, resulting in integrated maps of EM components. Combining these maps with geological and archaeological insights helped identifying areas of potential archaeological interest. This method was applied to the Torre Galli archaeological site in Calabria, Italy, known for its significance in Iron Age studies.

Based on comparisons with the findings of earlier excavations and results from a magnetic survey, the proposed procedure shows promise in improving the efficiency and accuracy of the FDEM method in identifying areas of archaeological interest. This suggests that automating the interpretation process could lead to a better cost management and time optimization in geophysical and archaeological studies.

How to cite: Capozzoli, A., Paoletti, V., Cella, F., La Manna, M., and Piegari, E.: Integrating FDEM data with K-Means clustering for improved archaeological site identification, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18082, https://doi.org/10.5194/egusphere-egu25-18082, 2025.

Artificial Intelligence (AI) is revolutionizing the field of geomorphology, offering a robust tool for objective and quantitative analyses. This pioneering study proposes an innovative framework based on Machine Learning clustering techniques, capable of classifying drainage patterns into multiple morphological classes. This work follows up on a related study in which an attempt at classifying 156 terrestrial and extraterrestrial (Mars and Titan) river networks was made. Rivers’ outlines are intrinsically noisy, difficult to isolate from the background, and can be ambiguous for the human eye. The previous works have been focused on accurately classifying patterns, using the expertise of morphologists, thus introducing a weak link, the human eye, in the chain. This time, a reliable, automatic, and scalable methodology has been obtained, leveraging computers’ precision, objectivity, and computational power. The HydroRIVERS dataset, a publicly available data bank containing vector data, was utilized in this study. All HydroRIVERS data layers are provided in a geographic projection (latitude/longitude), referenced to the WGS84 datum. Each data layer includes an attribute table with information on the morphometric characteristics of each river reach. The input parameters for the clustering models included morphometric features such as LENGTH_KM, DIS_AV_CMS, ORD_STRA, ORD_CLAS, and ORD_FLOW.
During a preliminary experiment, a local convexity test was conducted to determine the optimal number of clusters (k) to identify the best metric values. This test made sure that the number of clusters with the highest evaluation metric was selected, varying in a closed numeric interval. Each cluster corresponds to a specific river class. Significant results were obtained with k = 6, k = 8, k = 10, and k = 12. Subsequently, the K-Means algorithm was applied, grouping the dataset into distinct clusters based on the morphometric parameters. The results were remarkable, with 10 being the best value for k. The results indicate that the clustering algorithm is able to optimally separate the dataset, producing a high inter-cluster distance and a low intra-cluster distance. The dataset points along the features, as highlighted by the three principal components obtained by performing PCA on the final five-dimensional clustering resulting vector space, are well grouped in relatively small clusters, far away from each other. The next step involves using the centroids obtained from the analysis of the large dataset as a reference for classifying the 155 rivers. In general, the centroids obtained from this kind of Learning could be of great value to the scientific community, establishing a new and innovative way of discerning between different classes of rivers without having to manually analyze and inspect images. This approach promises efficient and accurate classification of both terrestrial and extraterrestrial drainage patterns.

How to cite: D'Aniello, M. and Donadio, C.: A novel approach using Machine Learning to objectively classify terrestrial and extraterrestrial river networks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19838, https://doi.org/10.5194/egusphere-egu25-19838, 2025.

EGU25-20509 | Orals | GI2.4 | Highlight

Making Sense of AI: The Important Role of Education and Communication 

Luis Azevedo Rodrigues

The big developments in Artificial Intelligence (AI) generated a mixture of fascination and apprehension, echoing historical responses to unknown phenomena. Like how medieval European societies interpreted certain natural events and instruments as magical, AI is currently perceived by many as a “black box,” blurring the lines between advanced technology and mystical force. This phenomenon fosters misconceptions and uncertainties about AI’s actual mechanisms, benefits, and risks. Consequently, it underscores the urgent need for science communicators and science museums to adopt an active role in enhancing the general public’s AI literacy and in debunking some of its enigmatic traits.

By drawing parallels with the Middle Ages—when objects such as mirrors and magnetite were often attributed supernatural capabilities—modern AI tools LLMs or image and video generators are frequently viewed as possessing a “magical” principle. The public’s limited grasp of how AI processes inputs and produces outputs further intensifies this impression. The lack of transparency (or “black box” effect) in deep learning algorithms, combined with the ambiguity of human language, has been shown to fuel both wonder and anxiety.

Science museums and communicators should have an active role by offering educational programs that demystify AI through different demographic and social activities as well as promoting the public debate. These initiatives could clarify AI’s underlying mathematical and computational principles, highlight practical examples of AI-driven applications, and examine ethical considerations surrounding its deployment. Public understanding of AI’s capabilities and limitations is crucial not only to temper undue fears but also to encourage informed engagement with emerging technologies.

How to cite: Azevedo Rodrigues, L.: Making Sense of AI: The Important Role of Education and Communication, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20509, https://doi.org/10.5194/egusphere-egu25-20509, 2025.

EGU25-292 | ECS | Posters on site | HS3.3

Influence of Temperature on Streamflow Dynamics: A Multi-Catchment Analysis Using the PCMCI+ Causal Discovery Algorithm 

Hossein Abbasizadeh, Petr Maca, and Martin Hanel

While precipitation is the primary driver of streamflow variability, temperature also plays a significant role. Temperature influences streamflow by modifying precipitation, evapotranspiration, and soil moisture. While this relationship is often studied using hydrological or black-box models, the causal effect of temperature dynamics on streamflow at the catchment scale is not fully understood. This study investigates the causal relationship between precipitation, temperature, and streamflow time series using the PCMCI+ causal discovery method. Having the causal structure, the total causal effect of temperature on stream flow is estimated. The analysis is conducted on CAMELS-GB (671 catchments) and LamaH (859 catchments) datasets to study the causal effects of temperature on streamflow across a wide range of catchments with different climate and physiographic characteristics. Preliminary results indicate that temperature significantly influences streamflow within a specific range, which changes over time for most catchments. The changes in the range within which the temperature has high causal effects on the temperature might be due to the shift in catchment storage and precipitation patterns, leading to a change in catchment response to temperature. These findings highlight the importance of identifying a relationship between temperature streamflow variability from a cause-and-effect perspective. This suggests that incorporating causal information can improve the modelling of the hydrological systems under changing climate. 

How to cite: Abbasizadeh, H., Maca, P., and Hanel, M.: Influence of Temperature on Streamflow Dynamics: A Multi-Catchment Analysis Using the PCMCI+ Causal Discovery Algorithm, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-292, https://doi.org/10.5194/egusphere-egu25-292, 2025.

EGU25-482 | ECS | Orals | HS3.3

Equifinality in River Discharge Prediction Revealed Through Explainable AI  

Qiuyang Chen, Simon Mudd, and Simon Moulds

River discharge prediction is critical for water resource management, yet equifinality—where multiple model configurations achieve similar accuracy—complicates process understanding. We explored this phenomenon using Long Short-Term Memory (LSTM) models trained on UK river basins, incorporating geomorphic descriptors derived from Digital Terrain Models and other environmental features from the CAMELS-GB dataset, including land cover, soil, and climate variables. Explainable AI techniques revealed that the models rely on different, yet equally effective, combinations of correlated features to achieve comparable performance. This variability underscores the complexity of hydrological systems and highlights the importance of integrating explainability and domain knowledge in machine learning to enhance model interpretability and robustness.

How to cite: Chen, Q., Mudd, S., and Moulds, S.: Equifinality in River Discharge Prediction Revealed Through Explainable AI , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-482, https://doi.org/10.5194/egusphere-egu25-482, 2025.

EGU25-1863 | ECS | Posters on site | HS3.3

Predicting Water Movement in Unsaturated Soil Using Physics-Informed Deep Operator Networks 

Qiang Ye, Zijie Huang, Qiang Zheng, and Lingzao Zeng

Accurate modeling of soil water movement in the unsaturated zone is essential for effective soil and water resources management. Physics-informed neural networks (PINNs) offer promising potential for this purpose, but necessitate retraining upon changes in initial or boundary conditions, posing a challenge when adapting to variable natural conditions. To address this issue, inspired by the operator learning with more universal applicability than function learning, we develop a physics-informed deep operator network (PI-DeepONet), integrating physical principles and observed data, to simulate soil water movement under variable boundary conditions. In the numerical case, PI-DeepONet achieves the best performance among three modeling strategies when predicting soil moisture dynamics across different testing areas, especially for the extrapolation one. Guided by both data and physical mechanisms, PI-DeepONet demonstrates greater accuracy than HYDRUS in capturing spatio-temporal moisture variations in real-world scenario. Furthermore, PI-DeepONet successfully infers constitutive relationships and reconstructs missing boundary flux condition from limited data by incorporating known prior physical information, providing a unified solution for both forward and inverse problems. This study is the first to develop a PI-DeepONet specifically for modeling real-world soil water movement, highlighting its potential to improve predictive accuracy and reliability in vadose zone modeling by combining data-driven approaches with physical principles.

How to cite: Ye, Q., Huang, Z., Zheng, Q., and Zeng, L.: Predicting Water Movement in Unsaturated Soil Using Physics-Informed Deep Operator Networks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1863, https://doi.org/10.5194/egusphere-egu25-1863, 2025.

EGU25-2661 | Posters on site | HS3.3

Using explainable artificial intelligence as a diagnostic tool  

Sheng Ye, Jiyu Li, Yifan Chai, Lin Liu, Murugesu Sivapalan, and Qihua Ran

Recent applications have demonstrated the strength of deep learning (DL) in information extraction and prediction. However, its limitations in interpretability have delayed its popularity for use in facilitating advancement of hydrologic understanding. Here we present a framework using explainable artificial intelligence (XAI) as a diagnostic tool to investigate distributed soil moisture dynamics within a watershed. Soil moisture and its movement generated by physically based hydrologic model were used to train a long short-term memory (LSTM) network, whose feature attribution was then evaluated by XAI methods. The aggregated feature importance presents abrupt rise in the model’s nodes located in riparian area, indicating threshold behavior in runoff generation and development of hydrologic connectivity at the watershed scale, which helps explain the rapid increase in streamflow. This work represents a demonstration of the potential of XAI to uncover underlying physical mechanisms and to help develop new theories from observed data.

How to cite: Ye, S., Li, J., Chai, Y., Liu, L., Sivapalan, M., and Ran, Q.: Using explainable artificial intelligence as a diagnostic tool , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2661, https://doi.org/10.5194/egusphere-egu25-2661, 2025.

EGU25-2740 | ECS | Orals | HS3.3

A General Framework for Integrating Neural Networks into Numerical Resolution Methods for Spatially Distributed Hydrological Models 

Ngo Nghi Truyen Huynh, Pierre-André Garambois, Benjamin Renard, and Jérôme Monnier
Machine learning (ML) methods have been utilized in hydrology for decades. Recently, hybrid approaches that combine data-driven techniques with process-based models have gained attention, highlighting the complementary strengths of ML and physical models. However, the explicability and adaptability of such hybrid models remain open questions. This work introduces a general framework for incorporating neural networks (NNs) and ML techniques into a regionalizable, spatially distributed hydrological model. As a case study, a simple NN is employed to correct internal fluxes within a conceptual GR hydrological model that allows analytical integration. The corresponding hybrid ordinary differential equation set is integrated with an implicit numerical scheme solved by the Newton-Raphson method. Implementation in Fortran-based code supports differentiability, enabling the computation of the cost gradient through a combination of an adjoint model and analytical NN gradients. Results over a large catchment sample show promising improvements in model accuracy and provide insights into hydrological behaviors through interpretable NN outputs. These findings demonstrate the framework's potential to advance hybrid hydrological modeling by enhancing explicability and adaptability. Additionally, the proposed framework offers flexibility for integration into other modeling chains and applications across diverse geophysical models.

How to cite: Huynh, N. N. T., Garambois, P.-A., Renard, B., and Monnier, J.: A General Framework for Integrating Neural Networks into Numerical Resolution Methods for Spatially Distributed Hydrological Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2740, https://doi.org/10.5194/egusphere-egu25-2740, 2025.

EGU25-3859 | ECS | Posters on site | HS3.3

Explainable convolutional neural network for flood susceptibility mapping in Southern Ontario   

Rahma Khalid and Usman T Khan

Flood susceptibility mapping (FSM) plays a crucial role in proactive flood risk management, particularly in light of increasing fluvial flooding events. Traditional FSM methods, such as physics-based and qualitative approaches, are hindered by either high computational demands or inherent uncertainty. To address this, machine learning (ML) models have become an increasingly popular FSM approach, though commonly cited as black-box approaches due to the difficulty associated with understanding their underlying mechanisms. In order to better understand the ML approaches used for FSM, this study uses the gradient-weighted class activation mapping (Grad-CAM) to interpret flood susceptibility predictions of a convolutional neural network (CNN) for the Don River watershed in Ontario, Canada. Grad-CAM is an explainable algorithm highlighting input regions that are influential to the output, aiding the user in understanding and visualizing model selected important features used to arrive at the prediction. Grad-CAM results are compared to the commonly used shapley additive explanation (SHAP) algorithm. SHAP is used to calculate the relative contribution of each input onto the output, and provides a benchmark for comparisons due to its popularity.

A two dimensional CNN with an architecture of two convolutional layers, two pooling layers and a fully connected layer is used to predict flood susceptibility. The inputs to the CNN include topographical and climactic variables across the entire watershed, with a 60-40% training and testing split respectively. The results of the CNN were compared against the floodplain map of the Don River. Using the area under curve- receiver operating characteristics (AUC-ROC) as a performance metric, the CNN exhibits high performance with an AUC-ROC of 0.96.

The study highlights the potential of CNNs for flood susceptibility mapping, as well as compares two explainable machine learning algorithms, helping to further their application within FSM. Explainable algorithms are essential to decision makers in flood risk management for proactive planning and resource allocation. Future work should explore expanding the scope to predict flood susceptibility at a nationwide level.

How to cite: Khalid, R. and Khan, U. T.: Explainable convolutional neural network for flood susceptibility mapping in Southern Ontario  , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3859, https://doi.org/10.5194/egusphere-egu25-3859, 2025.

Floods are the costliest hazard in Canada in terms of direct infrastructure damage. Flood susceptibility modelling (FSM) identifies flood hazard areas; input features are dependent on the study area and modelling methods, which affect the reliability and accuracy of FS maps. Typical features in FSM are static topographical inputs (digital elevation model, land use, wetness index, height above nearest drainage, etc.). Though meteorological variables have been included in FSM, they are often low temporal resolution (e.g. annual); seasonal meteorological variables are often not included. The 2023 Canadian National FS map was developed using machine learning (ML) ensembles, with features that include historical flood events and 30 years of climate data. This research initiates the update to the existing Canadian FS map by expanding the suite of input features used and comparing the impact of three feature selection methods (partial correlation, partial mutual information, combined neural pathway strength) on three types of ML algorithms: random forest, artificial neural network (ANN) and convoluted neural network (CNN). The expanded set of features includes geospatial indices and flood-specific meteorological data such as spring temperature, precipitation, and vapour pressure. Data from preceding seasons to specific flood events is also included. Preliminary findings from the feature selection methods show that including seasonal flood-specific meteorological data provides important information leading to better model performance. Model performances of the three algorithms were comparable. Random forest with extreme gradient boosting led to the highest model performance (AUC = 0.98, F1 = 0.94), followed by CNN (AUC = 0.0.96, F1 = 0.90). ANN ensemble with leave-one-out-cross-validation resulted in the lowest model performance (AUC = 0.91, F1 = 0.85). Results contribute to the development of an improved national FS map for Canada.

How to cite: Dunbar, K. E., McGrath, H., and Khan, U. T.: Enhancing flood susceptibility modelling in Canada: Integrating seasonal meteorological data, feature selection and machine learning approaches, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3871, https://doi.org/10.5194/egusphere-egu25-3871, 2025.

EGU25-4428 | ECS | Orals | HS3.3

Towards Physics-consistent Foundation Models for Flood Forecasting 

Qingsong Xu and Xiao Xiang Zhu

Effective flood forecasting is critical for informed decision-making and timely emergency response. Traditional physical models, which rely on fixed-resolution spatial grids and input parameters, often incur substantial computational costs, limiting their capacity to accurately predict flood peaks and provide prompt hazard warnings.  This paper introduces methods to ensure physical consistency in machine learning models, aiming to develop a fast, stable, accurate, cross-regional, and downscaled neural flood forecasting foundation model. Specifically, we present a Physics-embedded Neural Network, which integrates the momentum and mass conservations of flood dynamics into a neural network. Additionally, we combine this Physics-embedded Neural Network with a diffusion-based generative model, enhancing physical process consistency for long-term, large-scale flood forecasting. We also briefly introduce other models that integrate physics and machine learning, such as the FloodCast model by incorporating hydrodynamic equations into its loss function to maintain physical consistency, and the UrbanFloodCast model by learning physical consistency from urban flood dynamic data. The performance of these models will be analyzed using our proposed FloodCastBench dataset, a comprehensive collection of low-fidelity and high-fidelity flood forecasting dataset and benchmark. Results from the dataset demonstrate that incorporating physical consistency significantly enhances flood forecasting accuracy, demystifies the black-box nature of machine learning frameworks, and increases confidence in addressing dynamical systems. Finally, we propose a Spatiotemporal Foundation Model capable of forecasting floods across a variety of scales and regions.

How to cite: Xu, Q. and Zhu, X. X.: Towards Physics-consistent Foundation Models for Flood Forecasting, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4428, https://doi.org/10.5194/egusphere-egu25-4428, 2025.

EGU25-4845 | ECS | Posters on site | HS3.3

Using Causal Discovery to Identify Drivers and Controls of Streamflow in Large Sample Hydrology 

David Strahl, Sebastian Gnann, Karoline Wiesner, and Thorsten Wagener

Catchments are the fundamental units of hydrological analysis and integrate a vast number of physical, biological, and anthropogenic processes. Traditional hydrological modelling approaches, however, adopt a bottom-up perspective, aggregating small-scale physical principles to predict large-scale catchment behaviour. While effective for prediction, this approach can fall short in advancing our understanding of emergent processes and their interactions given the strong dependence on a priori assumptions. To address this gap, causal discovery algorithms offer a promising alternative by moving beyond simple correlation to directly identifying the dynamic causal structures emerging at the catchment scale. In this study, we applied the PCMCI+ algorithm to the CAMELS-US dataset in combination with a subsequent causal effect estimation. We explored how and to what extent dynamic causal structures can be learned from hydro-meteorological data alone, and which catchment properties and conditions influence their expression. We find that causal discovery in hydrology faces challenges due to non-stationarity, unsuitable conditional independence tests, and unmet methodological assumptions. Despite these limitations, our approach reconstructed physically plausible relationships controlled by meaningful catchment properties. These results highlight the potential of causal discovery in hydrology, where it could serve as a complementary framework for model evaluation studies or as an integral part of the model development process.

How to cite: Strahl, D., Gnann, S., Wiesner, K., and Wagener, T.: Using Causal Discovery to Identify Drivers and Controls of Streamflow in Large Sample Hydrology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4845, https://doi.org/10.5194/egusphere-egu25-4845, 2025.

EGU25-5052 | ECS | Posters on site | HS3.3

Can CNN-LSTM and lumped models improve (extreme) streamflow prediction of semi-distributed models? A comparative analysis of two hybrid frameworks 

Aseel Mohamed, Awad M. Ali, Ahmed Ali, Osama Hassan, Mohamed E. Elbasheer, and Mutaz Abdelaziz

Water resources management depends heavily on hydrological modeling for reservoir operation and risk mitigation, especially in data-scarce regions. Hybrid approaches that combine artificial intelligence and conceptual models offer great potential for accurate streamflow prediction. However, their implementation can be time-consuming and applied in different configurations. This study comprehensively compares two promising hybrid frameworks: the Conceptual-Data-Driven Approach (CDDA) and the Ensemble Approach. The analysis was conducted in the Upper Blue Nile Basin in Ethiopia over the period from 2002 to 2019. Six baseline models were developed, including CNN-LSTM (data-driven), NAM and HBV-Light (lumped), and SWAT+, WEAP, and HEC-HMS (semi-distributed). All models achieved NSE ≥ 0.85 during the validation period, with CNN-LSTM performing best (NSE = 0.94). Each model was integrated into the two hybrid frameworks using Random Forest (RF) or Artificial Neural Networks (ANN). Results showed that the Ensemble Approach outperformed CDDA by combining two conceptual models. ANN performed better than RF across both frameworks. Hybrid modeling significantly improved semi-distributed models, while lumped and data-driven models showed minimal benefits. In the Ensemble Approach, normal and extreme flows simulated using semi-distributed models performed best when supported by CNN-LSTM or lumped models. Our analysis also demonstrated the robustness of the Ensemble Approach for selecting the supporting model. These findings emphasize the value and feasibility of the Ensemble Approach for improving streamflow prediction and better supporting decision-making in data-scarce regions. Nevertheless, a thorough understanding of the opportunities in hybrid modeling requires further research with a specific focus on operational forecasting.

How to cite: Mohamed, A., M. Ali, A., Ali, A., Hassan, O., E. Elbasheer, M., and Abdelaziz, M.: Can CNN-LSTM and lumped models improve (extreme) streamflow prediction of semi-distributed models? A comparative analysis of two hybrid frameworks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5052, https://doi.org/10.5194/egusphere-egu25-5052, 2025.

EGU25-5727 | ECS | Posters on site | HS3.3

Using Explainable Artifical Intelligence (XAI) to Analyze the Behavior of Global Water Models 

Lea Faber, Karoline Wiesner, Ting Tang, Yoshihide Wada, and Thorsten Wagener

Model Intercomparison Projects in the Earth Sciences have shown, that the outputs of
Earth System Models often show large variations and can therefore give quite different results,
with no single model consistently outperforming others. Examples include Global Water
Models (GWMs), as well as Global Climate Models (GCMs). The high computational costs
of running such models make comprehensive statistical analyses challenging, a common issue
with many complex models today. Machine learning models have become popular surrogates
of slow process-based models, due to their computational speed, at least once trained. This
speed makes it possible to use techniques from Explainable AI (XAI) to analyze the behavior
of the surrogate model.
Here, we analyze long-term averages of the GWM ’Community Water Model’ (CWatM)
for different parts of the global domain for actual evapotranspiration Ea, total runoff Q and
groundwater recharge R. We train an artificial neural network on the model’s input and output
data and use three different strategies to assess the importance of input data: LassoNet for sub-
set selection and feature ranking, along with Sobol’ indices and DeepSHAP for interpretability.
Our results show that subset selection can effectively reduce model complexity before XAI
analysis. For some hydrological domains the number of relevant input
variables for a chosen output reduces to less then 15 variables out of 98 model inputs, while
others remain more complex, requiring many variables for performances with R2 > 0, 8.

How to cite: Faber, L., Wiesner, K., Tang, T., Wada, Y., and Wagener, T.: Using Explainable Artifical Intelligence (XAI) to Analyze the Behavior of Global Water Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5727, https://doi.org/10.5194/egusphere-egu25-5727, 2025.

EGU25-6278 | Orals | HS3.3

Training Surrogates with Knowledge and Data: A Bayesian Hybrid Modelling Strategy 

Anneli Guthke, Philipp Luca Reiser, and Paul Bürkner

Physics-based hydrological modelling provides great opportunities for risk assessment and water resources management. However, diagnostic model evaluation and quantitative uncertainty assessment remain a challenge: (1) Model choices, boundary conditions, and prior assumptions about input, parameter or data uncertainty might be hard to formulate or justify; (2) rigorous propagation of uncertainties struggles when the analysed model structure is not “true”, and (3) a full propagation of uncertainties is often computationally prohibitive for complex models.

Alternative approaches promote the extraction of information directly from data, thereby avoiding overly strict physics-based constraints and the pitfalls of uncertainty quantification. Challenges of these data-driven approaches include the lack (or difficulty of) explainability, transparency, and transferability to unseen scenarios.

To explore the frontier of where those two perspectives (should) converge, we investigate the potential of surrogate models (computationally cheaper, data-driven representations of complex models) as a binding link with several potential benefits: (1) they alleviate the computational burden and thereby allow for a fully Bayesian uncertainty analysis; (2) they are flexible enough to overcome structural deficits of the original complex model, thereby enabling a better predictive performance, and (3) being data-driven, we can elegantly fuse the information from available data into their training process.

Methodologically, we propose a weighted data-integrated training of surrogates via two competing approaches that differ technically, but also philosophically, and reveal complementing insights about the strengths and weaknesses of the physics-based model and about the additional information in the available data, thereby facilitating deeper system understanding and improved (hybrid) modelling. We demonstrate the proposed workflow on didactic examples and a real-world case study. We expect this approach to be generally useful for modelling dynamic systems, as it contributes to more realistic uncertainty assessment and opens up ways for model development.  

How to cite: Guthke, A., Reiser, P. L., and Bürkner, P.: Training Surrogates with Knowledge and Data: A Bayesian Hybrid Modelling Strategy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6278, https://doi.org/10.5194/egusphere-egu25-6278, 2025.

EGU25-7382 | Orals | HS3.3

Dynamical system neural network for hydrological modelling 

Derek Karssenberg

Neural networks are efficient and effective in predicting system states in hydrology. However, most current approaches lack hydrological flow partitioning, do not allow for training on measurements of multiple variables, or lack capability to tightly integrate physically-based components. To address these shortcomings I propose and evaluate an approach referred to below as Dynamical System Neural Network (DSNN). DSNN is a feedforward neural network with an architecture that resembles the organisation in components of the real-world system it represents. In hydrology, the DSNN represents each water flow (e.g. seepage, snow melt) by a collection of input, hidden, and output neural layers, where each input is the state of a hydrological storage (e.g. groundwater storage influencing seepage) or other variable (e.g. air temperature influencing snow melt). These components are interconnected to form a single neural network of the complete dynamical system considered, where all storages and flows are explicitly quantified. If physical understanding of a flow and its parameterization is available, a known formulation can be used as a replacement of a neural network component. The DSNN is applied forward in time, backpropagating gradients over all timesteps. It can be run in spatially lumped or semi-distributed mode. To demonstrate the approach, a DSNN is presented of the Austrian Dorfertal (Kals) Alpine catchment containing snow and subsurface water storages and associated flows including streamflow. The DSNN is trained, validated, and tested on daily streamflow over ~40 years. To explore the capability of the DSNN in estimating the magnitude and dynamics of internal system storages (snow water equivalent, subsurface water storage) and flows (evapotranspiration, sublimation, snowmelt, seepage), the DSNN is first trained and tested with streamflow data generated by a conceptual model. The DSNN turns out to be capable of reproducing - with a satisfactory level of precision - the system states and fluxes calculated by the conceptual model, with decreasing performance when measurement error is added to the artificially generated streamflow data before training. To explore its predictive performance, the DSNN is applied on measured streamflow data for the Dorfertal, comparing multiple DSNN setups that represent all flows as neural network components or only a subset of flows where remaining flows are represented with a standard conceptual model (e.g. linear reservoir). Preliminary results indicate that in predictive performance, in most setups, the DSNN outperforms a standard conceptual model trained on the same streamflow data, with NSE values for testing of 0.74 and 0.71, respectively. This preliminary result indicates DSNN to be a promising approach for blending process-based and neural network based modelling as well as for training (i.e. calibration) of neural network models on measurements of multiple hydrological variables as these are all explicitly represented by the DSNN and can thus be incorporated in the loss function (e.g. streamflow, snow depth, groundwater, evapotranspiration).

How to cite: Karssenberg, D.: Dynamical system neural network for hydrological modelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7382, https://doi.org/10.5194/egusphere-egu25-7382, 2025.

EGU25-7770 | Orals | HS3.3

Advancing distributed hydrological modeling with hybrid machinelearning 

Yi Zheng, Chao Wang, and Shijie Jiang
Accurately simulating large-scale water dynamics is important
for managing water resources, addressing climate change impacts, and
understanding hydrological variability. Despite advances in hydrological
modeling, simulating water fluxes and states at global or regional
scales remains challenging due to the complexity of distributed
processes and limited understanding of key components. Encoding physical
knowledge in deep neural networks (NNs) for differentiable modeling
offers a promising solution but has yet to be fully realized for
distributed hydrological models, especially for processes such as river
routing.
This study presents a novel differentiable modeling framework that
bridges physical and data-driven approaches for distributed hydrological
modeling. The framework encodes a large-scale hydrological model (i.e.,
HydroPy) as a neural network, incorporates an additional NN to map
spatially distributed parameters from local climate and land attributes,
and employs NN-based modules to represent poorly understood processes.
Multi-source observations are used to constrain the system in an
end-to-end manner, with the Amazon Basin as a case study to demonstrate
the framework’s applicability and effectiveness.
Results show that the developed model improves simulation accuracy by
30-40% compared to the original hydrological model. Replacing the
Penman-Monteith formulation with NN produces more realistic potential
evapotranspiration estimates. SHAP analysis of the NN parameterization
further reveals how climate and land attributes regulate the spatial
variability of key parameters. Overall, by integrating physical realism
with the flexibility of machine learning, this framework addresses
critical limitations of traditional hydrological models. It provides a
scalable, interpretable approach to advance large-scale hydrological
modeling and address pressing water and climate challenges.

How to cite: Zheng, Y., Wang, C., and Jiang, S.: Advancing distributed hydrological modeling with hybrid machinelearning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7770, https://doi.org/10.5194/egusphere-egu25-7770, 2025.

EGU25-9176 | ECS | Orals | HS3.3

GPU-Enabled Cell-to-Cell Routing in a High Resolution Hybrid Distributed Hydrological Model with Multi-Source Remote Sensing Data Assimilation: A Continental-Scale Computational Approach  

Mouad Ettalbi, Pierre-André Garambois, Ngo-Nghi-Truyen Huynh, Emmanuel Ferreira, and Nicolas Baghdadi

The integration of remote sensing observations into hydrological modeling frameworks presents a significant opportunity for improving spatial and temporal predictive capabilities across continental domains. This research introduces a novel hybrid distributed hydrological model that addresses key challenges in computational efficiency, by using a GPU-enabled computational infrastructure, and in predictive accuracy by assimilating multi-source remote sensing datasets, specifically satellite-based soil moisture and evapotranspiration, at a high spatial resolution (1km×1km) and temporal scale (hourly). The model addresses critical challenges in regional hydrological forecasting by leveraging advanced data assimilation techniques and machine learning methodologies.

The proposed hybrid modeling framework synthesizes physically-based distributed hydrologic modeling principles with data-driven machine learning approaches, facilitating a more comprehensive representation of land surface hydrological processes. A key innovation is the GPU-enabled cell-to-cell routing algorithm, which enables fast and efficient computational processing of complex hydrological connectivity and water movement across large spatial domains. By integrating remote sensing observations, the methodology enables enhanced initial condition specification and improved parameter estimation, particularly in regions characterized by sparse ground-based measurement networks.

Preliminary analytical results demonstrate significant improvements in model performance, particularly in capturing spatial and temporal variability of hydrological states and fluxes. The approach substantively advances current methodological capabilities in hydrological forecasting, offering a promising framework for developping enriched tensorial numerical solvers, addressing complex hydroclimatic prediction challenges in data-limited environments.

How to cite: Ettalbi, M., Garambois, P.-A., Huynh, N.-N.-T., Ferreira, E., and Baghdadi, N.: GPU-Enabled Cell-to-Cell Routing in a High Resolution Hybrid Distributed Hydrological Model with Multi-Source Remote Sensing Data Assimilation: A Continental-Scale Computational Approach , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9176, https://doi.org/10.5194/egusphere-egu25-9176, 2025.

EGU25-10256 | ECS | Orals | HS3.3

Uncovering the Dynamic Drivers of Floods through Interpretable Deep Learning 

Yuanhao Xu and Kairong Lin

The formation of floods, as a complex physical process, exhibits dynamic changes in its driving factors over time and space under climate change. Due to the black-box nature of deep learning, its use alone does not enhance understanding of hydrological processes. The challenge lies in employing deep learning to uncover new knowledge on flood formation mechanism. This study proposes an interpretable framework for deep learning flood modeling that employs interpretability techniques to elucidate the inner workings of a peak-sensitive Informer, revealing the dynamic response of floods to driving factors in 482 watersheds across the United States. Accurate simulation is a prerequisite for interpretability techniques to provide reliable information. The study reveals that comparing the Informer with Transformer and LSTM, the former showed superior performance in peak flood simulation (NSE over 0.6 in 70% of watersheds). By interpreting Informer’s decision-making process, three primary flood-inducing patterns were identified: precipitation, excess soil water, and snowmelt. The controlling effect of dominant factors is regional, and their impact on floods in time steps shows significant differences, challenging the traditional understanding that variables closer to the timing of flood event occurrence have a greater impact. Over 40% of watersheds exhibited shifts in dominant driving factors between 1981-2020, with precipitation-dominated watersheds undergoing more significant changes, corroborating climate change responses. Additionally, the study unveils the interplay and dynamic shifts among variables. These findings suggest that interpretable deep learning, through reverse deduction, transforms data-driven models from merely fitting nonlinear relationships to effective tools for enhancing understanding of hydrological characteristics.

How to cite: Xu, Y. and Lin, K.: Uncovering the Dynamic Drivers of Floods through Interpretable Deep Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10256, https://doi.org/10.5194/egusphere-egu25-10256, 2025.

Earth system data, measured by satellites and terrestrial stations and simulated by increasingly complex models, provide valuable information for identifying functional relationships within the Earth system. These relationships are essential for understanding complex interactions and predicting changes, for example, in climatic or ecological processes, but often only occur in certain spatiotemporal sections or within certain threshold values. With the increasing spatiotemporal resolution of remote sensing products and models, a manual analysis is impractical, and hypothesis-driven approaches can lead to undiscovered hidden relationships. Previous work proposed the SONAR (automated diScovery Of fuNctionAl Relationships) decision-tree algorithm to automatically search for functional relationships in earth system data without a-priori assumptions. We analyzed the proposed algorithm using artificially generated data to evaluate SONAR's functionality.  We tested if the choice of statistical indicator (Pearson’s r, Spearman’s ρ, Kendall’s τ, and Mutual Information) influences the functionality of the SONAR algorithm and which factors are important for the identification of functional relationships. Using 1512 synthetic data sets and the developed SAMPI (Similarity of A Manifested and Prototypical decision tree Indicator) coefficient, we demonstrate how the performance of the algorithm changes under different variations of the data sets - including the number of designated splits, the presence of interfering variables and the strength and nature of the underlying functional relationships. In particular, we show which statistical indicator provides the best results under these conditions. The results demonstrate that the SONAR algorithm is very versatile, especially when employing the most reliable statistical indicator. The SONAR algorithm could, therefore, have far-reaching applications, for example, in analyzing climatic patterns or investigating dependencies between environmental factors.

How to cite: Thöne, C., Bäthge, A., and Reinecke, R.: The effects of different statistical indicators in the new decision-tree-based SONAR algorithm for automated detection of functional relationships in Big Earth Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10444, https://doi.org/10.5194/egusphere-egu25-10444, 2025.

EGU25-11693 | ECS | Posters on site | HS3.3

Universal differential equations for estimating terrestrial evaporation 

Olivier Bonte, Diego G. Miralles, Akash Koppa, and Niko E. C. Verhoest

Terrestrial evaporation (E) is an essential climate variable, linking water, energy and carbon cycles. As E is influenced by the state of the atmospheric boundary layer, vegetation and soil, its modelling is a complex task, resulting in a myriad of simulation approaches. To combine the strong predictive skills of data-driven models with the interpretability and physical consistency of process-based models (PBMs), a new research field of differentiable modelling has emerged1

Here, we present a differentiable framework for E estimation, facilitating online training of NNs as intermediate PBM components. It is inspired by the GLEAM framework for estimating E, which applies offline training (i.e., outside the PBM) of neural networks (NNs) predicting evaporative stress2,3. Building upon the Julia SciML ecosystem’s implementation of universal differential equations4, a wide array of numerical methods are available for solving the PBM’s ordinary differential equations (ODEs) and calculating the parameter sensitivities5. In this way, the effect of the numerical methods on the obtained hybrid model can be investigated, moving beyond the direct automatic differentiation through explicit Euler solutions of ODEs as often applied in other hydrological hybrid modelling approaches. 

 

References

1Shen, C., Appling, A.P., Gentine, P. et al., Differentiable modelling to unify machine learning and physical models for geosciences, Nat. Rev. Earth. Environ., 4, 552–567, 2023, https://doi.org/10.1038/s43017-023-00450-9

2Koppa, A., Rains, D., Hulsman, P. et al., A deep learning-based hybrid model of global terrestrial evaporation, Nat. Commun., 13, 1912, 2022, https://doi.org/10.1038/s41467-022-29543-7

3Miralles, D. G., Bonte, O., Koppa, A. et al., GLEAM4: global land evaporation dataset at 0.1° resolution from 1980 to near present, preprint, 2024, https://doi.org/10.21203/rs.3.rs-5488631/v1 

4Rackauckas, C., Ma, Y.,  Martensen, J. et al., Universal differential equations for scientific machine learning, ArXiv, 2020, https://doi.org/10.48550/arXiv.2001.04385 

5Sapienza, F., Bolibar, J., Schäfer, F. et al., Differentiable Programming for Differential Equations: A Review, ArXiv, 2024, https://doi.org/10.48550/arXiv.2406.09699 

How to cite: Bonte, O., Miralles, D. G., Koppa, A., and Verhoest, N. E. C.: Universal differential equations for estimating terrestrial evaporation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11693, https://doi.org/10.5194/egusphere-egu25-11693, 2025.

EGU25-12068 | ECS | Posters on site | HS3.3

A Physics-Constrained Emulator for High-Resolution Soil Moisture  

Annie Y.-Y. Chang, Elena Leonarduzzi, Christian M. Grams, and Vincent W. Humphrey

Like much of Europe, Switzerland is increasingly experiencing severe summer droughts and heatwaves, prompting the mandate for an advanced national drought monitoring and early warning system. A key component of this initiative is the generation of gridded soil moisture estimates that are spatially distributed, extending beyond measurement stations.  Here, we present the concept of a novel physics-constrained land surface model emulator designed to produce high-resolution (e.g. finer than 250m), gridded soil moisture estimates up to 2m depth across Switzerland's diverse topography and climatic conditions. 

This framework aims to integrate multi-source datasets, including in-situ measurements, and reanalysis products, to train a machine learning based (e.g. Convolutional LSTM, or XGBoost) hybrid emulator that ensures physically consistent outputs. Compared to conventional dynamical land surface models, an emulator has the advantage of being more computationally efficient and less constrained by the specific requirements of a given numerical model (in terms of input variables and technical dependencies). To fulfil the needs of a very diverse user community, ranging from numerical weather prediction to agricultural decision-making, the emulator should be optimized for multi-scale applications, from climatological analysis, to near-real-time monitoring, and to medium-term forecasting.

How to cite: Chang, A. Y.-Y., Leonarduzzi, E., Grams, C. M., and Humphrey, V. W.: A Physics-Constrained Emulator for High-Resolution Soil Moisture , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12068, https://doi.org/10.5194/egusphere-egu25-12068, 2025.

EGU25-12928 | ECS | Orals | HS3.3 | Highlight

Global Vegetation Stress Drivers based on Hybrid Modelling and Explainable AI 

Fangzheng Ruan, Oscar M. Baez-Villanueva, Olivier Bonte, Akash Koppa, Wantong Li, Gustau Camps Valls, Yuting Yang, and Diego G. Miralles

Terrestrial evaporation (E) is a critical component of the water cycle, returning nearly 60% of continental precipitation to the atmosphere and dissipating approximately 50% of surface net radiation. A prevalent approach for estimating E involves computing a theoretical maximum, known as potential evaporation (Ep), and scaling it based on a multiplicative stress factor, often referred to as “evaporative stress” (S) or “transpiration stress” (St) when specifically applied to plant transpiration. Like stomatal or surface conductance, St is governed by a complex nonlinear interplay of environmental drivers such as soil moisture, air temperature, radiation, and atmospheric vapor pressure deficit. This complexity is not yet fully understood, which further hampers its accurate physical modelling and limits our ability to comprehend transpiration’s sensitivity to the changing environment.

The fourth generation of the Global Land Evaporation Amsterdam Model (GLEAM4) has yielded a global dataset of transpiration by integrating multi-source remote sensing data following a hybrid approach, in which Ep is computed based on a process-based model and St is calculated by employing deep neural networks. These neural networks are trained on global eddy covariance and sap flow measurements for both tall and short vegetation, and are informed by a set of environmental controls or biotic factors. These factors include soil moisture, vapor pressure deficit, atmospheric CO2 concentration, wind velocity, air temperature, downwelling shortwave radiation, LAI, and vegetation optical depth. Beyond the predictive capabilities of these deep neural networks, the relationships between environmental controls and St within these neural networks remain under exploration, leaving uncertainty as to whether GLEAM4 accurately represents real-world processes. To explore the relationships, we employ the SHapley Additive exPlanation (SHAP) method, which quantifies the marginal contributions of predictors to model predictions, offering insights into the relative importance of environmental drivers in determining St.

Our findings highlight dominant St drivers across various climatic regimes and ecosystems, revealing their contributions' temporal evolution. Additionally, we investigate how St responds to shifts in environmental conditions, including climate and vegetation changes, water stress, atmospheric aridity, and rising CO2 levels. Our study enhances global understanding of transpiration dynamics and provides critical insights into the impacts of diverse hydroclimatic drivers, thereby supporting broader applications within the hydrology and climate communities.

How to cite: Ruan, F., M. Baez-Villanueva, O., Bonte, O., Koppa, A., Li, W., Camps Valls, G., Yang, Y., and G. Miralles, D.: Global Vegetation Stress Drivers based on Hybrid Modelling and Explainable AI, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12928, https://doi.org/10.5194/egusphere-egu25-12928, 2025.

EGU25-13311 | ECS | Posters on site | HS3.3

Hybrid hydrological modelling of the biophysical impacts of earth’s greening on streamflow 

Georgios Blougouras, Alexander Brenning, Mirco Migliavacca, and Markus Reichstein

Vegetation plays an important but complicated role in modulating land-atmosphere interactions and the water cycle. Under global change, increasing vegetation greenness trends have been observed, which further complicate the control of vegetation in the earth system. Despite growing interest in the role of vegetation in the hydrological processes, large uncertainties still exist, particularly when it comes to the underexplored response of streamflow to vegetation greening. In this study, we explore the watershed-relevant biophysical controls of vegetation greening on streamflow. In order to do so, we develop a hybrid ecohydrological model. This model adheres to the water balance principles, while it simultaneously has a flexible structure that enables integrating physical insights from observational data. The multi-task learning optimization ensures physical consistency across a range of processes and temporal frequencies, which allows us to investigate the cascading impacts of vegetation changes across the water cycle, leading up to the streamflow as an end-process. Ecohydrological insights are directly derived from observational data, while physically meaningful model parameters reflect how ecosystem functions and hydrological processes respond to vegetation changes. We find that the marked change in streamflow can be attributed to vegetation change controls on diverse biophysical processes. Our research highlights the potential of hybrid models to capture complex earth system processes by exploiting multiple observational data streams, machine learning and physical constraints.

How to cite: Blougouras, G., Brenning, A., Migliavacca, M., and Reichstein, M.: Hybrid hydrological modelling of the biophysical impacts of earth’s greening on streamflow, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13311, https://doi.org/10.5194/egusphere-egu25-13311, 2025.

Cyanobacterial blooms have become more frequent and intense in Lake Superior since 2012, primarily due to increased nutrient loads, with phosphorus being the main limiting factor. To protect water quality, extensive monitoring of lakes and streams is crucial, but it is not cost-effective or practical to measure nutrients frequently across all ecosystems. This study presents a cost-effective, transferable solution using machine learning (ML) models to predict phosphorus concentrations and loads based on conventional water quality parameters like streamflow, dissolved oxygen, conductivity, turbidity, transparency, and total suspended solids. The research introduces an explainable hybrid ML framework combining probabilistic principal component analysis (P2CA) with several ML models, including Bagging Ensemble Learning, Boosting Ensemble Learning, Gaussian Process Regression, and Support Vector Regression, to enhance prediction accuracy. Results demonstrate that the P2CA-Boosting Ensemble Learning model consistently outperforms other approaches. To confirm its effectiveness, the developed model was tested with the same input data from a different river catchment, proving it works well in different environments. This study highlights the potential of combining P2CA with Boosting Ensemble Learning as a powerful tool for water quality management in streams and rivers.

How to cite: Kumar, A. and Zhang, K.: Development of a hybrid machine learning model to predict total phosphorus in streams over the north shore of Lake Superior, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13799, https://doi.org/10.5194/egusphere-egu25-13799, 2025.

EGU25-15436 | ECS | Orals | HS3.3

The Differentiable Distributed Regression Model (dDRM) Balancing Explainability and Predictive Performance 

Bjarte Beil-Myhre, Rajeev Shrestha, and Bernt Viggo Matheussen

The field of hydrology has undergone significant transformation over the past decade, driven by advancements in machine learning and data-driven techniques. A key breakthrough came from the work of Kratzert et al. (2018), who demonstrated that purely data-driven LSTM models could outperform traditional hydrological models in over 600 catchments across North America. However, while these models significantly improve predictive performance, they often sacrifice interpretability and explainability.

To address this trade-off, researchers have explored new approaches that merge physical principles with data-driven methods. One promising innovation is the concept of differentiable modeling, introduced by Chen et al. in 2022. This approach transforms physical models into differentiable functions, allowing neural networks to represent and learn model parameters. By doing so, differentiable modeling enhances flexibility while maintaining a foundation in physical principles.

This research presents a novel differentiable hydrological model called the Differentiable Distributed Regression Model (dDRM). The dDRM builds on the principles of differentiable modeling with the structure of a conceptually lumped model using a simplified representation of physics ("smooth" HBV model). Inspired by the simplicity of the LSTM model, which aggregates data at the catchment level rather than relying on a grid-based representation, we introduce four equally sized elevation zones instead of grid cells in the dDRM. These zones inherently reflect differences in hydrological processes, such as precipitation, temperature, and snowmelt dynamics, enabling the model to account for spatial heterogeneity while maintaining computational efficiency.

By leveraging the principles of differentiable modeling, the dDRM achieves a balance between explainability and predictive performance. To evaluate model performance, we tested the dDRM across sixty-three catchments in southern Norway, in a gauged setting. Only precipitation and temperature were used as input data. For benchmarking purposes, we also trained an LSTM model to the same catchments. 

Our results demonstrate that the dDRM outperforms the fine-tuned LSTM model in both daily predictions and cumulative runoff volumes. These findings underscore the potential of differentiable hydrological models to bridge the gap between performance and interpretability. By combining physical principles with data-driven techniques, the dDRM provides a pathway toward more effective and understandable forecasting tools in hydrology.

How to cite: Beil-Myhre, B., Shrestha, R., and Matheussen, B. V.: The Differentiable Distributed Regression Model (dDRM) Balancing Explainability and Predictive Performance, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15436, https://doi.org/10.5194/egusphere-egu25-15436, 2025.

Groundwater is a critical resource for drinking water supply, agriculture, and ecosystems in general. In regions facing water scarcity, such as Brandenburg (Germany), effective groundwater management is essential. This requires accurate assessments of groundwater dynamics, which data-driven models can deliver through efficient and reliable groundwater level (GWL) predictions. To effectively develop and apply data-driven models for groundwater level prediction, a deeper understanding of which and how the input features influence the groundwater level prediction is crucial.

Our primary objective is to assess the impact of the input features of a Deep learning (DL) model that predicts GWLs using feature attribution methods. Specifically, the influence of climatic features as well as different land use patterns is examined. This study employs a global DL model based on the Long Short-Term Memory (LSTM) architecture to predict seasonal GWLs for 16 weeks ahead. We utilize a comprehensive set of features, including dynamic features such as climatic variables (e.g., temperature, precipitation, relative humidity) and static features such as Corine land cover. By incorporating these, we aim to capture the complex interactions between climate, landuse and groundwater levels.

For the feature attribution itself, we apply the Shapley value sampling method. It analyses the effect of an alternation of an input feature to the respective chosen objective. The choice of that function is essential for the obtained results. We alternate the corresponding objective function in three distinct ways: first, by using the total change of the predicted GWL for the whole period of interest; second, per prediction horizon, i.e. per predicted week of the 16 week prediction; and third, through a decomposition into partial scale-respective signals of the period of interest using the discrete wavelet transform. Besides understanding which input features are most important for the predictive performance of the LSTM model, the results enable us to identify further aspects of the dynamics learned by the model. For example, if and when the model switches from extrapolation to prediction, and at which temporal scales different factors play a role; e.g. if forest vegetation is more important for seasonal or weekly effects on groundwater levels. This multi-faceted approach allows us to gain a deeper understanding of the factors influencing GWLs and their temporal dynamics, both for static and dynamic input features. Ultimately, feature attribution methods can enhance the awareness for reasonable land-use, hence, groundwater management and lead to better predictive models.

How to cite: Engel, M., Kunz, S., Wetzel, M., and Körner, M.: Multitemporal and Multiscale Feature Attribution Methods to Understand the Impact of Climatic and Land Use Features on the Prediction of Groundwater Levels, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15651, https://doi.org/10.5194/egusphere-egu25-15651, 2025.

The Integrated Multi-satellite Retrievals for GPM (IMERG) is a global satellite-based precipitation dataset that provides near real-time precipitation estimates by combining multiple satellite measurements. IMERG integrates microwave (MW) observations from low-orbit satellites with precipitation estimates inferred from the brightness temperature of geostationary infrared (IR) imagery. MW measurements provide accurate precipitation estimates due to their direct interaction with precipitation particles, while IR measurements offer broader spatial and temporal coverage by inferring precipitation from cloud-top brightness temperatures. Together, these complementary techniques balance precision and coverage to improve global precipitation monitoring. However, IR-based precipitation estimates are inherently less reliable due to the weak direct correlation between brightness temperature and precipitation. Conversely, MW-derived estimates are more accurate but spatially constrained by the limited footprint of low-orbit satellites. To investigate the contributing factors in IR precipitation error calibration, we leveraged ERA5 Land, a high-resolution reanalysis dataset that includes surface variables across nine domains, such as temperature, soil moisture, radiation, and vegetation indices. These variables offer a comprehensive lens for understanding the impact of the land surface on precipitation dynamics. We employed the XGBoost machine learning model to predict the errors in IR precipitation estimates relative to MW-derived benchmarks. Additionally, SHapley Additive exPlanations (SHAP) values were used to interpret the model’s predictions, uncovering how individual input features contribute to error correction.


Our findings indicate that the explainable machine learning model can correct the infrared (IR) precipitation estimates to resemble microwave (MW) products, achieving notable improvements across statistical metrics. In the preliminary analysis of 165 countries and territories, the XGBoost model’s calibration improved the RMSE in all validation datasets, with a median reduction of 19.89% and an average reduction of 22.5%. Similarly, the correlation coefficient improved, with a median increase of 18.43% and an average increase of 54.49%. Moreover, the spatial and temporal distributions of the variables' SHAP values show various patterns. The clustered spatial distribution may represent the local climate attributes in specific geographic regions, providing insights into how regional environmental factors influence precipitation estimates. Meanwhile, the temporal distribution may imply seasonal variation, which can help identify patterns in precipitation dynamics and refine IR-based calibration by accounting for temporal variability in precipitation processes. This study provides a robust framework for leveraging land surface variables to refine IR-based precipitation products. By integrating reanalysis data with machine learning models, we present a scalable solution for improving precipitation monitoring in data-sparse regions, particularly where MW observations are unavailable.

How to cite: Hung, H. T. and Wang, L.-P.: IRMerg: Enhancing Global Infrared Precipitation Estimates with Land Surface Variables and Contributing Factors Analysis Using Explainable Machine Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16863, https://doi.org/10.5194/egusphere-egu25-16863, 2025.

EGU25-16971 | ECS | Orals | HS3.3

The Extrapolation Dilemma in Hydrology: Unveiling the extrapolation properties of data-driven models 

Sanika Baste, Daniel Klotz, Eduardo Espinoza, Andras Bardossy, and Ralf Loritz

Long Short-Term Memory (LSTM) networks have shown strong performance in rainfall–runoff modelling, often surpassing conventional hydrological models in benchmark studies. However, recent studies raise questions about their ability to extrapolate, particularly under extreme conditions that exceed the range of their training data. This study examines the performance of a stand-alone LSTM trained on 196 catchments in Switzerland when subjected to synthetic design precipitation events of increasing intensity and varying duration. The model’s response is compared to that of a hybrid model and evaluated against hydrological process understanding. Our study reiterates that the stand-alone LSTM is characterised by a theoretical prediction limit, and we show that this limit is below the range of the data the model was trained on. We show that saturation of the LSTM cell states alone does not fully account for this characteristic behaviour, as the LSTM does not reach full saturation, particularly for the 1-day events. Instead, its gating mechanisms prevent new information about the current extreme precipitation from being incorporated into the cell states. Adjusting the LSTM architecture, for instance, by increasing the number of hidden states, and/or using a larger, more diverse training dataset can help mitigate the problem. However, these adjustments do not guarantee improved extrapolation performance, and the LSTM continues to predict values below the range of the training data or show hydrologically unfeasible runoff responses during the 1-day design experiments. Despite these shortcomings, our findings highlight the inherent potential of stand-alone LSTMs to capture complex hydro-meteorological relationships. We argue that more robust training strategies and model configurations could address the observed limitations, ensuring the promise of stand-alone LSTMs for rainfall–runoff modelling.

How to cite: Baste, S., Klotz, D., Espinoza, E., Bardossy, A., and Loritz, R.: The Extrapolation Dilemma in Hydrology: Unveiling the extrapolation properties of data-driven models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16971, https://doi.org/10.5194/egusphere-egu25-16971, 2025.

EGU25-18102 | ECS | Posters on site | HS3.3

In the application of physically-based and interpretable AI-based models for streamflow simulation 

Sara Asadi, Patricia Jimeno-Sáez, Adrián López-Ballesteros, and Javier Senent-Aparicio

Precise streamflow forecasting in river systems is crucial for water resources management and flood risk assessment. This study focuses on the Tagus Headwaters River Basin (THRB) in Spain, a key hydrological basin providing essential water for urban, industrial, and irrigation purposes. Additionally, a significant portion of its water resources is transferred to the Segura River Basin through the Tagus-Segura water transfer, Spain’s most extensive hydraulic infrastructure. Given that nearly all available water in the THRB is allocated for these demands, precise streamflow forecasting is vital. For streamflow estimation in this basin, we evaluated the Soil and Water Assessment Tool (SWAT+), a physically-based model, and three AI-based models: support vector regression (SVR), feed-forward neural network (FFNN), and long short-term memory (LSTM) models, across four gauging stations within the THRB. For the AI-based models, rainfall and time-lagged runoff data were used as input data. Additionally, an ensemble machine learning technique was evaluated, using the outputs of both physically-based and AI-based individual models as inputs for the ensemble model. The results show that the AI-based models and the ensemble machine learning technique significantly outperformed the SWAT+ model. While the precision of the AI-based models was considerably higher than that of the SWAT+ model, the application of the ensemble technique enhanced the precision of the AI-based models by 18 to 26% during the calibration period and 4.1 to 9.2% during the validation period. Furthermore, the Shapley Additive Explanations (SHAP) methodology was used to explore how each model contributes to the predictions in the ensemble technique. This work was supported by the Spanish Ministry of Science and Innovation, under grants PID2021-128126OA-I00.

How to cite: Asadi, S., Jimeno-Sáez, P., López-Ballesteros, A., and Senent-Aparicio, J.: In the application of physically-based and interpretable AI-based models for streamflow simulation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18102, https://doi.org/10.5194/egusphere-egu25-18102, 2025.

EGU25-18468 | ECS | Posters on site | HS3.3

An automated machine learning framework for multi-depth soil moisture prediction using hydro-meteorological datasets 

Vidhi Singh, Abhilash Singh, and Kumar Gaurav

 Soil moisture, one of the essential climate variables, forms a fundamental bridge between hydro-meteorological processes and influence climate dynamics. It is extremely variable and is driven by numerous hydrological, agricultural and ecological factors. Soil moisture subsequently impacts soil forming processes, root zone water availability, infiltration rates, runoff, groundwater storage and vegetation-soil interaction. Despite its significant contribution in hydro-ecological interaction, its variability at subsurface is not yet explored adequately. Precise estimation of soil moisture at various depths is crucial because it affects water retention characteristics and modulates the vertical and lateral movement of water within the soil profile. This subsurface information is integral to understanding recharge rates, groundwater interactions, and the overall water balance within a catchment. In this study, we present an automated machine learning framework designed to predict soil moisture at multiple depths of 10 cm, 20 cm, 30 cm, and 40 cm leveraging Bayesian optimization. We collected data from our hydrological observatory set up constituting an automatic weather station, a pan evaporimeter and a soil moisture recorder. To evaluate model performance, we categorized the dataset into four scenarios (S1, S2, S3, and S4), with each subsequent scenario incorporating a greater number of observations and rainfall events. We used 11 input features to train this AutoML model by integrating several hydrological and meteorological variables with in-situ soil moisture data. Among the predictor variables, humidity, dew point, and rainfall emerged as the most influential factors driving soil moisture variability. The model was trained to calculate the performance metrices for the entire dataset and for subsets containing only rainfall instances. Our optimized model demonstrated superior performance, with an R² of 0.88–0.99 and RMSE < 0.022 for the overall dataset, and R² of 0.76–1.00 with RMSE < 0.06 for rainfall-specific data across all soil moisture depths.

How to cite: Singh, V., Singh, A., and Gaurav, K.: An automated machine learning framework for multi-depth soil moisture prediction using hydro-meteorological datasets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18468, https://doi.org/10.5194/egusphere-egu25-18468, 2025.

EGU25-19057 | ECS | Posters on site | HS3.3

Inference of catchment areas from modeled discharge dynamics 

Fedor Scholz, Christiane Zarfl, Thomas Scholten, and Martin V. Butz

The delineation of catchment areas from elevation is a fundamental step in lumped process-based models (PBMs). Most machine learning (ML) approaches for rainfall-runoff modeling spatially aggregate inputs to represent basin-wide processes. Elevation-based lumping, however, disregards both human interventions such as drainages and underground hydrological flows, which can lead to significant model inaccuracies. In this work, we employ DRRAiNN (Distributed Rainfall-Runoff Artificial Neural Network) – a fully distributed neural network architecture – to infer catchment areas directly from observed precipitation and discharge dynamics without prior delineations.

As a first evaluation of the potential to infer actual catchment areas with DRRAiNN, we trained the model on relatively sparse data from 2006 until 2015: Radolan-based hourly precipitation data as input with a spatial resolution of 4x4 km and only daily discharge measurements from 17 stations in the Neckar river basin as target output. Elevation and solar radition were given as additional parameterization input. As DRRAiNN is fully differentiable, we were then able to infer station-specific attribution maps via backpropagation through space and time. To evaluate the alignment between the inferred attribution maps and elevation-based catchment areas, we compute the Wasserstein distance between attributions inside and outside the catchment boundaries. A higher distance indicates better agreement. The results show that DRRAiNN learns to propagate water in a physically plausible manner. Further, we reveal deviations that indicate additional water flows that are undetectable from elevation data alone. Our findings thus suggest that DRRAiNN captures key rainfall-runoff dynamics while avoiding the limitations of lumped models.

The quantitative evaluations alongside qualitative comparisons underscore the model’s potential for uncovering hidden hydrological processes. We show that catchment area estimates can be inferred from relatively little discharge data, which may, in the future, potentially be substituted by satellite data. As a result, DRRAiNN may be applicable in ungauged catchments. Given actual discharge measurements or discharge estimations, DRRAiNN can be used to analyze the hydrological dynamics of surface and subsurface runoff as well as baseflow esimations and has the potential to uncover unexpected and unknown runoff dynamics that would not be detectable otherwise.

How to cite: Scholz, F., Zarfl, C., Scholten, T., and Butz, M. V.: Inference of catchment areas from modeled discharge dynamics, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19057, https://doi.org/10.5194/egusphere-egu25-19057, 2025.

EGU25-20878 | ECS | Orals | HS3.3

High-Resolution Differentiable Models for Operational National and Global Water Modeling and Assessment 

Yalan Song, Chaopeng Shen, Haoyu Ji, and Farshid Rahmani

Continental and global water models have long been trapped in slow growth and inadequate predictive power, as they are not able to effectively assimilate information from big data. While Artificial Intelligence (AI) models greatly improve performance, purely data-driven approaches do not provide strong enough interpretability and generalization. One promising avenue is “differentiable” modeling that seamlessly connects neural networks with physical modules and trains them together to deliver real-world benefits in operational systems. Differentiable modeling (DM) can efficiently learn from big data to reach state-of-the-art accuracy while preserving interpretability and physical constraints, promising superior generalization ability, predictions of untrained intermediate variables, and the potential for knowledge discovery. Here we demonstrate the practical relevance of a high-resolution, multiscale water model for operational continental-scale and global-scale water resources assessment. (https://bit.ly/3NnqDNB). Not only does it achieve significant improvements in streamflow simulation compared to the established national- and global water models, but it also produces much more reliable depictions of interannual changes in large river streamflow, freshwater inputs to estuaries, and groundwater recharge. 

How to cite: Song, Y., Shen, C., Ji, H., and Rahmani, F.: High-Resolution Differentiable Models for Operational National and Global Water Modeling and Assessment, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20878, https://doi.org/10.5194/egusphere-egu25-20878, 2025.

EGU25-935 | ECS | Orals | HS3.4

Characterizing possible failure modes: Insights from LSTM-Based Streamflow Predictions 

Sarth Dubey, Pravin Bhasme, and Udit Bhatia

Long Short-Term Memory (LSTM) networks have become popular for streamflow prediction in hydrological systems due to their ability to model sequential data. However, their reliance on lumped catchment representation and climate summaries often limits their capacity to capture spatial heterogeneity in rainfall patterns and evolving rainfall trends, both of which are critical for hydrological consistency. This study explores the limitations of LSTM-based streamflow predictions by employing a distributed conceptual hydrological model, SIMHYD, coupled with Muskingum-Cunge routing, to generate synthetic datasets representing diverse hydroclimatic conditions. These datasets are designed to replicate rainfall-runoff dynamics across selected catchments from all 18 ecoregions in CAMELS-US and key Indian river basins, providing a robust testbed for evaluating model performance under controlled conditions. The pre-trained LSTM model is tested against synthetic discharge data, enabling direct comparisons to assess its ability to simulate realistic hydrological responses. Performance is evaluated using multiple metrics, including Nash-Sutcliffe Efficiency (NSE), Kling-Gupta Efficiency (KGE), Percent Bias (PBIAS), and mean peak timing errors, to identify systematic deviations. Results reveal that LSTM models struggle with spatially variable and temporally shifting rainfall patterns, leading to inaccuracies in peak flow timing, magnitude, and overall discharge volumes. These issues highlight vulnerabilities in current LSTM-based flood forecasting systems, particularly in their ability to generalize across diverse climatic conditions and regions. This study also characterizes specific failure pathways, such as underestimation of extreme events and poor temporal coherence in hydrographs, which are critical for operational forecasting. By diagnosing these limitations, the study provides a framework for integrating process-based hydrological knowledge with data-driven techniques to improve model robustness. The findings underscore the importance of using synthetic datasets and diverse diagnostic tools to evaluate and enhance the reliability of LSTM-based models, paving the way for hybrid approaches capable of addressing the complexities of real-world hydrological systems.

How to cite: Dubey, S., Bhasme, P., and Bhatia, U.: Characterizing possible failure modes: Insights from LSTM-Based Streamflow Predictions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-935, https://doi.org/10.5194/egusphere-egu25-935, 2025.

EGU25-1644 | ECS | Posters on site | HS3.4

Graph-enhanced Neural Operator for Missing Velocities Infilling in River Surface Velocimetry 

Xueqin Chen, Hessel Winsemius, and Riccardo Taormina

Measuring river surface velocity enables river discharge estimation, a fundamental task for hydrologists, environmental scientists, and water resource managers. While traditional image-based velocimetry methods are often effective, they struggle to produce complete velocity fields under complex environmental conditions. Poor lighting, reflective glare, lack of visible surface features, or excessive turbulence can all result in regions where feature tracking fails, leading to gaps in the resolved velocity field. Addressing these gaps through the reconstruction of missing velocity measurements is an important research challenge. Recently, researchers have employed deep learning to address various hydrology problems, demonstrating promising improvements. In this work, we propose a neural operator-based model to address the challenge of missing velocities in river surface velocimetry. Specifically, our model is based on the Fourier neural operator with a graph-enhanced lifting layer. It is trained on the river surface velocimetry reconstruction task using a self-supervised paradigm. Once trained, it can be used to infer missing velocities in unseen samples. Experiments conducted on a dataset collected from a river in the Netherlands demonstrate our approach’s ability to accurately infill missing surface velocities, even when faced with large amounts of missing data. We attribute this robustness to the neural operator’s ability to learn continuous functions, which enhances our model’s capacity for high-level feature representation and extraction. Our findings suggest that the reconstructed velocity fields produced by our model can act as reliable ground truth data for deep learning-based methods. In the future, we aim to improve our model’s performance and generalization by incorporating additional data collected from a wider range of rivers and under varying environmental conditions.

How to cite: Chen, X., Winsemius, H., and Taormina, R.: Graph-enhanced Neural Operator for Missing Velocities Infilling in River Surface Velocimetry, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1644, https://doi.org/10.5194/egusphere-egu25-1644, 2025.

EGU25-2616 | ECS | Posters on site | HS3.4

Dual-Source Adaptive-Fusion Transfer Learning for Hydrological Forecasting in Data-scarce Catchments 

Yuxuan Gao and Dongfang Liang

Historical records observed at hydrological stations are scarce in many regions, bringing significant challenges to the hydrological predictions for these regions. Transfer learning (TL), increasingly applied in hydrology, leverages knowledge from data-rich catchments (sources) to enhance predictions in data-scarce catchments (targets), providing new insight into data-scarce region predictions. Most existing TL approaches pre-train models using large meteoro-hydrological datasets to improve overall generalizability to target catchments. However, the predictive performance for specific catchments would be constrained due to irrelevant source data inputs and the lack of effective source fusion strategies. To address these challenges, this study proposes the Dual-Source Adaptive Fusion TL Network (DSAF-Net), which utilizes a pre-trained dual-branch feature extraction module (DBFE) to extract knowledge from two carefully selected source catchments, minimizing noise and redundancy associated with larger datasets. A cross-attention fusion module is then incorporated to dynamically identify key knowledge of the target catchment and adaptively fuse complementary information. This fusion module is embedded after each layer in the DBFE to enhance multi-level feature integration. Results demonstrate that DSAF-Net achieves superior prediction accuracy to single-source TL and large dataset TL strategies. These findings highlight the potential of DSAF-Net to advance hydrological forecasting and support water resource management in data-scarce regions.

How to cite: Gao, Y. and Liang, D.: Dual-Source Adaptive-Fusion Transfer Learning for Hydrological Forecasting in Data-scarce Catchments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2616, https://doi.org/10.5194/egusphere-egu25-2616, 2025.

EGU25-2761 | ECS | Posters on site | HS3.4

Forecasting Reservoir Inflows Using Regionally Trained and Finetuned LSTM Models: A Case Study with CAMELS-DE 

Gregor Johnen, Andre Niemann, Patrick Nistahl, Alexander Dolich, and Alexander Hutwalker

The increasing frequency of extreme hydrological events, such as floods and droughts, poses significant challenges for operators of drinking water reservoirs in maintaining a balance between water supply and demand. While the security of supply typically requires high water levels to meet consumer demands throughout the year, ensuring flood protection, on the contrary, necessitates that reservoir storage is kept partially free to accommodate high inflows. Accurate inflow forecasting is essential for making risk-based operational decisions, including the timely release of water from drinking water reservoirs to mitigate flood risks. While deep learning approaches, particularly Long Short-Term Memory (LSTM) networks, have become prevalent in rainfall-runoff modeling, most existing studies focus on small, homogeneous datasets limited to single hydrological basins. This study leverages the newly published CAMELS-DE dataset to develop a regionally trained and finetuned LSTM model encompassing 1,582 catchments across Germany. We apply this regional model to five small catchments upstream of drinking water reservoirs and compare its performance against basin-specific LSTM models. Our findings demonstrate that the regionally trained LSTM model significantly improves the accuracy of inflow estimates, especially when finetuned to our target catchments. This is highlighting its potential for enhancing reservoir management strategies in the face of climate change.

How to cite: Johnen, G., Niemann, A., Nistahl, P., Dolich, A., and Hutwalker, A.: Forecasting Reservoir Inflows Using Regionally Trained and Finetuned LSTM Models: A Case Study with CAMELS-DE, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2761, https://doi.org/10.5194/egusphere-egu25-2761, 2025.

EGU25-4247 | ECS | Posters on site | HS3.4

From multiple meteorological forecasts to river runoff: Learning and adjusting real-time biases to enhance predictions 

Oliver Konold, Moritz Feigl, Christoph Klingler, and Karsten Schulz

Deep learning models such as the Long Short Term Memory Network (LSTM) are capable of representing rainfall-runoff relationships and outperform classical hydrological models in gauged and ungauged settings (Kratzert et al., 2018). Previous studies have shown that combining multiple precipitation data in a single LSTM significantly improves the accuracy of simulated runoff, as the neural network learns to combine temporal and spatial patterns of inputs (Kratzert et al., 2021). However, every operational runoff forecasting setting requires meteorological forecasts. Nearing et al. (2024) have developed a global runoff forecast model based on an LSTM, with an ECMWF forecasting product as additional input over the forecast horizon. Compared with observed or reanalysis meteorological input data, forecasting products generally have a lower accuracy, with different reliabilities between various forecasting products. This is where the synergies of several meteorological forecasts combined with historical observational and reanalysis data can be used in a single deep learning model.

This study investigates how well LSTMs can predict runoff when trained on (1) multiple archived meteorological forecasts and (2) a combination of multiple archived meteorological forecasts and reanalysis data. All meteorological input data are aggregated to the catchments of the LamaH-CE dataset (Klingler, Schulz and Herrnegger, 2021). Runoff predictions are evaluated for a 24 hours forecasting horizon.  Preliminary analyses indicate that the coupling of reanalysis data and forecasting products from different sources improves the accuracy of operational runoff forecasting, suggesting that the model is able to learn and adjust real-time biases in forecasting data.

 

Klingler, C., Schulz, K. and Herrnegger, M. (2021) ‘LamaH-CE: LArge-SaMple DAta for Hydrology and Environmental Sciences for Central Europe’, Earth System Science Data, 13(9), pp. 4529–4565. DOI: 10.5194/essd-13-4529-2021.

Kratzert, F., Klotz, D., Brenner, C., Schulz, K. and Herrnegger, M. (2018) ‘Rainfall–runoff modelling using Long Short-Term Memory (LSTM) networks’, Hydrology and Earth System Sciences, 22(11), pp. 6005–6022. DOI: 10.5194/hess-22-6005-2018.

Kratzert, F., Klotz, D., Hochreiter, S. and Nearing, G.S. (2021) ‘A note on leveraging synergy in multiple meteorological data sets with deep learning for rainfall–runoff modeling’, Hydrology and Earth System Sciences, 25(5), pp. 2685–2703. DOI: 10.5194/hess-25-2685-2021.

Nearing, G., Cohen, D., Dube, V., Gauch, M., Gilon, O., Harrigan, S., Hassidim, A., Klotz, D., Kratzert, F., Metzger, A., Nevo, S., Pappenberger, F., Prudhomme, C., Shalev, G., Shenzis, S., Tekalign, T.Y., Weitzner, D. and Matias, Y. (2024) ‘Global prediction of extreme floods in ungauged watersheds’, Nature, 627(8004), pp. 559–563. DOI: 10.1038/s41586-024-07145-1.

How to cite: Konold, O., Feigl, M., Klingler, C., and Schulz, K.: From multiple meteorological forecasts to river runoff: Learning and adjusting real-time biases to enhance predictions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4247, https://doi.org/10.5194/egusphere-egu25-4247, 2025.

EGU25-4387 | Orals | HS3.4

Runoff Forecasting in Unmeasured Catchments and Rapid Flash Flood Prediction Based on Deep Learning. 

binlan zhang, qingsong xu, and chaojun ouyang

Runoff forecasting is a long-standing challenge in hydrology, particularly in unmeasured catchments and rapid flash flood prediction. For unmeasured catchment forecasting, we introduce the encoder-decoder-based dual-layer long short-term memory (ED-DLSTM) model[1]. This model fuses static spatial granularity attributes with temporal dynamic variables to achieve streamflow forecasting at a global scale. ED-DLSTM reaches an average Nash efficiency coefficient (NSE) of 0.75 across more than 2000 catchments from historical datasets in the United States, Canada, Central Europe, and the United Kingdom. Additionally, ED-DLSTM is applied to 150 fully ungauged catchments in Chile, achieving a high NSE of 0.65. The interpretability of the transfer capacities of ED-DLSTM is effectively tracked through the cell state induced by adding a spatial attribute encoding module, which can spontaneously form hydrological regionalization effects after performing spatial coding for different catchments.

Moreover, rapid flood prediction with daily resolution is challenged to capture changes in runoff over short periods. To address this, we also propose a benchmark evaluation for runoff and flood forecasting based on deep learning (RF-Bench) at an hourly scale. We introduce the Mamba model to hydrology for the first time. The benchmark also includes Dlinear, LSTM, Transformer, and its improved versions (Informer, Autoformer, Patch Transformer). Results indicate that the Patch Transformer exhibits optimal predictive capability across multiple lead times, while the traditional LSTM model demonstrates stable performance, and the Mamba model strikes a good balance between performance and stability. We reveal the attention patterns of Transformer models in hydrological modeling, finding that attention is time-sensitive and that the attention scores for dynamic variables are higher than those for static attributes.

Our work [2,3] provides the hydrological community with an open-source, scalable platform, contributing to the advancement of deep learning in the field of hydrology.

 

[1] Zhang, B., Ouyang, C., Cui, P., Xu, Q., Wang, D., Zhang, F., Li, Z., Fan, L., Lovati, M., Liu, Y., Zhang, Q., 2024. Deep learning for cross-region streamflow and flood forecasting at a global scale. The Innovation 5, 100617. https://doi.org/10.1016/j.xinn.2024.100617

[2] Zhang, B., Ouyang, C., Wang, D., Wang, F., Xu, Q., 2023. A PANN-Based Grid Downscaling Technology and Its Application in Landslide and Flood Modeling. Remote Sensing 15, 5075. https://doi.org/10.3390/rs15205075

[3] Xu, Q., Shi, Y., Bamber, J.L., Ouyang, C., Zhu, X.X., 2024. Large-scale flood modeling and forecasting with FloodCast. Water Research 264, 122162. https://doi.org/10.1016/j.watres.2024.122162

How to cite: zhang, B., xu, Q., and ouyang, C.: Runoff Forecasting in Unmeasured Catchments and Rapid Flash Flood Prediction Based on Deep Learning., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4387, https://doi.org/10.5194/egusphere-egu25-4387, 2025.

EGU25-4500 | ECS | Posters on site | HS3.4

Leveraging Graph Neural Networks for water level prediction 

George Koutsos, Panagiotis Kossieris, Vasiliki Thomopoulou, and Christos Makropoulos

Accurate water level prediction is essential for flood risk management, water resources management, inland water transportation and climate resilience. Traditional statistical methods, such as autoregressive models, and physically-based hydrological simulations, have been widely used in water level forecasting. However, these approaches often struggle to capture complex, dynamic, and nonlinear interactions in a hydrological system, particularly those affected by climate change. In recent years, machine learning models have emerged as a promising alternative, offering improved predictive accuracy and adaptability across varying environmental conditions. A special type of such models is the Graph Neural Network (GNN), which focuses especially on the reproduction of spatial dependencies, and hence it can be employed to capture the spatial dynamic of the hydrologic/hydraulic system, by treating hydrological networks as graph structures (e.g. nodes as gauges). Going one step further, GNN models can be combined with sequence-based machine learning techniques, such as the Long short-term memory (LSTM) neural network, to capture simultaneously the spatial and temporal dynamics of the system. In this work, we develop and assess a series of advanced hybrid-graph structured machine learning models (such as GNN-LSTM) to make hydrometric predictions across a long river channel. The developed models will be assessed on the basis of alternative performance metrics and against a series of traditional benchmark statistical and machine learning models such as ARIMA and LSTM respectively. As a test case, we exploit data from 19 water level gauges in the Red River of the North, which spans 885 km, serving the natural boundary between North Dakota and Minnesota and has experienced several severe historical flood events.

How to cite: Koutsos, G., Kossieris, P., Thomopoulou, V., and Makropoulos, C.: Leveraging Graph Neural Networks for water level prediction, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4500, https://doi.org/10.5194/egusphere-egu25-4500, 2025.

EGU25-4683 | ECS | Posters on site | HS3.4

Classification-Enhanced LSTM Model for Predicting River Water Levels 

Jiaming Luo

River level predicting underpins the management of water resource projects, steers navigational activities in rivers, and protects the lives and properties of riverside communities, etc. Traditionally, hydrological-hydraulic coupled models have been at the forefront of simulating and predicting river levels, achieving notable success. Despite their utility, these models encounter limitations due to the exhaustive demand for various data types—often difficult to obtain—and the ambiguity in determining downstream boundary conditions for the hydraulic model. Responding to these limitations, this study utilizes Long Short-Term Memory (LSTM) model, a deep learning technique, to predict river levels using upstream discharges. Three approaches were used to further enhance the accuracy and reliability of our model. Firstly, we incorporated historical water level data at or downsteam of the predicted station as input, secondly, we classified the datasets based on physical principles, and thirdly, we employed data augmentation techniques. These methods were evaluated within the Jingjiang-Dongting river-lake system in China. It achieves high prediction accuracy of water level and can mitigate the impact of input inaccuracies. The incorporation of water level data as input and the Classification-Enhanced LSTM model that segregates the input data according to rising and recession trends of water level,significantly improve prediction accuracy under extreme water level conditions compared with other deep learning approaches. The proposed model uses easily accessible data to predict water levels, offering enhanced robustness and new strategies for improving prediction accuracy under extreme conditions. It is applicable for predicting water levels at any hydrological station along a river and can enhance the prediction accuracy of hydraulic models by proving a robust downstream boundary condition.

How to cite: Luo, J.: Classification-Enhanced LSTM Model for Predicting River Water Levels, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4683, https://doi.org/10.5194/egusphere-egu25-4683, 2025.

EGU25-6294 | Orals | HS3.4

Improving generalization of soil moisture prediction using self-supervised learning 

Lijun Wang, Liangsheng Shi, and Shijie Jiang

Accurate soil moisture prediction is increasingly important due to its critical role in water resource management and agricultural sustainability under global climate change. While machine learning models have achieved high accuracy in soil moisture prediction, their ability to generalize to different environmental and meteorological conditions remains a significant challenge. Existing models often perform poorly when applied to conditions that differ from their training data, highlighting the need for approaches that improve generalization while effectively capturing underlying soil moisture dynamics.

In this study, we propose a novel soil moisture prediction model that combines self-supervised learning with a Transformer architecture. The performance of the model was compared with the widely used Long Short-Term Memory (LSTM)-based approach to evaluate its ability to generalize. The proposed model outperformed the baseline in tasks such as capturing extreme soil dryness, adapting to unobserved meteorological humidity conditions, and forecasting soil moisture dynamics at untrained depths. Further analysis revealed that the model’s success stems from its capability to learn comprehensive representations of underlying soil moisture processes. These results highlight the potential of advanced deep learning methods to improve prediction and our process understanding of soil hydrology in a changing climate.

How to cite: Wang, L., Shi, L., and Jiang, S.: Improving generalization of soil moisture prediction using self-supervised learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6294, https://doi.org/10.5194/egusphere-egu25-6294, 2025.

EGU25-7020 | ECS | Posters on site | HS3.4

Improving high-flow forecasting using dynamic multimodal feature fusion 

Konstantina Theodosiadou, Thomas Rodding Kjeldsen, and Andrew Barnes

This study evaluates a new approach to improving streamflow forecasting with deep learning: it focuses on the novel application of a dynamic multimodal feature fusion mechanism that adapts fusion operations based on the data's characteristics. Two baseline Long Short-Term Memory (LSTM) architectures are used, applying two dynamic fusion methods: dynamic operation-level fusion and attention-based fusion, to combine heterogeneous and multisource hydrometeorological data. The models are used for univariate (single flow gauge) and multivariate (multi-gauge) streamflow forecasting approaches. Applying these four approaches to the Severn Basin in the UK, known for long medium- to high-flow periods and shorter low-flow intervals, shows that the dynamic operation-level fusion consistently improved over the attention-based fusion in key performance metrics. In the multivariate case, Nash-Sutcliffe Efficiency (NSE) improved by 1.43%, Mean Absolute Error (MAE) decreased by 1.73%, Mean Absolute Scaled Error (MASE) dropped by 1.82%, and high-peak MAE decreased by 3.36%. For the univariate case, NSE improved by 1.44%, MAE decreased by 4.02%, MASE dropped by 3.89%, and high-peak MAE improved by 2.8%. In addition, multivariate models were considerably faster than univariate models, with training and inference times reduced by 74.57% and 73.81%, respectively. The multivariate models showed a 2.75% increase in NSE and a 72.04% decrease in MASE, indicating they captured better the hydrologic variability than the univariate models. Conversely, univariate models had a 20.59% lower MAE, a 21.17% lower high-peak MAE, and greater stability as indicated by tighter interquartile ranges, suggesting better error minimisation and more reliable predictions. Notably, in two river stations all models underperformed due to rapid flow variability and flashy hydrological responses in smaller catchment areas, suggesting in the future the use of higher-resolution climatic data. Overall, the study shows the potential of new dynamic multimodal fusion techniques, navigating the operational trade-offs between speed, stability, and accuracy across multi and uni-variate training strategies in streamflow forecasting. Nonetheless, the need for an optimal operational balance remains, suggesting further refinement of fusion techniques and focusing on minimising uncertainty.

How to cite: Theodosiadou, K., Rodding Kjeldsen, T., and Barnes, A.: Improving high-flow forecasting using dynamic multimodal feature fusion, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7020, https://doi.org/10.5194/egusphere-egu25-7020, 2025.

EGU25-7094 | ECS | Posters on site | HS3.4

Enhancing River Nutrient Predictions with Extreme Weather Indices and DL-Physical Hybrid Structures for Improved Interpretability 

Jiayi Tang, Leyang Liu, Kwok Chun, and Ana Mijic

Accurate nutrient predictions are crucial for river water quality management. While deep learning (DL) has shown promise in various Earth science applications, challenges such as data scarcity and limited interpretability hinder its use in river nutrient predictions. Building on insights into the physical dynamics of nutrients, this research investigates how incorporating extreme weather indices as additional input data, which are often overlooked in current DL-based nutrient prediction, could affect model performance. Additionally, we aim to improve model interpretability by developing hybrid DL-physical structures and identify the optimal structure for predicting nutrient indicators. 
 
The study proposes an assessment workflow and demonstrates its application by predicting dissolved inorganic nitrogen (DIN) and soluble reactive phosphorus (SRP) concentrations at the outlet of the Salmons Brook catchment, UK, where nutrient observations are scarce. The workflow includes two key decisions: selecting the input dataset and defining the DL-physical hybrid structure, each with two options. Comparing multiple predictions generated from all decision combinations enables the evaluation of the impacts of extreme weather events and different hybrid structures. 
 
The simulations demonstrate that incorporating extreme weather indices as additional inputs enhanced performance for both nutrient indicators, particularly in capturing extreme values. Overall, the choice of input dataset had a greater impact on the simulations than the hybrid structure, highlighting the importance of careful input selection and preprocessing in DL model development. Integrating results from a physical model into a DL model can improve simulation interpretability by introducing nutrient-related physical processes. In addition to the hybrid structure, incorporating insights into the physical behaviour of nutrients further enhances the interpretability of DL-based predictions, which is crucial for gaining the trust of domain experts, especially when validating results. 

How to cite: Tang, J., Liu, L., Chun, K., and Mijic, A.: Enhancing River Nutrient Predictions with Extreme Weather Indices and DL-Physical Hybrid Structures for Improved Interpretability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7094, https://doi.org/10.5194/egusphere-egu25-7094, 2025.

EGU25-7205 | ECS | Orals | HS3.4

Evaluating uncertainty in probabilistic deep learning models using Information Theory 

Manuel Alvarez Chaves, Hoshin Gupta, Uwe Ehret, and Anneli Guthke

Deep learning methods in hydrology have traditionally focused on deterministic models, limiting their ability to quantify prediction uncertainty. Recent advances in generative modeling have opened new possibilities for probabilistic modelling in various applied fields, including hydrological forecasting (Jahangir & Quilty, 2024). These models learn to represent underlying probability distributions using neural networks, enabling uncertainty quantification through sampling in a very flexible framework.

In this submission we introduce vLSTM, a variational extension of the traditional long short-term memory (LSTM) architecture that quantifies predictive uncertainty by adding noise sampled from a learned multivariate Gaussian distribution to perturb the model’s hidden state. The vLSTM preserves the traditional LSTM’s state-space dynamics while introducing a probabilistic component that enables uncertainty quantification through sampling. Unlike mixed-density networks (MDNs) which directly model the distribution of the target variable, vLSTM’s uncertainty is obtained by perturbations to the hidden state, providing a novel approach to probabilistic prediction. In rainfall-runoff modeling, vLSTM offers a different mechanism for uncertainty quantification to the well established MDN models (Klotz et al., 2022). This approach enriches the existing toolkit of uncertainty methods in deep learning while maintaining the simplicity of sampling for probabilistic predictions.

To rigorously evaluate probabilistic predictions across different model architectures, we develop new information-theoretic metrics that capture key aspects of how uncertainty is handled by a particular model. These include the average prediction entropy H(X), which quantifies model confidence, and average relative entropy DKL(pq), which measures the average alignment between the predicted distribution of a model and a target, among others. The proposed metrics take advantage of non-parametric estimators for Information Theory which have been implemented in the easy to use UNITE toolbox (https://github.com/manuel-alvarez-chaves/unite_toolbox). By expressing these metrics in compatible units of bits (or nats), we enable direct comparisons between different uncertainty measures. We apply these metrics to our newly introduced vLSTM and the existing MDN models to show strengths and weaknesses of each approach. This information-theoretic framework provides a unified language for analyzing and understanding predictive uncertainty in probabilistic models.

References

  • Jahangir, M. S., & Quilty, J. (2024). Generative deep learning for probabilistic streamflow forecasting: Conditional variational auto-encoder. Journal of Hydrology, 629, 130498. https://doi.org/10.1016/j.jhydrol.2023.130498
  • Klotz, D., Kratzert, F., Gauch, M., Keefe Sampson, A., Brandstetter, J., Klambauer, G., Hochreiter, S., & Nearing, G. (2022). Uncertainty estimation with deep learning for rainfall–runoff modeling. Hydrology and Earth System Sciences, 26(6), 1673–1693. https://doi.org/10.5194/hess-26-1673-2022

How to cite: Alvarez Chaves, M., Gupta, H., Ehret, U., and Guthke, A.: Evaluating uncertainty in probabilistic deep learning models using Information Theory, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7205, https://doi.org/10.5194/egusphere-egu25-7205, 2025.

EGU25-7468 | ECS | Posters on site | HS3.4

Empirical Evidence of the Importance of Data Recency in LSTM-Based Rainfall-Runoff Modeling  

Qiutong Yu and Bryan Tolson

Deep learning (DL)-based hydrological models, particularly those using Long Short-Term Memory (LSTM) networks, typically require large datasets for effective training. In the context of large-scale rainfall-runoff modeling, dataset size can refer to either the number of watersheds or the length of the training period. While it is well established that training a regional model across more watersheds improves performance (Kratzert et al., 2024), the benefits of extending the training period are less clear.

Empirical evidence from studies such as Boulmaiz et al. (2020) and Gauch et al. (2021) suggests that longer training periods enhance LSTM performance in rainfall-runoff modeling. This improvement is attributed to the need for extensive datasets to ensure proper model convergence and the ability to capture a wide range of hydrological conditions and events. However, these studies neglected the influence of data recency (or data recentness), which is critical for operational applications that forecast current and future hydrological conditions. In the context of climate change and anthropogenic interventions, the assumption of stationarity (i.e., that historical patterns reliably represent future conditions) may no longer hold for hydrological systems (Shen et al., 2022). Consequently, the selection of training periods should account for potential non-stationarity, as more recent data may better reflect current rainfall-runoff dynamics. Intriguingly, Shen et al. (2022) found that calibrating hydrologic models to the latest data is a superior approach compared to using old data, and completely discarding the oldest data can even improve the performance in streamflow prediction.

This study aims to address two research questions: (1) As the number of watersheds increases, is it still necessary to train LSTM models on decades of historical observations? (2) Can LSTM models achieve comparable performance using shorter training periods focused on more recent data? Specifically, we examine whether models trained on recent data outperform those trained on older data and explore how different temporal partitions of historical records affect predictive skill.

This study leverages a comprehensive dataset comprising streamflow records from over 1,300 watersheds across North America, representing diverse climatic and hydrological regimes, with streamflow data spanning 1950 to 2023. Training periods are designed to isolate the effects of temporal data recency while keeping period lengths consistent. This approach enables a systematic comparison of model performance using exclusively older (e.g., pre-1980) versus exclusively recent data (e.g., post-1980). This research provides evidence-based recommendations for selecting training data while balancing computational costs, data availability, and prediction accuracy.

 

References

Boulmaiz, T., Guermoui, M., and Boutaghane, H.: Impact of training data size on the LSTM performances for rainfall–runoff modeling, Model Earth Syst Environ, 6, 2153–2164, https://doi.org/10.1007/S40808-020-00830-W/FIGURES/9, 2020.

Gauch, M., Mai, J., and Lin, J.: The proper care and feeding of CAMELS: How limited training data affects streamflow prediction, Environmental Modelling and Software, 135, https://doi.org/10.1016/j.envsoft.2020.104926, 2021.

Kratzert, F., Gauch, M., Klotz, D., and Nearing, G.: HESS Opinions: Never train an LSTM on a single basin, Hydrology and Earth System Science, https://doi.org/10.5194/hess-2023-275, 2024.

Shen, H., Tolson, B. A., and Mai, J.: Time to Update the Split-Sample Approach in Hydrological Model Calibration, Water Resour Res, 58, e2021WR031523, https://doi.org/10.1029/2021WR031523, 2022.

How to cite: Yu, Q. and Tolson, B.: Empirical Evidence of the Importance of Data Recency in LSTM-Based Rainfall-Runoff Modeling , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7468, https://doi.org/10.5194/egusphere-egu25-7468, 2025.

EGU25-7531 | ECS | Orals | HS3.4

Predicting total organic carbon loads in river using a mass-conserving LSTM integrated with QUAL2E kinetics 

Hyemin Jeong, Byeongwon Lee, Younghun Lee, and Sangchul Lee

The increasing complexity of water pollution and its impact on aquatic ecosystems necessitates the accurate prediction of water pollutant loads for effective river management. Total Organic Carbon (TOC), a key indicator of organic pollution levels, is central to assessing ecosystem health and informing water treatment strategies. However, conventional process-based modeling methods, while capable of providing precise water quality predictions, require extensive input data and significant computational resources, limiting their practical application. Consequently, alternative modeling approaches, particularly those leveraging artificial intelligence, have been explored. Recent advancements in deep learning have improved predictive modeling in environmental sciences. These approaches have showed effectiveness in hydrological applications, such as streamflow forecasting, by capturing complex nonlinear relationships within environmental systems. Despite these advancements, a notable limitation of these models is their difficulty in maintaining physical consistency, specifically in adhering to the principle of mass balance—a fundamental concept in both hydrology and water quality modeling. In this study, we evaluate a Mass-Conserving Long Short-Term Memory network integrated with QUAL2E kinetics (MC-LSTM-QUAL2E) to predict TOC loads in river systems. By incorporating representations of decay and reaeration processes within a mass-conserving neural network framework, this model combines data-driven prediction capabilities with the requirements of physical consistency. A key component of this framework is the trash cell, designed to simulate TOC transformations based on QUAL2E dynamics. Within the trash cell, TOC decay and reaeration are modeled using parameters kdecay​ and kreaeration​, which are determined by environmental variables such as temperature, pH, dissolved oxygen, total nitrogen, and total phosphorus. The QUAL2E module updates the trash state at each timestep to account for TOC losses due to decay and gains from reaeration, ensuring mass conservation.  The MC-LSTM-QUAL2E model was compared to a conventional LSTM model using environmental variables, including temperature, pH, dissolved oxygen, and nutrient levels, as inputs. The analysis used data from 2012 to 2020, with the period from 2012 to 2017 designated for training and 2018 to 2020 for tests. Model performance was assessed using metrics such as Nash-Sutcliffe Efficiency (NSE), Kling-Gupta Efficiency (KGE), Root Mean Square Error-observations standard deviation Ratio (RSR), and Percent Bias (PBIAS). By maintaining mass balance and incorporating QUAL2E dynamics, the model provides reliable predictions of TOC loads in river systems and offers insights into associated biochemical and hydrological processes.

How to cite: Jeong, H., Lee, B., Lee, Y., and Lee, S.: Predicting total organic carbon loads in river using a mass-conserving LSTM integrated with QUAL2E kinetics, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7531, https://doi.org/10.5194/egusphere-egu25-7531, 2025.

EGU25-8045 | Posters on site | HS3.4

Developing a Deep Learning-Based Super-Resolution Urban Flood Model: Towards Scalable and Reliable Hydrological Predictions 

Hyeonjin Choi, Hyuna Woo, Minyoung Kim, Hyungon Ryu, Junhak Lee, Seungsoo Lee, and Seong Jin Noh

Integrating deep learning techniques into hydrology has opened a new way to improve urban flood modeling, with various solutions being developed to address urban flood problems driven by climate change and urbanization. However, predicting urban inundation in near real-time for large urban areas remains challenging due to computational demands and limited data availability. This work proposes a deep learning-based super-resolution framework that enhances the spatial resolution of process-based urban flood modeling outputs using convolutional neural networks (CNNs) while improving computational efficiency. This study investigates the interaction between deep learning model architecture and the underlying physical processes to improve prediction accuracy and robustness in urban pluvial flood mapping. The methodology will be applied to various urban flood scenarios, including extreme rainfall events and hurricane-induced flooding, and its performance will be evaluated through quantitative indicators and sensitivity analyses. The applicability and scalability of this model will also be discussed. In particular, strategies to enhance model reliability and integrate additional hydrological information under extreme conditions will be explored. The study will further address uncertainty estimation in deep learning-based super-resolution models and scalability challenges associated with super-resolution approaches for large-scale flood simulations. The findings aim to demonstrate the potential of deep learning as an innovative tool in hydrological modeling and to enable more effective flood risk management strategies.

How to cite: Choi, H., Woo, H., Kim, M., Ryu, H., Lee, J., Lee, S., and Noh, S. J.: Developing a Deep Learning-Based Super-Resolution Urban Flood Model: Towards Scalable and Reliable Hydrological Predictions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8045, https://doi.org/10.5194/egusphere-egu25-8045, 2025.

EGU25-9088 | ECS | Orals | HS3.4

Projecting storm surge extremes with a deep learning surrogate model 

Emiliano Longo, Andrea Ficchì, Sanne Muis, Martin Verlaan, and Andrea Castelletti

Sea level rise and increasing coastal flood risks demand the development of accurate and efficient coastal risk models capable of generating large ensembles of projections to support robust adaptation strategies. The latest IPCC report emphasizes the importance of projecting storm surge changes and their associated uncertainties, alongside mean sea level rise. However, the high computational cost of storm surge simulations continues to limit the feasibility of generating large ensembles.
Artificial Intelligence (AI) is emerging as a promising alternative to simulate storm surge scenarios with significantly reduced computational costs. Despite recent advancements, key challenges remain in accurately representing extreme events and ensuring robust model extrapolation under changing climate conditions. While AI-based surrogate models have been proposed in the literature, gaps persist in understanding their performance limits for extreme events in future scenarios, hindering their application in climate adaptation planning.
To address these challenges, we developed a deep learning (DL) surrogate model of the physics-based Global Tide and Surge Model (GTSM). The DL model is trained using reanalysis data (ERA5) and historical scenarios from the Coupled Model Intercomparison Project Phase 6 (CMIP6) High Resolution Model Intercomparison Project (HighResMIP). Our analysis focuses on the DL model's performance in simulating extreme storm surge events, validated against GTSM outputs for both historical reanalysis and future projections, with a case study along the New York coastline.
To enhance the surrogate model’s performance for extreme events, we explore various loss functions, including a customized quantile loss function, and test alternative DL architectures with different input configurations. Results demonstrate that the quantile loss improves the model's accuracy for extremes compared to standard loss functions such as mean square error. Additionally, fine-tuning DL models with specific Global Climate Model forcing fields improves the alignment of AI-predicted storm surge trajectories with GTSM outputs, even under diverse spatiotemporal resolutions and model setups.
These findings highlight the critical importance of selecting appropriate loss functions and training datasets to ensure robust performance over extreme events and projected future scenarios. Our globally applicable framework, relying solely on open-source data, offers a promising pathway to scalable and efficient storm surge projections, with implications for robust coastal adaptation planning.

How to cite: Longo, E., Ficchì, A., Muis, S., Verlaan, M., and Castelletti, A.: Projecting storm surge extremes with a deep learning surrogate model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9088, https://doi.org/10.5194/egusphere-egu25-9088, 2025.

EGU25-10650 | ECS | Posters on site | HS3.4

The relationship between theoretical maximum prediction limits of the LSTM and network size 

Daniel Klotz, Sanika Baste, Ralf Loritz, Martin Gauch, and Frederik Kratzert

Machine learning is increasingly important for rainfall–runoff modelling. In particular, the community started to widely adopt the Long Short-Term Memory (LSTM) network. One of the most important established best practices  in this context is to train the LSTMs on a large number of diverse basins  (Kratzert et al., 2019; 2024). Intuitively, the reason for adopting this practice is that training deep learning models on small and homogeneous data sets (e.g., data from only a single hydrological basin) leads to poor generalization behavior — especially for high-flows. 

 

To examine this behavior, Kratzert et al. (2024) use a theoretical maximum prediction limit for LSTMs. This theoretical limit is computed as the L1 norm (i.e., the sum of the absolute values of each vector component) of the learned weight vector that relates the hidden states to the estimated streamflow. Hence, for random vectors we could simply obtain larger theoretical limits by increasing the size of the network (i.e., the  number of parameters). However, since LSTMs are trained using gradient descent, this relationship is more intricate. 

 

This contribution explores the relationship between the theoretical limit and the network size. In particular, we will look at how increasing the network size in untrained models increases the prediction limit and contrast it to the scaling behavior of trained models.



How to cite: Klotz, D., Baste, S., Loritz, R., Gauch, M., and Kratzert, F.: The relationship between theoretical maximum prediction limits of the LSTM and network size, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10650, https://doi.org/10.5194/egusphere-egu25-10650, 2025.

EGU25-11240 | ECS | Orals | HS3.4

Exploring the Transferability of Knowledge In Deep Learning-Based Streamflow Models Across Global Catchments 

Jamal Hassan, John Rowan, and Nandan Mukherjee

Accurate streamflow prediction is critical for flood forecasting and water resource management, particularly in data-scarce regions like Central Asia (CA), where traditional hydrological models struggle due to insufficient discharge data. Deep learning models, such as Long Short-Term Memory (LSTM), have demonstrated the potential for global hydrologic regionalization by leveraging both climate data and catchment characteristics. We used a transfer learning (TL) approach to improve streamflow predictions by first pretraining LSTM models on catchments from data-rich regions like Switzerland, Scotland, and British Columbia (source regions). These deep learning models were then fine-tuned on the data scarce target region (CA basins). This approach leverages the knowledge gained from the source regions to adapt the model to the target region, enhancing prediction accuracy despite the data scarcity in CA. Incorporating lagged streamflow alongside ERA-5 climate data boosted prediction accuracy, particularly in snowmelt and glaciers influenced basins like Switzerland (median NSE=0.707 to 0.837), British Columbia (median NSE= 0.775 to 0.923) and CA (median NSE=0.693 to 0.798). K-Means algorithm was applied to categorize catchments from four global locations into five clusters (labeled 0–4) based on their specific attributes. The predictive performance of fine-tuned LSTM model has significantly enhanced when leveraging a pre-trained model with cluster 2, as demonstrated by higher median metrics (NSE=0.958, KGE=0.905, RMSE=10.723, MSE=115.055) compared to both the locally trained model (NSE=0.851, KGE=0.792, RMSE=20.377, MSE=415.579) and individual basin-based training approaches (NSE=0.69, KGE=0.692, RMSE=25.563, MSE=676.110). These results highlight the effectiveness of pretraining the LSTM model on diverse clusters (0, 1, 2, and 4) before fine-tuning on the target region (cluster 3). Moreover, pretraining the LSTM model with clusters 0 and 4 resulted in enhanced performance by increasing the number of basins, whereas the impact was minimal or even declined when using clusters 1 and 2, as well as when all basins from the four clusters were included. These findings demonstrate the feasibility of transfer learning in addressing data scarcity challenges and underscore the importance of diverse and high-quality training data in developing robust, regionalized hydrological models. This approach bridges the gap between data-rich and data-scarce regions, offering a pathway to improved flood prediction and water resource management.

How to cite: Hassan, J., Rowan, J., and Mukherjee, N.: Exploring the Transferability of Knowledge In Deep Learning-Based Streamflow Models Across Global Catchments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11240, https://doi.org/10.5194/egusphere-egu25-11240, 2025.

EGU25-11682 | ECS | Posters on site | HS3.4

Improving AI-based discharge forecasting through hydrograph decomposition and data assimilation 

Bob E Saint Fleur, Eric Gaume, and Nicolas Akil

Effective discharge forecasting is critical in operational hydrology. This study explores novel methods to improve forecast accuracy by combining data assimilation techniques and hydrograph decomposition. Traditional rainfall-runoff modeling, including AI-based approaches, typically simulates the entire discharge signal using a single model. However, runoff is generated by multiple processes with contrasting kinetics, which a single-model approach may fail to capture adequately. This study proposes using hydrograph decomposition to separate baseflow and quickflow components, training specific forecasting models for each component individually, and then merging their outputs to reconstruct the total discharge signal. This approach is expected to enhance forecast accuracy for both floods and droughts, identifying long-term dependencies governing baseflow to improve seasonal low-flow forecasts. Experiments will be conducted using a subset of the CAMELS dataset.

How to cite: Saint Fleur, B. E., Gaume, E., and Akil, N.: Improving AI-based discharge forecasting through hydrograph decomposition and data assimilation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11682, https://doi.org/10.5194/egusphere-egu25-11682, 2025.

EGU25-12110 | Orals | HS3.4

Increasing the Accuracy and Resilience of Streamflow Forecasts through Data Augmentation and High Resolution Weather Inputs 

David Lambl, Simon Topp, Phil Butcher, Mostafa Elkurdy, Laura Reed, and Alden K. Sampson

Accurately forecasting streamflow is essential for effectively managing water resources. High-quality operational forecasts allow us to prepare for extreme weather events, optimize hydropower generation, and minimize the impact of human development on the natural environment. However, streamflow forecasts are inherently limited by the quality and availability of upstream weather sources. The weather forecasts that drive hydrological modeling vary in their temporal resolutions and are prone to outages, such as the ECMWF data outage in November of 2023. 

Here, we present HydroForecast Short Term 3 (ST-3), a state-of-the-art probabilistic deep learning model for medium-term (10-day) streamflow forecasts. ST-3 combines long short-term memory architecture with Boolean tensors representing data availability and dense embeddings for processing of the information in these tensors. This architecture allows for a training routine that implements data augmentation to synthesize varying amounts of availability of weather inputs. The result is a model that 1) makes accurate forecasts even in the case of an upstream data outage, 2) achieves higher accuracy by leveraging data of varying temporal resolutions including regional weather inputs with shorter lead times than the most common medium term weather inputs, and 3) generates individual forecast traces for each individual weather source, facilitating inference across regions where weather data availability is limited. 

Initial results across CAMELS sites in North America indicate that the incorporation of near-term high resolution weather data increases early horizon forecast KGE by nearly 0.25 with meaningful improvements in metrics seen across our customers’ operational sites. Validation metrics across individual weather sources, as well as model interrogation through integrated gradients highlights a high level of fidelity in the model’s learned physical relationships across forecast scenarios.

How to cite: Lambl, D., Topp, S., Butcher, P., Elkurdy, M., Reed, L., and Sampson, A. K.: Increasing the Accuracy and Resilience of Streamflow Forecasts through Data Augmentation and High Resolution Weather Inputs, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12110, https://doi.org/10.5194/egusphere-egu25-12110, 2025.

Machine learning (ML) models have transformed our ability to perform reasonably-accurate, large-scale river discharge modeling, opening new opportunities for global prediction in ungauged basins. These ML models are data-hungry, and results have conclusively shown that ML techniques do best when a single ML model is trained on all basins in the dataset. This is contrary to inuitions from the hydrological sciences, where individual basin calibration traditionally provides the best forecasts. 

 

We bridge this gap between intuitions from traditional ML and hydrology by pre-training a single global model on basins in the worldwide Caravan dataset (~6000 basins), and then fine-tune that model on individual basins. This is a well-known practice within ML, and for us serves the purpose of producing models aimed at high-quality local prediction problems while still capturing the advantages of large-sample training. We show that this leads to a significant skill improvement. 

 

We have also conducted analysis of geophysical and hydrological regimes that benefit most from fine-tuning. These results point to how flood forecasting and water management agencies and operators can expect to fine-tune large, pretrained models on their own local data, which may be proprietary and not part of large, global training datasets.

 

This work illustrates how local agencies like national hydromet agencies or flood forecasting agencies might be able to leverage machine learning based hydrological forecast models while also maximizing the value and information of local data by tailoring large, pretrained models to their own local context. This is an important step in allowing local agencies to take ownership of these global models, and directly incorporate local hydrological understanding to improve performance.

How to cite: Ryd, E. and Nearing, G.: Fine Flood Forecasts: Calibrating global machine learning flood forecast models at the basin level through fine-tuning., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13027, https://doi.org/10.5194/egusphere-egu25-13027, 2025.

EGU25-13844 | ECS | Orals | HS3.4

AIFAS: Probabilistic Global River Discharge Forecasting 

Mohamad Hakam Shams Eddin, Yikui Zhang, Stefan Kollet, and Juergen Gall

Hydrological models are vital in river discharge to issue timely flood warnings and mitigate hydrological risks. Recently, advanced techniques in deep learning have significantly enhanced flood prediction by improving the accuracy and efficiency of forecasts, enabling more reliable early warnings and decision-making in flood risk management. Nevertheless, current applications of deep learning methods are still more restricted to local-scale models or in the best case on selected river points at a global scale. Many studies also lack spatial and topological information for training deep learning models, which can limit their generalization ability when applied to large regions with heterogeneous hydrological conditions. In addition, the lack of probabilistic forecasting impedes the quantification of uncertainty in flood predictions. Here we present the Artificial Intelligence Flood Awareness System (AIFAS) for probabilistic global river discharge forecasting. AIFAS is a generative AI model that is trained with long-term historical reanalysis data and can provide grid-based global river discharge forecasting at 0.05°. At the core of our model are the built-in vision module upon state space model (SSM) [1] and the diffusion-based loss function [2]. The vision SSM allows the model to connect the routing of the channel networks globally, while the diffusion loss generates ensembles of stochastic river discharge forecasts. We evaluate the AIFAS forecast skill against other state-of-the-art deep learning models, such as Google LSTM [3], climatology baseline, persistence baseline, and operational GloFAS forecasts [4]. The impact of different hydrometeorological products that drive AIFAS performance on different forecasting lead times will also be discussed. Our results show that the new forecasting system achieves reliable predictions of extreme flood events across different return periods and lead times.

References

[1] Gu, A., and Dao, T.: Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752, 2023.

[2] Ho, J., Jain, A., and Abbeel, P.: Denoising diffusion probabilistic models. Advances in neural information processing systems, 33, 6840-6851, 2020.

[3] Nearing, G., Cohen, D., Dube, V. et al. Global prediction of extreme floods in ungauged watersheds. Nature 627, 559–563 2024.

[4] Harrigan, S., Zsoter, E., Cloke, H., Salamon, P., and Prudhomme, C.: Daily ensemble river discharge reforecasts and real-time forecasts from the operational Global Flood Awareness System, Hydrol. Earth Syst. Sci., 27, 1–19, 2023.

How to cite: Shams Eddin, M. H., Zhang, Y., Kollet, S., and Gall, J.: AIFAS: Probabilistic Global River Discharge Forecasting, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13844, https://doi.org/10.5194/egusphere-egu25-13844, 2025.

EGU25-14004 | ECS | Posters on site | HS3.4

Learning Basin Similarity Through Combined Deep Learning and Random Forest Approaches for Improved Parameter Transfer in Ungauged Basins 

Rojin Meysami, Qiutong Yu, Bryan Tolson, Hongren Shen, and Rezgar Arabzadeh

Parameter regionalization for ungauged basins remains a critical challenge in hydrological modeling. While traditional approaches rely on physical catchment descriptors or spatial proximity, and recent machine learning applications have focused primarily on direct streamflow prediction, there remains significant potential to leverage machine learning for improved parameter transfer strategies. This study explores novel approaches that combine Long Short-Term Memory (LSTM) networks and Random Forest (RF) models to predict basin similarity and optimize parameter transfer for physically-based hydrologic models. Using case studies from British Columbia's Fraser River Basin and Ontario's Great Lakes region, we test multiple methodologies for integrating deep learning with traditional parameter transfer approaches. Our primary benchmark is established through an exhaustive parameter transfer experiment using the Raven hydrological model, where parameters from each potential donor basin were transferred to every possible receiver basin across 10 independent trials. This benchmark represents the best achievable KGE via parameter transfer methods. Our framework employs a regional LSTM model to capture complex streamflow dynamics and characterize basin similarity, then explores various RF-based approaches to predict optimal donor-receiver basin pairs for parameter transfer. These methods are evaluated against both the exhaustive transfer benchmark and emerging machine learning approaches. Results indicate that thoughtfully combining deep learning and random forest techniques can capture nuanced relationships between basin characteristics and hydrological response similarity, advancing the state-of-the-art in parameter regionalization for ungauged basins while maintaining physical interpretability.

How to cite: Meysami, R., Yu, Q., Tolson, B., Shen, H., and Arabzadeh, R.: Learning Basin Similarity Through Combined Deep Learning and Random Forest Approaches for Improved Parameter Transfer in Ungauged Basins, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14004, https://doi.org/10.5194/egusphere-egu25-14004, 2025.

EGU25-14223 | Orals | HS3.4

On the advancing frontier of deep learning in hydrology:  a hydrologic applications perspective 

Andy Wood, Laura Read, Grey Nearing, Juliane Mai, Chris Frans, Martyn Clark, and Florian Pappenberger

In the last decade, the realization that certain deep learning (DL) architectures are particularly well-suited to the simulation and prediction of hydrologic systems and their characteristic memory-influenced dynamics has led to remarkable rise in DL-centered hydrologic research and applications.  Numerous new datasets, computational and open software resources, and progress in related fields such as numerical weather prediction have also bolstered this growth.  Advances in DL for hydrologic forecasting research and operations is likely the most eye-catching and intuitive use case, but DL methods are now also making inroads into more process-intensive hydrologic modeling contexts, and among groups that have been skeptical of their potential suitability despite performance-related headlines. Nevertheless, even in the forecasting context, and despite offering new strategies and concepts to resolve long-standing hurdles in hydrologic process-based modeling efforts, the uptake of DL-based systems in many public-facing services and applications has been slow. 

This presentation provides perspective on the ways in which DL techniques are garnering interest in traditionally process-oriented modeling arenas -- from flood and drought forecasting to watershed studies to hydroclimate risk modeling – and on sources of hesitancy.  Clear pathways, momentum and motivations for DL approaches to supplant process-based models exist in some applications, whereas in others, governing interests and constraints appear likely to restrict DL innovations to narrower niches.  Concerns over explainability have been a common topic, but less discussed questions about fitness or adequacy for purpose and institutional requirements can also be influential.  Drawing from relevant hydrologic modeling programs, projects and initiatives in the US and elsewhere, we aim to provide a real-world status update on the advancing frontier of deep learning in applied hydrologic science and practice.  

How to cite: Wood, A., Read, L., Nearing, G., Mai, J., Frans, C., Clark, M., and Pappenberger, F.: On the advancing frontier of deep learning in hydrology:  a hydrologic applications perspective, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14223, https://doi.org/10.5194/egusphere-egu25-14223, 2025.

EGU25-15152 | Orals | HS3.4

Evaluation of LSTM Model for Stochastic Discharge Simulation 

Sonja Jankowfsky, Kanneganti Gokul, Shuangcai Li, Arno Hilberts, and Anongnart Assteerawatt

This study evaluates the capacity of a Long Short-Term Memory (LSTM) model trained on a diverse river discharge dataset from over 4,000 USGS gauges across the United States with the aim to generate extremely long stochastic discharge simulations. 

The LSTM model (Kratzert et al., 2022) was trained using 30 years of NLDAS v2 forcings, which were split into 10-year periods for training, validation, and testing respectively. Sixty percent of the gauges had a Nash Sutcliffe Efficiency (NSE) larger than 0.4 in the validation period, and ten percent had an NSE larger than 0.8, which was considered sufficient to proceed with applying the model using stochastic precipitation.  

The stochastic simulations are evaluated in terms of the model’s ability to capture peak discharges. The stochastic return period (RP) curves were evaluated against those from the historical time period and the observed discharge. For most of the gauges, the stochastic RP curves are in line with the historical RP curves, and for all of the gauges, the stochastic RP curves discharge of the extreme return period extend far beyond the discharge of the historical time period, showing the capacity of the model to extrapolate beyond the training dataset. 

This capacity, which is usually lacking in single-basin trained models, most likely results from training on a large dataset with a wide range of climatic conditions and variability as suggested by Kratzert et al. (2024). These findings underscore the robustness and versatility of the LSTM model in long-term stochastic discharge simulations, highlighting its potential for broader hydrological applications. 

Kratzert, F., Gauch, M., Nearing, G., & Klotz, D. (2022). NeuralHydrology — A Python library for Deep Learning research in hydrology. Journal of Open Source Software, 7(71), 4050. https://doi.org/10.21105/joss.04050

Kratzert, F., Gauch, M., Klotz, D., and Nearing, G. (2024). HESS Opinions: Never train an LSTM on a single basin. Hydrology and Earth System Sciences (HESS), Volume 28, Issue 17, published on September 12, 2024.  https://doi.org/10.5194/hess-28-4187-2024.

How to cite: Jankowfsky, S., Gokul, K., Li, S., Hilberts, A., and Assteerawatt, A.: Evaluation of LSTM Model for Stochastic Discharge Simulation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15152, https://doi.org/10.5194/egusphere-egu25-15152, 2025.

EGU25-15171 | ECS | Posters on site | HS3.4

Learning shallow water equations with physics-informed Deep Operator Network (DeepONet) 

Robert Keppler, Julian Koch, and Rasmus Fensholt

Physics-informed neural networks are an optimization-based approach for solving differential equations and have the potential to significantly speed up the modelling of complex phenomena, which conventionally is achieved via expensive numerical solvers. We present a Physics-Informed Deep Operator Network (DeepONet) framework for solving two-dimensional shallow water equations with variable bed topography under given boundary and initial conditions. While traditional physics-informed neural networks can solve differential equations on meshless grids using prescribed conditions, they require retraining for each new set of initial and boundary conditions. Our approach uses a DeepONet to learn the underlying solution operator rather than individual solutions, which provides an enhanced generalizability, making the DeepONet a feasible candidate for real world applications. The framework combines the advantages of neural networks with physical laws, effectively handling the complexities of varying bed topography and wet-dry transitions. We demonstrate that our DeepONet approach achieves comparable accuracy to classical numerical methods while significantly reducing inference time. In our modelling experiments we investigate the sensitivity of hyperparameter values and network architecture as well as the potential of introducing an additional data loss, emulating the availability of additional observational data on water levels or inundation extent.  This acceleration in computation speed makes the method particularly valuable for time-critical applications such as flood forecasting. The results establish physics-informed DeepONets as a promising alternative to traditional numerical solvers for shallow water systems, offering a balance between computational efficiency and solution accuracy.

How to cite: Keppler, R., Koch, J., and Fensholt, R.: Learning shallow water equations with physics-informed Deep Operator Network (DeepONet), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15171, https://doi.org/10.5194/egusphere-egu25-15171, 2025.

EGU25-15229 | ECS | Posters on site | HS3.4

Cryosphere Data and Its Value for Deep Learning Hydrological Simulations 

Corinna Frank, Jan Philipp Bohl, Manuela Brunner, Martin Gauch, and Marvin Höge

Deep learning models have been successfully applied to simulate streamflow in mountain catchments. While these mostly lumped models have demonstrated the ability to learn processes such as snow accumulation and melt that are crucial for streamflow generation in these regions, they still show deficiencies in simulating streamflow during the melting period. This suggests a misrepresentation of melting dynamics encoded within these models. We hypothesize that the sets of lumped meteorological variables (such as air temperature, precipitation, PET) and static attributes currently used to train and drive these models are not sufficient to describe the melting processes. 

To enhance the representation of snow and ice-related processes, we thus propose to incorporate additional data on snow and ice cover, such as Snow Covered Area, Snow Water Equivalent, and glacier mass within the respective basin. We assess (1) how much additional value can be extracted from cryosphere data to improve the representation of cryosphere related processes and (2) how the added value varies across different geographies and catchment types. In a lumped Long Short-Term Memory (LSTM) setup covering a large sample of catchments in different European mountainous regions, we compare different data integration methods with respect to their uncertainty reduction for streamflow simulation and their limitations for model applications.
Our findings provide insights into optimizing model configurations and data usage and offer practical guidance for ultimately improving the accuracy of streamflow simulations in mountainous, snow-influenced regions. 

How to cite: Frank, C., Bohl, J. P., Brunner, M., Gauch, M., and Höge, M.: Cryosphere Data and Its Value for Deep Learning Hydrological Simulations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15229, https://doi.org/10.5194/egusphere-egu25-15229, 2025.

EGU25-15350 | ECS | Posters on site | HS3.4

Graph-Based Representations in Hydrological Modeling: Comparing SWAT+ and Graph Neural Networks for Water Management Systems 

Moritz Wirthensohn, Ye Tuo, and Markus Disse

Extreme hydrological events such as droughts and floods are expected to become more frequent and severe according to climate change projections, making effective water management very important to mitigate environmental and socio-economic impacts. In this context, advanced hydrological modeling tools are essential for understanding and managing water systems. The Soil and Water Assessment Tool (SWAT+), a process-based and semi-distributed eco-hydrological model, has become very popular for simulating hydrological processes and water management scenarios, especially with its improved water allocation and reservoir modules. At the same time, Graph Neural Networks (GNNs), a deep learning model, have shown potential for modeling complex relationships in networked systems. Both SWAT+'s water allocation module and GNNs use graph-like structures to model water systems. The goal of this study is to systematically compare the structural components of these two approaches and provide insights into potential integration.

Using the Upper Isar River Basin's complex water management system as a case study, we examine how SWAT+ and GNNs can be used to model it. We perform a component-wise analysis, focusing on how these models can represent nodes, edges, and attributes in a networked water management system. While this study focuses on structural rather than performance comparisons, we anticipate that our results will highlight the strengths and limitations of each approach. SWAT+ is expected to excel at incorporating domain-specific knowledge and explicitly representing management actions. GNNs could provide advantages in learning complex patterns from data and faster simulations for larger catchments.

The findings could open the way for hybrid approaches that combine traditional hydrological models' strengths with GNNs' learning capabilities. This could lead to more robust and adaptable water management tools to deal with the growing complexity of hydrological systems caused by climate change and human intervention.

How to cite: Wirthensohn, M., Tuo, Y., and Disse, M.: Graph-Based Representations in Hydrological Modeling: Comparing SWAT+ and Graph Neural Networks for Water Management Systems, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15350, https://doi.org/10.5194/egusphere-egu25-15350, 2025.

Accurate reservoir outflow simulation is crucial for modeling streamflow in reservoir-regulated basins. In this study, we introduce a knowledge-guided Long Short-Term Memory model (KG-LSTM) to simulate the outflow of reservoirs-Fengshuba, Xinfengjiang, and Baipenzhu in the Dongjiang River Basin, China. KG-LSTM is built on the standard hyperparameters-optimized-LSTM and the loss function considering reservoir operation knowledge, while traditional reservoir model level pool scheme (LPS) is used as a benchmark model. Model uncertainty is analyzed using the bootstrap method. We then propose a hybrid approach that combines KG-LSTM with the Three-parameter monthly hydrological Model based on the Proportionality Hypothesis (KG-LSTM-TMPH) for streamflow simulation. The propagation of inflow errors to outflow simulations is studied across the three reservoirs. Results indicate that LSTM-based models greatly outperform LPS in all three reservoirs, with KG-LSTM demonstrating superior capability in capture reservoir outflow dynamics compared to the standard LSTM model. In the multi-year regulated Xinfengjiang Reservoir, KG-LSTM improves Nash-Sutcliffe efficiency (NSE) from 0.59 to 0.64, and reduces root mean squared error (RMSE) from 55.59 m³/s to 54.84 m³/s during the testing period. KG-LSTM shows reduced model uncertainty, decreasing the relative width (RW) from 0.55 to 0.51 in the Xinfengjiang Reservoir and from 0.48 to 0.44 in the Baipenzhu Reservoir, while demonstrating limited change in the Fengshuba Reservoir. For streamflow simulation, KG-LSTM-TMPH performs best across all four stations, achieving NSE values of approximately 0.87, 0.88, 0.91, and 0.92 at Longchuan, Heyuan, Lingxia, and Boluo stations, respectively. In the dry season, KG-LSTM-TMPH demonstrates substantial improvement over LSTM-TMPH, increasing R² by +0.11 and reducing RMSE by -4.22 m³/s at Heyuan station. Inflow errors impact outflow most significantly for the Xinfengjiang Reservoir in April and May, for the Fengshuba Reservoir throughout the year (particularly in April, May, July, and August), and for the Baipenzhu Reservoir primarily in July and August. This study enhances reservoir outflow modeling by integrating reservoir operation knowledge with deep learning. The hybrid KG-LSTM-TMPH approach shows practical potential for streamflow simulation in reservoir-regulated basins, offering valuable applications for water resource management.

How to cite: Wang, D.: A Knowledge-guided LSTM reservoir outflow model and its application to streamflow simulation in reservoir-regulated basins, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16156, https://doi.org/10.5194/egusphere-egu25-16156, 2025.

EGU25-16414 | Posters on site | HS3.4

Exploring the Limits of Spatial Generalization Ability in Deep Learning Models for Hydrology 

Benedikt Heudorfer, Hoshin Gupta, and Ralf Loritz

State-of-the-art deep learning models for streamflow prediction (so-called Entity-Aware models, EA) integrate information about physical catchment properties (static features) with climate forcing data (dynamic features) from multiple catchments simultaneously. However, recent studies challenge the notion that this approach truly leverages generalization ability. We explore this issue by conducting experiments running Long-Short Term Memory (LSTM) networks across multiple temporal and spatial in-sample and out-of-sample setups using the CAMELS-US dataset. We compare LSTMs equipped with static features with ablated variants lacking these features. Our findings reveal that the superior performance of EA models is primarily driven by meteorological data, with negligible contributions by static features, particularly in spatial out-of-sample tests. We conclude that EA models cannot generalize to new locations based on provided physical catchment properties. This suggests that current methods of encoding static feature information in our models may need improvement, and that the quality of static features in the hydrologic domain might be limited. We contextualize our results with observations made in the broader deep learning field, which increasingly grapples with the challenges of (lacking) generalization ability in state-of-the-art deep learning models.

How to cite: Heudorfer, B., Gupta, H., and Loritz, R.: Exploring the Limits of Spatial Generalization Ability in Deep Learning Models for Hydrology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16414, https://doi.org/10.5194/egusphere-egu25-16414, 2025.

EGU25-19190 | ECS | Orals | HS3.4

Application of Attention-Based Graph Neural Networks for Spatial Distribution Prediction of Streamflow 

Xian Wang, Xuanze Zhang, and Yongqiang Zhang

Accurate streamflow estimation is crucial for effective water resource management and flood forecasting. However, physics-based hydrological models fail to respond promptly to rapid hydrological events due to lack efficiency in model calibration and computing time for large-scale catchment , while existing deep learning models tend to neglect the physical processes of runoff transfer, failing to account for the spatial and temporal dependencies inherent in runoff dynamics. In this study, we propose a topological process-based model that integrates Graph Attention Networks (GAT) to capture the spatial topology of runoff transfer and Long Short-Term Memory (LSTM) networks to simulate the temporal transfer between upstream and downstream runoff. The model was applied to the Yangtze River Basin which is the largest river basin in China to predict streamflow at 10 km spatial resolution. Validation results show that our model achieves a median Nash-Sutcliffe Efficiency (NSE) value of 0.783 at secondary outlet stations across the basin and effectively simulates the streamflow peak due to flooding. Additionally, the model is capable of simulating the spatial distribution of daily streamflow for an entire year within 10 seconds, providing a significant computational speedup compared to physical process-based river confluence models. This work represents a step towards more efficient and responsive prediction of extreme hydrological events using deep learning model.

How to cite: Wang, X., Zhang, X., and Zhang, Y.: Application of Attention-Based Graph Neural Networks for Spatial Distribution Prediction of Streamflow, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19190, https://doi.org/10.5194/egusphere-egu25-19190, 2025.

EGU25-19716 | ECS | Orals | HS3.4

Leveraging Machine Learning to Uncover and Interpret Relevant Catchment Features  

Alberto Bassi and Carlo Albert

Recent advances in catchment hydrology [Kratzert et al., 2019–2021] demonstrate the superiority of LSTMs over traditional conceptual models for streamflow prediction in large-sample datasets. LSTMs achieve better streamflow accuracies by leveraging information from diverse hydrological behaviors. These models are enriched with static catchment attributes, which, when combined with meteorological drivers, play a critical role in streamflow formation. Augmenting LSTMs with these attributes further enhances their performance compared to vanilla LSTMs, underscoring the importance of these attributes for accurate streamflow predictions. Building on this, a recent study [Bassi et al., 2024] employed a conditional autoencoder to reveal that most of the relevant catchment information for streamflow prediction can be distilled into two features, with a third feature being beneficial for particularly challenging catchments. In this work, we directly derive a minimal set of catchment features from known attributes by passing them through the encoder and subsequently comparing streamflow predictions against state-of-the-art benchmarks [Kratzert et al., 2021]. Our findings indicate that while the intrinsic dimension of 26 commonly used attributes is four, only two features suffice for accurate streamflow prediction. This aligns closely with the findings of Bassi et al. (2024), suggesting that nearly all relevant information for streamflow prediction is encapsulated in known attributes. Finally, we provide an interpretation of these two machine-learning-derived features using information theory techniques.

How to cite: Bassi, A. and Albert, C.: Leveraging Machine Learning to Uncover and Interpret Relevant Catchment Features , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19716, https://doi.org/10.5194/egusphere-egu25-19716, 2025.

EGU25-453 | ECS | Orals | SM2.2

Enhancing seismic monitoring with Virtual Seismic Arrays: Application of a deep learning framework 

Jana Klinge, Sven Schippkus, and Céline Hadziioannou

Seismic arrays are essential for collecting and analyzing seismic data, significantly enhancing our understanding of geophysical processes such as the localization of seismic sources. We introduce the concept of Virtual Seismic Arrays, where the array recordings are predicted from a single reference station, removing the need for continuous deployment of all array stations. This work builds on the research by Klinge et al. (2025), which introduced a Deep Learning approach using encoder-decoder networks to learn and predict transfer properties between two seismic stations. By training the algorithm on data of the Gräfenberg array in the secondary microseism frequency band, we develop models that effectively capture the transfer characteristics between a chosen reference station and each of the other stations within the array. To evaluate how well the models represent the underlying wave propagation, we use beamforming and apply it to both the original data from all stations and the corresponding predictions generated by the models. We assess two scenarios: one where the dominant backazimuths and slownesses are consistent with the training dataset, and another where the models are applied to data from different conditions. Our results show strong agreement between the predicted and original beamforming results, demonstrating the potential of Virtual Seismic Arrays for future application.

How to cite: Klinge, J., Schippkus, S., and Hadziioannou, C.: Enhancing seismic monitoring with Virtual Seismic Arrays: Application of a deep learning framework, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-453, https://doi.org/10.5194/egusphere-egu25-453, 2025.

EGU25-739 | ECS | Posters on site | SM2.2

Optimizing neural network architectures and using clustering to detect seismic events in noisy ocean bottom seismometer data 

Alicia Ximena Cortés Barajas, Marco Calò, Erik Molino Minero Re, and Francesca Di Luccio

Detecting seismic events is essential for monitoring tectonic and volcanic activity, especially in marine environments where noise makes analysis particularly challenging. This study introduces a method that combines Evolutionary Neural Architecture Search (ENAS) with the third generation of the Non-dominated Sorting Genetic Algorithm (NSGA-III) to design and optimize neural networks for seismic event detection using Ocean Bottom Seismometers (OBS) data.

In this work we developed a methodology to analyze heavily noisy data recorded by the TYDE OBS experiment in the southern Tyrrhenian Sea, Italy. In 2000, 14 seismic stations were deployed on the seafloor and around the Aeolian Islands recording data for about 6 months. Stations consisted of wide-band Ocean Bottom Seismometers (OBS) and Hydrophones (OBH).

The preprocessing pipeline includes feature extraction with Discrete Wavelet Transform (DWT) and dimensionality reduction using Principal Component Analysis (PCA), which reduces over 6000 coefficients to just 55 while preserving 95% of the variance. Applied to 90-second overlapping windows, this approach has achieved strong results, with F1 scores exceeding 90% in balanced noisy datasets.

Building on these results, this study explores unsupervised clustering to group similar seismic events and identify possible false positives through anomaly detection. By using adaptive clustering methods that determine the optimal number of clusters based on the data, this approach aims to enhance reliability while providing additional insights into the detected seismic events.

This work highlights the potential of automated tools to complement traditional seismic monitoring methods, balancing accuracy and model complexity while improving efficiency in noise-heavy environments.

How to cite: Cortés Barajas, A. X., Calò, M., Molino Minero Re, E., and Di Luccio, F.: Optimizing neural network architectures and using clustering to detect seismic events in noisy ocean bottom seismometer data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-739, https://doi.org/10.5194/egusphere-egu25-739, 2025.

EGU25-2833 | ECS | Orals | SM2.2

Unsupervised exploration of seismic activity at Mount Fuji, Japan 

Adèle Doucet, Léonard Seydoux, Nobuaki Fuji, Yosuke Aoki, and Jean-Philippe Métaxian

Mount Fuji volcano, located 100~km away from Tokyo, directly threatens over 30~million people. It last erupted in 1707, and has remained dormant since then. Seismicity --and particularly Low-Frequency Earthquakes (LFE)-- is to now the primary indicator of processes occurring beneath the volcano and is usually linked to fluid movement. Yet, these signals are usually manually picked and classified as such, without the ability to formally define them for automatic detection systems. 

Our goal is to develop an automatic method to detect and classify LFEs, among other seismic events at Mount Fuji using the continuous seismic records from 2008 at 11 stations. First, we use the CovSeisNet software to detect events by analyzing the wavefield coherence, derived from the network covariance matrix width. Over one year of continuous data, the wavefield coherence shows distinct patterns that correspond to various event types, including LFEs and tectonic earthquakes. To enable interpretation, we apply a manifold learning algorithm (UMAP) to reduce the dimension of the coherence patterns into two dimensions to ease the interpretation. We name this low-dimensional representation a "coherence atlas" where each point represents a time window of seismic data, grouped by similarity. This automatic approach enables not only the detection but also the classification of seismic events, as compared with the Japan Meteorological Agency catalog. Moreover, the atlas helps identify previously unrecorded events and facilitates the definition of new event classes. By autonomously mapping and classifying seismic activity beneath Mount Fuji, this method offers unprecedented insights into its activity and allows us to detect new events that had been hidden in the manually prepared catalog.

How to cite: Doucet, A., Seydoux, L., Fuji, N., Aoki, Y., and Métaxian, J.-P.: Unsupervised exploration of seismic activity at Mount Fuji, Japan, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2833, https://doi.org/10.5194/egusphere-egu25-2833, 2025.

EGU25-2836 | ECS | Orals | SM2.2

PEGSGraph: A Graph Neural Network for Fast Earthquake Characterization Based on Prompt ElastoGravity Signals 

Céline Hourcade, Kévin Juhel, and Quentin Bletery

State-of-the-art earthquake early warning systems use the early records of seismic waves to estimate the magnitude and location of the seismic source before the shaking and the tsunami strike. Because of the inherent properties of early seismic records, those systems systematically underestimate the magnitude of large events, which results in catastrophic underestimation of the subsequent tsunamis. Prompt elastogravity signals (PEGS) are low-amplitude, light-speed signals emitted by earthquakes, which are highly sensitive to both their magnitude and focal mechanism. Detected before traditional seismic waves, PEGS have the potential to produce unsaturated magnitude estimates faster than state-of-the-art systems. Accurate instantaneous tracking of large earthquake magnitude using PEGS has been proven possible through the use of a Convolutional Neural Network (CNN). However, the CNN architecture is sub-optimal as it does not allow to capture the geometry of the problem. To address this limitation, we design PEGSGraph, a novel deep learning model relying on a Graph Neural Network (GNN) architecture.
PEGSGraph accurately estimates the magnitude of synthetic earthquakes down to Mw~7.6-7.7 and determines their focal mechanisms (thrust, strike-slip or normal faulting) within 70 seconds of the event's onset, offering crucial information for predicting potential tsunami wave amplitudes. Our comparative analysis on Alaska and Western Canada data shows that PEGSGraph outperforms PEGSNet, providing more reliable rapid magnitude estimates and enhancing tsunami warning reliability.

How to cite: Hourcade, C., Juhel, K., and Bletery, Q.: PEGSGraph: A Graph Neural Network for Fast Earthquake Characterization Based on Prompt ElastoGravity Signals, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2836, https://doi.org/10.5194/egusphere-egu25-2836, 2025.

The mechanics of slow frictional creep in landslides remains debated and only a few detailed seismic studies have been conducted in landslide-prone areas. To illuminate basal slip processes for a slow moving landslide, we deployed a dense 80 node seismic array at Oak Ridge Earthflow in California’s Diablo Range for one to two months during the rainy seasons of 2023 and 2024, both winters where decimeters of landslide displacement occurred at Oak Ridge. Simultaneously, GNSS receivers, strain meters, and piezometers were deployed at the same site. During our deployments, various types of very small signals were recorded by the seismometers. These events were local, detected only by nearby stations sited within about 100 m of each other. The cause of these events remains unclear, whether due to shear slip at the base of the earthflow or other sources, such as water movement or animal activity. To investigate the cause of these signals, and evaluate the role of stick-slip motion and shear localization, we automatically detected the events and analyzed their spatiotemporal distribution. We used quakephase (Shi et al., 2024) to identify the phases of the very small signals. The primary challenge with automatic picking in our dataset is the long processing time due to high sampling rates. To address this issue, we applied array signal processing, covseisnet (Seydoux et al., 2016), to extract signal candidates based on the coherence of dominant frequencies across the seismic array, followed by automatic picking. This approach successfully and efficiently identified specific signals we believe are associated with earthflow motion. These signals are not continuously observed but concentrate within specific time periods. We focused on events in these time periods, utilizing scattering networks and matched-filter techniques for more detailed classification. By combining our results with other temporal data, such as pore fluid pressure, precipitation, temperature, and displacement, we will discuss the causes of these signals to better understand the mechanism of the earthflow motion.

How to cite: Iwasaki, Y., Schwartz, S., and Finnegan, N.: Classification of Small Seismic Signals Associated with the Oak Ridge Earthflow in California Using a Combination of Machine Learning and Array Signal Processing, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5384, https://doi.org/10.5194/egusphere-egu25-5384, 2025.

EGU25-5429 | ECS | Posters on site | SM2.2

Deep-learning-based dual model with an iterative prediction process for the improvement of missing well-log predictions  

Jaesung Park, Jina Jeong, and Mbarki Sinda

Traditional well-log analysis often involves incomplete datasets, which reduces the accuracy of petrophysical assessments. This study thus introduces an innovative dual-model approach that integrates a conditional variational autoencoder (CVAE) with a long short-term memory (LSTM) to predict missing shear-slowness (DTS) data and other well-log data. Utilizing well-logs and the corresponding lithological sequence from the Volve oil field in the North Sea, the proposed model demonstrates excellent prediction capabilities when facing multiple types of missing well-logs. Our findings reveal that the CVAE-LSTM model not only enhances DTS prediction accuracy but also adapts to the inherent variability of geological formations. It outperforms traditional autoencoder and standalone LSTM models across a range of metrics, including correlation coefficients, the root mean squared error, and Kolmogorov–Smirnov statistics, validating the predictive accuracy of the proposed model and the alignment of the statistical distributions for predicted and actual logs. The robustness of the proposed model is further highlighted by its ability to maintain its high performance despite the absence of key well-log data such as compressional slowness and the neutron porosity index. This study demonstrates the effectiveness of advanced machine-learning techniques in overcoming the limitations associated with incomplete well-log datasets.

How to cite: Park, J., Jeong, J., and Sinda, M.: Deep-learning-based dual model with an iterative prediction process for the improvement of missing well-log predictions , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5429, https://doi.org/10.5194/egusphere-egu25-5429, 2025.

EGU25-6412 | ECS | Posters on site | SM2.2

Enhancing seismicity detection and characterization in the Val d’Agri region: insights into tectonic and induced processes using Deep Learning techniques 

Elisa Caredda, Andrea Morelli, Maddalena Errico, Giampaolo Zerbinato, Marius Paul Isken, and Simone Cesca

Monitoring microseismicity is fundamental to advancing our understanding of fault mechanics under natural and anthropogenic influences. Recent advancements in seismological methodologies, particularly those employing deep learning techniques, have significantly improved the detection of weak earthquakes while preserving high levels of precision and reliability.

This study aims to enhance the detection and characterization of seismicity in the Val d’Agri region (Southern Italy) by implementing advanced deep learning-based methodologies, focusing on understanding the tectonic and anthropogenic influences driving seismic activity. The Val d’Agri region is a tectonically active area of considerable scientific and industrial relevance, hosting Europe’s largest onshore oil reservoir and an artificial lake. By employing state-of-the-art deep learning and full waveform earthquake detection methods we identified and located seismic events over a three-year period, achieving a twofold increase in detected events compared to the manually revised bulletin, with a recall rate of ~95%.

Spatial and temporal analyses, based on a density-based clustering approach, revealed distinct seismic clusters. The seismicity is mostly concentrated along the Monti della Maddalena fault system in the southwestern region, characterized by shallow earthquakes (5–7 km depth), while the northeastern and northwestern areas exhibit sparser and deeper activity (15–20 km depth). High-resolution event localization illuminated fault geometries and spatial distributions with high detail. Additionally, our dataset highlights a temporal correlation between seismicity rates and the filling and emptying phases of the Pertusillo artificial reservoir.

Our findings underscore the utility of automated workflows in improving seismic monitoring and fault characterization, providing critical insights into tectonic processes and reservoir-induced seismicity.

How to cite: Caredda, E., Morelli, A., Errico, M., Zerbinato, G., Isken, M. P., and Cesca, S.: Enhancing seismicity detection and characterization in the Val d’Agri region: insights into tectonic and induced processes using Deep Learning techniques, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6412, https://doi.org/10.5194/egusphere-egu25-6412, 2025.

EGU25-6730 | Posters on site | SM2.2

Readapting PhaseNet to Laboratory Earthquakes: AEsNet, a Robust Acoustic Emission Picker Illuminating Seismic Signatures of Different Fault Gouge Materials 

Marco Maria Scuderi, Giulio Poggiali, Federico Pignalberi, and Giacomo Mastella

Laboratory acoustic emissions (AEs), resembling small-scale earthquakes, provide vital insights into frictional instability mechanics. Recent advancements in acoustic monitoring technology allow for the rapid collection of thousands of AE waveforms within minutes, highlighting the critical need for efficient detection and analysis methods. This study presents a deep learning model designed to automatically detect AEs in laboratory shear experiments.

Our dataset comprises approximately 30,000 manually identified AE waveforms obtained under varying experimental boundary conditions using two fault gouge materials: Min-U-Sil quartz gouge and glass beads. We modified the PhaseNet model, originally designed for detecting seismic phases in natural earthquakes, by optimizing its architecture and training process to develop AEsNet—an advanced AE detection model that consistently outperforms existing picking methods for Min-U-Sil quartz gouge and glass beads.

To assess the model's generalizability across different boundary conditions and materials, we employed transfer learning, examining performance relative to training dataset size and material diversity. Results indicate that while model performance remains consistent across varying boundary conditions, it is notably influenced by the specific material type due to distinct frequency characteristics inherent to each material. This sensitivity stems from the distinct frequency characteristics of AEs, reflecting the microphysical processes unique to each granular material. Consequently, models trained on one material do not transfer effectively to another.

However, rapid fine-tuning of AEsNet substantially improves its performance, outperforming a similarly fine-tuned PhaseNet model pre-trained on natural earthquakes. This highlights the necessity of tailoring models to the specific features of laboratory-generated AEs, aligning with observations in transfer learning applications for natural seismicity.

In summary, our deep learning approach effectively enhances AE detection across diverse laboratory settings, enabling the creation of reliable AE catalogs that deepen our understanding of fault mechanics. This advancement facilitates the development of reliable AE catalogs, significantly contributing to the understanding of fault mechanics in controlled experimental environments.

How to cite: Scuderi, M. M., Poggiali, G., Pignalberi, F., and Mastella, G.: Readapting PhaseNet to Laboratory Earthquakes: AEsNet, a Robust Acoustic Emission Picker Illuminating Seismic Signatures of Different Fault Gouge Materials, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6730, https://doi.org/10.5194/egusphere-egu25-6730, 2025.

EGU25-7736 | ECS | Posters on site | SM2.2

Toward a Multi-Station Deep Learning Framework for Enhanced Earthquake Early Warning 

Jorge Antonio Puente Huerta, Christian Sippl, and Vaclav Kuna

Earthquake Early Warning (EEW) systems are vital for providing timely alerts in seismically active regions, potentially reducing damage and saving lives. However, achieving both rapid and reliable alerts remains a significant challenge. Recent advances in deep learning (DL) and established workflows for picking, associating, and locating events offer complementary paths to improved performance. In this work, we propose to investigate a multi-station deep learning framework that can be integrated with existing event-location pipelines or used directly to estimate ground shaking (e.g., peak ground acceleration, PGA). By fusing raw seismic waveforms with station metadata (e.g., location, sensor characteristics) in an end-to-end manner, the approach aims to capture both local site conditions and regional propagation effects. As an initial step, we will establish baseline performance using simpler neural networks (e.g., CNNs, LSTMs), then expand to more advanced models to evaluate potential gains in accuracy and speed. Preliminary findings indicate that aggregating real-time signals from multiple stations can outperform single-station methods in both alert timing and predictive reliability. Ultimately, our goal is to develop an adaptable, data-driven EEW pipeline that accommodates either direct shaking forecasts or event-based parameter estimation, enabling seamless integration into larger-scale monitoring networks and enhancing the timeliness of earthquake alerts.

How to cite: Puente Huerta, J. A., Sippl, C., and Kuna, V.: Toward a Multi-Station Deep Learning Framework for Enhanced Earthquake Early Warning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7736, https://doi.org/10.5194/egusphere-egu25-7736, 2025.

EGU25-9276 | Orals | SM2.2

Uncovering environmental and other exotic seismic sources with machine learning 

Clément Hibert, Joachim Rimpot, Camille Huynh, Charlotte Groult, Jean-Philippe Malet, Germain Forestier, Jonathan Weber, Camille Jestin, Vincent Lanticq, Floriane Provost, Antoine Turquet, and Tord Stangeland

Seismology, beyond the study of earthquakes, has become an indispensable tool for understanding environmental changes, offering unique insights into a wide range of phenomena and natural risks, from slope instabilities to glacial dynamics and hydrological hazards. However, the sheer volume and complexity of modern seismic datasets, amplified by the emergence of dense seismic networks and technologies such as Distributed Acoustic Sensing (DAS), pose significant challenges. Recent advances in Artificial Intelligence (AI) and Machine Learning (ML) have revolutionized our ability to analyze these datasets, enabling a deeper exploration of seismic data to find rare and exotic environmental seismic sources. 

Supervised learning approaches have been successfully applied to create large-scale instrumental catalogs of landslides and other environmental processes, at different  spatio-temporal scales, from short-term datasets recorded on dense local seismic stations networks, to chronicles spanning decades on seismic networks covering whole regions of the world (Alaska, Alps, Greenland). These techniques achieve high detection rates and robust classification of seismic events, even for low-magnitude or rare signals that traditional methods might overlook. Supervised learning approaches also allow us to advance our capability to estimate physical properties from seismic waves, such as the use of machine  learning to infer mass and kinematics of slope instabilities, which provide critical inputs for understanding the dynamics of these events and their associated hazards. These methodologies not only allow us to document environmental processes more exhaustively but also open up possibilities for studying poorly understood or previously undetectable seismic sources. Going beyond supervised learning, we have developed workflows based on self-supervised and unsupervised approaches to analyze continuous seismic data, uncovering unexpected patterns and revealing hidden environmental seismic sources recorded by dense seismic stations networks. Distributed Acoustic Sensing represents another frontier, turning fiber optic cables into dense seismic networks. By combining DAS with innovative AI-driven methods, we have demonstrated the potential to detect and classify low-magnitude earthquakes and anthropogenic sources, even in noisy environments, paving the way for real-time seismic monitoring on unprecedented scales.

By applying these AI-driven approaches, we are enhancing the field of environmental and exotic sources seismology, improving our ability to analyze vast seismic archives, and offering new tools to monitor, understand, and mitigate geohazards in a changing environment. This talk will highlight the latest methodological advances and showcase how they are applied to various geological and environmental contexts, from landslides, avalanches and glaciers in the Alps to fiber optic networks at different scales, underscoring the far-reaching implications of AI for seismological sources identification.

How to cite: Hibert, C., Rimpot, J., Huynh, C., Groult, C., Malet, J.-P., Forestier, G., Weber, J., Jestin, C., Lanticq, V., Provost, F., Turquet, A., and Stangeland, T.: Uncovering environmental and other exotic seismic sources with machine learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9276, https://doi.org/10.5194/egusphere-egu25-9276, 2025.

EGU25-9795 | ECS | Posters on site | SM2.2

Explainable artificial intelligence for short-term data-driven aftershock forecasts  

Foteini Dervisi, Margarita Segou, Brian Baptie, Piero Poli, Ian Main, and Andrew Curtis

The catastrophic nature of earthquakes drives the need for understanding seismic events, as well as for providing forecasts of when these are likely to occur. Due to the clustering nature of earthquakes, large magnitude events often trigger aftershocks that occur close to the mainshock in both space and time. In this study, we use a convolutional neural network to develop a data-driven spatiotemporal model to forecast next-day seismicity in an attempt to provide information that can contribute to answering one of the most pressing questions: whether a larger magnitude earthquake is to be expected after an intermediate magnitude event. We design our test to estimate expected seismicity within one day after earthquakes of magnitude four and above. We assemble a comprehensive dataset of earthquake catalogues from diverse tectonic regions to achieve a representative sample of input data and use it to create weekly spatiotemporal sequences of seismicity consisting of daily maps. Leveraging the predictive power of deep learning, our model uncovers complex patterns within this large dataset to produce next-day expected seismicity rate and magnitude forecasts in regions of interest. We use gradient-weighted class activation mapping (Grad-CAM) to provide visual explanations of the produced forecasts. We evaluate the performance of our forecasting model using data science and earthquake forecasting metrics and compare against persistence, which assumes no change between consecutive days, echoing typical experimental setups of forecasting models. Furthermore, we use a time series forecasting foundation model to generate next-day aftershock forecasts on the same dataset and compare these results against those produced by the convolutional neural network. We find that deep learning approaches are a promising solution for producing short-term aftershock forecasts, providing valuable insights that can contribute to better earthquake preparedness and response and be integrated with disciplinary statistics and physics-based forecasts.

How to cite: Dervisi, F., Segou, M., Baptie, B., Poli, P., Main, I., and Curtis, A.: Explainable artificial intelligence for short-term data-driven aftershock forecasts , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9795, https://doi.org/10.5194/egusphere-egu25-9795, 2025.

EGU25-10208 | Posters on site | SM2.2

The development of AI-based earthquake early warning system in China 

Jindong Song, Yefei Ren, Hongwei Wang, and Ruizhi Wen

China is a country prone to earthquakes. Earthquake early warning (EEW) is one of the important means to mitigate earthquake disasters. To mitigate the damage caused by destructive earthquakes, we are currently developing an artificial intelligence (AI)-based EEW system in China. By utilizing emerging technologies such as artificial intelligence and big data analysis, we have developed a complete set of methods for continuous measurement of EEW parameters based on AI. This AI-based method achieves the research goal of improving the accuracy and timeliness of EEW parameters measurement in the entire workflow of EEW system, including waveform interference elimination, earthquake event recognition, P-wave picking, magnitude estimation, seismic damage zone prediction, and so on. Presently, the offline testing results of this AI-based method on earthquake data in the Sichuan-Yunnan region of China show that the AI-based magnitude estimation reaches ±1 magnitude estimation error 4 s earlier than the existing EEW system. Meanwhile, some modules such as AI-based magnitude estimation and waveform interference elimination have been running online in the Fujian Earthquake Agency in China. In the future, China is expected to establish and improve an AI-based EEW system, and further reduce the casualties and economic losses caused by earthquakes through AI-based EEW system.

How to cite: Song, J., Ren, Y., Wang, H., and Wen, R.: The development of AI-based earthquake early warning system in China, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10208, https://doi.org/10.5194/egusphere-egu25-10208, 2025.

EGU25-10901 | Posters on site | SM2.2

Machine-learning-based operational tsunami warning from light-speed elasto-gravity signals 

Quentin Bletery, Gabriela Arias, Kévin Juhel, Céline Hourcade, Andrea Licciardi, Adolfo Inza, Martin Vallée, and Jean-Paul Ampuero

Earthquake early warning (EEW) systems implemented worldwide use early seismic records of P waves to rapidly detect, locate, and estimate the magnitude (Mw) of potentially damaging earthquakes. These systems are well known to saturate for large magnitude events, which results in dramatic underestimation of the subsequent tsunamis. Alternative approaches based on different signals have been proposed to rapidly estimate the magnitude of large events, but these approaches are much slower (taking 10 to 20 minutes for first warning). Prompt elasto-gravity signals (PEGS) are light-speed gravitational perturbations induced by large earthquakes that can be recorded by broadband seismometers. They have tremendous potential for early warning but their extremely small amplitudes (on the order of 1 nm/s2) have challenged their possible operational use. We designed a deep learning approach to rapidly estimate the magnitude of large earthquakes based on PEGS. We applied this approach to the seismic networks operating in Japan, Chile, Alaska and Peru. We will present the performances obtained in these different contexts. In Alaska, the approach has proven capable to reliably estimate the magnitude of Mw ≥ 7.6 earthquakes (without saturation) in less than 2 minutes, outperforming state-of-the-art tsunami early warning algorithms. Motivated by these performances, we initiated a first implementation of an operational tsunami warning system based on PEGS in Peru. We will present the simulated real-time performance of this system. 

How to cite: Bletery, Q., Arias, G., Juhel, K., Hourcade, C., Licciardi, A., Inza, A., Vallée, M., and Ampuero, J.-P.: Machine-learning-based operational tsunami warning from light-speed elasto-gravity signals, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10901, https://doi.org/10.5194/egusphere-egu25-10901, 2025.

EGU25-11335 | ECS | Orals | SM2.2

Unsupervised Clustering and Pattern Identification from Continuous Seismic and Strainmeter Data in Tectonic and Volcanic Settings 

Zahra Zali, Patricia Martínez-Garzón, Grzegorz Kwiatek, Gregory Beroza, Fabrice Cotton, and Marco Bohnhoff

We developed a deep learning model for automatic dimensionality reduction and feature extraction from time series. The model employs an encoder-decoder architecture with skip connections, enabling efficient compression and reconstruction of input data while preserving essential features. These features are used for unsupervised clustering enabling anomaly detection, and pattern recognition.

We initially developed the model to analyze seismic data from the 2021 Geldingadalir volcanic eruption in Iceland, successfully identifying a weak yet important pre-eruptive tremor that commenced three days before the eruption. Advancing the architecture with additional layers and skip connections allowed for highly accurate input reconstruction. The latter version, named AutoencoderZ, demonstrated its ability to process different data types. We applied AutoencoderZ to investigate low-frequency patterns preceding the 2023 MW 7.8 Kahramanmaraş Earthquake in Türkiye. The model identified tremor-like episodes linked to anthropogenic activities at cement plants near the earthquake’s epicenter. Additionally, we applied AutoencoderZ to strainmeter data from the Sea of Marmara, achieving accurate reconstructions and enabling the detection of distinct tectonic-related signals.

This study highlights AutoencoderZ’s potential as a powerful tool for uncovering patterns in continuous geophysical data, providing valuable insights for monitoring and interpreting seismic and strainmeter signals.

How to cite: Zali, Z., Martínez-Garzón, P., Kwiatek, G., Beroza, G., Cotton, F., and Bohnhoff, M.: Unsupervised Clustering and Pattern Identification from Continuous Seismic and Strainmeter Data in Tectonic and Volcanic Settings, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11335, https://doi.org/10.5194/egusphere-egu25-11335, 2025.

EGU25-11459 | ECS | Posters on site | SM2.2

Automating Seismic Event Discrimination: A Comparative Study of Convolutional Neural Networks and Vision Transformers 

Valentin Kasburg, Marcel van Laaten, Markus Zehner, Jozef Müller, and Nina Kukowski

To date, the discrimination of seismic events recorded in seismic networks is often performed manually by experts, classifying events into categories such as earthquakes, quarry blasts, or mining events. While recent studies have shown that deep learning algorithms, particularly Convolutional Neural Networks (CNNs), can efficiently and accurately distinguish between different types of seismic events, their application for automated seismic event discrimination remains limited. This limitation arises from several factors, including the absence of globally applicable models that maintain high precision for local seismic networks, the scarcity of data required for fine-tuning Deep Learning (DL) models, and the lack of interpretability in the decision-making processes of these black-box models.

In this contribution, we explore the use of Vision Transformers (ViTs) as a novel approach for automating seismic event discrimination. To assess their potential for accuracy and explainability, we applied CNNs and ViTs to classify seismic events such as earthquakes, quarry blasts, and mining events. For this purpose, we pretrained the models on openly available seismic event data from Utah and Northern California and then fine-tuned and tested them on data from the Seismic Network of the Ruhr-University Bochum (RuhrNet) and the Thuringian Seismic Network (TSN).

Our findings reveal that ViTs can analyze the entire spectrogram of a seismic event in a coherent manner, offering superior generalizability in pattern recognition compared to CNNs. In addition to achieving high discrimination accuracy, the attention weights of ViTs provide insights into the models’ decision-making process, offering a transparent and interpretable explanation of the underlying mechanisms driving its classifications.

How to cite: Kasburg, V., van Laaten, M., Zehner, M., Müller, J., and Kukowski, N.: Automating Seismic Event Discrimination: A Comparative Study of Convolutional Neural Networks and Vision Transformers, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11459, https://doi.org/10.5194/egusphere-egu25-11459, 2025.

EGU25-13238 | ECS | Orals | SM2.2

A Waveform-Based Graph Neural Network Approach for Microseismic Monitoring 

Matteo Bagagli, Francesco Grigoli, and Davide Bacciu

In this work, we introduce HEIMDALL, a grapH-based sEIsMic Detector And Locator specifically designed for microseismic applications. Building on recent progress in deep learning (DL), HEIMDALL employs spatiotemporal graph-neural networks to detect and locate seismic events in continuous waveforms. It simultaneously associates and provides preliminary locations by leveraging the output probability functions of the graph-neural network over a dense, three-dimensional grid (0.1 km spacing). By integrating detection and location within a single framework, HEIMDALL aims to address persistent challenges in microseismic data analysis, such as accurately associating wavefront arrivals and enabling consistent and robust event localization in complex geothermal regions. To train our models, we utilize synthetics generated using Green’s function available in the area, in combination with a small fine-tuning over a subset of real data. This approach allows us to achieve homogeneous coverage of the study area while addressing nuances that inevitably arise across synthetic and real domains.

Our evaluation focuses on data collected at the Hengill Geothermal Field in Iceland as part of the COSEIMIQ project (December 2018 to August 2021). Specifically, we analyzed one month of continuous seismic recordings from December 2018 and a brief sequence on February 3, 2019, which occurred in the middle of the geothermal plant. The dataset also features frequent burst sequences, providing an ideal testbed for advanced detection and location algorithms. By benchmarking HEIMDALL against multiple approaches, we reveal both the strengths and limitations inherent in our novel method and in more conventional workflows used in observational seismology.

Ultimately, we highlight the importance of continued innovation in ML-based workflows for induced seismicity monitoring at enhanced geothermal system (EGS) sites, where the capacity to detect and accurately locate a large number of microseismic events can be critical for operational safety and resource management.

How to cite: Bagagli, M., Grigoli, F., and Bacciu, D.: A Waveform-Based Graph Neural Network Approach for Microseismic Monitoring, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13238, https://doi.org/10.5194/egusphere-egu25-13238, 2025.

EGU25-13430 | ECS | Posters on site | SM2.2

Towards the robust Clustering of various cryogenic signal types using Seismic Array Information 

Antonia Kiel, Vera Schlindwein, and Conny Hammer

Ice mass loss in polar regions is a major contributor to sea level rise driven by climate change. To better predict ice mass loss due to calving and melting, it is essential to monitor ice dynamics by linking observed seismic signatures to physical processes such as meltwater infiltration into crevasses or crack formation caused by high tides at the grounding line. However, current knowledge about the distinct patterns of icequake types remains limited.

To address this gap, approximately 15 years of continuous seismic data from the Watzmann Array near the Neumayer Station in Antarctica are analyzed to automatically cluster seismic recordings. This analysis involves the automatic extraction of seismic events and the application of beamforming to each event. As a result, directional information is incorporated and the local noise is significantly reduced.

In the following, clustering methods, combined with techniques like dynamic time warping and feature extraction, are employed to categorize seismic events into distinct groups representing different icequake types. A key focus of this work is on leveraging dynamic time warping to cluster seismic waveforms directly, prioritizing the identification of physical properties inherent in the signals rather than relying solely on features extracted through machine learning. This approach ensures that the obtained clustering reflects the true underlying source processes rather than being limited to abstract feature representations.

In a follow-up study, these clusters can be related to environmental factors and directional information. Finally, with this we hope to shed some light on the hidden source processes of observed icequake types.

How to cite: Kiel, A., Schlindwein, V., and Hammer, C.: Towards the robust Clustering of various cryogenic signal types using Seismic Array Information, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13430, https://doi.org/10.5194/egusphere-egu25-13430, 2025.

Recent advances in machine learning (ML) have enabled significant progress in geoscience by capturing complex relationships and enhancing predictive skills. However, the success of many ML algorithms in data-rich settings does not seamlessly transfer to climate and atmospheric applications, where observational datasets are often limited. This underscores the need for methods that deliver high predictive accuracy under data-scarce conditions while retaining interpretability.

Here, we compare various ML approaches with the Nonlinear AutoRegressive Moving Average model with eXogenous inputs (NARMAX) in typical small-data climate applications, such as seasonal weather forecasting and Greenland Blocking Index (GBI) prediction. NARMAX, a transparent, interpretable, parsimonious and simulatable (TIPS) framework, demonstrates robust performance and avoids common pitfalls such as overfitting and unstable predictions when data are scarce. Notably, it achieves superior or competitive forecast accuracy for small or limited data conditions, underscoring its practical value in operational climate science. By adopting a sparse system identification approach, NARMAX yields model structures that readily reveal key predictors and their relative contributions, providing valuable physical and statistical insights into climate variability.

Our findings illustrate how NARMAX bridges the gap between purely data-driven modelling (focusing on prediction) and mechanistic modelling (focusing on physical insights), offering a clear pathway for refining model strategies and deepening our understanding of climate dynamics. We propose that NARMAX and similar methods play an inherently powerful role for both small and large data modelling problems and meanwhile serve as potent components to potentially improve the explainability of ML methods. By showcasing both interpretability and predictive efficacy, this work encourages the adoption of machine learning methods that best meet the needs for specific data modelling tasks in climate science and beyond.

How to cite: Sun, Y., Wei, H.-L., Hanna, E., and Luu, L.: Small Data for Big Tasks in Seasonal Weather Forecasting: A Balanced Perspective on Interpretability and Predictability of NARMAX and Machine Learning Methods, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13666, https://doi.org/10.5194/egusphere-egu25-13666, 2025.

Substorms are a magnetospheric phenomenon which causes high geomagnetic and auroral disturbances. It is widely accepted that substorm activity is controlled by solar wind conditions. It is, however, difficult to predict substorms deterministically because of the complex physical processes underlying substorm occurrences. We propose a framework for modelling time series of event occurrences controlled by external forcing. In this framework, occurrences of external-driven events are represented with a non-stationary Poisson process, and its intensity, which corresponds to the occurrence rates per unit time, is described with a simple machine learning model, the echo state network, which is fed with forcing variables. The echo state network is trained by maxmising the likelihood given the event time series data. 

We apply this approach for analysing time series of substorm onsets identified from Pi2 pulsations, which are irregular geomagnetic oscillations associated with substorm onsets. We train the echo state network to well describe the response of substorm activity to solar-wind conditions. We then examine the characteristics of the substorm activity by feeding synthetic solar-wind data into the echo state network. The results show what solar wind variables effectively contribute to the substorm occurrence. 

Our echo state network model is also useful for examining the statistical properties of the substorm occurrence rate. For example, we can evaluate what mainly controls the seasonal and UT variations of substorm activity. There are two explanations for the seasonal and UT variations. One explanation is that the seasonal and UT variations is controlled by the inner product between the solar-wind magnetic field and the Earth's dipole axis. The other is that the variations are due to the angle between the solar-wind flow and the Earth's dipole axis. Since these two effects are related with different input variables in our echo state network model, we can examine the contribution of each effect to the substorm occurrence frequency. The result shows that the seasonal and UT variations are mostly dependent on the angle between the solar-wind flow and the Earth's dipole axis. 

How to cite: Nakano, S., Kataoka, R., and Nose, M.: Modelling of time series of external-driven events with echo state network and its application to substorm activity analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14041, https://doi.org/10.5194/egusphere-egu25-14041, 2025.

EGU25-14779 | ECS | Posters on site | SM2.2

Deep Learning-Based Surface-Related Multiple Suppression in Shallow Arctic Seismic Data 

Hyeji Chae, Daeun Na, Seung-Goo Kang, and Wookeen Chung

Seismic data recorded in shallow water in the Arctic Ocean contain not only primary reflections but also surface-related multiples with strong amplitudes and short-period characteristics. These multiples generate false stratigraphic boundaries on stacked seismic sections, thereby reducing the accuracy of geological interpretation. Therefore, the attenuation of multiples is an essential step in seismic data processing for accurate geological interpretations. Recently, with the advancement of deep learning technology, research on suppressing surface-related multiples using deep learning networks (such as U-Net and stacked BiLSTM) has been actively proposed.

Firstly, surface-related multiple suppression algorithms using U-Net and stacked BiLSTM were applied to Arctic field data respectively. Each algorithm was designed to predict surface-related multiples by using input data that contained both primary reflections and surface-related multiples. Fractional Fourier transform (FrFT) and continuous wavelet transform (CWT), which represent time-series data in the time-frequency domain, were applied to synthetic data and used as input data feature for each network. Finally in order to suppress the surface-related multiples for seismic data in shallow depth Arctic Ocean, the proper methods (network architectures, input data feature) are suggested.

 

Acknowledgments

This research was supported by Korea Institute of Marine Science & Technology Promotion (KIMST) funded by the Ministry of Oceans and Fisheries, Korea (RS-2023-00259633).

 

How to cite: Chae, H., Na, D., Kang, S.-G., and Chung, W.: Deep Learning-Based Surface-Related Multiple Suppression in Shallow Arctic Seismic Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14779, https://doi.org/10.5194/egusphere-egu25-14779, 2025.

EGU25-15039 | ECS | Posters on site | SM2.2

Machine learning and feature extraction for detecting transient signals in GNSS time series 

Martín Sepúlveda, Marcos Moreno, and Matthew Miller

Advances in the processing of Global Navigation Satellite System (GNSS) positioning data and the increasing densification of geodetic networks have provided an unprecedented opportunity to detect and analyse transient deformation signals, including Slow Slip Events (SSE). These events, characterised by very slow rupture and durations of days to months, are often associated with areas of low coupling and sometimes show clear recurrence patterns. Despite their importance in subduction zones, reliable detection of SSEs remains an ongoing challenge. The sheer volume of GNSS data, combined with high noise levels and the subtle nature of these signals, requires efficient and robust methods capable of rapidly processing large datasets.

To overcome these challenges, we propose a method that relies on feature extraction techniques and machine learning to improve the detection and analysis of possible SSEs. Specifically, we use the TSFRESH algorithm to extract relevant features from GNSS time series, coupled with supervised machine learning classification techniques. Preliminary results of our current model, trained on synthetic data and validated through various performance tests, demonstrate high detection capabilities and accuracy. We further validated the model using a collection of GNSS time series from the Cascadia subduction zone with a single-station method scaled to the entire network, where the model showed satisfactory performance in detecting possible SSEs compared to similar work. Future efforts will focus on improving the robustness and generalisation of the model to new data, and refining methods for estimating the slip and duration of each possible SSE.

How to cite: Sepúlveda, M., Moreno, M., and Miller, M.: Machine learning and feature extraction for detecting transient signals in GNSS time series, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15039, https://doi.org/10.5194/egusphere-egu25-15039, 2025.

EGU25-15129 | ECS | Posters on site | SM2.2

Auto-Picking of First-Break Refractions in Arctic Ocean Seismic Data Using Stacked BiLSTM Networks 

Seoje Jeong, Hyeji Chae, Seung-Goo Kang, Sung-Ryul Shin, and Wookeen Chung

 In seismic exploration data from the Arctic Ocean, refractions are recorded earlier than direct waves due to the shallow depths and subsea permafrost with high velocity. These refraction signals could be utilized for estimating the velocity, thickness, and depth of the subsea permafrost. However, it is very challenging work to pick the accurate first arrivals of seismic data in the Arctic Ocean because of many factors such as ambient noise and etc. Therefore, identifying first-break refractions is crucial and can be performed by manual or automated picking methods. Various semi-automatic techniques have been developed to identify first-break refractions, but these methods are often sensitive to pulse variations and require parameter tuning. Recently, deep learning-based methods have also been explored, but their reliance on training data often results in inconsistent performance, making it essential to generate training data optimized for the target environment. 

This study presents a recurrent neural network-based algorithm optimized for Arctic Ocean environments to automatically identify first-break refractions. To effectively classify first-break refractions, a stacked bidirectional long short-term memory (BiLSTM) network was constructed to iteratively learn bidirectional long-term dependencies by utilizing the temporal patterns of time-series data. Additionally, the training data were generated by creating velocity models that reflect the subsurface properties of subsea permafrost, enabling the generation of first-break refraction label data. The proposed network demonstrated superior performance in identifying first-break refractions from noisy data, achieving over 95% accuracy in numerical experiments and field tests. Field data applications demonstrated that the proposed network achieves high accuracy in classifying first-break refractions, validating its robustness and adaptability.

 

 

Acknowledgments

This research was supported by Korea Institute of Marine Science & Technology Promotion (KIMST) funded by the Ministry of Oceans and Fisheries, Korea (RS-2023-00259633).

How to cite: Jeong, S., Chae, H., Kang, .-G., Shin, .-R., and Chung, W.: Auto-Picking of First-Break Refractions in Arctic Ocean Seismic Data Using Stacked BiLSTM Networks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15129, https://doi.org/10.5194/egusphere-egu25-15129, 2025.

EGU25-16157 | ECS | Posters on site | SM2.2

Real Time Estimation of Earthquake Location and Magnitude Using Large Language Models 

Aurora Bassani, Daniele Trappolini, Giulio Poggiali, Elisa Tinti, Fabio Galasso, Chris Maron, and Alberto Michelini

Estimation of earthquake parameters has always been a focus for seismologists. Efficient and rapid determination of earthquake location and magnitude is essential for mitigating the potential hazards associated with seismic shaking. Nowadays, Earthquake Early Warning Systems (EEWS) are implemented in most earthquake-prone areas, with the system varying according to the specific needs. Although methods for their estimation exist, many still lack a fast enough process, which is crucial for reducing the waiting time before issuing a warning.

Here, we propose a novel model to enhance multi-station EEWS using Large Language Models (LLM). We adopt a pre-trained LLM and fine-tune it on a customized version of INSTANCE (The Italian Seismic Dataset for Machine Learning), thus eliminating the need to develop and train a tailor-made architecture. The model uses stations with P-wave arrival times up to 5 s apart from the first recorded one, and, for each seismic trace, it exploits a very small time window around the P-wave arrival time (0.21 s), thus effectively reducing warning latency.

Comparative analysis against the automatic method employed by the Italian National Institute of Geophysics and Volcanology (INGV) demonstrates that our model achieves comparable performance in magnitude estimation and superior accuracy in epicenter, hypocenter and origin time prediction. For instance, the LLM-based model achieves average errors of 6.3 km, 11.1 km, and 1.1 s for epicenter, hypocenter, and origin time estimation, respectively, in contrast to 8.6 km, 15.0 km, and 1.8 s for the INGV automatic solution resulting in an average improvement of more than 26% for all parameters.

We study the validity of our model by assessing its ability using P- and S-waves to predict magnitude, and show that in this case study the S-waves are not strictly necessary for accurate predictions.

How to cite: Bassani, A., Trappolini, D., Poggiali, G., Tinti, E., Galasso, F., Maron, C., and Michelini, A.: Real Time Estimation of Earthquake Location and Magnitude Using Large Language Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16157, https://doi.org/10.5194/egusphere-egu25-16157, 2025.

EGU25-17400 | ECS | Orals | SM2.2

Deep learning to investigate post-seismic evolution of fault zone elastic properties 

Gabriele Paoletti, Daniele Trappolini, Elisa Tinti, Fabio Galasso, Cristiano Collettini, and Chris Marone

Fault zone properties evolve dynamically during the seismic cycle due to stress changes, microcracking, and wall rock damage. Understanding these changes is vital to gaining insights into earthquake preparation and post-seismic processes. The latter include fault healing, which refers to the recovery of mechanical and elastic properties in fault zones after seismic and aseismic fault slip. Despite its importance, detecting and characterizing fault healing through seismic signals remains a challenge due to the subtle nature of these changes.

In this study, we investigate the potential of deep learning techniques, specifically a 4-layer Convolutional Neural Network (CNN), to characterize post-seismic evolution by analyzing raw seismic waveforms recorded after the largest event (Mw 6.5,  30 October) of the 2016 Central Italy seismic sequence. These data provide a unique opportunity to examine fault zone dynamics. A key aspect of our approach is the hypothesis that ray paths traversing highly impacted areas of the fault zone contain richer information about its temporal evolution. To test this hypothesis, we examined seismic waves from two clusters — DHwS, located in the hanging wall beneath the hypocentral region, and C1, situated in the footwall. They represent contrasting ray trajectories as recorded on seismic stations MC2 and MMO1. Seismic waves recorded at MC2 pass through heavily damaged fault regions, which are likely to reveal evolving fault properties, whereas MMO1 predominantly captures paths that skirt or in the case of C1 completely miss these impacted areas, serving as a comparative baseline.

We assessed temporal variations in elastic properties using binary classification tests on normalized, raw seismic waveforms of events before and after a reference date. This date was arbitrarily selected within the temporal range of the analyzed seismicity and serves solely as a neutral point of comparison. Our hypothesis is that if the CNN can achieve good classification performance, it implies the presence of time-evolving properties in the fault zone, potentially linked to healing processes or other time-dependent factors.

To further validate these findings, we employed adversarial training, a technique designed to disentangle time-dependent effects from structural changes. By introducing controlled label noise into one cluster during training, we isolated the influence of confounding factors such as seasonal variations. Preliminary results suggest that adversarial training enhances the model's robustness and provides valuable insights into the time-evolving properties of the fault zone.

Deep learning offers significant potential for analyzing spatiotemporal changes in elastic properties and thus the evolution of fault zone properties over the seismic cycle. By detecting subtle temporal and structural changes, we hope to gain a deeper understanding of fault dynamics and post-seismic processes.

How to cite: Paoletti, G., Trappolini, D., Tinti, E., Galasso, F., Collettini, C., and Marone, C.: Deep learning to investigate post-seismic evolution of fault zone elastic properties, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17400, https://doi.org/10.5194/egusphere-egu25-17400, 2025.

EGU25-19236 | ECS | Orals | SM2.2

High Resolution Generative Waveform Modeling Using Denoising Diffusion 

Kadek Hendrawan Palgunadi, Andreas Bergmeister, Andrea Bosisio, Laura Ermert, Maria Koroni, Natanaël Perraudin, Simon Dirmeier, and Men-Andrin Meier

Reliable synthesis and prediction of seismic waveforms play an important role in evaluating seismic hazards and designing earthquake-resilient structures. However, current methods, such as ground motion models and physics-based simulations, are often limited in fully capturing the complexity of seismic wave propagation, at higher frequencies (>5 Hz). Some of these limitations can potentially be overcome through machine learning techniques. In earthquake engineering, machine learning models have been used for predicting peak ground accelerations and Fourier spectra responses. To model entire waveforms, extensive efforts to generate seismic waveforms have employed advanced machine learning techniques, such as generative models, with most previous approaches relying on generative adversarial networks (GANs). In contrast to these earlier models, this study presents an efficient and extensible generative framework to produce realistic high-frequency seismic waveforms, compared to GANs. Our approach encodes spectrograms of the waveform data into a lower-dimensional sub-manifold using an autoencoder, and a state-of-the-art diffusion model is subsequently trained to generate these latent embeddings. Conditioning is currently performed on key parameters: earthquake magnitude, recording distance, site conditions, and faulting style. The resulting generative model can synthesize waveforms with frequency content up to 50 Hz, from which several scalar ground motion statistics, such as peak ground motion amplitudes, spectral accelerations, or Arias intensity can be directly derived. We validate the quality of the generated waveforms using standard seismological benchmarks and performance metrics from image generation research. Our openly available model produces high-frequency waveforms that align with real data across a wide range of input parameters, including regions where observations are sparse, and accurately reproduces both median trends and variability of empirical ground motion statistics. Our generative waveform model can be potentially used to perform seismic hazard where broadband data are often required such as to train earthquake early warning model. Given the increasing number of generative waveform models, we emphasize that they should be openly accessible and included in community efforts for ground motion model evaluations.

How to cite: Palgunadi, K. H., Bergmeister, A., Bosisio, A., Ermert, L., Koroni, M., Perraudin, N., Dirmeier, S., and Meier, M.-A.: High Resolution Generative Waveform Modeling Using Denoising Diffusion, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19236, https://doi.org/10.5194/egusphere-egu25-19236, 2025.

EGU25-21456 | Orals | SM2.2

Prediction of measured earthquake waveforms from synthetic data: a two-step deep learning approach 

Alexander Bauer, Jan Walda, and Conny Hammer

Single-station waveforms of teleseismic earthquakes are highly complex, because they are a superposition of numerous phases corresponding to different wave types and propagation paths. Moreover, data recorded at single stations is contaminated by noise, which often has similar or larger amplitudes than the arrivals of teleseismic earthquakes, especially in densely-populated areas. For high precision research facilities, for example in the field of particle physics or gravity wave detection, a precise knowledge of the seismic wavefield generated by teleseismic earthquakes can be critical for the calibration of experiments. However, the density of seismological stations is often sparse, particularly in regions with low seismic hazard such as Northern Germany.

To overcome this limitation, we introduce a deep learning scheme for the prediction of very-low-frequency earthquake waveforms from synthetic data at arbitrary locations within the metropolitan area of Hamburg, Germany. For this aim, we propose to train a convolutional neural network (CNN) to predict the measured earthquake waveforms from their synthetic counterparts. While synthetic earthquake waveforms can be conveniently generated for arbitrary coordinates and moment tensors with Instaseis and the IRIS synthetics engine (Syngine), the amount of available measured waveforms is constrained by the availability of seismological stations and their installation date. In this work, we use measured data from a station in Bad Segeberg, north of Hamburg, which has been measuring continuously since 1996. During first experiments, we trained a CNN on data from earthquakes larger than M6.0 and obtained reasonable initial results. However, the number of such earthquakes is limited and the measured waveforms used as labels partly contained noise of considerable amplitude, which caused the neural network to predict unwanted noise.

In order to increase the amount of earthquakes in the training data and mitigate their contamination with noise, we propose a two-step approach. In the first step, we generate a large number of noise-free synthetic waveforms and contaminate them with artificially generated noise that has the same characteristics as the noise measured at the station in Bad Segeberg. With this dataset, we train a first CNN to denoise the synthetic earthquake waveforms. In the second step, we apply the trained neural network to the actual earthquake waveforms measured in Bad Segeberg to denoise them. We then train a second CNN to translate synthetic earthquake waveforms to the denoised measured ones. Results for earthquakes not part of the training data demonstrate that the second CNN provides convincing estimates of measured earthquake waveforms, not only for the station in Bad Segeberg, but also for stations in Hamburg. This can be seen as a first step towards a three-dimensional prediction of the earthquake wavefield without the need for densely-distributed stations.

How to cite: Bauer, A., Walda, J., and Hammer, C.: Prediction of measured earthquake waveforms from synthetic data: a two-step deep learning approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21456, https://doi.org/10.5194/egusphere-egu25-21456, 2025.

EGU25-469 | ECS | Posters on site | CR6.8

Advancing snow grain classification for snow micro-penetrometer signals using machine learning 

Jil Lehnert, Marie Hofmann, Julia Kaltenborn, Martin Schneebeli, and Christoph Mitterer

The layered nature of snow is a key characteristic of the seasonal alpine snowpack. In fact, snow stratigraphy influences all physical processes e.g., mechanical or thermal behavior. In order to describe these physical processes precisely, a profound and objective representation of the snow stratigraphy is paramount. The Snow-Micro Penetrometer (SMP) is a rod-driven snow penetrometer that provides resistance-force profiles across snow depth, offering an objective method to measure vertical snow stratigraphy. These submillimeter-scale profiles facilitate the derivation of a micro-mechanical snow model. These derivatives have the potential to initialize complex, physics-based snow cover models (e.g., SNOWPACK). While many parameters for snowpack simulations can be derived directly, determining grain type remains challenging due to the absence of a clear physical correlation. To address this, machine learning (ML) approaches have been investigated. However, prior ML models are limited in their number of snow grain type classes and datasets, which prevents the operational use of these models. Recently, Kaltenborn et al. introduced Snowdragon, a ML benchmark for automated classification and segmentation of SMP profiles. The current version of Snowdragon is trained on SMP profiles collected during the MOSAiC expedition and contains only specific non-standardized grain types typically observed for snow on Arctic sea ice. In this work, we re-trained the supervised models of the Snowdragon benchmark on Alpine snow. To enable the usage of Snowdragon for a broader community, we adapted the classification of grain types according to the international standard for seasonal snow. Our dataset comprises 52 manually labeled SMP profiles recorded in Alpine snow in Switzerland. Previously identified high-performing ML models were re-trained without additional hyperparameter tuning and subsequently evaluated. We found that the ML model Random Forest performed best but nevertheless had difficulties in recognizing faceted crystals, similar to the other models. Additionally, all models react sensitive to minor force changes in the SMP profiles, often leading to predictions of alternating micro-classes between two grain types. These preliminary results demonstrate the feasibility of this approach for grain type classification, but underscore the limitations posed by the small dataset size. Future work will focus on expanding the training dataset and developing a robust interface for operational use of the prediction output. This work marks a step toward more reliable and generalizable snow grain classification of SMP signals for operational use, like snowpack modeling and avalanche assessment.

How to cite: Lehnert, J., Hofmann, M., Kaltenborn, J., Schneebeli, M., and Mitterer, C.: Advancing snow grain classification for snow micro-penetrometer signals using machine learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-469, https://doi.org/10.5194/egusphere-egu25-469, 2025.

EGU25-4229 | ECS | Orals | CR6.8

Deep learning-based rock glacier mapping using Earth observation data 

Vanessa Streifeneder, Benjamin Aubrey Robson, Daniel Hölbling, Elena Nafieva, Zahra Dabiri, Emma Hauglin, and Lorena Abad

Rock glaciers are tongue-shaped complex landforms that indicate current or past permafrost conditions. They are commonly found in high-latitude and/or high-elevation environments and consist of poorly sorted angular debris and ice-rich sediments formed by gravity-driven creep. In the Austrian Alps, it is estimated that over 5700 rock glaciers exist (Kellerer-Pirklbauer et al., 2022). Knowing the location, extent and characteristics of rock glaciers is important for several reasons. These include estimating their hydrological importance as a water resource (e.g., for alpine huts) and assessing the geohazard potential because of the destabilisation of rock glaciers due to climate change. Unlike other cryosphere features, such as snow and glaciers, rock glaciers are spectrally inseparable from the surrounding terrain. This makes them difficult to automatically detect and delineate from Earth observation (EO) data. As a result, rock glaciers are usually mapped by labour-intensive, subjective manual interpretation of EO data. This often leads to inhomogeneous, incomplete, and inconsistent mapping. Therefore, there is a need for automated and efficient methods to map rock glaciers. This can be achieved by using globally applicable satellite data sets such as Sentinel-2.   

Modern machine learning methods, such as deep learning (DL), provide new opportunities to automate mapping tasks and address the challenges of detecting rock glaciers from EO data. However, research on DL-based rock glacier mapping remains limited, and there is no consensus on the best-suited parameters for this application. In addition, features with surface textures similar to rock glaciers, such as landslides, avalanche deposits, or fluvial deposits, may be misclassified by DL models. Hence, a thorough investigation of the DL model architectures and input data types is necessary to determine the most effective approach for mapping rock glaciers. In the project “ROGER - EO-based rock glacier mapping and characterisation”, we test different DL models (e.g. Unet, DeepLABV3) with different settings (backbones, input layers (including optical imagery and DEM-derived information)) to identify the most suitable model for rock glaciers delineation in Austria. We evaluate the performance, robustness, and reliability of the different DL models for automated EO-based mapping of rock glaciers in different study areas in Austria, and quantify the accuracy of the results in comparison with reference data.

Through our study, we aim to make a substantial contribution to cryospheric research by evaluating methods for the automated identification of rock glaciers, thereby enhancing our understanding of the potential of DL to efficiently map complex natural phenomena using EO data. The results will also contribute to increase the trustworthiness of DL methods, which is critical for various applications and particularly in communicating and explaining results to stakeholders and decision makers. 

 

Kellerer-Pirklbauer, A., Lieb, G.K., Kaufmann, V. (2022). Rock Glaciers in the Austrian Alps: A General Overview with a Special Focus on Dösen Rock Glacier, Hohe Tauern Range. In: Embleton-Hamann, C. (eds) Landscapes and Landforms of Austria. World Geomorphological Landscapes. Springer, Cham.

How to cite: Streifeneder, V., Robson, B. A., Hölbling, D., Nafieva, E., Dabiri, Z., Hauglin, E., and Abad, L.: Deep learning-based rock glacier mapping using Earth observation data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4229, https://doi.org/10.5194/egusphere-egu25-4229, 2025.

EGU25-4329 | ECS | Posters on site | CR6.8

From Decision Trees to Deep Learning: Enhanced Supraglacial Lake Detection in Antarctica 

Celia A. Baumhoer, Jonas Koehler, and Andreas Dietz

With the ongoing effects of global warming, supraglacial meltwater in polar regions plays a critical role in ice sheet dynamics, influencing global sea levels. In Antarctica, the accumulation of meltwater on ice surfaces not only reduces albedo—accelerating melting in a self-reinforcing cycle—but also drives processes such as meltwater injection and basal lubrication, with possible destabilizing effects for ice sheets. Monitoring the seasonal evolution and dynamics of supraglacial lakes is essential for understanding these processes, yet the vast and remote nature of the Antarctic ice sheet presents significant challenges. Spaceborne remote sensing offers the best solution, providing continuous, large-scale, and long-term observations. However, extracting reliable information from optical and synthetic aperture radar (SAR) data remains complex due to limitations in spatial transferability, cloud cover, polar night, and the spectral similarities of frozen lakes with surrounding ice. The Sentinel mission bridges these gaps, enabling the combination of optical and SAR data to achieve the best possible accuracy for mapping and monitoring supraglacial lakes.

This study evaluates whether a deep learning-based mapping approach outperforms a pixel-based Random Forest (RF) classification algorithm for supraglacial lake (SGL) detection in Antarctica. As a benchmark, we utilized an RF model trained on 14 regions and 24 input channels, including Sentinel-2 spectral bands, spectral indices, and topographic variables. To work toward a circum-Antarctic, operational SGL mapping product, we reduced the input channels by selecting the four most important features identified by the RF approach and trained a convolutional neural network (CNN) on partially labeled data from 16 Sentinel-2 scenes, including more images with cloud cover. Both models were validated using the same 16 test areas across eight Antarctic ice shelves.

The RF approach achieved a producer’s accuracy, user’s accuracy, and F1 score of 0.750, 0.945, and 0.837, respectively, whereas the CNN-based workflow achieved scores of 0.915, 0.912, and 0.913, respectively. In scene-specific comparisons, the CNN outperformed the RF approach in 13 of the 16 validation scenes. Key advantages of the CNN approach include its ability to detect lakes under thin clouds and over floating ice, resulting in less fragmented lake area estimates and requiring fewer input features. However, challenges persist in transition zones between lakes and slush, where spectral details outweigh the benefits of shape-based detection.

How to cite: Baumhoer, C. A., Koehler, J., and Dietz, A.: From Decision Trees to Deep Learning: Enhanced Supraglacial Lake Detection in Antarctica, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4329, https://doi.org/10.5194/egusphere-egu25-4329, 2025.

EGU25-5204 | ECS | Posters on site | CR6.8

Mapping of Glaciers in the Poiqu Basin (Central Himalaya) Using U-Net and Transfer Learning 

Farzaneh Barzegar and Tobias Bolch

Monitoring of glaciers is crucial as they are an important source of freshwater, an indicator of global warming, and a contributor to sea level rise. Accurate delineation of glaciers plays a crucial role in glacier monitoring and remote sensing is the most appropriate tool to map glaciers.

Existing glacier inventories have shortcomings such as unavailability in recent years and data quality. Traditional glacier mapping methods using remote sensing often rely on spectral band ratio techniques or manual digitizing. However, glacier boundaries achieved from manual digitizing are highly affected by human errors. Moreover, in the band ratio technique challenges arise in mapping debris-covered glaciers as traditional optical methods fail to distinguish debris-covered ice from surrounding rock due to their spectral similarities. Therefore, automatic mapping of glaciers is still challenging.

Advanced deep learning methods have demonstrated significant advancements in automatic glacier mapping. However, the potential of state-of-the-art deep learning methods in glacier mapping has not yet been fully explored. When it comes to deep learning, one of the challenges is the amount of training data. With the low amount of training data, the results won't be of the desired accuracy. However, it is still possible to obtain good results using a lower amount of training data and the transfer learning technique.

This study focuses on glacier mapping in Poiqu Basin (Central Himalaya), using U-Net and transfer learning. To this purpose, Sentinel-2 images and the Shuttle Radar Topography Mission (SRTM) Digital Elevation Model are deployed.

The results indicate that transfer learning leads to considerably better results than training the deep learning network from scratch. Moreover, trying different backbones does not considerably affect the results. This study highlights the efficiency of the transfer learning technique, emphasizing its potential and effectiveness in regions with limited training data.

How to cite: Barzegar, F. and Bolch, T.: Mapping of Glaciers in the Poiqu Basin (Central Himalaya) Using U-Net and Transfer Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5204, https://doi.org/10.5194/egusphere-egu25-5204, 2025.

EGU25-5944 | ECS | Posters on site | CR6.8

Neural network emulators of high resolution melt processes under Antarctic ice shelves 

Helen Ockenden, Clara Burgard, Nicolas Jourdain, and Pierre Mathiot

To make accurate projections of future sea level rise, small-scale ice-sheet and ice-shelf processes must be included in global climate models. Since high-resolution fully-coupled ice-sheet--ocean models are computationally expensive, multi-centennial simulations use lower resolution grids combined with simple parameterizations of the ice-ocean interface. However, these simple parameterizations do not fully reproduce observed melt patterns and have low sensitivity to warmer conditions. Instead, neural networks can be used to improve models by emulating the ice-ocean interactions simulated by high resolution models. We present a framework for training neural networks to emulate small-scale Antarctic basal melt processes within a global low-resolution model (here the NEMO ocean model). We employ a multi-layer perceptron which is trained with a variety of model simulations on a grid with quarter degree resolution, and aim to assess the performance of the neural network, particularly in warmer conditions representative of potential future climate states. This simple framework provides a springboard for future work using more complex architectures, and offers the potential to run computationally affordable long-period global simulations while still capturing crucial ice-shelf--ocean interactions.  

How to cite: Ockenden, H., Burgard, C., Jourdain, N., and Mathiot, P.: Neural network emulators of high resolution melt processes under Antarctic ice shelves, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5944, https://doi.org/10.5194/egusphere-egu25-5944, 2025.

EGU25-8017 | Orals | CR6.8

A first view of the EO-driven digital twin for ice sheets 

Sebastian Simonsen, Nanna Karlsson, and The DTC Team

The response of ice sheets and shelves to climate change profoundly influences global human activities, ecosystems, and sea-level rise. As such, ice sheets are a vital component of the Earth system, making them a cornerstone for developing a future Digital Twin Earth. Here, we present the initial steps toward an Earth Observation (EO)-driven Digital Twin Component (DTC) for Ice Sheets, marking an effort to understand and predict the behaviour of the Greenland Ice Sheet and Antarctic ice shelves under user-defined “what-if” scenarios.

To meet the diverse needs of stakeholders, DTC Ice Sheets will adopt a modular design comprising 10 Artificial Intelligence/Machine Learning (AI/ML) and Data Science modules. All targeted four initial use cases that will drive the development of DTC Ice sheets. These initial use cases are: (1) Greenland Hydropower Potential: By modelling and monitoring ice sheet hydrology and meltwater runoff, the DTC ice sheets will evaluate Greenland’s renewable energy opportunities and provide actionable insights for sustainable hydropower development. (2) EU Sea Level Response Fingerprint: The DTC Ice Sheets will deliver region-specific insights into how ice sheet mass loss will contribute to global sea level rise, focusing on the implications for coastal infrastructure across Europe. (3) State and Fate of Antarctic Ice Shelves: Through detailed stability analysis, the DTC Ice Sheets will investigate the vulnerability of Antarctic ice shelves to climatic and oceanic changes, shedding light on their role in regulating ice sheet mass loss and global sea level. (4) Enhanced Surface Climate: Leveraging EO data and climatology, the DTC Ice Sheets will improve understanding of surface climate interactions, advancing predictions of feedback loops between ice sheets, the atmosphere, and the ocean.

The DTC Ice sheet implementation on the DestinE Core Service Platform (DESP) will consist of interconnected modules to serve the use cases. Still, it will also, when fully implemented, provide a holistic view of an ice sheet digital twin. Hence, DTC Ice Sheets aims to provide high-resolution insights into ice sheets' past, present, and future states, align with stakeholders, and foster interdisciplinary collaboration by interfacing with other thematic Digital Twin Earth systems, such as ocean and coastal processes. The DTC ice sheets will empower stakeholders to explore What-if scenarios to address climate change's impacts and feedback mechanisms. All are found in current state-of-the-art EO data of ice sheets. 

How to cite: Simonsen, S., Karlsson, N., and DTC Team, T.: A first view of the EO-driven digital twin for ice sheets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8017, https://doi.org/10.5194/egusphere-egu25-8017, 2025.

EGU25-8433 | ECS | Posters on site | CR6.8

Mapping and analyzing ice shelf damage using multisensor imagery and machine learning 

Kaian Shahateet, Romain Millan, Lucie Bacchin, Cyrille Mosbeux, and Trystan Surawy-Stepney

The floating ice shelves surrounding Antarctica play a crucial role in regulating ice sheet mass loss by providing mechanical buttressing, which regulates ice discharge into the ocean. The recent collapses of Larsen-A and B along with the rapid retreat of Thwaites were followed by accelerated ice discharge, underlining the importance of understanding the full dynamic of this process. Enhanced damage through fracturing has recently been shown to play a critical role in ice shelf weakening, reducing its buttressing capacity and potentially accelerating their collapse. Despite its significance, the processes governing ice shelf damage remain poorly understood. Damage manifests itself as large crevasses, rifts, and shearing regions clearly visible on satellite imagery. Historically, the mapping of fractures has been challenging due to the labor-intensive nature of manual delineation. Rapid advancements in machine learning, however, have revolutionized damage mapping, enabling the automatic detection of damage features. Although SAR backscatter imagery from ESA's Sentinel-1 has been the primary source of data in recent studies, it suffers from limited temporal coverage (2013-present), which does not capture the entire damage dynamic of ice shelves that destabilized in the early 2000s. Other available products, also  exhibited significant discrepancies with modeled changes in ice viscosity, suggesting that critical features of ice damage are not fully captured (e.g. basal fracturing). To address these gaps, this study presents a novel methodology leveraging multisensor optical imagery and supervised/semi-supervised machine learning algorithms to identify damage features. A U-Net algorithm was trained on manually annotated images from 10 acquisitions from the USGS/NASA's Landsat satellite, across diverse Antarctic ice shelves. These annotations represented various types of damage to ensure broad applicability. The model was then refined using a human-in-the-loop approach with additional Landsat and Sentinel imagery datasets, enhancing prediction accuracy. We demonstrate the capability of our model to map comprehensively the evolution of damage in the Amundsen Sea Embayment, one of Antarctica's most vulnerable regions, from the 1990s to the present.  The results are compared with existing damage products derived from machine learning and radon transform methods using Sentinel-1 SAR images, on the period 2013 to present. We map the dynamic evolution of surface and basal fractures, along with their morphological characteristics such as maximum length and area, and compare this evolution with dynamical changes over the same time period. We complement our analysis by comparing our result to damage modeling using an ice flow model on the Pine Island ice shelf. We use the Shallow Shelf Approximation within the Elmer/ice model to invert for damage and ice viscosity evolution since 1992, by assimilating a long record of satellite-derived surface flow velocity and thickness. We finally analyze the spatial correlations between modeled and observed damaged and draw conclusions on the features of importance regarding ice sheet stability through time. We demonstrate the potential of multisensor optical imagery, which offers broader temporal coverage dating back to the 1970s, to address critical gaps in understanding ice shelf damage and its evolution.

How to cite: Shahateet, K., Millan, R., Bacchin, L., Mosbeux, C., and Surawy-Stepney, T.: Mapping and analyzing ice shelf damage using multisensor imagery and machine learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8433, https://doi.org/10.5194/egusphere-egu25-8433, 2025.

EGU25-8973 | ECS | Posters on site | CR6.8

Machine learning for prediction of sea ice stability 

Alessandro Cotronei, Claudio Gallicchio, and Rune Graversen

Arctic sea ice, the vast body of frozen water near the North Pole, has been in steady decline since satellite observations began. While state-of-the-art models attempt to project future scenarios, they often show significant discrepancies, even though the sea ice system is generally considered to decline linearly with rising temperatures. Machine learning models, although they may lack the ability to fully explain the underlying physical processes, offer a complementary approach. By training these models on existing data, we can generate plausible future predictions that are less influenced by the biases inherent in traditional modeling methods. In this study, we evaluate several machine learning architectures to identify the most effective ones. Using the best-performing model, we explore the stability and potential hysteresis behaviors of the Arctic sea ice system.

How to cite: Cotronei, A., Gallicchio, C., and Graversen, R.: Machine learning for prediction of sea ice stability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8973, https://doi.org/10.5194/egusphere-egu25-8973, 2025.

EGU25-9315 | ECS | Orals | CR6.8

Ice Floe Data Augmentation Using Diffusion Models 

Justin Bunker, Martin S. J. Rogers, Louisa van Zeeland, Jeremy Wilkinson, and Mark Girolami

The monitoring of ice floe is essential for mapping marine ecosystems, ensuring safe ship navigation, and ice hazard forecasting. Satellite imagery, such as Synthetic Aperture Radar (SAR), is a prime candidate for capturing information related to ice floes, due to the ability to discern sea ice conditions in this imagery in cloudy or poor lighting conditions. This SAR imagery can then be passed along to image processing algorithms to extract quantities of interest such as floe size distribution (FSD). Whilst considerable research has used fully supervised machine learning models in this domain, such models require an abundant amount of annotated data for training. The time-consuming, subjective, and costly process of annotating limits the amount of available data that can be used during training and, thus, reduces the performance of the trained model. To alleviate this problem, we turn towards the burgeoning field of generative modeling to create synthetic labeled data.

An important class of generative models, known as diffusion models, has been shown to be particularly efficient. Over the years, a rich plethora of techniques and architectures have been developed to enable these diffusion models to provide realistic samples from an approximate distribution of the training data. Moreover, such models can also be conditioned by additional information, such as texts or images, offering an interesting degree of flexibility to explore and enhance the sampling process. More pertinently, diffusion models have been employed to generate synthetic images of semi-natural areas captured by drones, as well as satellite imagery of rural and urban scenes. However, to date, their application to SAR imagery of the cryosphere remains unexplored.

In this work, we describe a process whereby we use a diffusion model, namely a Denoising Diffusion Probabilistic Model, to model the joint distribution over the space of SAR images and their corresponding labels. In addition to standard error metrics, we use FSD to demonstrate that the synthetic SAR data is consistent with the real data. Furthermore, we show that using a dataset composed of both the real data and the synthetic data results in better performance for segmentation modeling. Additional experiments are performed to show performance as a function of the amount of real and synthetic data. 

How to cite: Bunker, J., Rogers, M. S. J., van Zeeland, L., Wilkinson, J., and Girolami, M.: Ice Floe Data Augmentation Using Diffusion Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9315, https://doi.org/10.5194/egusphere-egu25-9315, 2025.

EGU25-10699 | ECS | Posters on site | CR6.8

Creating a Pan-Arctic Retrogressive Thaw Slump Dataset with Harmonized Sentinel-2 Data and Deep Learning Methods 

Jonas Küpper, Tobias Hölzer, Todd Nicholson, Luigi Marini, Lucas von Chamier, Sonja Hänzelmann, Ingmar Nitze, Anna Liljedahl, and Guido Grosse

In a rapidly changing permafrost environment driven by climate change and anthropogenic disturbances, tracking geomorphological dynamics is a crucial task, not only to provide hazard monitoring, but also to evaluate climatological feedback processes. Yet, the impact on rapid permafrost disturbances on the Earth system is still uncertain, making the availability of reliable, long term data a very important building block to understand the interconnections and feedbacks between several environmental subsystems. 

Specifically, Retrogressive Thaw Slumps (RTS) are a major mass-wasting phenomenon and a rapid disturbance in ground-ice rich permafrost landscapes. They can mobilize large quantities of formerly frozen ground and consequently sediment, carbon, and nutrients. Once initiated they can grow and develop broader erosion disturbances. Over years and decades they can undergo polycyclic behaviour of initialization, growth, stabilization, and re-activation. The spatial distribution and temporal dynamics of RTS are generally poorly quantified so far on a pan-arctic scale, except for some regions covered by more intensive research. 

Multiple methods and data are used to map permafrost disturbances like RTS, including in-situ mapping. However, due to the remoteness and reduced accessibility, earth observation data is the primary source of RTS inventories. While RTS mapping is also done manually utilizing expert knowledge from high-resolution remote sensing imagery, machine learning techniques are increasingly used to segment permafrost features from satellite images. However, due to the requirement to process large amounts of data and also the reduced availability of suitable image data, especially in the high-latitudes, these datasets still often lack the temporal and spatial coverage to derive insights related to the recent global environmental changes. Current advancements in artificial intelligence based inference methods make feature segmentation now much more feasible and efficient, so activities for mapping RTS based on high resolution PlanetScope images and deep-learning methods, such as the DARTS dataset, already cover large RTS affected regions. Nevertheless, a full pan-arctic coverage over multiple time-steps is still lacking, thus far. 

To expand the existing body of RTS inventories, we use a convolutional neural network to detect these permafrost features from Sentinel 2 imagery to create a multi-year dataset of detected thaw slumps in the circumpolar arctic. The comparison with existing manually labelled and automatically derived high resolution thaw slump inventories provides a quantifiable verification to estimate uncertainties. This is crucial for evaluating Sentinel-2 as a high resolution dataset with favourable properties in terms of data availability and processing requirements compared to commercial and access restricted VHR imagery. Our work can underpin downstream tasks to extend RTS classification, understanding trigger mechanisms and improve vulnerability mapping. Also, time series of RTS disturbance data may be used for the temporal and spatial correlation with climate reanalysis and atmospheric datasets for large scale climate change impact modelling and feedback evaluation over the permafrost domain. Additionally, the open architecture of the processing pipeline can be used to implement near real-time monitoring services based on the Sentinel-2 data release stream for public access. We present ongoing work on the RTS segmentation dataset and current key downstream results.

How to cite: Küpper, J., Hölzer, T., Nicholson, T., Marini, L., von Chamier, L., Hänzelmann, S., Nitze, I., Liljedahl, A., and Grosse, G.: Creating a Pan-Arctic Retrogressive Thaw Slump Dataset with Harmonized Sentinel-2 Data and Deep Learning Methods, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10699, https://doi.org/10.5194/egusphere-egu25-10699, 2025.

EGU25-11168 | ECS | Posters on site | CR6.8

Bridging machine learning and physics-based models for improving snow water equivalent predictions in the northern hemisphere 

Oriol Pomarol Moya, Derek Karssenberg, Walter W. Immerzeel, Philip Kraaijenbrink, Madlene Nussbaum, Siamak Mehrkanoon, and Isabelle Gouttevin

Snow water equivalent (SWE) is an important component of the hydrological cycle but still faces large uncertainties in its quantification due to its high temporal and spatial variability. While machine learning (ML) has been applied to multiple domains in hydrology, its use for SWE prediction has been hindered by limited observational training data beyond the local scale. Hybrid models that integrate simulated data from physics-based models with a ML setup may overcome this lack of observations, outperforming both physics-based models and conventional ML approaches in data-scarce regions.

In this project, we tested two different hybrid ML setups that predict the daily change in SWE using Crocus snow model simulations together with data from ten meteorological and snow observation stations throughout the northern hemisphere containing 7-20 years of data. The first setup follows a common post-processor approach where the outputs and state variables from Crocus are fed as additional predictors to the ML model at each time step. The second setup follows the concept of data augmentation, where Crocus is used to simulate SWE for stations for which no observations are available. These simulations are then fed as additional data points to the ML model, but are weighted in the loss function to control their influence during training.

The obtained results show that the post-processor approach is best suited for predicting SWE in years excluded during training. However, when predicting SWE in untrained stations the data augmentation setup achieves the largest increase in performance, reducing the root mean squared error by 22% compared to Crocus and by 42% compared to the measurement-based ML model. A feature importance analysis reveals that the hybrid model predictions are influenced the most by the current SWE status, incoming radiation, snowfall and air temperature. These results showcase the potential of hybrid models for predicting variables that suffer from data scarcity such as SWE.

How to cite: Pomarol Moya, O., Karssenberg, D., Immerzeel, W. W., Kraaijenbrink, P., Nussbaum, M., Mehrkanoon, S., and Gouttevin, I.: Bridging machine learning and physics-based models for improving snow water equivalent predictions in the northern hemisphere, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11168, https://doi.org/10.5194/egusphere-egu25-11168, 2025.

EGU25-11399 | ECS | Posters on site | CR6.8

Towards quantifying ice contents in mountain permafrost environments 

Julie Røste and Andreas Kääb

In its special report on Ocean and Cryosphere in a Changing Climate (SROCC) from 2019, the Intergovernmental Panel on Climate Change (IPCC) highlights clear knowledge gaps concerning the extent and ice content of permafrost in mountain regions. We present results from a study on the distribution of mountain permafrost that includes an improved understanding of its characteristics and an estimation of the sub-surface ice reserves in mountainous regions under climate scenarios. We explore the feature space of mountain permafrost using a range of statistical and machine learning techniques in an uncertainty-aware setting. This space consists of topographic and climatic features such as topographic masks, elevation models, potential incoming solar radiation, seasonal ground temperatures and snow accumulation. We combine these features with existing inventories of rock glaciers, as these are good visible indicators of mountain permafrost, and in addition typically ice-rich. Based on such datasets we create a data-driven model to predict the probability of potential rock glaciers occurrence in order to obtain a first estimate of ground ice content. In addition, output from a numerical permafrost model, the CryoGrid community model, provides synthetic observations. We further investigate the vulnerability of these potentially ground-ice rich areas under climate change by including forcing data from climate models based on various RCP scenarios.

How to cite: Røste, J. and Kääb, A.: Towards quantifying ice contents in mountain permafrost environments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11399, https://doi.org/10.5194/egusphere-egu25-11399, 2025.

EGU25-11686 | Orals | CR6.8

Generative Model-Based Downscaling of the Surface Mass Balance of the Greenland Ice Sheet 

Nils Bochow, Philipp Hess, and Alexander Robinson

The surface mass balance (SMB) is projected to become the main driver of mass changes for the Greenland Ice Sheet (GrIS) by the end of this century. Therefore, it is crucial to have realistic projections of the SMB for future estimates of mass loss and sea-level rise.

To date, estimates of the surface mass balance are most often provided by either (i) stand-alone parameterization schemes, such as positive degree days (PDD) or energy balance approaches, (ii) direct outputs from Earth system models (ESMs), or (iii) regional climate models (RCMs) forced by boundary conditions from ESMs. Each of these approaches has its disadvantages. Stand-alone parameterization schemes are often overly simplified and unable to capture smaller-scale processes at the surface. ESMs often provide forcing fields that are too coarse compared to the resolution required for ice sheets. Meanwhile, regional climate models are expensive to run and computationally slow.

In this study, we address these issues by employing a generative model-based approach to realistically downscale the SMB directly from ESM fields to a 5 km resolution. We train a diffusion-based model on historical and future SMB fields from the regional climate model MAR. This allows us to generate high-resolution SMB fields in a fraction of the time required by a regional climate model. We condition our diffusion model on an initial estimate of the SMB derived from ESMs. Specifically, we add noise to the initial ESM estimate and subsequently de-noise the SMB field at different noise levels. By selecting the noise level during inference, we can effectively choose the spatial scale at which ESM features should be preserved.

Our approach enables fast, simple, and probabilistic downscaling of the SMB and potentially other climate fields.

How to cite: Bochow, N., Hess, P., and Robinson, A.: Generative Model-Based Downscaling of the Surface Mass Balance of the Greenland Ice Sheet, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11686, https://doi.org/10.5194/egusphere-egu25-11686, 2025.

EGU25-12902 | Posters on site | CR6.8

Label-free ice floes segmentation in SAR images for floe size distribution in the Antarctic 

Louisa van Zeeland, Martin S. J. Rogers, Nick Hughes, Ben R. Evans, Oliver Strickson, Gaëlle Veyssière, Andrew Fleming, Scott Hosking, and Jeremy Wilkinson

Sea ice is a crucial component of the polar marine environment. A contiguous piece of sea ice is called an ice floe, and the size variation in these floes across a region is described as the floe size distribution (FSD). Analysis of FSD provides information on the physical processes associated with sea ice dynamics, which is needed for calibrating and validating numerical sea ice models. For example, the size and shape of sea ice floes is predominantly controlled by wind and ocean wave conditions, thus the FSD metric provides crucial insight into these environmental conditions. Consequently, the automatic detection of floes, and hence FSD, is required to improve our understanding of these conditions over large spatial-temporal scales. Here, we present a method to automatically segment sea ice from Synthetic Aperture Radar (SAR) images for downstream applications. Our method uses an autoencoder architecture, minimising dual losses concurrently to guide the training on a large number of SAR images.

For machine learning (ML) to assist in automatic labelling of sea ice, traditional supervised learning models require the provision of a sufficiently large, labelled dataset to train the model. Manual interpretation and identification of sea ice in satellite imagery is a time consuming and tedious process, frustrating the development of annotations over large spatial areas. Additionally, manually labelled data are subject to unintentional human variability thus potentially introducing bias. It is not a scalable solution.

Feature learning or representation learning is a ML technique that automatically guides its own training to extract useful information without the need for labelled data. Instead of using optical images as many other works done on FSD with supervised learning techniques, we use SAR images here with representation learning. Using SAR images allows us to monitor sea ice conditions year-round, including during periods of polar darkness and cloudy conditions, where the detection of sea ice conditions in optical images is problematic. As this autoencoder model does not require labelled data, it can be scaled both spatially and temporally. It also has the potential to be extended to detect other features and to learn beyond ice-water segmentation.

How to cite: van Zeeland, L., Rogers, M. S. J., Hughes, N., Evans, B. R., Strickson, O., Veyssière, G., Fleming, A., Hosking, S., and Wilkinson, J.: Label-free ice floes segmentation in SAR images for floe size distribution in the Antarctic, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12902, https://doi.org/10.5194/egusphere-egu25-12902, 2025.

EGU25-14145 | ECS | Orals | CR6.8

Cryo2S1: Mapping sea ice radar freeboard in Sentinel-1 SAR imagery from CryoSat-2 using deep learning 

Andreas Stokholm, Jack Christopher Landy, Roberto Saldo, Tore Wulf, Anton Korosov, and Sine Munk Hvidegaard

Sea ice is critical to map for safe and efficient maritime navigation, to mitigate ship trapping and capsizing. Sea ice is also vital to monitor to assess the state of the changing climate and a critical component in climate and weather models, reflecting sunlight towards space and acting as an insulating material between the ocean surface and the atmosphere.

Professional sea ice analysts at national ice services map sea ice based on Synthetic Aperture Radar (SAR) images acquired by satellites, such as the Sentinel-1 (S1) satellite constellation. The ice analysts manually interpret the SAR images using their in-depth knowledge and experience to create sea ice charts with information on the sea ice conditions.

A challenge for the S1 5.4 GHz SAR measurements is that the radar wave does not penetrate deep into the sea ice and is scattered/reflected by the surface. Therefore, the SAR images provide information primarily about the sea ice surface, useful for identifying and classify sea ice conditions. The charts describe, among others, the sea ice’s stage of development - the type of sea ice - an indicator of its thickness. The manual charting process apply sea ice classes, defined by the International Ice Charting Working Group (IICWG) on behalf of the World Meteorological Organization (WMO). Considerable uncertainties are associated with the ice classes that can vary from, e.g. 30-200cm or 70-120cm in thickness. Deep-learning models that produce stage-of-development information from S1 radar images exist but has the same inherent limitations of the sea ice charts in the model outputs.

Current state-of-the-art sea ice thickness retrieval methods relies on altimeter satellites, such as the CryoSat-2 (CS2) satellite. The distance between the ocean and the sea ice is measured, known as the sea ice freeboard. For a Ku-band radar altimeter like CS2, it is assumed that the radar response penetrates the snow and returns from the sea ice surface. As the true penetration is unknown, and the radar wave propagation is delayed when the signal passes through snow, the measured quantity is known as the radar freeboard.

The sea ice thickness can be estimated with an accuracy of 20-40% using the radar freeboard by calculating the sea ice's buoyancy based on snow and ice density estimates, and auxiliary snow depth information. However, CS2 only measures 1600m across the orbit and can thus only monitor sea ice thickness in the Arctic monthly - insufficient for many applications, such as maritime navigation, and leaves data record gaps. S1 SAR on the other hand, cover 400km in Extra Wide mode across the orbit with repeating coverage every week.

Here, we present our preliminary results of circumventing the limitations of CS2 and S1 by training supervised deep-learning convolutional neural network (CNN) models to recognise sea ice textures in S1 SAR images and assign sea ice radar freeboard estimates acquired by CS2. This approach transfers information acquired by CS2 to S1, which we call Cryo2S1. A Cryo2S1 dataset is curated, containing several thousand collocated S1 SAR images and along-track CS2 measurements during 2020-2021.

How to cite: Stokholm, A., Landy, J. C., Saldo, R., Wulf, T., Korosov, A., and Hvidegaard, S. M.: Cryo2S1: Mapping sea ice radar freeboard in Sentinel-1 SAR imagery from CryoSat-2 using deep learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14145, https://doi.org/10.5194/egusphere-egu25-14145, 2025.

EGU25-14736 | ECS | Orals | CR6.8

Automated Analysis of Snowpack Stratigraphy NIR Images Using Deep Learning 

Marlena Reil, Olya Mastikhina, Jennifer Marks, Karla Felix Navarro, Mohammad Reza Davari, Lars Mewes, Julia Kaltenborn, and David Rolnick

Snowpacks are important elements of the Earth’s cryosphere and are composed of layers with unique physical properties. Snow stratigraphy, the study of distinct snow layers and their properties, provides essential data for climate modeling, water resource management, and avalanche prediction. However, existing methods for characterizing snowpacks with near-infrared (NIR) photography are based on manually segmenting layers from images, which is a laborious and time-consuming task. In this work, we develop an approach to automate snowpack layer segmentation based on fine-tuning Segment Anything (SAM), a state-of-the-art deep learning segmentation model. We use a small set of expert-labeled NIR snowpack images and explore different task representations. We approach the problem through the lens of 1) edge detection, which focuses on detecting snowpack layer boundaries and 2) region detection, which focuses on predicting the area occupied by the layers.  Our results indicate that deep learning segmentation is promising for automating the segmentation of snowpacks. This ultimately leads to facilitating snow stratigraphy analysis to improve applications such as avalanche forecasting and snowpack modeling.

How to cite: Reil, M., Mastikhina, O., Marks, J., Navarro, K. F., Davari, M. R., Mewes, L., Kaltenborn, J., and Rolnick, D.: Automated Analysis of Snowpack Stratigraphy NIR Images Using Deep Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14736, https://doi.org/10.5194/egusphere-egu25-14736, 2025.

EGU25-15293 | ECS | Posters on site | CR6.8 | Highlight

Glacier Area Change Assessment over 2015-2023 in the European Alps with Deep Learning 

Codrut-Andrei Diaconu, Jonathan L. Bamber, and Harry Zekollari

Glacier retreat is a key indicator of climate change and requires regular updates of the glacier area. The most recent inventory for the European Alps, released in 2020, showed that  glaciers retreated approximately 1.3% per year from 2003 to 2015. This ongoing retreat underscores the urgent need for accurate and efficient monitoring techniques.

Recent advancements in Deep Learning have led to significant progress in the development of fully automated glacier mapping techniques. In this work, we use DL4GAM, a multi-modal Deep Learning-based framework for Glacier Area Monitoring, to assess the change in glacier area in the European Alps over 2015-2023. The main data modality used for training is based on Sentinel-2 imagery, combined with additional features derived from a Digital Elevation Model, along with a surface elevation change map, which is particularly useful for debris-covered glaciers. The framework provides an area (change) estimate independently for each glacier, with uncertainties quantified using an ensemble of models. Region-wide, we estimate a retreat of -1.90 ± 0.71%, which is greater than the rate observed during the previous decade. Our estimates also present a significant inter-glacier variability which we analyze with respect to various topographical parameters such as slope, aspect, or elevation.

Several challenges persist, including model limitations, data availability issues, and the impact of debris, cloud cover, and seasonal snow. We discuss these challenges, the design choices made to address them, and the remaining open issues.

How to cite: Diaconu, C.-A., Bamber, J. L., and Zekollari, H.: Glacier Area Change Assessment over 2015-2023 in the European Alps with Deep Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15293, https://doi.org/10.5194/egusphere-egu25-15293, 2025.

EGU25-15407 | ECS | Posters on site | CR6.8

Temporal Evolution of the Petermann Glacier Surface Elevation with Implicit Neural Representation in High Spatiotemporal Resolution using CryoSat-2 Data 

Peter Naylor, Andreas Stokholm, Nikolaos Dionelis, Natalia Havelund Andersen, and Sebastian Bjerregaard Simonsen

Global warming threatens to cause irreversible planetary changes and is accelerated in the polar regions, warming at nearly four times the global average. Warmer temperature exacerbates ice sheet ice loss, increasing the freshwater discharge into oceans and contributing to rising sea levels and regional changes in ocean salinity, threatening a collapse of ocean currents. The number of humans living below sea level is projected to rise by 73% by the turn of the century. Therefore, accurately determining the ice loss and the freshwater discharge is paramount to enable decision-makers to take necessary actions.

 

Ice sheet ice loss can be estimated using a satellite altimeter, measuring the spatial ice sheet surface height at many time instances. The apparent elevation change can be converted into mass change by accounting for bedrock movement and snow/firn processes. An obstacle in utilising satellite altimeter data is the unstructured nature of the data points resulting from elevation observations at different time instances. We propose to treat these altimeter data as cloud points in the space-time domain and utilise implicit neural representation (INR) to encode the target field as a continuous function varying both in time and space. Compared to traditional interpolation methods such as trilinear interpolation or kriging, the INR method can capture non-linearities and long-term trends while providing a compact encoding of the target field, allowing for scalable dissemination of the product.

 

We present a feasibility study of utilising INR to reconstruct the surface elevation of the Petermann glacier, northwest Greenland, from CryoSat-2 radar altimeter elevation observations. We carried out many model training experiments, consisting of ablation studies on additional loss terms as well as model architectures (SIREN, RFF, KAN and MFN) and hyperparameters (number and width of layers and loss term weights), to find the best combination. The main difficulty is correctly capturing the glacier temporal dynamics. In addition, we trained models with varying quantities of data (5 months, 1 year, 2 years and 12.5 years) to investigate whether more data improved the model performance. Results are evaluated using Operation IceBridge (OIB) LIDAR, and GeoSAR elevation measurements. OIB allows for evaluation of model elevation over a large temporal and geographical area, whereas GeoSAR allows for comparing high resolution elevation data on a single day over a small area.

 

Results indicate that we achieve the best performance using the SIREN INR architecture coupled with high temporal and spatial loss weights. In addition, models perform best when using CryoSat-2 data from the entire 12.5 year time frame. The models perform particularly well in regions with high data point density but struggle at the outer rims of the ice sheet where the point density is low. The feasibility study presents a promising direction in modelling the spatiotemporal evolution of the ice sheet at a sub kilometre resolution with a daily temporal time step using INR. We foresee these methods being applicable to many geoscience applications with irregular data sampling in space and time.

How to cite: Naylor, P., Stokholm, A., Dionelis, N., Andersen, N. H., and Simonsen, S. B.: Temporal Evolution of the Petermann Glacier Surface Elevation with Implicit Neural Representation in High Spatiotemporal Resolution using CryoSat-2 Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15407, https://doi.org/10.5194/egusphere-egu25-15407, 2025.

EGU25-16155 | ECS | Orals | CR6.8

High-resolution mass balance reconstructions for Swiss glaciers using machine learning 

Marijn van der Meer, Harry Zekollari, Kamilla Hauknes Sjursen, Matthias Huss, Jordi Bolibar, and Daniel Farinotti

Glacier retreat poses significant environmental and societal challenges. Understanding the local impacts of climate drivers on glacier evolution is essential, with glacier mass balance being a central concept. This study uses the Mass Balance Machine (MBM; Sjursen et al., 2025), an open-source, data-driven model based on eXtreme Gradient Boosting (XGBoost) that reconstructs glacier mass balance with high spatiotemporal resolution at regional scales. Trained on point mass balance data from multiple glaciers, MBM captures both intra- and inter-glacier variability, enabling the identification of transferable patterns and applications to glaciers without direct observations. Here, we applied MBM to reconstruct the mass balance of Swiss glaciers. The model was trained using a comprehensive dataset of approximately 34,000 winter and annual point mass balance measurements from 35 Swiss glaciers in diverse climate settings from 1951 to 2023. Using MBM, we generated high spatial resolution reconstructions of seasonal and annual mass balance for these 35 glaciers. When validated on independent unseen glaciers, MBM demonstrated robust performance across spatial scales (point to glacier-wide) and temporal scales (monthly to annual). This study underscores how MBM can be effectively used in Switzerland to generalize across diverse glaciers and climatic conditions, highlighting the model's versatility and broad applicability.

How to cite: van der Meer, M., Zekollari, H., Sjursen, K. H., Huss, M., Bolibar, J., and Farinotti, D.: High-resolution mass balance reconstructions for Swiss glaciers using machine learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16155, https://doi.org/10.5194/egusphere-egu25-16155, 2025.

EGU25-16392 | ECS | Posters on site | CR6.8

Comparison of selected machine learning algorithms to derive glacier velocity maps 

Magdalena Łucka and Miłosz Sumara

Marine-terminating glacier dynamics play a crucial role in understanding the climate system. They connect large ice sheets, oceans, and the atmosphere; thus their changes might deliver important information about the relationship between those systems. One of the factors describing ice dynamics is velocity. Its changes can reflect the processes occurring on and underneath the ice sheet surface. Nowadays, that information is delivered mainly by remote-sensing sensors, including satellite radar images (SAR), which provide timely and continuous data even in isolated areas. Plenty of offset-based algorithms already exist to deliver reliable velocity maps based on satellite products. However, these methods require setting a bunch of processing parameters, and they are usually suitable for only one sensor type. This study investigates possible machine learning solutions for finding corresponding areas on satellite images in order to provide velocity maps in an alternative way. In this work, SAR datasets from Sentinel-1 satellite were used to test two machine learning approaches for glacier velocity retrieval. The first approach is based on utilising convolutional neural networks (CNN) to select similar areas on the image pairs. The input data consist of only two coregistered SAR intensity images, which are augmented in the next processing step. As the model output, the most similar image patch is returned. After selecting corresponding image patches, the offsets in both image axes are determined and calculated into velocity values based on a pixel size and temporal baseline. The second approach investigates the possibility of applying the LightGlue image matching technique to the analysis of SAR data in order to detect similar features and determine their movement. The same input products are used, and methods performance and reliability are assessed. Both techniques are tested on two glaciers with different ice dynamics and locations: one in Greenland and one in Svalbard. The methods are compared in terms of efficiency, information density, and velocity values reliability. The final maps are validated by offset-tracking results processed for the same input images.

How to cite: Łucka, M. and Sumara, M.: Comparison of selected machine learning algorithms to derive glacier velocity maps, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16392, https://doi.org/10.5194/egusphere-egu25-16392, 2025.

Permafrost degradation and its impact on carbon cycling in the Arctic necessitate innovative approaches for monitoring and understanding freeze-thaw dynamics. The zero-curtain, a critical period wherein a subsurface phase-state threshold of 0°C is maintained, significantly influences permafrost degradation and carbon release. Understanding these dynamic processes is vital for predicting drivers of change and formulating strategies to address mitigation and intervention methods and the broader implications on global climate systems. The generation of Circumarctic zero-curtain maps and subsidence products leverages advanced radar polarimetry from Sentinel-1 C-band inSAR and UAVSAR L-band inSAR (NISAR) data as well as thermal imagery derived from MODIS and ASTER (SBG-TIR). Radar backscatter intensity, interferometry, and polarimetric decomposition were applied to detect and infer surface deformation and subsurface moisture content. We utilized eigenvalue decomposition and matrix algebra to extract coherent scatterers and compute ground displacement, i.e., thaw subsidence. To better resolve the zero-curtain with subsurface thermal gradients, water flow, and thermal conductivity in permafrost regions, we reconciled energy balance, derived probabilistic phase transitions, and computed molecular momentum and quantum tunneling before training and validation. To generate Circumarctic zero-curtain maps, coherency-masked radar data was down-sampled using wavelet transform and kriging interpolation, while in situ-calibrated thermal data was up-sampled with bilinear interpolation. We examined freeze-thaw dynamics and trained a robust hybridized ensemble learning framework (GeoCryoAI) with in situ subsurface temperature and soil moisture content measurements at depth. The GeoCryoAI architecture integrates teacher forcing to support in situ learning reinforcement, multimodal data harmonization for validation and scaling efforts, multidimensional memory-encoded convolutional-layered (ConvLSTM3D) hybrid stacking to capture spatiotemporal dependencies, and physics-informed modules rooted in mathematical theory, thermodynamic principles, and quantum mechanics. These methods introduce key relationships and real-world dynamics to the modeling framework (e.g., heat equation, Stefan-Boltzmann law, Stefan’s equation, Darcy’s law, Fourier’s law, Schrödinger’s law, Heisenberg uncertainty principle) while also resolving complex optimization problems with database searching and integer factorization. In this study, we integrated empirical and theoretical perspectives to gain insight into the permafrost carbon feedback with novel applications. By exploring the elusive nature of the zero-curtain phenomenon across the Circumarctic with various quantification methods and preparing an efficient, robust pipeline and scalable framework for NISAR harmonization, pre-processing, simulations and forecasts, and 12-day high-resolution analysis-ready maps and science products, this study leverages multimodal data resources, high-performance computing infrastructure, a novel quantum-driven classification scheme, a novel hybridized dynamical feedback ensemble learning framework, and provide contemporary resources that inform, engage, and promote high-impact cross-disciplinary research across the northern latitudes.

How to cite: Gay, B. and Miller, C.: Advancing Arctic Science in the NISAR Era: Mapping Circumarctic Zero-Curtain Dynamics with inSAR Polarimetry, Thermal Imaging, and Quantum-Enhanced AI, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17518, https://doi.org/10.5194/egusphere-egu25-17518, 2025.

EGU25-17904 | Orals | CR6.8

Probabilistic forecasts of interannual September Arctic sea ice extent with data-driven statistical models 

Lauren Hoffman, François Massonnet, and Annelies Sticker

he widespread impacts of declining Arctic sea ice cover necessitate accurate and reliable predictions of Arctic sea ice. Up to now, much emphasis has been placed on either predictions at the sub-seasonal to seasonal timescales, or projections at the multi-decadal time scales, and less so on predictions at the seasonal to interannual time scales that are key for planning and infrastructure upgrade. Internal variability is a dominant source of uncertainty in predicting Arctic sea ice on seasonal to interannual timescales. However, initialized predictions conducted with dynamical climate models are of little use today, since these models exhibit biases and long-term drift that lead to poor skill beyond the seasonal time scale. In this study, we test and develop several statistical models in the form of transfer operators and neural networks to forecast probabilistic state transitions of the internal variability in Arctic September sea ice extent. Both the transfer operators and neural networks are trained on a large database of state transitions available from the CMIP6 archive. The models show comparable skill to other numerical and statistical models included annually in the Sea Ice Outlook for the predictions of September sea ice extent initialized in June, July, and August. While both statistical model types are able to make accurate and reliable predictions for many initialization months, the model performance is characterized by the spring predictability barrier and decreases for predictions initialized in March--May. The statistical models show skill beyond simple persistence when it comes to predicting sea ice extent trends at the interannual time scale. In particular, predictions initialized in July 2000 are able to reproduce the 2000-2010 accelerated decline in September sea ice extent, and predictions initialized in July 2012 capture the 2012-2024 slow-down in sea ice decline.

How to cite: Hoffman, L., Massonnet, F., and Sticker, A.: Probabilistic forecasts of interannual September Arctic sea ice extent with data-driven statistical models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17904, https://doi.org/10.5194/egusphere-egu25-17904, 2025.

EGU25-18155 | Posters on site | CR6.8

On emulating Sea Ice in the Finite Element Sea Ice-Ocean Model (FESOM) 

Florent Birrien, Nils Hutter, and Nikolay Koldunov

In recent years, Artificial Intelligence (AI) has been a game-changer in climate modeling, providing innovative and adaptable approaches while improving accuracy and computational efficiency. For instance, hybrid models can preserve the robustness of physical modeling while emulating components that are computationally expensive or poorly represented. The relatively two-dimensional and localized nature of sea ice makes it an ideal candidate for AI-based emulation, offering a solution to the significant computational burden it imposes on ocean models. Here, we present a sea ice emulator for the Finite Element Sea Ice-Ocean Model (FESOM), capable of predicting the evolution of sea ice thickness (SIT), concentration (SIC), and drift (SID) on timescales ranging from weeks to months.

First, an adaptive U-Net-based model is trained to predict sea ice state (SIT, SIC, SID) increments at one (or multiple) lead times ahead, using corresponding atmospheric forcing and past and current sea ice states. The model is driven by multi-decadal series of daily to sub-daily atmospheric forcing and 2D sea ice and ocean outputs from FESOM, which have been preprocessed and re-interpolated onto a regular grid. To ensure scalability, training sequences are divided into chunks, managed by a custom mapper that balances their usage during training and supports compatibility with multi-GPU configurations. The model is trained by minimizing a penalized mean square error loss function, with an adaptive learning rate controlled via a dedicated scheduler, until convergence. The quality and accuracy of the training process are systematically assessed prior to inference.

Emulation of sea ice can then be performed using recursive inference of the trained models for rollouts spanning from some weeks to a year. Subsequent sea ice states are occasionally clipped into their physical range in order to prevent non-physical behaviors. Rolling predictions can be eventually generated daily or weekly along the test sequences, similar to operational forecasting.

Apart from SIT, SIC, and SID maps, metrics including Integrated Ice Edge Error, root mean square error, mean ice thickness, and Sea Ice Extent are implemented to evaluate the quality of the prediction in comparison to the actual FESOM outputs and some predefined baselines. The emulator demonstrates robust predictions up to 100 days, while still maintaining a realistic representation of various sea ice states beyond this time. Both training and inference are scalable and have been deployed on GPUs, although rolling predictions can be run on a single CPU without incurring prohibitive costs. Computation times for both steps will be estimated, along with the time required for a standard FESOM simulation including sea ice, to assess the potential gain in computational efficiency.

How to cite: Birrien, F., Hutter, N., and Koldunov, N.: On emulating Sea Ice in the Finite Element Sea Ice-Ocean Model (FESOM), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18155, https://doi.org/10.5194/egusphere-egu25-18155, 2025.

EGU25-18799 | ECS | Posters on site | CR6.8

Leveraging Self-Supervised Learning for Sea Ice Segmentation in the Arctic to Reduce on Labelling 

Jacob Seston, William D. Harcourt, Georgios Leontidis, Brice Rea, Matteo Spagnolo, and Lauren McWhinnie

The rapid decline of Arctic sea ice driven by climate change poses significant challenges and opportunities for global shipping, ecosystems, and coastal communities. Understanding and mapping sea ice variability is crucial for assessing its implications on navigability and ensuring maritime safety in this dynamic region. One of the most significant challenges in applying machine learning (ML) to cryospheric sciences is the reliance on large quantities of human-labelled data, which is both costly and time-intensive to produce, particularly in remote and harsh environments like the Arctic. This contribution addresses this challenge by leveraging self-supervised learning (SSL) techniques and Convolutional Neural Network (CNN) to reduce the dependency on labelled data while maintaining high levels of model performance. We used the well-known UNet model, a CNN designed for pixel-wise segmentation tasks, and integrate BYOL (Bootstrap Your Own Latent), an SSL technique that leverages unlabelled data to learn features without requiring explicit labels. BYOL trains the model to match representations of the same image under different transformations, allowing it to learn useful features from unlabelled data without needing explicit labels.

We apply these models to Sentinel-1 SAR imagery in the Canadian Arctic Archipelago, a region of critical importance due to its role in global shipping routes, where sea ice variability directly impacts navigability and maritime safety.

We created binary ice and open water labels to serve as a benchmark for evaluating model performance. Early preliminary results suggest that using BYOL reduces the labelling requirement by approximately 50% compared to models trained without self-supervised pretraining. By pretraining the UNet model on unlabelled Sentinel-1 SAR imagery and fine-tuning it for sea ice segmentation, this approach demonstrates how leveraging unlabelled data can significantly minimise the need for human annotation while maintaining robust segmentation accuracy. These methods optimise the use of limited labelled datasets, enabling efficient and scalable models that potentially generalise to sea ice segmentation tasks where high-quality labels are often scarce or imprecise. These techniques enhance the adaptability of ML models, allowing them to be applied to new datasets and tasks with minimal retraining, further reducing the computational and data requirements. By reducing reliance on labelled data, this approach improves efficiency and opens up possibilities for tackling broader challenges, such as real-time ice monitoring, assessing shipping route viability, and conducting long-term trend analysis.

How to cite: Seston, J., Harcourt, W. D., Leontidis, G., Rea, B., Spagnolo, M., and McWhinnie, L.: Leveraging Self-Supervised Learning for Sea Ice Segmentation in the Arctic to Reduce on Labelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18799, https://doi.org/10.5194/egusphere-egu25-18799, 2025.

ESSI2 – Data, Software and Computing Infrastructures across Earth and Space Sciences

EGU25-1298 | Posters on site | ESSI2.3

The HAWK: Platform for Earth Observation Data Processing 

Julien Petiton, Gouillon Flavien, Simeon Mathilde, and Marie-Laure Frery

The CNES (Centre National d'Études Spatiales) has developed the HAWK [1] (High resolution Altimetry WorKspace) framework as part of its expertise in supporting Earth observation missions. This platform, actually tailored for the SWOT [2] (Surface Water and Ocean Topography) mission, operates on the CNES High-Performance Computing (HPC) infrastructure and is strongly linked with the SWOT routine mission center.

 

HAWK provides a collaborative environment for hydrological and oceanographic data analysis, enabling remote access to extensive datasets without local downloads. Currently operational, the platform supports SWOT expertise activities and offers plethora of user tutorials [3].

 

Future developments aim to generalize HAWK for use with multiple Earth observation missions, expanding its accessibility and impact within the scientific community to promote EO data. We also have for objective to offer open training session to the communities.

 

References

 

[1] SWOT Mission, https://cnes.fr/projets/swot

[2] high-resolution-altimetry-workspace-hawk, HIGH RESOLUTION ALTIMETRY WORKSPACE (HAWK) – GEODES

[3] Microwave expertise center: Providing an efficient framework for microwave data exploration,   https://www.proceedings.com/76012.html (IEEE Catalog No.: CFP24IGA-USB    ISBN: 979-8-3503-6031-8)

How to cite: Petiton, J., Flavien, G., Mathilde, S., and Frery, M.-L.: The HAWK: Platform for Earth Observation Data Processing, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1298, https://doi.org/10.5194/egusphere-egu25-1298, 2025.

Research into the effects of changing Arctic climate has been limited by large-scale in situ data availability because of the remoteness and harsh climate of the Arctic. Only recently, therefore, are hourly-to daily measurements covering most Arctic regions publicly available, but scattered in various local databases.

In this project, we obtained in situ weather data from all major Arctic regions from publicly available data sources across the Arctic with focus on the period 1990-2023. The data set, which contains 719 unique locations from 14 data sources and covers all Arctic regions, has been restructured and -formatted into a standardized data format, combined with metadata about location and elevation. It was further quality checked by running it through five optional modules of increasingly user-involved judgement-based checks. We supply the code involved in import and standardization, and the modular quality check, as well as the standardized, but unchecked data set, and the final, quality checked, data set.

The data set has the potential to benefit pan-Arctic in situ research opportunities as e.g. validation and ground truthing of modelling efforts.

How to cite: Rasmussen, L. H., Markussen, B., and Ditlevsen, S.: Compiling, normalizing and quality checking a pan-Arctic dataset og in situ weather observations from 1990-2023 collected from publicly available data sources, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3749, https://doi.org/10.5194/egusphere-egu25-3749, 2025.

EGU25-4993 | Orals | ESSI2.3

Blue-Cloud 2026 project - Deploying BEACON data lakes for harmonizing ocean data access for Virtual Research Environments 

Dick M. A. Schaap, Peter Thijsse, Tjerk Krijger, and Robin Kooyman

In order to provide users with fast and easy access to multidisciplinary data originating from large collections, MARIS has developed a software system called BEACON that can, on the fly with high performance, extract specific data based on the user’s request. This software has been customised and deployed in the Blue-Cloud2026 project and several other European projects and is designed to return one single harmonised file as output, regardless of whether the input contains different data types. In January 2025, BEACON 1.0.0 was made publicly available as an open-source software, allowing everyone to set-up their own BEACON ‘node’ to enhance the access to their data or use existing BEACON nodes from well-known data infrastructures such as Euro-Argo or the World Ocean Database for fast and easy access to harmonized data subsets. More technical details, example applications and general information on BEACON can be found on the website https://beacon.maris.nl/.

Within the context of Blue-Cloud2026, BEACON is deployed to provide access to harmonised subsets from Blue Data Infrastructures for the WorkBenches (WB) that aim to generate harmonised and validated data collections of Essential Ocean Variables (EOVs). To this end a set of monolithic BEACON nodes were set-up for relevant data collections such as the WOD, CMEMS Cora, Euro-Argo and more. Developments are well underway for parallel deployment of these BEACON instances and related notebooks at the D4Science e-infrastructure as part of the Blue-Cloud VRE, giving access to all users registered as Blue-Cloud users. 

Going one step further, the output from multiple monolithic BEACON instances are combined into one merged BEACON node for each WB. Work is ongoing for a structural mapping from each monolithic BEACON to the target Common Metadata Profile as defined by the WB teams. These mappings will be used in the BEACON queries to retrieve and load contents ‘as-is’ from monolithic BEACON instances into the merged BEACON instances, giving a common structure for variables, units, values, quality flags, and common metadata profile fields. The structured metadata and data will be supplemented by additional metadata data as available for each of the monolithic BEACON instances.

This presentation will cover an introduction of the Blue-Cloud 2026 project and  developments of the merged BEACON nodes, explaining how it can practically serve as data lakes for many VRE applications and how it is extendable to other domains. By using examples from the WBs, the reduction in time and effort spent for the researchers to collect the data are highlighted. 

How to cite: Schaap, D. M. A., Thijsse, P., Krijger, T., and Kooyman, R.: Blue-Cloud 2026 project - Deploying BEACON data lakes for harmonizing ocean data access for Virtual Research Environments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4993, https://doi.org/10.5194/egusphere-egu25-4993, 2025.

EGU25-7885 | ECS | Orals | ESSI2.3

The Isotope Virtual Research Environment developed within ITINERIS Project: Isotope Studio 

Paolo Di Giuseppe, Simona Gennaro, Erico Perrone, Samuele Agostini, Irene Tunno, Eugenio Trumpy, Andrea Rielli, Ilaria Baneschi, Chiara Boschi, Irene Cornacchia, Maddalena Pennisi, Matteo Salvadori, Eleonora Regattieri, Simone Vezzoni, Andrea Dini, and Antonello Provenzale

Applications of conventional isotopes (e.g., H, O, C, N, and S), as well as non-conventional (e.g., B, Li, Fe, Cu, Zn, and Mg) and radiogenic isotopes (e.g., Sr, Nd, and Pb), offer unique opportunities to evaluate deep geological processes, environmental processes and their interactions within the Critical Zone (CZ). Human-induced climate change represents one of the most pressing environmental challenges of the twenty-first century; in the light of this, isotopic composition analysis provides an effective means to investigate it. Numerous studies have highlighted the fundamental role of isotope geochemistry in understanding environmental systems to critical zone processes, and the volume of research in this field continues to grow. Considering this, a comprehensive inventory of stable and radiogenic isotopes has become essential for tracking processes involving fluids, minerals, rock evolution, and origin, as well as examining interactions in soils, plants, and other reservoirs. Currently, data and information are unevenly distributed across various sources and institutions, leading to challenges in data recovery and integration. To address this gap, the ITINERIS Project (PNRR) has initiated Work Package 8.9, which focuses on developing the ISOTOPE Virtual Research Environment (VRE). A VRE can be described as an online environment offering remote and shareable disk space (workspace), catalogues and several customized tools for data processing. This initiative represents a pioneering step toward establishing Italy's first comprehensive national VRE service, encompassing a national database on stable isotopes. The Isotope VRE integrates tools for data analysis, interpretation, and modelling, enabling researchers and stakeholders to access coordinated information and advanced analytical tools. Some examples of data modelling are here reported: i) data plotting; ii) ternary diagrams; iii) mixing models. Initial results demonstrate the significant potential of the Isotope VRE in advancing our understanding of Earth system processes. Additional mathematical modelling approaches are under development, further enhancing the platform's capabilities. The Isotope VRE aims to provide the scientific community with a comprehensive virtual research environment for isotopic data sharing, analysis, and interpretation. This platform will empower researchers to investigate environmental processes with a suite of powerful tools, fostering new insights and applications in geochemistry and beyond.

How to cite: Di Giuseppe, P., Gennaro, S., Perrone, E., Agostini, S., Tunno, I., Trumpy, E., Rielli, A., Baneschi, I., Boschi, C., Cornacchia, I., Pennisi, M., Salvadori, M., Regattieri, E., Vezzoni, S., Dini, A., and Provenzale, A.: The Isotope Virtual Research Environment developed within ITINERIS Project: Isotope Studio, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7885, https://doi.org/10.5194/egusphere-egu25-7885, 2025.

EGU25-9099 | Orals | ESSI2.3

The ENVRI-Hub: Advancing Multidisciplinary Collaboration and FAIR Data Integration in Environmental Research 

Ulrich Bundke, Angeliki Adamaki, Daniele Bailo, Magdalena Brus, Claudio Dema, Dario De Nart, Federico Drago, Marta Gutierrez David, Anca Hienola, Andreas Petzold, Alex Vermeulen, and Zhiming Zhao

Addressing complex environmental and climate challenges requires integrated approaches that connect data, services, and data analysis and management tools across disciplines. The ENVRI-Hub represents a transformative success story in fostering such integration. As the central gateway of the European Environmental Research Infrastructures (ENVRIs), the ENVRI-Hub also bridges disciplinary boundaries by enabling seamless access to interoperable datasets and web services across the Earth system domains - atmosphere, marine, ecosystems, and solid earth. The ENVRI-Hub acts as the CLuster Open Science Competence Centre (CLOCC) for the European ENVRIs

Cluster, offering a virtual hub dedicated to fostering research excellence through training and knowledge transfer. 

The ENVRI-Hub serves as a gateway for researchers to find, access, and use high-quality, FAIR (Findable, Accessible, Interoperable, and Reusable) data tailored for multi- and inter-disciplinary studies. Its Virtual Research Environments (VREs) will allow users to conduct scientific analysis directly within the hub, using datasets related to variables essential for climate and environmental studies, promoting efficiency and reproducibility. By fostering the provision of open data, coupled with advanced computational tools to e.g. process big data and efficiently operate on cloud services, the ENVRI-Hub empowers researchers to develop innovative methodologies and accelerate progress in climate science and environmental monitoring.

This presentation will highlight the technical underpinnings of the ENVRI-Hub that enable machine-to-machine (M2M) communication and interoperability, fostering collaborations between data providers, scientists, and e-infrastructures. We will showcase examples that demonstrate how the ENVRI-Hub can catalyse interdisciplinary research, enhance data integration, and support the development of climate and environmental models.

How to cite: Bundke, U., Adamaki, A., Bailo, D., Brus, M., Dema, C., De Nart, D., Drago, F., Gutierrez David, M., Hienola, A., Petzold, A., Vermeulen, A., and Zhao, Z.: The ENVRI-Hub: Advancing Multidisciplinary Collaboration and FAIR Data Integration in Environmental Research, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9099, https://doi.org/10.5194/egusphere-egu25-9099, 2025.

EGU25-9682 | Posters on site | ESSI2.3

The AquaINFRA research data infrastructure: Knowledge generation through FAIR open data and reproducible computational workflows 

Markus Konkol, Simon Jirka, Henning Sten Hansen, Kaori Otsu, Sami Domisch, Merret Buurman, Vanessa Bremerich, Astra Labuce, Pekka Latvala, Juha Oksanen, and Björn Grüning

The primary aim of the EU-funded AquaINFRA [1] initiative is the development of a research data infrastructure to help marine and freshwater scientists generate new knowledge for restoring healthy oceans, rivers, and lakes. Several use cases representing pan-Europe, the Baltic Sea, and the North Sea define the scope for implementing the infrastructure and demonstrating its potential. An essential goal of the project is to address high-quality, FAIR, and open multi-disciplinary data. In addition, the infrastructure focuses on making python- and R-based data analyses accessible as reusable tools, services, and workflows. Finally, a specific objective is to develop a platform that is compliant with the European Open Science Cloud (EOSC) Interoperability Framework as an overarching research infrastructure.

In this contribution, we provide an overview of the recent developments in AquaINFRA and show its realization in a use case. The key components of the AquaINFRA research data infrastructure are the Data Discovery and Access Service (DDAS), the AquaINFRA Interaction Platform (AIP), and the Virtual Research Environment (VRE). 

The DDAS [2] is the backend of the infrastructure and based on a federated metadata search mechanism sending requests to selected remote metadata providers on the fly. Access to harmonized metadata is accomplished by leveraging community standards, such as the OGC API family including OGC API Records, OGC API Features, and OGC API Coverages. Another component of the DDAS is the Ontology Search that suggests alternative search terms for the keyword entered by the user. 

The AIP [3] is the central gateway to the project and provides a search interface to find, access, and reuse aquatic digital resources. It is built on top of the DDAS and allows users to search for data, services, and software. Several features are provided to refine the search query, for instance, using a bounding box and selecting specific data providers. 

The purpose of the VRE is to have a set of web-based applications facilitating the reuse of existing tools as well as the contribution of newly developed tools, and connecting them to create readily shareable and reproducible workflows. Hence, the VRE is composed of three modules: 

  • MyBinder is provided as a virtual lab to let users engage with an existing analysis written in R or Python in a pre-defined computational environment.
  • OGC API Processes are provided as a web API service allowing researchers to make remote requests and integrate these into their analysis.
  • Access to the Galaxy platform is provided to create reproducible computational workflows. To achieve that, the analysis scripts developed in the case studies are integrated into the Galaxy platform by wrapping the OGC API Processes as Galaxy tools.

Besides the research data infrastructure demonstrating the usefulness of FAIR open data and reproducible computational workflows, the project outcomes are expected to foster collaboration across borders, data infrastructures, and disciplines.

This project has received funding from the European Commission’s Horizon Europe Research and Innovation programme. Grant agreement No 101094434.

1) Project website: https://aquainfra.eu/ 

2) DDAS: https://vm4072.kaj.pouta.csc.fi/ddas

3) AIP: https://aquainfra.dev.52north.org/ 

How to cite: Konkol, M., Jirka, S., Hansen, H. S., Otsu, K., Domisch, S., Buurman, M., Bremerich, V., Labuce, A., Latvala, P., Oksanen, J., and Grüning, B.: The AquaINFRA research data infrastructure: Knowledge generation through FAIR open data and reproducible computational workflows, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9682, https://doi.org/10.5194/egusphere-egu25-9682, 2025.

EGU25-10262 | Orals | ESSI2.3

Notebook-as-a-VRE (NaaVRE): collaborative virtual labs to build digital twins of ecosystems 

Gabriel Pelouze, Spiros Koulouzis, Koen Greuell, Nafiseh Soveizi, and Zhiming Zhao

Solving many environmental challenges requires connecting a variety of datasets with advanced statistical or AI models, and to access distributed computing resources to create workflows or digital twins. Digital twins are particularly challenging because they require the integration of many datasets and models, resulting in complex interactions. Furthermore, researchers rely on interdisciplinary collaboration to build workflows and digital twins. To that end, they need to discover, reuse, and sometimes modify existing assets. While this can be supported by Virtual Research Environments (VREs), most existing solutions are tailored to a specific domain or use case, making it difficult to integrate external resources or to support the complex model composition required by digital twins.

To address these limitations, we have developed Notebook-as-a-VRE (NaaVRE), a VRE solution built on top of JupyterLab. NaaVRE enables researchers to create computational blocks by containerizing the cells of notebooks, to organize them into workflows, and to manage the full experimental cycle, including data and workflow sharing. The tool includes features such as metadata-driven resource discovery, workflow automation, and compatibility with external repositories. Designed for cloud infrastructures, NaaVRE provides cost-efficient and scalable solutions to support digital twin development.

Using NaaVRE, we build customized virtual labs to address specific scientific problems. These gather models, data access tools, workflows, and documentation into a shared space, where a community of users can develop, share and reuse them. We present a framework to guide the development and operations of virtual labs throughout their entire lifecycle.

We showcase NaaVRE and by building customized virtual labs for scientific data processing workflows and prototype digital twins. Those virtual labs are managed by LifeWatch ERIC and the University of Amsterdam following the aforementioned development framework. They include characterizing ecosystem structures with LiDAR data, tracking bird migration using radar, mapping invasive species, deriving essential variables in the context of the ENVRI-Hub NEXT project, and developing digital twins of ecosystems in the context of the Dutch NWO LTER-LIFE project.

How to cite: Pelouze, G., Koulouzis, S., Greuell, K., Soveizi, N., and Zhao, Z.: Notebook-as-a-VRE (NaaVRE): collaborative virtual labs to build digital twins of ecosystems, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10262, https://doi.org/10.5194/egusphere-egu25-10262, 2025.

EGU25-11023 | Orals | ESSI2.3

Jupyter Notebooks in European Plate Observing System (EPOS) 

Jan Michalek, Kety Giuliacci, Alessandro Spinuso, Luca Trani, Daniele Bailo, Rossana Paciello, Ian Neut van der, Joanna Kocot, and Dedalo Marchetti and the EPOS IT Team

The European Plate Observing System (EPOS) addresses the problem of homogeneous access to heterogeneous digital assets in geoscience within the European tectonic plate. EPOS is a European Research Infrastructure Consortium (ERIC) since 2018, with the goal of building long-term and sustainable infrastructure for solid Earth science. The EPOS Data Portal was launched into the operational phase in April 2023 and is introducing new ways for cross-disciplinary research, especially for data discovery. Currently the EPOS Data Portal, a metadata and semantic-driven system for integrating Data, Software and services,  provides access to data and data products from ten different geoscientific areas: Seismology, Near Fault Observatories, GNSS Data and Products, Volcano Observations, Satellite Data, Geomagnetic Observations, Anthropogenic Hazards, Geological Information and Modelling, Multi-scale laboratories and Tsunami Research. The presentation shows the achievements of the EPOS community, demonstrates features of the Portal user interface and also the underlying architecture of the whole system and online processing environment. The IT system as such is shared as Open Source and is under continuous development and improvement, following the SDLC methodology: Shape Up.

This presentation focuses on the integration of Jupyter Notebooks into EPOS through Virtual Research Environment (VRE) which allows advanced processing of datasets provided already through EPOS Data Portal. Examples of Jupyter Notebooks covering various scientific multidisciplinary use cases introduce typical data processing workflows and visualizations for efficient use of services collected through EPOS. Expansion of the EPOS-DCAT-AP metadata model by new entities for managing software components and its utilization opens new possibilities for data access and paves the road for integration into other e-infrastructures. 

How to cite: Michalek, J., Giuliacci, K., Spinuso, A., Trani, L., Bailo, D., Paciello, R., Neut van der, I., Kocot, J., and Marchetti, D. and the EPOS IT Team: Jupyter Notebooks in European Plate Observing System (EPOS), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11023, https://doi.org/10.5194/egusphere-egu25-11023, 2025.

EGU25-11500 | Posters on site | ESSI2.3

Radar Aeroecology in the cloud. A Virtual Lab for continental scale Aeroecological analysis 

Berend-Christiaan Wijers, Gabriel Pelouze, Spiros Koulouzis, Koen Greuell, and Zhiming Zhao

The airspace is becoming increasingly crowded. High-rises, wind farms and airports all contribute to conflict with aerial organisms. Information about the movements of organisms in the air is required to identify stop-over sites, migratory routes, and patterns. This can inform mitigation of conflicts by, for example, wind-turbine curtailments or early warning systems for aviation. Weather radars, that continuously monitor the sky across continents, can be used to study movements of birds, bats, and insects. However, for continental scale analysis, large volumes of data are required to be processed and analyzed, which often rely on institute-specific tools and computational resources. This severely hampers collaborative efforts because of the initial investment of time and resources to gain access to existing computing infrastructure.

Here we show a Radar Aeroecology Virtual Lab (RA-VL) which uses the Lifewatch ERIC infrastructure to facilitate collaboration and re-use of infrastructure and tools. By providing RA-VL, we aim to facilitate collaboration between ornithological institutes. This Virtual Lab (VL) will reduce the initial investment of acquiring access and expertise to computational resources and provide immediate access to tools built by domain experts. These tools are then run in the cloud leveraging the performance and flexibility of cloud computing. The VL is shipped with the data management plan used by the University of Amsterdam's Animal Movement Ecology group (UvA IBED-TCE AME) to provide an out of the box solution for managing large datasets. RA-VL is currently capable of accessing, processing, managing and visualizing data from the The Royal Netherlands Meteorological Institute's (RNMI) open Radar Data repository. The VL has multi-language support, and has well known libraries such as bioRad in R and xradar in Python installed.  Furthermore, it uses vol2bird for processing biological echoes found in Polar Volume files to Vertical Profiles. Notebook as a Virtual Research Environment (NaaVRE) provides a user-friendly interface to leverage functionality for parallel processing, data organization and visualization. Over the past four years, the RA-VL has seen a substantial amount of developments. By applying the Readiness Framework, key points were identified and updated to bring this VL from a Development-Version to a Demo-Version. By bringing the VL to the next Readiness Level we hope to receive critical feedback from the community to improve and prepare the lab for the next level.

How to cite: Wijers, B.-C., Pelouze, G., Koulouzis, S., Greuell, K., and Zhao, Z.: Radar Aeroecology in the cloud. A Virtual Lab for continental scale Aeroecological analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11500, https://doi.org/10.5194/egusphere-egu25-11500, 2025.

NSF Unidata is a community-focused data and software facility, funded primarily by the United States National Science Foundation. NSF Unidata’s mission is to provide the data services, tools, and cyberinfrastructure leadership that advance Earth Systems science, enhance educational opportunities, and broaden participation. NSF Unidata’s hallmark has been democratizing access to data and tools by providing open and free access to all its resources.

As the enabler of a broad community, NSF Unidata

 

  • Acquires and distributes data to facilitate Earth system education and research
  • Develops software for accessing, managing, analyzing, visualizing, and effectively using those data
  • Provides comprehensive support to users
  • Conducts training workshops on Unidata software packages
  • Facilitates advancement of standards, conventions, and interoperability
  • Provides leadership in geosciences cyberinfrastructure and fosters technological change
  • Advocates on behalf of the university community on data issues and negotiates data agreements
  • Fosters community interaction and engagement to promote sharing of data, tools, and ideas
  • Grants equipment awards to universities to enable and enhance participation in Unidata

 

An integrated approach that transcends disciplinary and geographic boundaries is needed to understand and address societally important environmental problems such as weather prediction, climate change, and the water cycle. Similarly, an Earth Systems Science approach that employs inquiry-based learning is recommended for teaching geoscience. Advances in the Earth system science are possible only through state-of-the-art, robust, and flexible data and software infrastructure, transparent and seamless access to high-quality data from diverse sources, along with requisite tools and services to analyze, synthesize, visualize, interpret, and use the data effectively.

 

The university community conceived and established the NSF Unidata program more than forty years ago to meet the needs meteorology departments, specifically to acquire and distribute real-time weather data to U.S. universities, together with the necessary tools for data analysis and visualization.

 

While the program’s primary mission of serving the academic community remains unchanged through the years, the user base has broadened, and its activities and portfolio of products and services have grown as the discipline and community needs have evolved.

 

 

Over the past four decades, NSF Unidata has experienced a gradual but natural evolution from a program focused primarily on synoptic scale meteorology to one that serves a broader geosciences community. Unidata has attracted a broader community because it has been successful in providing tools and services that are open, free, interoperable, extensible, and platform independent. The robustness and quality of Unidata tools and services have resulted in their use beyond a community of several dozen universities in the U.S. to now several thousand organizations in academia, research and operational sectors in over 150 countries. In the process, NSF Unidata has matured into a cornerstone, cloud-based data and software facility upon which the Earth system science community and other stakeholders have come to rely.

In this talk, I’ll present NSF Unidata’s evolution over the past forty years, how we are reimagining our future activities in delivering a comprehensive suite of products and services to meet the current and emerging needs of the Earth Systems Science research and education community, and the lessons learned as the facility evolved.

How to cite: Ramamurthy, M.: NSF Unidata: The evolution of a community-focused data and software facility over a four-decade period, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13649, https://doi.org/10.5194/egusphere-egu25-13649, 2025.

EGU25-13888 | ECS | Orals | ESSI2.3

Facilitating the development of Machine Learning-based Digital Twin applications for extreme weather events 

Donatello Elia, Emanuele Donno, Matteo Bunino, Massimiliano Fronza, Davide Donno, Gabriele Padovani, Sandro Fiore, and Andrea Manzi

With the increasing availability of higher-resolution weather and environmental data as well as advances in Machine Learning (ML) algorithms, data-driven approaches have emerged over the last few years as innovative and fast-computing solutions for addressing detection and prediction of extreme weather events, like storms and wildfires. 

Designing, training and deploying ML models is not trivial and can result in a time consuming process. An integrated software infrastructure for supporting and automating the different steps of the workflow, from weather/climate data gathering and preparation, ML model configuration and training, to deployment of the trained model for detection and prediction applications is required. In this regard, solutions for tracking training metrics and provenance information are crucial components for reproducibility of the results. Besides the software components, HPC infrastructures for handling distributed training over multiple GPUs are also needed to speed up the process.

In the context of the EU-funded interTwin project we are implementing ML-powered Digital Twin (DT) applications for the analysis of extreme events (i.e., Tropical Cyclones and wildfires). The interTwin project is designing and developing a generic Digital Twin Engine (DTE) for supporting DTs from different scientific domains. The DTE provides a software and computing infrastructure for simplifying the creation and management of complex DT workflows. 

This contribution, in particular, will present how the interTwin DTE is supporting the different workflow stages, from model training to their execution, of ML-based DT applications for the detection and prediction of extreme events.

How to cite: Elia, D., Donno, E., Bunino, M., Fronza, M., Donno, D., Padovani, G., Fiore, S., and Manzi, A.: Facilitating the development of Machine Learning-based Digital Twin applications for extreme weather events, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13888, https://doi.org/10.5194/egusphere-egu25-13888, 2025.

EGU25-15036 | Orals | ESSI2.3

How can the Complex Citation be implemented in the IPCC AR7 using existing frameworks and interactive notebooks? 

Martina Stockhause, Lina Sitz, Charlotte Pascoe, Deborah Agarwal, James Ayliffe, Justin Buck, Joan Damerow, José Manuel Gutiérrez, Birgit Hassler, Forrest M. Hoffman, Graham Parton, Jared Lewis, Molly McRae, Shelley Stall, and Lesley Wyborn

The introduction of FAIR data practices into the assessment process of the Intergovernmental Panel on Climate Change (IPCC) enhances the transparency of the Assessment Reports (ARs). The approach focuses on the figure creation process and the reproducibility of figures within the reports. The process of creating figures for IPCC reports makes it particularly challenging to credit all contributors. A single figure can incorporate data from observations, models, and publications, originating from numerous sources and generated by multiple authors, each contributing to different parts of the figure. Additionally, it is not always straightforward to include all those involved in the creation process, as the final data (used to produce the figure) often result from assessments and post-processing of other data (input or intermediate data). Among the plans and recommendations for the current Seventh Assessment cycle is the use of the Complex Citation approach (Stockhause et al., 2024), which has been developed in the RDA Complex Citation Working Group (Agarwal et al., 2024).

Some chapters of the ARs use established community frameworks for the generation of their figures. These are currently harmonized in the CMIP Rapid Evaluation Framework (REF) project for the AR7 Fast Track simulations. Others use interactive and shareable Jupyter notebooks. The IPCC needs to support and harmonize both ways of working. Further, the authors’ additional effort to meet the metadata and data requirements needs to be minimized. Therefore, the collaboration with both projects, REF and Jupyter, is planned to offer broad support for the ways of working of a large number of authors. 

This contribution analyzes how the Complex Citation approach, with the aims of giving credit to input providers and enabling the traceability of all digital objects used to create figures, can be implemented in REF and Jupyter and integrated in the AR7 process to support the IPCC’s FAIR and open data approach.

References:
Agarwal, D., Ayliffe, J., J. H. Buck, J., Damerow, J., Parton, G., Stall, S., Stockhause, M., & Wyborn, L. (2024). Complex Citation Working Group Recommendation. Zenodo. https://doi.org/10.5281/zenodo.14106602 
Stockhause M, Huard D, Al Khourdajie A, Gutiérrez JM, Kawamiya M, Klutse NAB, et al. (2024) Implementing FAIR data principles in the IPCC seventh assessment cycle: Lessons learned and future prospects. PLOS Clim 3(12): e0000533. https://doi.org/10.1371/journal.pclm.0000533 

Links:
REF project: https://wcrp-cmip.org/cmip7/rapid-evaluation-framework/ 
Jupyter project: https://jupyter.org/ 

How to cite: Stockhause, M., Sitz, L., Pascoe, C., Agarwal, D., Ayliffe, J., Buck, J., Damerow, J., Gutiérrez, J. M., Hassler, B., Hoffman, F. M., Parton, G., Lewis, J., McRae, M., Stall, S., and Wyborn, L.: How can the Complex Citation be implemented in the IPCC AR7 using existing frameworks and interactive notebooks?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15036, https://doi.org/10.5194/egusphere-egu25-15036, 2025.

EGU25-15633 | Posters on site | ESSI2.3

ATMO-ACCESS: Advancing Atmospheric Research with Integrated cross-RI Virtual Services 

Ariane Dubost, Sabine Philippin, Misha Faber, Cathrine Lund Mhyre, Alex Vermeulen, Valérie Thouret, and Véronique Riffault

ATMO-ACCESS is a research infrastructure pilot project funded under the Horizon 2020 program (2021–2025) that addresses the needs of distributed atmospheric research infrastructures (RIs), including ICOS ERIC (Integrated Carbon Observing System), ACTRIS ERIC (Aerosol, Clouds, and Trace Gases Research Infrastructure), and IAGOS (In-flight Global Observing System). The goal of ATMO-ACCESS is to develop sustainable solutions for access to distributed atmospheric research facilities. Specific activities are directed to provide virtual, physical, remote and hybrid access to users world-wide. 

The project has notably developed innovative online services, leveraging the expertise of these three infrastructures to provide virtual access to advanced digital resources. These services include data archiving, integrated data products, analysis tools, and online training resources, facilitating the integration of several infrastructures.

These initiatives strengthen the ability of scientific communities and stakeholders to effectively exploit the data and tools available to meet the challenges associated with the study of climate, air quality and the atmosphere.



How to cite: Dubost, A., Philippin, S., Faber, M., Lund Mhyre, C., Vermeulen, A., Thouret, V., and Riffault, V.: ATMO-ACCESS: Advancing Atmospheric Research with Integrated cross-RI Virtual Services, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15633, https://doi.org/10.5194/egusphere-egu25-15633, 2025.

EGU25-16205 | Posters on site | ESSI2.3

Empowering Environmental Research Through the EOSC EU Node: FAIR Data and Scalable Services 

Mark Dietrich and Marta Gutierrez

The EOSC EU Node is a key enabler of multidisciplinary and multinational research, promoting the use of Findable, Accessible, Interoperable, and Reusable (FAIR) data principles. As the first European-level node of the EOSC Federation, the platform provides researchers with secure access to interoperable data, computational resources, and managed services tailored for complex research workflows.

This presentation will highlight  environmental and climate science success stories of researchers leveraging EOSC EU Node services and demonstrate the potential of the platform to address challenges in data access, integration, and reuse. Examples include the use of Interactive Notebooks for data visualisation and analysis, and File Sync and Share for secure, collaborative data management across global teams. These use cases demonstrate how the EOSC EU Node fosters seamless integration of data and services, accelerates knowledge sharing, and supports reproducible research.

By bridging thematic and regional infrastructures, the EOSC EU Node is empowering researchers to tackle pressing global challenges, including climate change, biodiversity loss, and disaster risk reduction. The presentation will showcase the platform’s role in driving collaboration, innovation, and impactful research outcomes in the era of open science.

Keywords: FAIR data, climate science, environmental research, EOSC Federation, cloud services, open science.

How to cite: Dietrich, M. and Gutierrez, M.: Empowering Environmental Research Through the EOSC EU Node: FAIR Data and Scalable Services, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16205, https://doi.org/10.5194/egusphere-egu25-16205, 2025.

EGU25-16973 | Posters on site | ESSI2.3

Svalbard Integrated Arctic Earth Observing System, regionally distributed, multinational and multidisciplinary Research Infrastructure 

Heikki Lihavainen, Ilkka Matero, Eleanor Jones, Christiane Hübner, Daan Kivits, and Richard Ashley

Climate warming in Arctic is 3-4 faster than in rest of the globe. Even in Arctic context Svalbard is a hotspot of climate change but also international research collaboration and geopolitics. Svalbard is under Norwegian sovereignty but it has very international research environment based partly on Svalbard Treaty. This creates an ambiguous and continuously changing operational landscape as geopolitical tensions increases.  

Svalbard Integrated Arctic Earth Observing System (SIOS) is an international consortium of 29 research institutions from 10 countries with research interests and infrastructure in and around Svalbard. Mission of SIOS is to make ESS data in and round Svalbard available through the SIOS data management system following FAIR principles, decrease the environmental footprint of science by pooling resources and enabling new technology, sharing infrastructure and facilitating interdisciplinary research collaboration.

SIOS focuses on long-term monitoring of parameters that are important to understand the Arctic in the context of global environmental change. However, the observing system is dynamic and is developed continuously by so called SIOS Science Wheel concept. A State of Environmental Science in Svalbard reports (SESS) is in the core of the Science Wheel. The SESS report is an arena for open sharing of ideas and discussions on measures that should be taken to enable scientists to provide observations needed to gain a comprehensive view of the Earth System of Svalbard and the Arctic in general. The report summarises the state of current knowledge of key Earth System Science parameters and analyses how these parameters influence one another. It combines the long-term monitoring data that form the core of the observing system with innovative monitoring and research.

SIOS Access program has been developed to foster excellent Arctic science and to facilitate research infrastructures to function at ideal capacity. SIOS has built its own training programs which for example help field scientists to utilize different level of remote sensing in their research and planning field campaigns.

In this presentation we will share our experiences on building SIOS, multinational and multidisciplinary research infrastructure, Science wheel success stories, challenges and lessons learned and way forward towards IPY 2032-2033.

How to cite: Lihavainen, H., Matero, I., Jones, E., Hübner, C., Kivits, D., and Ashley, R.: Svalbard Integrated Arctic Earth Observing System, regionally distributed, multinational and multidisciplinary Research Infrastructure, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16973, https://doi.org/10.5194/egusphere-egu25-16973, 2025.

EGU25-17213 | Posters on site | ESSI2.3

Downstream VRE for multidisciplinary applications: Land and Marine domain toolboxes. 

Rachele Franceschini, Catalina Reyes, Alessandro Altenburge, Giuliana Rossi, and Alessandra Giorgetti

As part of the ITINERIS project, funded by the NextGenerationEU programme (2022-2025), the downstream effects of climate and environmental change are being investigated. The Virtual Research Environment VRE (Assante et al., 2021) on downstream impacts of environmental change is dedicated to the use of Research Infrastructures providing tools for the visualisation, analysis, and data sharing. The Downstream VRE hosted by the D4Science infrastructure (Assante et al., 2019) and toolboxes have been developed for the marine and terrestrial domains.

The marine domain toolbox will take advantage of the available data in order to generate an integrated dataset for temperature, salinity, pH and CO2 in the gulf of Trieste (Italy) with a focus on the National Institute of Oceanography and Applied Geophysics – OGS data for the last 10 years as use case. Once the data is harvested a subsequent data validation, quality control and merging will be performed using erddap-navigator, a web application that allows the user to visualize, assign quality control flags, analyze and merge data. The integrated dataset will then be used to calculate climate change indicators, such as ocean acidification and ocean carbon cycle budget.

The land domain toolbox aims to analyse areas subject to hydrogeological hazards (landslide phenomena). In this context, a geoserver and a geonetwork have been implemented containing maps at regional level of the NE-Italy region of Friuli Venezia Giulia. At the local level, different monitoring systems (interferometric radar, 1 GPS, 2 extensometers and 2 inclinometers, date coordinator) were installed at the Passo della Morte in Forni di Sotto to detect possible ground instabilities. The instruments provide geodata in various formats that define trends in slope displacement through interpretation, e.g. of time series. Each product has its own description and can be downloaded.

 

Acknowledgements

The work has been funded by EU - Next Generation EU Mission 4 “Education and Research” - Component 2: “From research to business” - Investment 3.1: “Fund for the realisation of an integrated system of research and innovation infrastructures” - Project IR0000032 – ITINERIS - Italian Integrated Environmental Research Infrastructures System - CUP B53C22002150006.

The authors acknowledge the Research Infrastructures participating in the ITINERIS project with their Italian nodes: ACTRIS, ANAEE, ATLaS, CeTRA, DANUBIUS, DISSCO, e-LTER, ECORD, EMPHASIS, EMSO ,EUFAR ,Euro-Argo, EuroFleets, Geoscience, IBISBA, ICOS, JERICO, LIFEWATCH, LNS, N/R Laura Bassi, SIOS, SMINO.

 

References

Assante, M., Candela, L., Castelli, D., Cirillo, R., Coro, G., Frosini, L., Lelii, L., Mangiacrapa, F., Pagano, P., Panichi, G., & Sinibaldi, F. (2019). Enacting open science by D4Science. Future Generation Computer Systems, 101, 555–563. https://doi.org/10.1016/j.future.2019.05.063https://doi.org/10.5281/ZENODO.10070443

Assante, M., Candela, L., & Pagano, P. (2021). Blue-Cloud D4.4 Blue Cloud VRE Common Facilities (Release 2). https://doi.org/10.5281/ZENODO.10070443

How to cite: Franceschini, R., Reyes, C., Altenburge, A., Rossi, G., and Giorgetti, A.: Downstream VRE for multidisciplinary applications: Land and Marine domain toolboxes., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17213, https://doi.org/10.5194/egusphere-egu25-17213, 2025.

EGU25-17240 | Posters on site | ESSI2.3

Implementation of a Virtual Research Environment (VRE) to the study of forest environments 

Sergi Costafreda-Aumedes, Paolo Tagliolato, Riccardo Giusti, Maurizio Iannuccilli, Giorgio Matteucci, Francesco Mazzenga, Alessandro Messeri, and Alessandro Oggioni

The monitoring and study of forests is essential to understand their condition, dynamics and to adopt optimal management to ensure their sustainability. Long-term studies require the installation of permanent field sites with sensors and equipment and collection of data, also from other sources, which are often difficult to obtain both in terms of discovery and standardization.

In this framework, Virtual Research Environments (VRE) are online research platforms that allow easy access to the available FAIR data, to find smart solutions and to support decision making. The Virtual Research Environment for Essential Variables (VRE-EVs), created within the ITINERIS NextGeneration EU project (https://itineris.cnr.it) and hosted on the D4Science infrastructure (https://itineris.d4science.org), offers several services to registered users to develop open and reproducible science. The VRE-EV aims to enable virtual environmental research in the perspective of the two global frameworks of Essential Biodiversity Variables (EBVs; e.g., phenology, species distribution) and Essential Climate Variables (ECVs; e.g., surface air temperature, precipitation, relative humidity), which are known to be critical for plant and soil biogeochemical processes (e.g., tree growth, soil mineralization, water fluxes, litter decomposition). This is very interesting in experimental sites with long-term monitoring, such as that installed in the 3000-ha forest environment of Collelongo - Selva Piana (https://deims.org/9b1d144a-dc37-4b0e-8cda-1dda1d7667da), one of the founding sites of the italian ICP Forests network and also part of the eLTER and AnaEE international research infrastructures. The main study site is a pure mature beech forest (Fagus sylvatica L.) with trees over 125 years old.   This study describes the VRE-EVs and demonstrates, through a use case, how data from heterogeneous sources, made easily accessible within the VRE-EVs, are useful to analyse Forest environments Essential Variables. 

We propose the use of an interactive application (Shiny App) specifically developed within the VRE-EVs, an RStudio platform, to integrate the functions provided by the ReLTER package (10.1016/j.ecoinf.2024.102915) and by all other ITINERIS project facilities, with the aim of merging different datasets available in European repositories (e.g., Copernicus Land and Climate Services, European Environment Agency), international data publishers (e.g., Pangaea, Zenodo), other essential variables online repositories, and in-situ data. The joint analysis of the different datasets available through the VRE-EV allows the improvement of ecological, ecophysiological processes and carbon fluxes of the Collelongo beech forest in response to global changes.

How to cite: Costafreda-Aumedes, S., Tagliolato, P., Giusti, R., Iannuccilli, M., Matteucci, G., Mazzenga, F., Messeri, A., and Oggioni, A.: Implementation of a Virtual Research Environment (VRE) to the study of forest environments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17240, https://doi.org/10.5194/egusphere-egu25-17240, 2025.

EGU25-18357 | Posters on site | ESSI2.3

Use of Geospatial tools and techniques for enhancing capacity for climate change-related planning and management in Malta 

Christopher Gauci, Abdal Belaama, Daniel Fenech, Josianne Vassallo, George Buhagiar, and Emanuele Colica

Malta, as a small island state, faces increasing challenges from climate change due to its vulnerability to climate impacts. This study investigates the application of geospatial tools and techniques to enhance Malta’s capacity for climate change-related planning and management, aligning with Sustainable Development Goal (SDG) 13: Climate Action. The methodology integrates historic cartographic resources, as detailed in our previous work (Tranchant et al., 2024), with contemporary approaches such as UAV photogrammetry and dataset comparisons using software like Cloud Compare. These datasets are augmented by ground-truthing data, including geotechnical monitoring via tilt plates and ground monitoring nails - both deliverables from a previous project, Coastal Satellite-Assisted Governance (SAGE). The collected data will be compiled into a unified geodatabase to enhance disaster risk reduction efforts through real-time monitoring of climate-enhanced risk levels. The tools and insights, where permissible, will be shared with stakeholders beyond government and academia to promote education and public awareness. While the study does not directly aim to mitigate the effects of climate change, it strengthens the Maltese government’s capacity to proactively evaluate and respond to its impacts, particularly with respect to coastal dynamics. Future efforts will focus on developing an open-source, WebGIS-based A-DiNSAR monitoring system for ground deformation. This system aims to replicate the Copernicus European Ground Motion Service while leveraging higher-resolution datasets to achieve greater precision at localized scales, such as monitoring ground movement and infrastructural stability in cliffside and coastal zones. By addressing areas most susceptible to ground movement and stability issues due to climate change, the study enhances Malta’s resilience to climate impacts, aligning with the objectives of SDG 13: Climate Action.

How to cite: Gauci, C., Belaama, A., Fenech, D., Vassallo, J., Buhagiar, G., and Colica, E.: Use of Geospatial tools and techniques for enhancing capacity for climate change-related planning and management in Malta, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18357, https://doi.org/10.5194/egusphere-egu25-18357, 2025.

EGU25-18524 | Orals | ESSI2.3

CCP: A Cloud Computing Platform for VREs in Earth Sciences 

Alfredo Oliviero, Marco Lettere, Andrea Dell'Amico, and Pasquale Pagano

The Cloud Computing Platform (CCP) was developed under the aegis of D4Science [1]. D4Science is an operational digital infrastructure co-funded by the European Commission, and represents a significant advancement in supporting the FAIR (Findable, Accessible, Interoperable, and Reusable) principles, open science, and reproducible data-intensive science. D4Science has evolved to harness the "as a Service" paradigm, offering web-accessible Virtual Research Environments (VREs) [2] that have also been instrumental in facilitating science collaborations [3] with a particular focus on Earth observation, Earth science, and marine and agricultural environments. These environments simplify access to datasets while concealing underlying complexities, and include functionalities such as a cloud-based workspace for file organisation, a platform for large-scale data analysis, a catalogue for publishing research results, and a communication system rooted in social networking practices.

A key component for enabling large-scale, affordable, and reproducible computation and data analysis is CCP: a cloud computing platform specifically designed for VREs and Open Science.

CCP enables researchers to import, execute, and share methods ranging from statistical analysis to image classification, from AI models to 3D reconstruction, from data format conversion to pattern searching in DNA sequences, while embodying FAIR (Findable, Accessible, Interoperable, and Reusable) principles.

By leveraging container technology, an API-based design, and adherence to standards such as the OGC Processes API [4], CCP supports high interoperability, flexibility and integrability in scientific workflows. Methods can be written in any programming language (Python, Julia, R, etc) and executed either via dedicated web UIs or programmatically from virtually any development environment (command line, custom applications, Galaxy workflows, Jupyter notebooks, RStudio, etc). Code generators are provided to ease the integration into common scientific tools.

CCP can be deployed on container orchestration platforms, such as Docker Swarm or Kubernetes, which can leverage specialized hardware configurations (e.g., HPC clusters or GPU-enabled nodes) depending on the policies and resources available, thereby offering flexible and scalable computational environments per the needs of each community.

Automatic provenance management captures the complete history of a method's execution for reproducibility and accountability, according to common provenance models (Prov-O, RO-crate).  Re-submitting executions can be as simple as clicking on a shared link.

CCP has been integrated into several VREs, many related to Earth science including the Blue-Cloud [5] virtual laboratories and demonstrators and ITINERIS [6].

Keywords:
Virtual Research Environments, Open Science, Cloud Computing, FAIR, Provenance Management, Data Analysis
 
Acknowledgements:
This work is supported by the European Community’s HEU Program under the scheme ‘HORIZON-INFRA-2022-EOSC-01’, grant agreement #101094227 ‘Blue-Cloud 2026: A federated European FAIR and Open Research Ecosystem for oceans, seas, coastal and inland waters [5]’
 

References

1. M. Assante et al. (2019) “Enacting open science by D4Science”. Future Gener. Comput. Syst. 101: 555-563 10.1016/j.future.2019.05.063

2. L. Candela, D. Castelli and P. Pagano(2023) “The D4Science Experience on Virtual Research Environments Development". in IEEE Computing in Science & Engineering, doi: 10.1109/MCSE.2023.3290433

3. M. Assante et al. (2023) “Virtual research environments co-creation: The D4Science experience”. Concurrency Computat Pract Exper. 2023; 35(18):e6925. doi:10.1002/cpe.6925

4. https://ogcapi.ogc.org/processes/

5. https://cordis.europa.eu/project/id/101094227

6. ITINERIS, Italian Integrated Environmental Research Infrastructures System, Funded by EU - Next Generation EU PNRR- Mission 4 “Education and Research” https://itineris.cnr.it/

How to cite: Oliviero, A., Lettere, M., Dell'Amico, A., and Pagano, P.: CCP: A Cloud Computing Platform for VREs in Earth Sciences, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18524, https://doi.org/10.5194/egusphere-egu25-18524, 2025.

EGU25-19176 | ECS | Posters on site | ESSI2.3

Virtual Research Environment for studying Critical Zone (CZ) dynamics and isotope fingerprinting: an approach for geoscientific data management and modeling within the ITINERIS project (PNRR, Italy). 

Simona Gennaro, Pasquale Bove, Francesca Caparrini, Paolo Di Giuseppe, Silvio Marta, Erico Perrone, Andrea Dini, Samuele Agostini, Eugenio Trumpy, and Antonello Provenzale

Recent advancements in laboratory implementation and instrumentation have led to the generation of increasingly abundant data, with the need for greater collaboration and data sharing. To address this challenge, the expansion of e-infrastructures and the development of Virtual Research Environments (VREs) have become essential.

VREs provide an integrated ecosystem for data collection, analysis, and publication, following the Open Science principles, of transparency, inclusion, integrity, and collaboration. VREs are based on the D4Science e-infrastructure, which promotes collaboration and cooperative work among the scientific communities and the stakeholders identified by the researchers.

In the framework of the ITINERIS Project, the new comprehensive Italian Research Infrastructures (RIs) hub in the geoscientific and environmental fields, several multidisciplinary teams are developing thematic VREs for studying the entire Earth System, by combining field and lab measurements, data analysis, and modelling tools across all the environment domains.

Among the VREs, the “Critical Zone (CZ) VRE” is specifically designed for collecting datasets and information from the Critical Zone Observatories (CZOs, active in Italy and abroad) primarily managed by Italian research teams and including tools for data visualization and analysis, as well as models useful for studying the complex dynamics of the Critical Zone.

The Critical Zone represents the thin layer between the unweathered bedrock and the top of the vegetation canopy, where “rock meets life”. It includes rocks, soil, water, microbiota, vegetation and fauna, along with the services they provide to humankind and all the processes supporting terrestrial ecosystems and the soil-vegetation-atmosphere interactions.

D4Science-enabled Critical Zone VRE offers a set of tools supporting all the steps of the research lifecycle, from data collection to data analysis, and visualization. Data collection and dataset assembly are fostered by the Collaborative Storage Framework, which promotes teamwork among users and offers a collaborative space to share digital objects. For data analysis, the Critical Zone VRE is equipped with an Analytics Engine Framework, which includes Cloud Computing Platforms (CCPs), as well as the DataMiner. Additionally, the Critical Zone VRE is equipped with RStudio 4 and JupyterLab. These tools enable the development of specific codes (in various free-license programming languages) and models that can be launched directly from the VRE to analyze and visualize data. Data publishing of research outcomes is also facilitated by the development of metadata and spatial data catalogues. In particular, the catalogues help to organize and make research outcomes available to the broader scientific and multidisciplinary community.

Further improvements in studying Critical Zone components and dynamics are essential, and in this case, valuable support can be gained through the interaction between the Critical Zone VRE and other VREs. An example is the interaction with the Isotope VRE, which contains a dedicated web application for analysis and modelling (called “Isotope Studio”) and aims to represent the first Italian database on environmental isotopes, allowing researchers and environmental managers to interpret and model bio-geochemical processes in the framework of the Environmental Sciences.

How to cite: Gennaro, S., Bove, P., Caparrini, F., Di Giuseppe, P., Marta, S., Perrone, E., Dini, A., Agostini, S., Trumpy, E., and Provenzale, A.: Virtual Research Environment for studying Critical Zone (CZ) dynamics and isotope fingerprinting: an approach for geoscientific data management and modeling within the ITINERIS project (PNRR, Italy)., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19176, https://doi.org/10.5194/egusphere-egu25-19176, 2025.

EGU25-19342 | Orals | ESSI2.3

Data exchange and integration: Use cases for EMODnet Chemistry 

Alessandra Giorgetti, Chiara Altobelli, Dick M.A. Schaap, Luminita Buga, Lotta Fyrberg, Julie Gatti, Neil Holdsworth, Athanasia Iona, Martin Mørk Larsen, Reiner Schlitzer, Ann Kristin Østrem, Marilena Tsompanou, Charles Troupin, and Karin Wesslander

The European Marine Observation and Data Network (EMODnet) is the gateway to multidisciplinary marine in-situ data and data products and has been funded by the European Commission for 15 years. In particular, EMODnet Chemistry collects and makes freely available nearly 1,300,000 million metadata entries and related datasets on seawater quality in various matrices.The aim of this contribution to the EGU is to illustrate this long-term initiative of the European Commission and its wealth of use cases. These show how data fairness can enable faster and more accurate modelling and solutions to pressing global environmental emergencies. 
The EMODnet Chemistry data infrastructure supports the development of evidence-based knowledge on eutrophication, ocean acidification and contaminants, including marine litter. The measurement data are accessible via a data discovery and access service and are regularly aggregated, harmonised and validated to create thematic data collections and associated data products. Subsets of the data collections can be downloaded via the webODV explorer and extractor tool, which also allows users to create customisable data analyses and visualisations. The functioning of EMODnet Chemistry relies heavily on SeaDataNet: a pan-European marine data management infrastructure involving 110 national oceanographic data centres, which has developed consolidated services, standards and best practises. 
Over the years, EMODnet Chemistry has collected dozens of success stories about different types of data providers and data users who were willing to open up their data and use them for many purposes. Data providers include marine research institutes, environmental agencies, government marine managers from EU Member States dedicated to marine monitoring and/or marine science, ICES, the Copernicus Marine Service (CMEMS) and citizen scientists.
In terms of data users, the European Environment Agency, the EC Joint Research Centre and most of the Regional Sea Conventions have made extensive use of EMODnet Chemistry data for the implementation of the European Union's marine policy. Researchers and CMEMS use this data source to develop tools, data products and models to assess the state of the environment and trends. More recently, the partners of the Horizon Europe Blue Cloud 2026 project, which supports the implementation of the European Open Science Cloud, have used EMODnet Chemistry data together with data from CMEMS and the World Ocean Database. The objective is to develop a toolbox to create customisable, validated datasets on Essential Ocean Variables of eutrophication and assess the consistency of the information. Based on these data sources the Project is also developing tools for calculating online metocean information and indicators of the environmental quality of the Mediterranean and global oceans. Finally, EMODnet together with CMEMS is providing the data backbone for EDITO: the core infrastructure of the European Digital Twin of the Ocean, which aims to facilitate the development of applications for the digital twin.

In conclusion, EMODnet’s work, although focused on European Union data sources, is increasingly relevant to support the implementation of agreements and data services at a global level. This contribution will continue and hopefully be expanded in the coming years to increase the global marine knowledge

How to cite: Giorgetti, A., Altobelli, C., Schaap, D. M. A., Buga, L., Fyrberg, L., Gatti, J., Holdsworth, N., Iona, A., Larsen, M. M., Schlitzer, R., Østrem, A. K., Tsompanou, M., Troupin, C., and Wesslander, K.: Data exchange and integration: Use cases for EMODnet Chemistry, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19342, https://doi.org/10.5194/egusphere-egu25-19342, 2025.

EGU25-19423 | Orals | ESSI2.3 | Highlight

Geoportal Technology: A Customizable Framework for Managing and Publishing Georeferenced Research Objects 

Francesco Mangiacrapa, Gian Luca Vannini, and Pasquale Pagano

The Geoportal technology is an open-source framework designed to address the challenges of managing and publishing georeferenced research objects. These objects are “complex” documents (called Projects), consisting of various files and metadata and are characterized by geospatial and temporal features.

The Geoportal platform allows user communities to fully customize the data model for their specific instance, enabling them to define the structure, content, and workflow of the research objects to be managed. This highly configurable framework can be adapted to diverse application scenarios, which may vary in terms of:

  • the types of projects to be supported;
  • the workflows guiding the publication, updating, and accessibility of objects based on access policies.

Moreover, this technology emphasizes interoperability and scalability, facilitating the organization and sharing of geospatial, temporal, and multimedia data across multiple domains. By adhering to international standards such as OGC’s Web Map Service (WMS) and Web Feature Service (WFS), the Geoportal ensures seamless integration with other systems, making it a powerful tool for collaborative research.

Developed on the D4Science infrastructure, the Geoportal integrates Web-GIS technologies, offering tools for spatial and temporal data analysis and enabling dynamic visualizations of geodata. Its capabilities include real-time data harvesting, interactive mapping, and seamless integration of up-to-date research outputs.

The system includes two primary interfaces:

  • Data-Entry Interface: Allows users to create, edit, and validate new data entries through a moderation process, ensuring that all entries are reviewed and approved by a supervisor.
  • Data-Viewer Interface: Provides public access to published datasets via an intuitive cartographic map and timeline, enabling historical navigation of the data.

An API further extends the platform’s functionality, supporting automated data retrieval and specialized queries, making it an ideal solution for large-scale data management and integration.

As a case study, we present the Dataset for the National Geoportal for Archaeology (D4GNA), developed for the Italian Ministry of Culture. D4GNA leverages Geoportal technology to create a comprehensive, scalable environment for managing archaeological data, supporting a wide range of geospatial, textual, and multimedia formats. This application exemplifies Geoportal’s capabilities, focusing on the collection, management, and publication of archaeological data within the National Geoportal for Archaeology (GNA) .

Since its launch, D4GNA has cataloged over 1,150 archaeological investigations across Italy and 15 abroad, all licensed under Creative Commons CC-BY 4.0. Future enhancements, including the assignment of Digital Object Identifiers (DOIs) and integration with D4Science’s Catalog service, will enable advanced search functionality and provide statistical insights into the data.

This study demonstrates the transformative potential of Geoportal technology in enabling efficient management, publication, and sharing of georeferenced research objects. By offering a flexible and interoperable framework, Geoportal supports diverse domains such as environmental studies, geology, and climate research, facilitating the integration of geospatial, temporal, and multimedia data. Its adherence to FAIR principles and integration with international standards ensures interoperability and scalability, promoting open access and collaboration in scientific research. This highlights Geoportal as a powerful tool for advancing scientific discovery and preserving valuable research data across disciplines.

Keywords

Georeferenced research objects, Web-GIS, Interoperability, D4GNA

How to cite: Mangiacrapa, F., Vannini, G. L., and Pagano, P.: Geoportal Technology: A Customizable Framework for Managing and Publishing Georeferenced Research Objects, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19423, https://doi.org/10.5194/egusphere-egu25-19423, 2025.

EGU25-1664 | Orals | ESSI2.7

Enhancing Collaboration Across NASA’s Science Repositories and Stakeholders: Findings From the 2024 Data Repositories Workshop 

Kaylin Bugbee, Deborah Smith, Rebecca Ringuette, Robert Downs, Thomas Morgan, Daniel Berrios, Samrawit Gebre, Lorella Angelini, Steve Hughes, Vandana Desai, Alex Young, and Charley Haley

NASA has a long history of collecting and openly sharing scientific data to help users better understand the sun, the Earth, the solar system and the universe. Over 40 repositories across five broad scientific disciplines work to archive, manage and care for these valuable NASA assets. To improve interdisciplinary and transdisciplinary science, NASA has developed a scientific data and information governance strategy. The strategy focuses on collaborative approaches to build a more connected and cooperative stewardship community while recognizing the diversity of domain specific data needs. In this presentation, we will share NASA’s vision for connected scientific data and information governance in addition to findings and lessons learned from the 2024 Open Source Science Data Repositories Workshop.

How to cite: Bugbee, K., Smith, D., Ringuette, R., Downs, R., Morgan, T., Berrios, D., Gebre, S., Angelini, L., Hughes, S., Desai, V., Young, A., and Haley, C.: Enhancing Collaboration Across NASA’s Science Repositories and Stakeholders: Findings From the 2024 Data Repositories Workshop, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1664, https://doi.org/10.5194/egusphere-egu25-1664, 2025.

EGU25-2408 | Posters on site | ESSI2.7

Palaeogeomorphic reconstruction of the South China Sea: coupling tectonics, climate, and ocean dynamics 

Ze Liu, Pengcheng Shu, Pengcheng Wang, Yanxi Li, Syed Wajid Hanif Bukhari, Lulu Zhang, Ruibo Wang, and Kaiyue Lu

Throughout its extensive history, Earth’s surface has undergone dramatic transformations, accompanied by significant changes in climate, environment, and resources. Paleogeomorphology is the result of the interaction between deep Earth tectonic processes and surface processes. The coupling process between active tectonics and geomorphological evolution such as earthquakes, volcanic activities, glacial and fluvial processes represents a key interface. Research on paleogeomorphology is closely related to the interactions between the deep and shallow layers of Earth’s interior and is a crucial component of multi-layered, multi-scale Earth system science. Therefore, the reconstruction of global paleogeography and paleogeomorphology during geological time has long been a focus of interest among geologists. The South China Sea, as a sensitive region to global climate and oceanic environmental changes, has experienced significant changes in its topography and geomorphology due to its complex geological structure and dynamic hydrological processes. This region has become a key subject of Earth system science research. The evolution of its three-dimensional geological environment is deeply influenced by tectonic activity, climatic fluctuations, and ocean dynamics, making its changes highly complex and uncertain, which traditional methods fail to resolve accurately. Therefore, it is essential to approach the reconstruction of the paleogeomorphological evolution of this multi-phase tectonic region from a quantitative perspective. In recent years, with the accumulation of Earth system data and the application of machine learning methods in complex system modeling, machine learning-based Earth system simulation has gradually emerged as a new frontier method, particularly in paleoelevation reconstruction. The application of this technology significantly enhances the precision and predictive capabilities of simulations for geological evolution and environmental changes in the South China Sea region. However, using Sr/Ca and Mg/Ca ratios in paleoelevation reconstruction has certain limitations, primarily due to interference from factors such as geological background and climate change, with weaker environmental responses in extreme environments such as arid or cold regions. To improve the accuracy of paleoelevation reconstructions, multiple geochemical indicators (such as δ¹⁸O, δD, and Δ₄₇) can be combined, and experimental calibration can be applied to separate the effects of climate and elevation, thereby optimizing existing models. By combining three-dimensional evolution models, it is possible to reconstruct the geomorphological evolution of the region, explore the tectonic factors of the South China Sea's opening and closure, the arc-continent collision that led to the Taiwan orogeny, and the impacts of surface factors such as paleoclimate and sea-level fluctuations on the northern South China Sea's hydrological systems, topography, and basin sedimentation. This further reveals the intrinsic mechanisms between the complex geological evolution of the South China Sea and oceanic dynamical processes.

How to cite: Liu, Z., Shu, P., Wang, P., Li, Y., Bukhari, S. W. H., Zhang, L., Wang, R., and Lu, K.: Palaeogeomorphic reconstruction of the South China Sea: coupling tectonics, climate, and ocean dynamics, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2408, https://doi.org/10.5194/egusphere-egu25-2408, 2025.

EGU25-2760 | Orals | ESSI2.7

Enhancing Semantic Interoperability in Earth System Sciences: TheRole of EarthPortal 

Christelle Pierkot and Guillaume Alviset

The study of the Earth system involves a wide range of disciplines that are increasingly collaborate to address global challenges like climate change. In France, Data Terra, the national e-infrastructure supports this effort by managing data for five key Earth system components: Ocean, Atmosphere, Solid Earth, Continental Surface, and Biodiversity. Alongside distributed infrastructure services, Data Terra provides discovery, access, and dissemination tools to enable researchers to effectively conduct their scientific investigations. Among these resources, EarthPortal, a FAIR-compliant semantic artefact catalogue, promotes the use of controlled vocabularies and ontologies, enhancing semantic interoperability across Earth sciences disciplines.

Developed within the EOSC FAIR-IMPACT project using OntoPortal technology, EarthPortal aligns with European and national recommendations for sharing and reusing semantic artefacts in interdisciplinary research contexts. EarthPortal is a thematic catalogue and semantic repository specifically tailored for Earth and environmental sciences, hosting artefacts like SKOS-controlled vocabularies and OWL ontologies, organized into categories and thematic groups with filtering options. Beyond providing access to semantic artefacts, EarthPortal offers advanced tools to support third-party applications. Key functionalities include (i) the annotator, which suggests relevant terms based on textual or keyword input ; (ii) the recommender, which identifies pertinent semantic artefacts corresponding to the provided input ; and (iii) the mappings tool, which generates, stores, and visualizes relationships between different semantic artefacts. Through its REST API, EarthPortal facilitates seamless integration with external applications, enabling users to access semantic artefacts and leverage these tools directly within their workflows.

This presentation highlights EarthPortal’s functionalities including its user interface and API capabilities. Using a concrete application example, we will demonstrate how the tool enhances the semantic interoperability of data. Specifically, we will illustrate the integration of EarthPortal with EaSy Data, the French data repository for Earth and environmental sciences. This connection enriches metadata with semantic annotations, improving discoverability and user experience.

Additionally, we will explore new developments, including the federation of EarthPortal with other OntoPortal Alliance platforms (e.g., BiodivPortal, AgroPortal) to enable cross-domain interoperability by facilitating the discovery of semantic artefacts across domains, and providing users with a unified interface to access diverse thematic resources. Finally, we will outline future developments, including the creation of a generic connector for the GeoNetwork catalogue, which will allow the direct use of EarthPortal's semantic artefacts and tools, enhancing metadata editing, search functionalities and data harvesting processes.

How to cite: Pierkot, C. and Alviset, G.: Enhancing Semantic Interoperability in Earth System Sciences: TheRole of EarthPortal, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2760, https://doi.org/10.5194/egusphere-egu25-2760, 2025.

The paradigm of FAIR (Findable, Accessible, Interoperable, Reusable) implies that due credit for scientific results can and will be provided when such results are properly referenced and cited in scientific literature. This applies also to open research code as part of scientific output.

This presentation provides a summary of a four year analysis for a cohort of open source software projects, and their DOI-based citation in scientific publications, based on metadata from Crossref (crossref.org). The results indicate that the mere availability of the required underlying technical infrastructure is a requirement for but does not suffice to ensure acceptance in science and common practice. In addition, organisational changes of best practices are needed, for software project communities, but also journals and publishers. Once all these factors are established, the results show positive trends of acceptance und usage across scientific communities, to reference fundamental open scientific software by Digital Object Identifers (DOI). For this study, a cohort of open source software projects was approached to mint DOI as an emerging good practice for a high visibility publication which was published in 2021 (Springer Handbook of Geographic Information, https://doi.org/10.1007/978-3-030-53125-6_30). The cohort comprised a representative range of open geospatial software projects of that time, including core libraries such a GDAL and PROJ, desktop GIS such as GRASS GIS, QGIS and gvSIG, but also infrastructure and web based applications like PostGIS, rasdaman and actinia, among others. The majority of the software project cohort is federated in the OSGeo Foundation (OSGeo.org). OSGeo membership requires for software projects to implement quality standards for project, code and community management, including the use of a software repository. For this,most of the projects use GitHub (github.com/osgeo), which allowed for automated software release deposit in the Zenodo Open Access repository (Zenodo.org), resulting in the minting of an initial concept DOI and further version DOI for following version releases by the software projects. In the following, the open digital infrastructure Crossref  for persistent cross-platform citations linking in online academic journals was anually queried for citations related to the cohort, and additional OSGeo projects which elected to mint DOI independently. The current results show a modest use of DOI based citations for the majority of the cohort but a significant and growing citation footprint for core geospatial libraries such as PROJ (proj.org) and GDAL (gdal.org), which are used across all scientific fields. The DOI-based citations of GDAL registered by CrossRef show a continuing growth rate of over 100% anually since 2022. An additional outcome of the analysis is that for some Open Access publications, full metadata forwarding to Crossref, which is a requirement for DOI-based FAIR software citation, still remains to be implemented.

How to cite: Löwe, P.: Open Geospatial Software Citation: Status, Patterns and Trends – A Crossref-based data analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2872, https://doi.org/10.5194/egusphere-egu25-2872, 2025.

EGU25-3315 | Orals | ESSI2.7

ESS Data Publication at a "General-Purpose" Supercomputing Centre 

Johannes Munke, Alexander Wellmann, Mallika Muralidharan, Christin Henzen, and Stephan Hachinger

At the Leibniz Supercomputing Centre (LRZ) of the Bavarian Academy of Sciences and Humanities, we have set up a portal and system to publish large datasets from simulations, assign DOIs, and present landing pages. This "LRZ FAIR Data Portal" - currently in demonstrator status - is based on InvenioRDM, with the idea of using this framework as a presentation layer. The login and data-upload possibilities typical for repositories are disabled in this setup. Instead, the portal presents metadata together with a dataset-specific link to the GLOBUS (www.globus.org) data-transfer service, where LRZ is connected. By getting themselves a GLOBUS login for free, users can thus reliably copy the data to many other supercomputing centres or download them to their laptop. This system shall make datasets FAIR (Findable, Accessible, Interoperable, Reusable) that are produced at LRZ and are too large to be moved to institutional, community, or general-purpose research-data repositories.

We are currently developing a mechanism to automatically ingest metadata from LRZ storage systems into InvenioRDM. The idea of this mechanism is that users who wish to publish their data store metadata in a DataCite-centric format, and our mechanism scans these user's volumes for metadata to be published. The datasets are thus automatically presented in the portal. The necessary workflows are harmonized and developed with the partners from the Gauss Centre of Supercomputing (HLRS, JSC as the two other largest German supercomputing centres) within the InHPC-DE project. The InvenioRDM instance also provides an OAI-PMH interface, enabling the harvesting of metadata. This allows datasets stored with us to be discoverable in other services, such as earth-data.de. A corresponding exporter that filters datasets according to the Dewey Decimal Classification has been developed as part of the NFDI4Earth project (see NFDI4Earth Knowledge Hub: knowledgehub.nfdi4earth.de).

In the current demonstrator status of our LRZ FAIR Data Portal, the portal and a preliminary push mechanism are used for the publication of the first "friendly-user" datasets. In particular, an ERA5-based downscaled meteorological data suite (3 km resolution, 40 years timespan) has been published via the portal. We report on the experience publishing this dataset, and also further ESS-related datasets from other projects we are working on. The advantages and limitations of the approach are discussed in relation to our concept of a general-purpose data publication mechanism for huge datasets produced at a supercomputing centre. We also shed some light on differences and potential of domain-specific publication mechanisms (including the terrabyte satellite-data platform at LRZ), highlighting how shortcomings of publication approaches can be addressed and opportunities leveraged.

How to cite: Munke, J., Wellmann, A., Muralidharan, M., Henzen, C., and Hachinger, S.: ESS Data Publication at a "General-Purpose" Supercomputing Centre, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3315, https://doi.org/10.5194/egusphere-egu25-3315, 2025.

EGU25-3434 | Orals | ESSI2.7

Data Governance and Beyond: TPDC’s Role in Advancing Earth System Science 

Xiaoduo Pan, Xin Li, Min Feng, Xiaowei Nie, and Xuejun Guo

The Qinghai-Xizang Plateau, esteemed as the World’s Third Pole, plays a pivotal role on the global stage, shaping climate rhythms, safeguarding water supplies, and nurturing biodiversity across the Asian continent and beyond. Decoding the complex interplay within the Third Pole’s lithosphere, hydrosphere, cryosphere, biosphere, atmosphere, and anthroposphere hinges on the availability of scientific data. As a cornerstone of research in one of the world’s most critical ecosystems, The National Tibetan Plateau/Third Pole Environment Data Center (TPDC, https://data.tpdc.ac.cn) goes beyond traditional data governance by implementing innovative strategies for data sharing, accessibility, and interoperability. This approach not only accelerates the pace of scientific discovery but also fosters a collaborative environment where researchers from around the globe can contribute to our understanding of the complex interactions within the Earth’s systems, ultimately leading to more effective conservation and management of our planet’s resources.

TPDC stands as one of the pioneering 20 national data centers backed by China’s Ministry of Science and Technology since 2019. TPDC is steadfast in its mission to aggregate and harmonize a wealth of data resources concerning the Tibetan Plateau. Boasting the most comprehensive scientific dataset for the Third Pole and its adjacent areas, TPDC curates over 6,600 datasets spanning a multitude of fields, from terrestrial studies to human-environment interactions, atmospheric research, geology, cryospheric science, remote sensing, paleoenvironmental analysis, and more.

TPDC furnishes a cloud-based infrastructure that streamlines online data procurement, quality assurance, analysis, and visualization, thereby enhancing the accessibility of shared data. Adhering to the FAIR principles—findable, accessible, interoperable, and reusable—TPDC forges strategic alliances with international entities. It partners with the WMO to advance the Global Cryosphere Watch initiative and engages with ICIMOD on data swap, observational capabilities, skill development, and collaborative studies. As a preferred data repository for leading international journals like Nature, AGU, ESSD, and Elsevier, TPDC encourages researchers to publish their data in conjunction with scholarly articles. Furthermore, TPDC extends its data expertise to a host of international scientific endeavors, including TPE, GEWEX/GASS LS4P, and WCRP-CORDEX CPTP, bolstering the pursuit of knowledge for the World’s Third Pole." By the subcenter Qinghai and Xizang, TPDC also contribute to the regional sustainability.

In recent years, the TPDC team has made substantial contributions to data management and sharing within Earth system science. Their work, published in renowned journals such as Nature Geoscience, Nature Reviews Earth & Environment, Reviews of Geophysics, and Science Bulletin, underscores the critical need to advance data-sharing policies, enhance data assimilation techniques, and foster international collaboration. The team advocates for stronger policy, technological, and managerial actions to promote data sharing, stimulate scientific cooperation, and support the creation of a Digital Twin of Earth. Additionally, their research highlights the integration of advanced AI technologies and big data assimilation to tackle complex challenges in Earth system science, driving paradigm shifts from data-intensive science to robot scientists. Also they establish security mechanism for the security.Collectively, their efforts provide a robust framework for improving data governance, accelerating scientific progress, and enhancing global cooperation in Earth data sharing.

How to cite: Pan, X., Li, X., Feng, M., Nie, X., and Guo, X.: Data Governance and Beyond: TPDC’s Role in Advancing Earth System Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3434, https://doi.org/10.5194/egusphere-egu25-3434, 2025.

EGU25-4997 | Orals | ESSI2.7

SeaDataNet, panEuropean infrastructure for marine and ocean data management and its relations with EMODnet, Blue-Cloud2026, and DTO 

Dick M. A. Schaap, Athanasia Iona, Steven Piel, Marine Vernet, Serge Scory, Alessandra Giorgetti, and Alexandra Kokkinaki

SeaDataNet is a major pan-European infrastructure for managing and providing access to marine data sets and data products, acquired by European organisations from research cruises and other observational activities in European coastal marine waters, regional seas and the global ocean. Founding partners are National Oceanographic Data Centres (NODCs), major marine research institutes, and ICES. The SeaDataNet network, initiated in the nineties, expanded over time its network of data centres and infrastructure, during a series of dedicated EU RTD projects, and by engaging as core data management infrastructure and network in leading European Commission initiatives such as the European Marine Observation and Data network (EMODnet), Copernicus Marine Service (CMS), and the European Open Science Cloud (EOSC). These facilitated ongoing development and evolution of the SeaDataNet technical infrastructure, standards, tools, and services, while managing and further expanding a large network of connected data centres and data providers.

SeaDataNet develops, governs and promotes common standards, vocabularies, software tools, and services for marine data management, which are freely available from its portal and widely adopted and used by many communities and projects. A core SeaDataNet service is the CDI data discovery and access service which provides online unified discovery and access to vast resources of data sets, managed by 115+ connected SeaDataNet data centres from 34 countries around European seas, both from research and monitoring organisations. Currently, it gives access to more than 3 Million data sets, originating from 1000+ organisations in Europe, covering physical, geological, chemical, biological and geophysical data, and acquired in European waters and global oceans. Standard metadata and data formats are used, supported by an ever-increasing set of controlled vocabularies to mark up the metadata profiles in a semantically controlled way, resulting in rich and highly FAIR metadata and data sets. Services include online CDI catalogue, cloud-based data cache, and online shopping mechanism. FAIRness is further amplified by machine-to-machine services.  

SeaDataNet provides core services in EMODnet Chemistry, Bathymetry, Physics, and Ingestion for bringing together and harmonizing large amounts of marine data sets, which are used by EMODnet groups for generating thematic data products. Examples: Digital Terrain Model for all European seas (Bathymetry), and European aggregated and validated data collections for eutrophication, contamination, and marine litter (Chemistry). These products are very popular and find their way to many users from government, research, industry, and public. SeaDataNet is also back-office for EMODnet Ingestion, reaching out and achieving input from data providers that are not (yet) participating in the European data exchange.

SeaDataNet is well engaged in EOSC projects, such as EOSC-FUTURE and the Blue-Cloud project. A pilot deployed a versatile cyber platform federating multidisciplinary data repositories, analytical tools, and computing facilities for exploring and demonstrating the potential of cloud based open science for ocean sustainability. A further evolution is underway into a Federated European Ecosystem and EOSC Blue Node to deliver FAIR & Open data and analytical services, instrumental for deepening research of oceans, coastal & inland waters. This is also highly important for the Digital Twins of the Oceans (DTO) initiative.

How to cite: Schaap, D. M. A., Iona, A., Piel, S., Vernet, M., Scory, S., Giorgetti, A., and Kokkinaki, A.: SeaDataNet, panEuropean infrastructure for marine and ocean data management and its relations with EMODnet, Blue-Cloud2026, and DTO, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4997, https://doi.org/10.5194/egusphere-egu25-4997, 2025.

EGU25-6121 | Orals | ESSI2.7 | Highlight

 A Researcher’s Journey: Navigating Interdisciplinary Science with ENVRI RDIs and Tools 

Anca Hienola, Andreas Petzold, and Ulrich Bundke

Modern research environments are becoming increasingly complex, driven by the explosion of data volume and variety, the demands of interdisciplinary collaboration, and the proliferation of competing standards and emerging technologies. Adding to this complexity is a dramatic shift in the user community: a new generation of researchers with distinct skillsets, needs, and expectations has replaced the users for whom many Research Data Infrastructures (RDIs) were originally designed. The transition from an infrastructure-centric approach to a data-driven paradigm has created new challenges and opportunities that we are only beginning to fully understand.

This presentation follows the journey of Dr. Reese Arch (got it?), a fictional researcher tackling the challenge of integrating marine biodiversity and atmospheric datasets to model ecosystem responses to climate change. Through Reese’s experience, we explore how ENVRIs, ENVRI Hub, Virtual Research Environments (VREs), and tools can empower researchers while highlighting the barriers that still exist. Reese’s story captures the realities of modern science: successes like FAIR-compliant tools and semantic alignment, and frustrations such as overly complex workflows, and siloed systems.

A dedicated "Complaint Box" captures the key pain points voiced by researchers, from tool integration challenges to the cognitive load of navigating fragmented infrastructures. These frustrations highlight the gap between user needs and existing solutions, underscoring the importance of co-designed tools and community-driven innovation.

The narrative emphasizes how ENVRI is addressing these challenges by fostering interoperability, streamlining workflows, and empowering researchers through tailored VREs and support networks. Dr. Reese’s journey serves as a mirror for the research community, showcasing both the progress made and the need for collective action to bridge gaps in next-generation research infrastructures.

Join us to explore how ENVRI is turning the increasing complexity of research into an opportunity to create an interconnected, sustainable RDI ecosystem where interdisciplinary science thrives.

How to cite: Hienola, A., Petzold, A., and Bundke, U.:  A Researcher’s Journey: Navigating Interdisciplinary Science with ENVRI RDIs and Tools, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6121, https://doi.org/10.5194/egusphere-egu25-6121, 2025.

EGU25-6484 | Orals | ESSI2.7

Harmonizing NetCDF Metadata Workflows: A Collaborative Initiative for Enhanced Data Integration and Reusability 

Björn L. Saß, Romy Fösig, Emanuel Söding, Christof Lorenz, Sabine Barthlott, Philipp S. Sommer, Elke M. I. Meyer, Beate Geyer, Ulrich Loup, Corinna Rebmann, Katharina Loewe, Sibylle K. Haßler, Florian Obersteiner, Tobias Kerzenmacher, Angela Schäfer, Dorothee Kottmeier, Klaus Getzlaff, and Andrea Pörsch

In the pursuit of making data FAIR (Findable, Accessible, Interoperable, Reusable)1, datasets need to be well-described to enable both human and machine users to maximize the benefits towards knowledge discovery and innovation. Self-describing data enriched with metadata are essential to facilitate interoperability and reusability. Within the Earth System Science (ESS) community, the NetCDF data format has become the quasi-standard for storing multidimensional data, supported by general metadata standards such as the CF conventions (Climate and Forecast)2, the Attribute Convention for Data Discovery (ACDD)3 for global attributes and more specific ones like the AtMoDat project4 for atmospheric modeling.

NetCDF files can be self-describing if they contain all relevant metadata. In practice, NetCDF metadata is frequently incompatible because they are either non-standardized or not as detailed as required for repositories or data portals. We aim to address these discrepancies by establishing a standardized NetCDF data workflow that ensures the seamless integration of NetCDF data into downstream processes and enables the extraction of metadata by downstream applications.

The NetCDF Metadata Guideline Initiative is a collaborative effort from researchers, research data management, research data infrastructure and metadata experts from several research centers across Germany. We are supported by the Helmholtz DataHub, the central data infrastructure of the Research Field Earth and Environment within the Helmholtz Association. This initiative aims to develop harmonized data handling guidelines to unify the diverse sub-communities within both observational and modeling fields, building towards a unified and consistent infrastructure for environmental research data.

Our approach includes examining existing guidelines, followed by their integration and expansion to address diverse needs within Earth and Environment disciplines. We will produce a set of comprehensive guidelines designed to enhance data interoperability and reusability, along with tools to facilitate their adoption.

Key milestones include:

  • Reconciling attributes in former guidelines.

  • Implementation of a collaborative and public guidelines document.

  • Development of machine-readable templates and validation tools.

  • Provision of highly user-friendly tools that support scientists when entering metadata profiles based on the guidelines.

  • Integration of our enhanced NetCDF-profiles into selected downstream clients like the Earth Data Portal (EDP)5.

By harmonizing metadata practices, we aim to enhance the interoperability and accessibility of geoscientific data, facilitating more efficient data sharing and utilization across research domains.

The implementation of standardized metadata practices for NetCDF across the ESS communities, would enable data repositories such as PANGAEA® Data Publisher and the World Data Center for Climate (WDCC) to present metadata in compliance with established norms and integrate it into their specific schemas.

This presentation will outline the key challenges identified, the proposed solutions, and the anticipated impact on the geoscientific community. With this presentation we call for participation in this evolving initiative to create a common NetCDF metadata foundation.

1 Wilkinson et al., 2016: https://doi.org/10.1038/sdata.2016.18
2 https://cfconventions.org/
3 https://wiki.esipfed.org/Attribute_Convention_for_Data_Discovery_1-3
4 https://www.atmodat.de/
5 https://earth-data.de

How to cite: Saß, B. L., Fösig, R., Söding, E., Lorenz, C., Barthlott, S., Sommer, P. S., Meyer, E. M. I., Geyer, B., Loup, U., Rebmann, C., Loewe, K., Haßler, S. K., Obersteiner, F., Kerzenmacher, T., Schäfer, A., Kottmeier, D., Getzlaff, K., and Pörsch, A.: Harmonizing NetCDF Metadata Workflows: A Collaborative Initiative for Enhanced Data Integration and Reusability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6484, https://doi.org/10.5194/egusphere-egu25-6484, 2025.

The integration of scientific data across disciplines and stakeholders is crucial for advancing Earth system sciences (ESS) and ensuring data accessibility and usability. One of the primary challenges in this context is the terminological diversity found within different scientific fields, as well as between stakeholders with varying levels of expertise. Standardized metadata and effective metadata (MD) mappings are essential tools for overcoming this challenge, serving as a "translation" layer that enhances the visibility and usability of ESS data across language barriers and scientific domains. In our contribution we highlight the BITS project as a pioneering pilot initiative aimed at the integration of terminologies and services within ESS, with a focus on standardizing metadata and improving data interoperability. The project leverages key frameworks such as provided by NFDI4Earth, base4NFDI, and TS4NFDI, ensuring alignment with international best practices in data integration and sharing. By examining the role of standardized metadata in connecting diverse ESS stakeholders, we demonstrate how terminological harmonization and service integration can bridge disciplinary gaps, enhancing the accessibility, discoverability, and collaborative potential of ESS data. This opens new opportunities for data centers to contribute with their data to transdisciplinary research on sustainability challenges.

How to cite: Lammert, A., Martens, C., Löhden, A., and Anders, I.: Overcoming the challenges of terminological diversity within different scientific fields to enhance the accessibility, discoverability, and collaborative potential of ESS data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8663, https://doi.org/10.5194/egusphere-egu25-8663, 2025.

EGU25-9011 | Posters on site | ESSI2.7

Customising a generic repository with domain-specific metadata – the CREATIVE project  

Sibylle K. Hassler, Carlos Zuleta Salmon, Mirko Mälicke, Peter Braesicke, Jörg Meyer, and Erwin Zehe

RADAR4KIT is the generic repository for all scientific data at the Karlsruhe Institute of Technology (KIT). The metadata for datasets which are being published on this repository are also generic and do not necessarily offer the best description for datasets from specific domains such as the environmental sciences.

In the light of an increasing number of initiatives harvesting and aggregating datasets from different repositories which need more detailed descriptions to facilitate meaningful aggregation or filtering, the CREATIVE project provides a solution for domain-specific metadata in the generic RADAR4KIT repository. We developed customized templates and input masks for domain-specific metadata, which enhance the RADAR4KIT usability for the environmental sciences. These metadata schemas are harmonised with the schemas used by the NFDI4Earth, the National Research Data Infrastructure (NFDI) for Earth System Sciences (ESS), so the data which are harvested from RADAR4KIT can be found and processed there. Similarly, the metadata schemas are compatible to the virtual research environment V-FOR-WaTer, which is being developed at KIT in a collaboration between hydrologists and computer scientists and can be used for processing and visualisation of datasets.

Another key focus of the CREATIVE project is connecting the data stewards in environmental science at KIT, facilitating exchange and mutual support in research data management and publishing. Testing the domain-specific metadata templates with this user group and the involved initiatives and infrastructures will also give indication on how this approach can be generalised and transferred to other repositories and disciplines to support the cultural change towards sharing FAIR data at KIT in Germany and internationally.

How to cite: Hassler, S. K., Zuleta Salmon, C., Mälicke, M., Braesicke, P., Meyer, J., and Zehe, E.: Customising a generic repository with domain-specific metadata – the CREATIVE project , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9011, https://doi.org/10.5194/egusphere-egu25-9011, 2025.

EGU25-9589 | Posters on site | ESSI2.7

Characterizing the Diversity of Data Repositories in ESS, and the Role of re3data 

Claudia Müller and Stephan Frickenhaus

The NFDI4Earth (National Research Data Infrastructure for Earth System Sciences) initiative focuses on creating a sustainable research data infrastructure for Earth System Science (ESS) by aligning with the FAIR principles (Findable, Accessible, Interoperable, and Reusable). As part of the NFDI4Earth project, 150 research data repositories with highly diverse subject areas were collected to build the NFDI4Earth Service Catalogue. The consequence of diversity is that small and niche repositories lack visibility beyond their community. As researchers in ESS often cooperate on a multidisciplinary basis, there is also a need for discovering multidisciplinary repositories. Repository metadata is subject to constant change and needs to be individually and manually updated. Therefore, the need for a platform as a sustainable already existing data source was identified that can be harvested by NFDI4Earth. With re3data (Registry of Research Data Repositories), a global registry of research data repositories, such a platform is already available. Re3data operates now for more than a decade, is maintained through an international scientific network, and describes repositories with well-defined metadata fields. Nevertheless, we identified metadata fields with the potential to represent added value. These metadata fields include the “geographical extent of data”, the “maximum data upload size” and whether the repository is “unrestricted to external upload” (in contrast to hosting data from the maintaining institution only, or is restricted to project data). In a first approach, “maximum data upload size” and “unrestricted to external upload” were identified as metadata fields that would best add value, and characterize repositories further. In ESS scientists often deal with big data, such as in Remote Sensing and Satellite Imagery, Climate Modeling, or Environmental Monitoring, so that such information is important for guidance.

When we started this study, our hypothesis was that data repositories are unrestricted for any data upload, however, that there is a maximum data upload. Yet, our study showed that 70% of the repositories we evaluated in one or other way are restricted in their upload of data (in respect to a specific region, e.g. Australia Ocean Data Network Portal; topic, e.g. GEOFON; or institute affiliation, e.g. Geoportal BGR), and just 10% are unrestricted to the kind of data, but have a maximum upload size. We also identified 20% of the evaluated repositories as unrestricted to the kind of data and with no maximum upload size (e.g. PANGAEA). One result of this study is therefore that repositories have to be differentiated by their upload characteristic “restricted” and “unrestricted” in respect to the kind and size of the data. Highlighting this characteristic more clearly in the future should make it easier for users to distinguish multidisciplinary repositories from niche repositories. The next steps will be to discuss these findings with the wider community in ESS, identify further valuable metadata fields – some of which are domain-specific metadata fields - and progress the inclusion of these fields in re3data.

How to cite: Müller, C. and Frickenhaus, S.: Characterizing the Diversity of Data Repositories in ESS, and the Role of re3data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9589, https://doi.org/10.5194/egusphere-egu25-9589, 2025.

EGU25-9922 | Posters on site | ESSI2.7

Building a Digital Twin Application for Climate Extremes: Using the interTwin Digital Twin Engine 

Christian Pagé, Anne Durif, Paul Millar, Dijana Vrbanec, Matteo Bunino, and Rakesh Sarma

Weather Extremes and their impacts are getting a lot of attention lately, because their occurrence, severity and spatial coverage are increasing and will likely increase further towards the mid and end of the century. Many countries are experiencing significant impact of those extremes due to climate change. It becomes more and more important to better assess the change of characteristics of those extremes according to users and society needs.

It is a challenge to detect and characterize weather extremes for the future climate in all available and relevant climate simulations. A novel approach and methodology is being developed to detect and characterize the changes in weather extreme events using Artificial Intelligence (AI). This is a generic method based on Convolutional Variational Autoencoders (CVAE). This deep learning technique, that uses neural networks, can process large climate datasets much faster than traditional analytical methods.

Another big challenge is to develop on-demand real-world applications that users can manipulate to explore what-if scenarios. Data does not necessarily only come from one research infrastructure (RI), but can also come from several RIs because addressing climate extremes involves climate change impacts that depend on other relevant datasets. Developing robust Digital Twin applications takes a lot of development time. In the context of the interTwin project, a very flexible Digital Twin Engine (DTE) is being developed and implemented. It provides Core Components that can be used by several DT applications from very diverse scientific domains. Applications using AI techniques can also benefit from advanced capabilities using minimal development. It also provides almost automatically generic features and capabilities. This DTE framework acts as an accelerator in order to rapidly develop user-oriented DT applications in diverse scientific domains. 

In this presentation, the interTwin DTE will be presented, and it will be shown how it can be easily used to leverage an existing tool in order to create a DT application. Some results of the method applied on Global Coupled Climate Model datasets will be shown for several greenhouse gas scenarios, over Western Europe.

This project (interTwin) has received funding from the European Union’s Horizon Europe research and innovation programme under grant agreement N°101058386.

How to cite: Pagé, C., Durif, A., Millar, P., Vrbanec, D., Bunino, M., and Sarma, R.: Building a Digital Twin Application for Climate Extremes: Using the interTwin Digital Twin Engine, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9922, https://doi.org/10.5194/egusphere-egu25-9922, 2025.

EGU25-10084 | Posters on site | ESSI2.7

Galaxy Europe - An IT infrastructure for FAIR Data Analysis 

Jérôme Detoc and Marie Jossé

Galaxy Europe (https://usegalaxy.eu) is an ELIXIR-recommended open-source IT infrastructure (RIR) (https://galaxyproject.org/news/2023-12-14-elixir-rir-for-galaxy-europe) emphasizing interoperability and FAIR data analysis. This infrastructure, based on the Galaxy project (https://galaxyproject.org), supports multi-disciplinary, data-driven research.

Galaxy Europe enables users to:

  • Freely access thousands of tools, regularly enhanced and updated, from various research fields such as life, earth system, environment, or climate sciences. This includes tools for data import, organization, sharing, annotation, and export. Each tool can be plugged into a workflow.
  • Use interactive tools such as QGIS, RStudio, and JupyterLab directly on the platform.
  • Design, reproduce, (remotely) run, share, and publish analysis workflows using these batch and interactive tools, (with or) without programming skills.
  • Freely use huge compute and storage resources without any charge.
  • Have a dedicated subdomain designed with the necessary tools regarding one scientific domain. For instance, FAIR-EASE implemented a subdomain focusing on earth system sciences (earth-system.usegalaxy.eu) gathering ocean, land, atmospheric and biodiversity processing software.

Fully integrated into the work area, the Galaxy Training Network (https://training.galaxyproject.org) provides hundreds of free and open tutorials and learning pathways from over thirty scientific topics on data analysis, tool development, and workflow design. The Galaxy Training Network, thus, helps democratize the use of Galaxy, supporting the adoption of open science practices and promoting the reuse of tools and data

The two EOSC projects — EuroScienceGateway and FAIR-EASE — have been joining forces for two years to further improve Galaxy Europe and make it a unique solution for FAIR analysis of scientific data. FAIR-EASE adds numerous tools, workflows, and tutorials for multidisciplinary and interdisciplinary studies on Earth System sciences, creating an interdomain digital architecture for the integrated use of environmental data. EuroScienceGateway, on the other hand, provides a robust open infrastructure for data-driven research.

With inputs from these two projects, we will showcase how Galaxy facilitates access to data, tools, and workflows, while supporting the reuse of these resources across diverse research contexts. Additionally, RO-Crate provides a FAIR packaging solution for research objects, including data, methods, and software, by incorporating structured metadata that preserves workflows and their execution histories. This approach, adopted by projects as a practical implementation of the FAIR vision, enables workflows to be deposited into registries like WorkflowHub for broader accessibility. 

How to cite: Detoc, J. and Jossé, M.: Galaxy Europe - An IT infrastructure for FAIR Data Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10084, https://doi.org/10.5194/egusphere-egu25-10084, 2025.

EGU25-10288 | Posters on site | ESSI2.7

STAC for federated data access to high-volume ESM datasets in preparation for Exascale 

Kameswar Rao Modali, Karsten Peters-von Gehlen, Florian Ziemen, Carsten Hinz, Rajveer Saini, and Martin Schultz

Currently certain earth system models, due to their advanced modeling capabilities and improved computational power, can perform simulations at extremely high resolutions as close to a km. The data from these simulations act as drivers for many downstream scientific research applications as well as decision making tools that aid in policy making. These applications in turn depend on shared or standalone computational resources at HPC infrastructures. As a result the federated data access system design is required to revolve around a triad comprising of:

  • Data

  • Analysis Tools

  • Computing resources

Further, at each of the HPC infrastructures, depending on the earth system model, the format of the data being produced varies. Furthermore, each center has its own combination of storage tiers, each of which are subject to specific hardware constraints. Also, based on the focus of the scientific research, the data usage pattern differs. Hence the organization of the data at each data center for efficient discoverability within the federated data access should cater to :

  • Technicalities of the data ( format, size, file count etc.)

  • Usage pattern of the data

  • Constraints arising due to the Specifications and Limitations of the storage tiers.

Spatial Temporal Asset Catalogs (STAC) fundamentally cater to the discoverability of data corresponding to a specific geographic location associated with a particular time instance or duration. ESM data are a natural fit for such representation. In the present work we provide an overview of the application of STAC for the federated data access within the Warmworld project at the DKRZ and JSC HPC centers. We explain how each of the aforementioned factors at each data center have been addressed and display concrete benefits for data producers and reusers.

How to cite: Modali, K. R., Peters-von Gehlen, K., Ziemen, F., Hinz, C., Saini, R., and Schultz, M.: STAC for federated data access to high-volume ESM datasets in preparation for Exascale, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10288, https://doi.org/10.5194/egusphere-egu25-10288, 2025.

Cloud computing has become essential in Earth System Sciences (ESS) due to the increasing volume of data and the necessity for cross-disciplinary collaboration. Over recent years, this has led to a fragmented landscape of cloud computing solutions, with no clear consolidation emerging. Consequently, users face a diverse range of options, often designed to achieve similar objectives. Despite a strong interest in shifting from traditional workflows to cloud-based processing, user surveys reveal that the transition is hindered by a steep learning curve, significant fragmentation, and particularly, a lack of interoperability among platforms, data, and workflows (Wagemann et al., 2021; DiLeo et al., 2024).

It is already difficult to get an overview of existing solutions in the realm of cloud-based research data infrastructures (RDI) which usually include tools for data providers, platforms, API’s, clients, software and file formats. As a user, finding a way to create meaningful and interoperable end-to-end workflows is even more challenging. Especially, since not all combinations of software, API’s and data are available on all RDIs. Various projects exist with the goal of consolidation of choices in cloud native ESS to increase interoperability and decrease complexity for users, e.g. from the European Space Agency (EOEPCA – Earth Observation Exploitation Platform Common Architecture, APEx – Application Propagation Environment) or the Open Geospatial Consortium (Testbed-20 GeoDataCube API, Open Science Persistent Demonstrator).

This contribution will show our experiences in providing user-friendly computing and ESS services in an agile approach in the above-mentioned projects and discuss the challenges, which begin in finding a common language, in standardizing, combining or harmonizing existing solutions.

How to cite: Zellner, P. and Eberle, J.: Usability in ESS Data Cubes and Cloud Computing: Challenges to Converge in a Landscape of Diverse Solutions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10662, https://doi.org/10.5194/egusphere-egu25-10662, 2025.

EGU25-11112 | Posters on site | ESSI2.7

Cross-disciplinary user support network in NFDI4Earth 

Hela Mehrtens, Klaus Getzlaff, and Sören Lorenz

The national research data infrastructure NFDI4Earth builds a network of 65 partners to support research data management (RDM) in the earth system sciences. One of the currently established infrastructures is the user support network (USN).

The partner institutions of NFDI4Earth range from universities over research institutes to governmental agencies, all having different approaches to handle research data in their specific discipline, different support infrastructures and workflows and a different way to support users in research data handling. In most cases the local RDM support of an institution is the right place to go for researchers. But for cross-disciplinary projects it might be difficult for the single user to find the right contact person. A distributed support network can help in these cases involving several support persons dealing with one user service request.

Our ambition is to provide a high-quality data management support network for ESS researchers. In this network, already existing support institutions contribute jointly. We also collaborate with national consortium helpdesks of related disciplines to share experiences and offer an even wider support portfolio.

We present our support network, the workflows and internal organization, and address the challenges we are facing.

How to cite: Mehrtens, H., Getzlaff, K., and Lorenz, S.: Cross-disciplinary user support network in NFDI4Earth, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11112, https://doi.org/10.5194/egusphere-egu25-11112, 2025.

EGU25-11173 | Posters on site | ESSI2.7

User Scenarios in Action: how European Open Science Cloud services can help Earth System Scientists 

Anna-Lena Flügel, Beate Krüss, Heinrich Widmann, Hannes Thiemann, Fanny Adloff, and Stephan Kindermann

Researchers from around the world and across various disciplines are collaborating to address climate change, one of the most pressing global challenges. Producing, processing and handling huge data collections is an everyday challenge for Earth System scientists and data managers alike. Within the Horizon Europe project FAIRCORE4EOSC, we examine several newly developed FAIR enhancing services in the context of the European Open Science Cloud (EOSC) to address some of these data challenges in the ‘Climate Change’ case study. The overall goal is to assess how to improve the discoverability, reusability and traceability of climate data collections at several levels of granularity, thus enhancing the FAIRness of the ENES ( European Network for Earth System Modelling) Research Infrastructure data. The case study explores how to link provenance metadata, processing steps applied to the data, citation information as well as information about connected research activities to the data itself.

The ‘Climate Change’ case study investigates the benefits of integrating the following FAIRCORE4EOSC services: RAiD (Research Activity Identifier Service), PIDGraph and DTR (Data Type Registry). 

RAiDs are exemplarily used to provide an exhaustive research context for existing data collections, helping domain agnostic users with an aggregated view on related details - from data generation by the Earth System modellers up to publication of final assessment reports. RAiD metadata will be supplied to Open Science Graphs like the PIDGraph. The DTR offers the possibility to register and assign persistent identifiers to single and complex data types and achieves a machine-actionable standardization of type metadata that is used for some typical climate data objects. This is a prerequisite for machine-aided analytics and is of high priority due to the commonly large data volumes in climate science.

Additional efforts aim to level the pass to future developments, such as potentially extending existing Web Processing Services (WPS) to be able to publish RAiD details. To advance further, we propose using the DTR for data typing of STAC (SpatioTemporal Assets Catalogs) defined as FDOs (FAIR Digital Objects), paving the way for enhanced interoperability across data spaces in climate science. 

Our presentation will demonstrate the practical benefits of these new EOSC services for a climate research ecosystem paving the way for a more efficient, collaborative, and impactful Earth System Science community.

How to cite: Flügel, A.-L., Krüss, B., Widmann, H., Thiemann, H., Adloff, F., and Kindermann, S.: User Scenarios in Action: how European Open Science Cloud services can help Earth System Scientists, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11173, https://doi.org/10.5194/egusphere-egu25-11173, 2025.

EGU25-11892 | ECS | Posters on site | ESSI2.7

Enhancing Multiscale Geologic Data Collection, Sharing, and Reusability Through the Community-Driven StraboSpot Ecosystem 

Youseph Ibrahim, Julie Newman, Basil Tikoff, J Douglas Walker, Drew Davidson, Thomas Shipley, Ellen Nelson, and Claire Martin

StraboSpot is an open-source, US NSF-funded, FAIR-aligned, and community-driven data system that enables researchers to collect, store, and share spatially referenced geologic data across scales. StraboSpot's development has been a collaborative initiative, shaped through a series of community workshops initially focused on providing tools that enhance research in structural geology. Over time, other communities have expanded these efforts to include diverse geological subdisciplines. Community workshops have played a crucial role in establishing shared metadata standards, a unified vocabulary, and workflows tailored to meet the unique needs of each user community. This inclusive approach has been instrumental in creating an infrastructure that integrates diverse data types and promotes interdisciplinary collaboration while minimizing barriers to data entry.

The "Spot" approach provides a flexible framework for characterizing spatial areas across all scales, facilitating seamless integration of data ranging from micro-scale laboratory measurements to regional-scale field data. By allowing Spots to nest indefinitely within one another, users can preserve critical information about spatial context, scale, orientation, and inter-relationships. The StraboSpot ecosystem comprises three core applications: StraboMicro, a desktop tool designed for managing and contextualizing laboratory-derived micro-scale data; StraboField, a mobile application tailored for mapping and field-based research; and StraboExperimental, designed for data derived from rock deformation experiments. All data are tied to a shared database that preserves provenance and context. Additionally, researchers can link data to external repositories, such as EarthChem, further enhancing data interoperability.

In response to clear community priorities, StraboSpot is actively enhancing its functionality to support enhanced collaboration and improve data reusability. Current developments include the implementation of Group Workflows, which enable collaboration in both field and laboratory settings while maintaining data provenance and robust version control. Additionally, a Quality Assurance/Quality Control (QA/QC) framework is being developed to improve confidence in shared observational data. This QA/QC system will provide a transparent and systematic mechanism for evaluating data quality and completeness, facilitating the use of shared data sets within and between disciplines.

How to cite: Ibrahim, Y., Newman, J., Tikoff, B., Walker, J. D., Davidson, D., Shipley, T., Nelson, E., and Martin, C.: Enhancing Multiscale Geologic Data Collection, Sharing, and Reusability Through the Community-Driven StraboSpot Ecosystem, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11892, https://doi.org/10.5194/egusphere-egu25-11892, 2025.

The  Anthropocene is the time in which humanity has a greater influence and impact on our planet than all natural forces combined. The dominant influence of human forces on our planet’s land, water and atmosphere has already overstepped biophysical planetary boundaries, and is threatening to increase and worsen its conditions if politics, society and economy are not adjusting their forces. The problem we face in the Anthropocene can be summarised in the most recent and multiple international reports and publications spanning from the IPCC, IPBES, UNESCO, ENECE and the United Nations. Three ‘broad and deep’ transitions needed to address this huge societal problem are the energy transition, current use of space, and the total greenhouse emissions of the food system. How can disciplines become overarching in breaking loose big societal problems of climate change, biodiversity loss, soil and water contamination, social unrest, pandemics and inequality?

We are discerning two different types of transformations: 1. The three ‘broad and deep’ transitions, and 2. A call for transformation that is supported by a multi- to inter- to transdisciplinary theory of the Anthropocene. The distinction between the two transformations is between practice and theory: Is the theoretical transformation (2) needed to support the practical transformation (1)? How can disciplines become overarching and supporting to each other? How can the result of overarching and supporting disciplines contribute to potential solutions? In addition to being solution-oriented, the overarching umbrella will also have to focus on the total unruly problem (blockages, dilemmas, ambivalences, polarization etc.) in the societal domain.

At first we need to primarily understand how we as human beings have come across living in the Anthropocene. Second, we need to understand how the Anthropocene will develop further and what our options for action are. We need knowledge from many disciplines and a theory that can relate as much of that relatively reductionist knowledge as possible in a relatively coherent and holistic way. A theory on the (multi-level) natural and cultural (co-)evolution of complex adaptive systems might be able to relate or even to some extent unify insights from the various sub-disciplines in a reasonably coherent (and process-philosophical) way.

The current Anthropocene socio-ecological system is (resembles) a runaway super-organism, the question being to what extent it can still be tamed. The driving force of this super-organism seems to be a dysfunctional, potentially self-destructive capitalist/extractivist (infinite growth) ideology of survival of the fittest. To what extent might this evolving super-organism still be capable of self-reflection and self-direction, via (geopolitical) cultural evolution and self-domestication, mindfully directed? This last question probably cannot be answered purely theoretically, but will have to be brought to an answer empirically and in a transdisciplinary way. Towards a multi- to inter- to transdisciplinary Anthropocene theory is therefore an urgent need and a call for transformation that needs incremental steps and which needs eventually to be acknowledged by all academic, governmental, corporal and societal actors.

How to cite: Kluiving, S., van der Linde, L., and Lont, J.: Towards a multi- to inter- to transdisciplinary Theory of the Anthropocene - Review of overarching disciplines and research addressing planetary boundaries and social and humanitarian crises, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11942, https://doi.org/10.5194/egusphere-egu25-11942, 2025.

EGU25-12207 | Orals | ESSI2.7

The OGC API - Connected Systems: An emerging standard for interoperable sharing of observation data 

Simon Jirka, Christian Autermann, and Jan Speckamp

The Open Geospatial Consortium (OGC) API standards offer a modern approach to accessing geospatial data and services on the web, promoting interoperability and simplicity. These standards, developed by the OGC, improve upon older OGC standards by embracing web-centric practices and resource-oriented designs. They use a RESTful architecture, enabling developers to interact with geospatial resources through standardized HTTP mechanisms and JSON encoding, ensuring ease of integration, discoverability, and scalability.

The OGC API family comprises several key standards, each tailored for specific functionalities. Notable examples include OGC API - Features, succeeding the Web Feature Service (WFS); OGC API - Maps, succeeding the Web Mapping Service (WMS); and OGC API - Coverages, succeeding the Web Coverage Service (WCS). Together, they address a wide range of geospatial data types, creating a robust framework for interoperable geospatial applications.

However, a direct successor to SensorWeb standards, such as the Sensor Observation Service (SOS) or Sensor Planning Service (SPS), was missing from the OGC API suite. The SensorThingsAPI, developed earlier, follows a different architectural model. To address this gap, the OGC Connected Systems Standards Working Group (SWG) introduced a draft specification for the OGC API - Connected Systems. This standard focuses on managing descriptions of sensor systems, networks, and their data, building on established models such as the Sensor Model Language (SensorML), SWE Common, and Observations, Measurements, and Samples (OMS). It also aligns with contemporary standards like the Semantic Sensor Network Ontology (SOSA/SSN).

The OGC API - Connected Systems draft consists of two parts. Part 1 extends the OGC API - Features standard to manage static resources, such as systems, procedures, deployments, and sampling features. Systems encompass entities such as sensors, platforms, actuators, and processing components that produce data or receive commands. Procedures define the processes undertaken by systems, while deployments detail where and when systems are used. Sampling features represent real-world objects observed by systems. This part supports various data formats, allowing for both rich metadata descriptions in SensorML and simpler representations such as GeoJSON.

Part 2 addresses dynamic data, including datastreams of observations, control channels for sending commands to systems, and historical data on system events. Datastreams enable flexible grouping of observations, such as by sensor network or observed property. Control channels allow systems to receive commands, such as initiating measurements or altering states. Historical events, now managed as dedicated datastreams, avoid overloading system descriptions with excessive details.

The draft specification foresees integrating publish/subscribe patterns, such as MQTT, for managing data streams, control channels, and events. 

With this contribution we aim to provide an insight into how the emerging OGC API - Connected Systems standard provides a modern successor to Sensor Web technologies. We discuss how this initiative empowers data managers and scientists to efficiently exchange observational data and metadata, ensuring compatibility and interoperability across diverse applications.

How to cite: Jirka, S., Autermann, C., and Speckamp, J.: The OGC API - Connected Systems: An emerging standard for interoperable sharing of observation data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12207, https://doi.org/10.5194/egusphere-egu25-12207, 2025.

The ever-growing complexity of the multiple global challenges requires Earth System Sciences (ESS) to collaborate in interdisciplinary research and demands reliable and well-accessible research data as a foundation. NFDI4Earth, a consortium within the German National Research Data Infrastructure (NFDI), addresses these needs by empowering researchers to adopt FAIR principles, enabling the discovery, access, and use of ESS data provided by partners and other sources. Uniting 66 partner institutions in Germany - from universities and research organizations over public agencies to infrastructure providers - NFDI4Earth brings together a wide range of expertise in data management. This submission highlights (first) NFDI4Earth achievements after three years of NFDI-funding and reflects on lessons learned.

NFDI4Earth gives an active role to the ESS communities, allowing them to advance FAIRness in several ways. Through open calls for NFDI4Earth Pilots, Incubators and EduPilots the community was invited to formulate needs, innovative ideas for RDM and competence building in ESS. More than 40 projects were funded and led to a variety of data products, software solutions and educational resources, all openly available for reuse by researchers, covering a breadth of ESS subdisciplines. NFDI4Earth offered information, training, and outreach to the ESS community in various formats.

NFDI4Earth launched first prototype versions of new services: The User Support Network offers ESS-RDM-support and operates as a federated system across RDM-support-desks of the NFDI4Earth partner institutions. The Knowledge Hub builds the information backbone and offers metadata over a broad range of resources including datasets, learning resources, repositories, and software source code. The EduTrain Portal provides access to a range of curated open educational resources related to RDM in the ESS. Finally, the OneStop4All serves at the portal to allow discovering and easily accessing all the NFDI4Earth resources.

To further advance the implementation of FAIR principles NFDI4Earth started designing an overarching NFDI4Earth Architecture and developing the NFDI4Earth Label as a means of classifying ESS-related repositories. The NFD4Earth Label intends (1) to help researchers find the most suitable repositories for data access and data publication, and (2) provides repository operators with a self-assessment tool for documentation and evaluation. In parallel, the recent launch of the first version of the NFDI4Earth FAIRness and Openness Commitment invites individuals and organisations to declare their support to NFDI4Earth and their intent to promote the transition to FAIR RDM in the ESS.

To secure sustained operation and cooperation networks NFDI4Earth established close cooperation with important national public environmental data and geodata infrastructures. Relevant research communities have pledged to closely link their data infrastructures and service developments to NFDI4Earth. NFDI4Earth is internationally connected, notably through involvements in initiatives like the European Open Science Cloud (EOSC) and the Open Geospatial Consortium (OGC) contributing to the development, alignment, and adoption of global standards while fostering collaboration across borders.

NFDI4Earth is funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation), project number: 460036893.

 

How to cite: Bernard, L. and the NFDI4Earth Consortium: First Achievements and Experiences from building NFDI4Earth - a National Research Data Infrastructure for Earth System Sciences, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12287, https://doi.org/10.5194/egusphere-egu25-12287, 2025.

EGU25-13340 | Orals | ESSI2.7

Digital Ecosystem for Time Series Data Management in Earth System Science 

David Schäfer, Martin Abbrent, Nils Brinckmann, Florian Gransee, Joost Hemmen, Tobias Kuhnert, Ralf Kunkel, Christof Lorenz, Peter Lünenschloß, Bert Palm, Thomas Schnicke, Hylke van der Schaaf, and Jan Bumberger

Understanding and managing the Earth System requires sustainable, interdisciplinary approaches to data accessibility, integration, and processing. To address these challenges, we present a modular and scalable digital ecosystem designed to enhance earth data science and support multidisciplinary applications [1]. Adhering to the FAIR data as well as the FAIR research software principles, the system employs standardized interfaces, and open-source technologies to foster collaboration across disciplines, extending beyond Earth System Sciences.

The ecosystem comprises three core components: (i) the Sensor Management System (SMS) for detailed metadata registration and management [2]; (ii) time.IO, a platform for efficient storage, transfer, and real-time visualization of time series data [3]; and (iii) the System for Automated Quality Control (SaQC), which ensures data integrity through automated data analysis and quality assurance [4,5]. Developed, maintained, and distributed as dedicated projects, these components integrate seamlessly into a coherent time series data management system. Leveraging widely adopted solutions and standards such as the OGC SensorThings API, OGC SensorML and the EUDAT B2INST persistent identifier, the system ensures compatibility and integration across research infrastructures, software systems, and diverse disciplines.

This cloud-ready and highly adaptable ecosystem supports deployments from small-scale local research projects to large-scale international environmental monitoring networks. It provides a user centric solution for storing, analyzing, and visualizing data. The use of established metadata standards and the community-driven development of metadata schemes and semantic annotations ensure consistency, interoperability, and reusability of metadata and data formats across various applications The applicability of the proposed ecosystem for use cases from Earth System Sciences and its usability across all stages of a typical sensor data lifecycle will be demonstrated using Cosmic Ray Neutron Sensing data as an illustrative example.

By aligning user needs with sustainable software solutions, this ecosystem facilitates FAIR-compliant practices, supports scientific innovation, and promotes robust, transparent research in Earth System sciences.

 

References:

[1] Bumberger, J., Abbrent, M., Brinckmann N., Hemmen, J., Kunkel, R., Lorenz, C., Lünenschloß, P., Palm, B., Schnicke, T., Schulz, C., van der Schaaf, H., and Schäfer, D. (2025). Digital Ecosystem for FAIR Time Series Data Management in Environmental System Science. SoftwareX (accepted)

[2] Brinckmann, N., Alhaj Taha, K., Kuhnert, T., Abbrent, M., Becker, W., Bohring, H., Breier, J., Bumberger, J., Ecker, D., Eder, T., Gransee, F., Hanisch, M., Lorenz, C., Moorthy, R., Nendel, L. J., Pongratz, E., Remmler, P., Rosin, V., Schaeffer, M., Schaldach, M., Schäfer, D., Sielaff, D., & Ziegner, N. (2024). Sensor management system - SMS (1.17.1). Zenodo. https://doi.org/10.5281/zenodo.13329925

[3] Schäfer, D., Abbrent, M., Gransee, F., Kuhnert, T., Hemmen, J., Nendel, L., Palm, B., Schaldach, M., Schulz, C., Schnicke, T., & Bumberger, J. (2023). timeIO - A fully integrated and comprehensive timeseries management system (0.1). Zenodo. https://doi.org/10.5281/zenodo.8354839

[4] Schmidt, L., Schäfer, D., Geller, J., Lünenschloss, P., Palm, B., Rinke, K., Rebmann, C., Rode, M., & Bumberger, J. (2023). System for automated Quality Control (SaQC) to enable traceable and reproducible data streams in environmental science. Environmental Modelling & Software, 105809. https://doi.org/10.1016/j.envsoft.2023.105809

[5] Schäfer, D., Palm, B., Lünenschloß, P., Schmidt, L., Schnicke, T., & Bumberger, J. (2024). System for automated Quality Control - SaQC (v2.6.0). Zenodo. https://doi.org/10.5281/zenodo.5888547

How to cite: Schäfer, D., Abbrent, M., Brinckmann, N., Gransee, F., Hemmen, J., Kuhnert, T., Kunkel, R., Lorenz, C., Lünenschloß, P., Palm, B., Schnicke, T., Schaaf, H. V. D., and Bumberger, J.: Digital Ecosystem for Time Series Data Management in Earth System Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13340, https://doi.org/10.5194/egusphere-egu25-13340, 2025.

EGU25-15067 | Posters on site | ESSI2.7

Linking services to enhance multidisciplinary dataset usability with semantic metadata in the Geo-INQUIRE Project 

Kety Giuliacci, Rossana Paciello, Manuela Sbarra, Valerio Vinciarelli, Marco Salvi, Daniele Bailo, Jan Michalek, Agnieszka Mtupa-Ndiaye, and Franck Chanthaw

Geo-INQUIRE (Geosphere INfrastructures for QUestions into Integrated REsearch) addresses the pressing need for sustainable and interoperable research data infrastructures to tackle critical societal challenges. With a consortium of over 50 partners, the project integrates diverse data sources and services across geoscientific domains. By fostering FAIR (Findable, Accessible, Interoperable, and Reusable) principles, Geo-INQUIRE supports a cultural shift toward open and reusable science.

The complexity of modern societal challenges, such as climate change, resource management, and natural hazard mitigation, demands a multidisciplinary approach to research. Despite the importance of multidisciplinary research, significant challenges remain for users in accessing, integrating, and analyzing data across diverse scientific domains. The diversity of data formats, vocabularies, and structures often creates barriers to discovering and connecting related datasets, particularly when data stewardship services originate from different disciplines. 

In this contribution, we present a metadata approach designed to address challenges in data services interoperability and usability. A key challenge lies in improving the EPOS-DCAT-AP metadata model (https://epos-eu.github.io/EPOS-DCAT-AP/) to describe complex relationships between data services effectively. As of now, these are indeed unrelated, thus hindering users from connecting and querying related datasets. To address this, we introduced semantic information describing service input parameters and output response. 

The metadata approach is applied to a multidisciplinary use case involving two data services selected from the Geo-INQUIRE portfolio, provided by two different scientific domains: Anthropogenic Hazards and Geological Information and Modeling. The Anthropogenic Hazards domain provides access to data on human-induced seismic events in the form of a catalogue organized into episodes. Each episode consolidates industrial and geophysical data on anthropogenic activities in specific areas and in a given time period. 

The Geological Information and Modeling domain provides access to datasets on geological maps, boreholes (for water, oil, or gas extraction), 3D geological models, and mineral resources.

Specifically, linking seismicity data to borehole data allows users to build better subsurface models for more precise spatio-temporal analysis of relationships between industrial activities, subsurface geology, and seismic responses. For example, users can explore how specific borehole characteristics, such as lithology, correlate with seismicity patterns in a given region.

This work demonstrates how semantic metadata can significantly enhance the usability and interoperability of geoscientific data services. By explicitly defining relationships between services and leveraging metadata-driven automation, the approach significantly reduces users' efforts in multidisciplinary research.

How to cite: Giuliacci, K., Paciello, R., Sbarra, M., Vinciarelli, V., Salvi, M., Bailo, D., Michalek, J., Mtupa-Ndiaye, A., and Chanthaw, F.: Linking services to enhance multidisciplinary dataset usability with semantic metadata in the Geo-INQUIRE Project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15067, https://doi.org/10.5194/egusphere-egu25-15067, 2025.

The assessment of the exploration potential and high upfront exploration costs are significant barriers to the deployment of geothermal energy exploitation. The DEGREE project aims to enhance the success rate of such projects through advancing exploration methodologies and the development of a virtual digital underground laboratory. This laboratory leverages a digital twin of the target area, a geothermal test site in the East Eifel region of Germany. The project covers the entire workflow, from data collection and processing to geological and coupled hydro-thermo-mechanical modeling, as well as visualisation and analysis of the model results. This workflow will be designed to be fully automated, including the computation of individual data processing steps on distributed systems. The final model results will be accessible through a browser-based frontend, featuring three-dimensional visualisation and advanced analysis tools. This poster presents an overview of the project’s approach and the planned architecture of the virtual digital laboratory.

How to cite: Meeßen, C., Volk, M., and Castell, W.: Developing a virtual digital laboratory for geothermal exploration: data processing, distributed systems and three-dimensional model visualisation of a digital twin, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15245, https://doi.org/10.5194/egusphere-egu25-15245, 2025.

EGU25-16801 | Orals | ESSI2.7

Data management services for earth and climate communities  

Debora Testi, Yann Le Franc, Mark Van de Sanden, Sander Apweiler, and Rob Carrillo

In all scientific domains, there is a growing production of data from different sources which can be hard to replicate. To be scientifically valuable those data have to be preserved, accessed, and shared across researchers and communities following the FAIR principles. 

EUDAT’s vision is that data should be shared and preserved across borders and disciplines and its mission is to enable data stewardship within and between European research communities through the Collaborative Data Infrastructure. To meet this vision, EUDAT offers a set of services for managing data. EUDAT is the largest pan-European data infrastructure and is conceived as a network of collaborating, cooperating centres, combining the richness of numerous community-specific data repositories with the permanence and persistence of some of Europe’s largest scientific data and computing centres. Covering both access and deposit, from informal data sharing to long-term archiving, and addressing identification, discoverability and computability of both long-tail and big data, EUDAT services aim to address the full lifecycle of research data.

The services offered by EUDAT are designed to be generic and to support  researchers from a variety of scientific domains at the different stages of the Data Life Cycle. These include the exchange of data with team members via B2DROP, publishing of datasets with the assignment of DOIs via B2SHARE, the long term preservation of data with replication across sites via B2SAFE, the persistent identification of the datasets with B2HANDLE and the discovery of datasets with B2FIND. Over the years, different earth and environmental sciences use cases (for example from SeaDataNet, TOAR, EPOS, ENES) have been integrating EUDAT services as major components of their scientific workflows. 

As an example we can mention the work done in collaboration with ICOS Carbon Portal which aimed to extend the community platform into a benchmarking environment that controls the whole processing chain from selection and preparing the prior information datasets, running the models, followed by the benchmarking and analysis of the results. In order to achieve this EUDAT data services have been integrated in the ICOS workflow: B2SAFE for storing the data and staging them for analysis at computing platforms, data are assigned with persistent identifiers via the B2HANDLE service, and B2FIND harvests the ICOS Carbon portal for broader sharing of the data. The use case addresses all aspects from Findability, Accessibility, Interoperability, and Reproducibility principles, with a focus on reproducibility.

In this presentation, we will provide an overview of the EUDAT services suite and how the services can be integrated by the climate and earth science communities with the support of examples from current use cases. We will discuss with the participants to understand possible gaps in the service offering. 

How to cite: Testi, D., Le Franc, Y., Van de Sanden, M., Apweiler, S., and Carrillo, R.: Data management services for earth and climate communities , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16801, https://doi.org/10.5194/egusphere-egu25-16801, 2025.

The reputation of data providers and the accuracy of geodata and data sources are critical factors for the trustworthiness of environmental metrics and indicators to support policy making. In this context, the possibility of user-specific assessment of the quality of input data for composite indicator calculations using data-fitness-for-use and data-fitness-for-purpose approaches has gained importance. Data-fitness-for-use criteria describe the general suitability of a data set for further use and thus refer to the intrinsic quality of the data such as completeness, accuracy, consistency, and timeliness. Data-fitness-for-purpose criteria are more context-dependent and emphasise the suitability of data for a specific application according to the suitability of the user's requirements.

This contribution first analyses spatiotemporally dynamic input datasets for the derivation of environmental indicators to determine whether they contain sufficient quality information to enrich the indicators with data quality information. We consider two crop-type classifications for the derivation of a biodiversity indicator and phenological and meteorological data for the derivation of an extreme weather indicator. The available information on the quality of the input data represents established key metrics for the production-oriented assessment of thematic accuracy. A prerequisite for their calculation are reference data that are accepted as true and which are often not available for geodata derivatives such as composite indicators. If reference data are not available, we show that such overall input data accuracy metrics are not sufficient to derive quality metrics for subsequent products, such as composite indicators, as they lack information on the spatial distribution of accuracy.

Secondly, the structure of an open framework is discussed that allows the extension of geospatial quality standards such as ISO 19157-1 (2023). There, production-orientated thematic accuracy sub-elements such as “classification correctness or “quantitative attribute accuracy” are already included. In contrast, the aspects of “usability” for geodata remain undefined due to their heterogeneous nature. As a possible key sub-element, we propose spatial uncertainty metrics as an additional data layer, often derived as a by-product of modelling, which can support usability assessments and the communication of local and spatially aggregated indicator uncertainties.

In conclusion, the approach presented focuses on data quality, usability and standardisation, which is closely related to the FAIR principles, and emphasises the importance of making geospatial data and environmental indicators more reusable. We believe that this approach can significantly increase the value and utility of research objects in the earth and environmental sciences and foster their reuse in the context of science policy frameworks. The packaging of research objects together with quality assessments in a lightweight container format such as RO-Crate and Annotated Research Context facilitates in addition the (semi-)autonomous processing of these data by machines and thus their AI Readiness. 

 

ISO 19157-1 (2023). Geographic information: Data quality. International Organization for Standardization. Geneva, Switzerland. https://www.iso.org/standard/78900.html

How to cite: Möller, M., Weiland, C., and Martini, D.: Enhancing environmental indicator trustworthiness: A framework for user-specific quality assessment of spatial input data using data-fitness-for-purpose principles, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17188, https://doi.org/10.5194/egusphere-egu25-17188, 2025.

EGU25-19091 | ECS | Posters on site | ESSI2.7

Easier Access to ESM Data: Implementation at Jülich Supercomputing Centre 

Carsten Hinz, Sander Apweiler, Simon Grasse, Björn Hagemeier, Rajveer Saini, and Martin Schultz

Improvements in computational speed lead to better resolutions in Earth System Models (ESM) allowing them to resolve scales of a few kilometers. The volume of the resulting data greatly increases with the improvements in resolution and introduces a challenge to processing and storing these results.
While modern HPC systems provide petabyte-scale capacity for file storage, analyzing such data on local user systems can become a prohibitive bottleneck. Beyond the sheer demands on processing high-volume ESM data, there is also increasing demand to make them FAIR and in particular findable.

The goal of the “Easier” module of the Warmworld project aims to simplify the access to ESM data from different HPC centers, in particular the German Climate Computing Center (DKRZ) and the Jülich Supercomputing Centre (JSC). One aspect is the creation of a joined catalog following the SpatioTemporal Asset Catalogs (STAC) specification. This allows browsing data available at both centers. Details on the STAC implementation will be presented by Kameswar Rao Modali et al. at this General Assembly (EGU25-10288).

The catalog also contains links to access the data, either as a download or later as zarr-stream. The explicit implementation of the required REST-APIs depends on the infrastructure, hardware and software, of the data centers as well as the organization of the stored data.

As the first data backend at JSC, we set up a Fields DataBase (FDB), developed by ECMWF, to store the ESM results as multi-dimensional data cubes. For data retrieval, we provide a download service based on the MARS language for identifying data within the FDB and the UNiform Interface to COmputing REsources (UNICORE) for an automated access to our HPC system. The download service integrates the Helmholtz authentication and authorization infrastructure. This will allow a large number of institutions to access these services with the possibility to control access and resource use.

Our poster will provide details on the ongoing implementation and combination of the various tools and services used at JSC as well as further details on the implementation on our HPC systems. In addition, we will provide first information about the planned authorization for different data accesses and also show the performance of different components with a set of benchmarks.

How to cite: Hinz, C., Apweiler, S., Grasse, S., Hagemeier, B., Saini, R., and Schultz, M.: Easier Access to ESM Data: Implementation at Jülich Supercomputing Centre, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19091, https://doi.org/10.5194/egusphere-egu25-19091, 2025.

EGU25-19463 | Posters on site | ESSI2.7

The national research data infrastructure for Earth System Sciences NFDI4Earth: an approach to realize international interoperability 

Christiane Schmidt, Dominik C. Hezel, Ira Gerloff, Kirsten Elger, Valentina Protopopova-Kakar, Melanie Lorenz, Marcel Meistring, Jie D. Xu, Florian Ott, and Wolfgang zu Castell

Earth System Scientists (ESS) work in various research fields spanning from atmospheric research, to land use and oceanographic research using a multitude of data formats, vocabularies and sources for their data. However, data are mostly not presented in a concise and holistic way. Thus, ESS still are forced to deal with different interfaces, searching for relevant information and tools. The National Research Data Infrastructure for Earth System Sciences NFDI4Earth was created to address these issues. With building a platform connecting a large group of sources for data in Earth system sciences, including expert information and educational resources. The OneStop4All provides a user-friendly single-point-of-entry from which all resources can be addressed. This entry-point is enriched with a living handbook and educational resources and the User Support Network provides additional help or the user.

To realize the envisioned level of interoperability, agreements and partnerships have to be established at several levels, which will be the visualized on the poster presentation. On the political level, agreements on modes of participation, a common adoption of quality standards, as well as a proper embedding into the national and international ecosystem have to be constituted. At the same time, a high level of interoperability requires widely accepted identifiers such as DOI, IGSN and other PID services to identify objects across organizational boundaries. On an intermediate level, a jointly accepted data-centric architecture requires to agree on standardized interfaces, harmonization of approaches to metadata standards. Finally, on the technical level, systems need to be integrated, identification and access services have to provide a user experience with as few system disruptions as possible. 

Hereby, NFDI4Earth follows the principle of making use of what has been established by the consortium partners such as the DataHub of the Helmholtz Research Field Earth & Environment, as well as following the recommendations of international initiatives, like OneGeochemistry to ensure international connectivity. In particular, the approach is guided by the FAIR principles and full embracement of Open Science (see the NFDI4Earth Fairness and Openness Commitment). Being distributed by design, the approach is open to allow new contributors and members to join-in and engage in this joint endeavour.

How to cite: Schmidt, C., Hezel, D. C., Gerloff, I., Elger, K., Protopopova-Kakar, V., Lorenz, M., Meistring, M., Xu, J. D., Ott, F., and zu Castell, W.: The national research data infrastructure for Earth System Sciences NFDI4Earth: an approach to realize international interoperability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19463, https://doi.org/10.5194/egusphere-egu25-19463, 2025.

EGU25-19483 | Posters on site | ESSI2.7

Boosting interoperability of EMODnet chemistry aggregated datasets with a new point of access 

Maria Eugenia Molina Jack, Sebastian Plehan, Luminita Buga, Julie Gatti, Athanasia Iona, Martin M. Larsen, Ann Kristin Østrem, Matteo Vinci, Karin Wesslander, Dick Schaap, and Alessandra Giorgetti

European and global institutions are working hard to collect and share high quality environmental data and  information. This is because it is clear that the ocean knows no borders, and data sharing is necessary to make environmental policies based on knowledge that is informed by extensive data.

Fifteen years ago, the European DG MARE launched the EMODnet data infrastructure, which today provides a consolidated framework for open access to marine data, products and services. The institutions behind it work together by scientific thematic area to offer FAIR (Findable, Accessible, Interoperable and Reusable) information.

The implementation of the FAIR principles on the data is complex and is constantly being developed and improved.

The thematic discipline of EMODnet Chemistry focuses on data and information relevant to eutrophication, pollution by hazardous substances and litter. Each year, regional aggregated datasets on these themes are published, containing validated and unrestricted data.

There are already various access points to the aggregated data collections, which can be downloaded directly (https://emodnet.ec.europa.eu/en/chemistry#chemistry-data) or used for visualization and subsetting (https://emodnet-chemistry.webodv.awi.de/). So far, none of them is interoperable, but this is being sorted out by establishing an ERDDAP instance for the EMODnet Chemistry aggregated data collections (https://erddap.emodnet-chemistry.eu/erddap/index.html), starting with the eutrophication datasets.

ERDDAP was chosen because it is an open-source, simple and user-friendly tool that allows downloading data in different formats. In addition, the automation of requests is possible and it can be linked to other ERDDAPs.

Technical research and development work has been carried out to implement a sustainable workflow that can be reused annually with the release of each new version of the collections. Using a series of Python scripts, the data in the extended SeaDataNet ODV-ASCII format (containing metadata and data) are split into small chunks and converted to NetCDF format for uploading to ERDDAP. Using the newly developed tool “erddapcfg” (https://github.com/PlehanSebastian/erddapcfg) and a SQLite database, the ERDDAP metadata are compiled, ensuring that all metadata and data fields are given appropriate scientific meaning. Complementary, a web service with geoserver  provides  a WMS of the data stations.

The various thematic datasets differ in their structure, so each type of collection requires a specific development. In the case of the eutrophication datasets, the biggest challenge was the file size.

Once the work on the eutrophication datasets is completed, the next step will be to tackle the work on the microlitter datasets, considering the high relevance of this type of data on  a global scale.

How to cite: Molina Jack, M. E., Plehan, S., Buga, L., Gatti, J., Iona, A., Larsen, M. M., Østrem, A. K., Vinci, M., Wesslander, K., Schaap, D., and Giorgetti, A.: Boosting interoperability of EMODnet chemistry aggregated datasets with a new point of access, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19483, https://doi.org/10.5194/egusphere-egu25-19483, 2025.

EGU25-399 | ECS | PICO | ESSI2.10

Comparison of moment tensor inversion methods in a Bayesian framework 

Thomas Mancuso, Cristina Totaro, and Barbara Orecchio

Focal mechanism and moment tensor computation based on regional and local waveforms has become routine task in seismology. These tools are essential for understanding seismotectonic stress regimes and are among the most widely used data for stress inversion, playing a crucial role in identifying deformation zones and tectonically active structures at both local and regional scales (e.g., Totaro et al., GRL 2016; Martínez-Garzón et al., JGR SE 2016).

Many different and similar approaches are available to perform inversion for the double-couple, deviatoric or full moment tensor. However, a key aspect often not fully addressed is the estimation of moment tensor uncertainty. It can be mostly caused by measurement (e.g., data contamination by noise) and theory errors (e.g., mathematical simplifications), and can affect the accuracy of results limiting their interpretation. Over the past decades, considerable efforts have been made in this context, and Bayesian inference is increasingly being applied in moment tensor inversion problems due to the advantage of quantifying parameter uncertainties (Vasyura-Bathke et al., SRL 2020). The Bayesian approach allows for a thorough exploration of the solution space by using appropriate samplers (e.g., Del Moral et al., JRSS 2006) and generates an ensemble of solutions rather than a single optimal one, providing a measure of uncertainty based on the solution distribution.

In this study, we focused on testing the stability of double-couple solutions obtained using two recently developed open-source software packages: BEAT (Bayesian Earthquake Analysis Tool; Vasyura-Bathke et al., SRL 2020) and MCMTpy (Yin and Wang, SRL 2022). These moment tensor inversion algorithms are extremely useful for estimating source parameter uncertainties and the range of acceptable solutions. We applied them to the 2016 Mw 6.0 Amatrice mainshock and a Mw 3.2 earthquake from the same sequence occurred in Central Italy, in order to check the performance of the algorithms at different magnitude levels. We selected this region due to several reasons: it is characterized by active tectonics, it benefits from good azimuthal coverage of seismic stations, and it offers plenty of moment tensor solutions obtained using different approaches (e.g., Scognamiglio et al., BSSA 2009; Artale Harris et al., JGR SE 2022).

For these two earthquakes we compared the results obtained by BEAT and MCMTpy with solutions available in the main seismic catalogs to evaluate the overall coherence of the results and the possible improvements in resolution and robustness. Then, we focused on the performance evaluations by proposing a series of methodological tests which simulate different data setup as not-optimal network geometry, epicentral location errors, biases in the velocity model. By applying these tests on the selected algorithms, we (i) explored their stability, (ii) identified their limitations in resolving double-couple moment tensors and (iii) evaluated the related uncertainty estimates. By doing so, we provide a comprehensive understanding of how these algorithms perform in real-world scenarios and we also suggest an approach useful to verify and eventually compare the performance of moment tensor inversion algorithms also taking into account the uncertainty estimates.

How to cite: Mancuso, T., Totaro, C., and Orecchio, B.: Comparison of moment tensor inversion methods in a Bayesian framework, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-399, https://doi.org/10.5194/egusphere-egu25-399, 2025.

EGU25-1838 | ECS | PICO | ESSI2.10

Geophysical Analysis Of Soil Properties For Engineering-Geological Studies And Urban Planning 

Dimitri Akubardia, Tea Godoladze, Zurab Javakhishvili, Nato Jorjiashvili, Mikheil Tserodze, David Tsiklauri, and Giorgi Tatunashvili

The study area is Tbilisi, the capital of Georgia, which is the most densely populated part of the region and is undergoing rapid urbanization. Tbilisi is situated in a tectonically active and stressed region, characterized by significant seismic activity. Additionally, the area features various geological rock structures and complex topography; thus, the seismic effects of an earthquake will vary across different geological zones.

Given these factors, assessing the impact of natural hazards on building sites is a critical prerequisite for construction projects. To achieve this, it is necessary to analyze soil categories and physical-mechanical properties in accordance with building codes.

In the initial phase, our objective was to collect all available materials from geophysical and geological surveys conducted in Tbilisi. We created an online database that facilitated the selection of new research locations based on an engineering-geological map. Subsequently, we performed a geophysical survey at over 100 locations and generated a map of Vs30 points across Tbilisi.

Calculating the average shear wave velocity Vs for a specific depth range (top 30 meters) can be performed using various methods. We used the seismic refraction method and multi-channel analysis of surface waves (MASW), tailored to the geological area. Field data collection, processing, and interpretation were conducted according to ASTM standards. Seismic data were processed using the SeisImager and ParkSEIS software packages.

Following the guidelines outlined in the Georgian building code and Eurocode 8, we classified the ground category at each surveyed point.

How to cite: Akubardia, D., Godoladze, T., Javakhishvili, Z., Jorjiashvili, N., Tserodze, M., Tsiklauri, D., and Tatunashvili, G.: Geophysical Analysis Of Soil Properties For Engineering-Geological Studies And Urban Planning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1838, https://doi.org/10.5194/egusphere-egu25-1838, 2025.

EGU25-6731 | ECS | PICO | ESSI2.10

A new open-source Python toolbox for processing seismic surface wave data 

Ilaria Barone, Nathalie Roser, Alberto Carrera, and Adrian Flores Orozco

The use of open-source processing tools represents a strategic resource for the scientific community. The Open Science philosophy (https://www.unesco.org/en/open-science) promotes transparency, reproducibility and accessibility to data and source codes. This not only ensures continuous and collaborative development, but also increases the quality of proposed solutions.

Characterizing the near surface based on geophysical methods is of considerable interest for many disciplines, and the reliability and quality of the provided results is tied to the available processing resources. The surface wave analysis (SWA) of active seismic data is widely used to determine the shear wave velocities of a site. Several efforts have been made to create open-source tools for SWA, starting with the precursor Geopsy (Wathelet, 2005), continuing with the more recent SWIP (Pasquet and Bodet, 2017), MASWaves (Olafsdottir et al., 2018), and SWprocess (Vantassel and Cox, 2022). The classical procedure they propose is limited to a local 1D analysis on (moving) spatial windows, where homogeneous conditions are assumed. Although this is a robust approach, it does not highlight small-scale lateral variations.

In this talk, we introduce a new open-source tool under continuous development  for processing surface wave data. The Python-based library incorporates, in addition to the classical 1D analysis on moving windows, more advanced techniques such as the Multi-Offset Phase Analysis (MOPA; Strobbia and Foti, 2006) and the Tomography-like approach (Barone et al., 2021), which perform high-resolution 2D SWA for a more accurate identification of lateral velocity variations. The ultimate intent of our Python library is to contribute to further developing standards for processing and inversion of surface wave data in a proper 2D sense.

 

References

Barone I., Boaga J., Carrera A., Flores Orozco A. and Cassiani G., 2021. Tackling Lateral Variability Using Surface Waves: A Tomography-Like Approach. Surveys in Geophysics 42, no. 2, 317–38. https://doi.org/10.1007/s10712-021-09631-x

Olafsdottir E. A., Erlingsson S., and Bessason B, 2018. Tool for Analysis of Multichannel Analysis of Surface Waves (MASW) Field Data and Evaluation of Shear Wave Velocity Profiles of Soils. Canadian Geotechnical Journal 55, no. 2, 217–233. https://doi.org/10.1139/cgj-2016-0302

Pasquet S., and Bodet L., 2017. SWIP: An Integrated Workflow for Surface-Wave Dispersion Inversion and Profiling. GEOPHYSICS 82, no. 6, WB47–61. https://doi.org/10.1190/geo2016-0625.1

Strobbia C., and Foti S., 2006. Multi-Offset Phase Analysis of Surface Wave Data (MOPA). Journal of Applied Geophysics 59, no. 4, 300–313. https://doi.org/10.1016/j.jappgeo.2005.10.009

Vantassel J. P., and Cox B.R., 2022. SWprocess: A Workflow for Developing Robust Estimates of Surface Wave Dispersion Uncertainty. Journal of Seismology 26, no. 4, 731–56. https://doi.org/10.1007/s10950-021-10035-y

Wathelet M., 2005. Array recordings of ambient vibrations: surface-wave inversion. Ph.D. Thesis, University of Liège (Belgium)

How to cite: Barone, I., Roser, N., Carrera, A., and Flores Orozco, A.: A new open-source Python toolbox for processing seismic surface wave data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6731, https://doi.org/10.5194/egusphere-egu25-6731, 2025.

EGU25-8491 | PICO | ESSI2.10

Interactive Time-Dependent Seismic Hazard Assessment with SHAppE 

Andrew Redfearn and Kostas Leptokaropoulos

The Seismic Hazard Parameters Evaluation (SHAPE) toolbox (Leptokaropoulos and Lasocki, SRL, 2020, https://doi.org/10.1785/0220190319) has evolved into SHAppE, an interactive MATLAB app. SHAppE facilitates probabilistic assessment of seismic hazard parameters, including the mean return period (MRP) and exceedance probability (EP) of earthquake magnitudes, along with confidence intervals. Its interactive features support real-time analysis and visualization, making it suitable for researchers and practitioners analyzing time-dependent seismicity, such as aftershocks, stress triggering, and seismicity induced by human activities. 

SHAppE offers a graphical user interface (GUI) that simplifies parameter selection and data filtering, making it more accessible to users with limited programming experience. It supports four magnitude distribution models, the Unbounded and Truncated versions of the Gutenberg-Richter law and non-parametric Kernel density estimation. The app is demonstrated through case studies from regional datasets (e.g., Song Tranh 2 reservoir in Vietnam) and global catalogues (ISC), showcasing its utility in monitoring seismic responses and evaluating hazard mitigation measures. All input parameters, output data, and results are systematically archived to ensure thorough experiment tracking and facilitate reproducibility. SHAppE provides an intuitive platform, suitable for research and teaching time-dependent, probabilistic seismic hazard analysis.

How to cite: Redfearn, A. and Leptokaropoulos, K.: Interactive Time-Dependent Seismic Hazard Assessment with SHAppE, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8491, https://doi.org/10.5194/egusphere-egu25-8491, 2025.

EGU25-9026 | PICO | ESSI2.10

 Near-surface characterisation with B3AM: case studies of 3C ambient noise beamforming from geothermal sites across Europe 

Katrin Löer, Gabin Simonet, Heather Kennedy, Claudia Finger, and Thomas Hudson

The Matlab toolbox B3AM (B3AMpy for Python) for three-component beamforming of ambient noise data provides a means to characterise the seismic (noise) wavefield and image near-surface seismic properties quickly and cheaply. Provided with three-component array data, B3AM outputs dispersion curves for pro-/retrograde Rayleigh and Love waves, estimates of wavefield composition and propagation direction as a function of frequency, and can be extended for surface wave anisotropy analysis. We present recent results from seismic array data gathered at geothermal sites in the Netherlands, the UK, and Switzerland using B3AM or B3AMpy.

For the geothermal site Kwintsheul (NL), we derive a shear-velocity profile for the first 500 meters, updating an existing profile based on P velocities and regional vp/vs estimates. Comparing dispersion curves from beamforming to those from cross-correlation interferometry, we find that the Rayleigh first higher mode seems to provide most of the energy in the considered frequency range and that the fundamental mode can only be recovered using the beamforming scheme but not from interferometry.

Using a nodal seismic data set collected at the Eden geothermal project (Cornwall, UK), we investigate the anisotropy of the ambient noise wavefield and relate it to faults and fractures in the area. With the additional module AssessArray we estimate the effect array geometry and source distribution have on observed anisotropy. AssesArray synthesises a data set by computing (vertical component) phase shifts at each station location corresponding to a wavefield excited by a single source or multiple sources distributed randomly around the array. We then beamform the data set as we do for real data (although for 1 component only) and analyse the variation in velocity and number of detections as a function of azimuth and frequency. We find that the array design introduces frequency dependent anisotropy as well as apparent dominant directions of wave energy that align with the maximum aperture of the array. Further, we find that the number of sources used in creating the synthetic wavefield affects the observed anisotropy. In general, we observe a larger magnitude of anisotropy for a larger number of sources, i.e., for a more complex wavefield, whereas apparent anisotropy is small or not detectable for fewer sources or a single source, respectively.

For the GeoHEAT project, which explores a joint analysis of passive seismic and borehole geo-radar data for characterising and monitoring fractured geothermal systems, we implemented and tested the beamforming workflow for a novel nodal data set from the Kanton of Thurgau (CH). Besides dispersion analysis and source directionality, we consider wavefield composition and classify time windows with respect to their dominant wave type to inform and improve Green’s function recovery for ambient noise cross-correlation tomography.

How to cite: Löer, K., Simonet, G., Kennedy, H., Finger, C., and Hudson, T.:  Near-surface characterisation with B3AM: case studies of 3C ambient noise beamforming from geothermal sites across Europe, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9026, https://doi.org/10.5194/egusphere-egu25-9026, 2025.

EGU25-10682 | PICO | ESSI2.10

Automatic workflow for detection, localization, and calculation of spectral parameters for induced seismicity on the Episodes Platform 

Lukasz Rudzinski, Jakub Kokowski, Joanna Kocot, and Hubert Siejkowski

Recently the automatic processing of seismological data recorded in areas with anthropogenic seismicity has become an important issue, as the number of available recordings has increased significantly over the past few years. The problem is also addressed within the DT-Geo Project WP8: Anthropogenic Geophysical Extremes, where a specific workflow is being developed for the automatic processing of induced seismicity-related waveforms. The workflow is designed as a set of independent applications implemented inside the interactive, publicly available Episodes Platform (https://episodesplatform.eu/). The applications created for the workflow include:

  • A tool for picking the first arrivals of seismic waves using neural network-based solutions available within the SeisBench library,
  • A phase association tool that employs the PyOcto algorithm,
  • Location procedures already existing within the Episodes Platform.
  • An application for calculating spectral parameters using the spectral fitting method,

Ultimately, the workflow will enable fully automated processing of raw and continuous seismic data, including event detection, localization, and spectral parameter calculation. The workflow can be used on the Episodes Platform either with data collected from various Episodes, which are geophysical datasets related to regions where induced seismicity has been observed, or with datasets uploaded by the user.

This work is supported by Horizon Europe grant DT-Geo 101058129 and a project co-financed by the Minister of Science Republic of Poland under contract no. 2024/WK/05. We gratefully acknowledge Polish high-performance computing infrastructure PLGrid (HPC Centers: ACK Cyfronet AGH) for providing computer facilities and support within computational grant no. PLG/2024/017279.

How to cite: Rudzinski, L., Kokowski, J., Kocot, J., and Siejkowski, H.: Automatic workflow for detection, localization, and calculation of spectral parameters for induced seismicity on the Episodes Platform, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10682, https://doi.org/10.5194/egusphere-egu25-10682, 2025.

Fracture surface morphology influences important rock joint behavior, such as shear strength, fluid flow, contaminant transport, and heat transfer. Digitizing a fracture surface in the laboratory or from a drill core is state of the art – its quantitative assessment is not. There are many suggestions and comparisons of roughness parameters, each highlighting a different morphological feature. However, models and experiments dealing with surface roughness often present only a single quantity – if any at all. This makes those experiments and simulations difficult to reproduce and compare.

On the other hand, temperature profiles along boreholes, fractures, or mine shafts can provide a tremendous amount of information. For example, the determination and monitoring of water and heat fluxes, as well as heat generation mechanisms, are possible through the analysis of such temperature-depth profiles. Hence, understanding complex hydraulic systems using temperature as a tracer is possible with comparatively simple measurement devices. However, the analysis and processing of such profiles are so far primarily based on experience and individual data perception.

This work presents two toolboxes developed to standardize data-driven analysis of geophysical data: (1) FSAT – A fracture surface analysis toolbox; (2) TDprof – Algorithm-based segmentation of temperature-depth profiles. Both toolboxes provide easy access to common methods of data analysis in their field. This includes well-documented open-source code, maintenance of the code base, videos, guides, and manuals.

Building on the experience with these two toolboxes for geophysical data analysis, this contribution highlights the differences, additional efforts needed, and potential benefits of going the extra mile in delivering (re-)usability to the scientific community, while being “low-key” on continuous maintenance.

How to cite: Heinze, T.: A fracture surface analysis toolbox and a temperature-depth profiler – toolboxes for standardized geophysical data analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12479, https://doi.org/10.5194/egusphere-egu25-12479, 2025.

EGU25-14527 | ECS | PICO | ESSI2.10

2D Bayesian transdimensional inversion for b-value variations 

Catalina Morales-Yáñez, Roberto Benavente, Phil Cummins, Malcolm Sambridge, and Rhys Hawkins

The 2D Bayesian transdimensional inversion methodology is a data-driven methodology that allows for multiple solutions and does not need regularization. This is because Bayesian transdimensional inversion allows the retrieval of the parameters and the number of parameters needed to explain the data simultaneously. It also has intrinsic parsimony, meaning simple solutions will be chosen over complex ones. For all these reasons, it is a perfect tool to retrieve the spatial b-value variation. 
The b-value corresponds to the slope of the Gutenberg–Richter law, which relates the number of earthquakes with their magnitude. Several authors agree that the changepoints of the b-value (i.e., the places where the b-value varies) show more valuable information than the value by itself. In particular, the spatial changes in the b-value in seismicity catalogs have been associated with different stresses, fluid processes, geological structures, and earthquake hazard estimation. 
Given this parameter's importance, robustly retrieving and characterizing b-values and their changepoints is essential. In general, most of the methodologies to retrieve the b-value fix the spatial window of the seismic catalog (i.e., binning) and/or use optimization methods to obtain the values, introducing methodological bias in the solutions. For this reason, we use the Bayesian transdimensional approach to objectively estimate b-value variations along two arbitrary dimensions. This implementation allows a self-defined seismic domain according to the seismic catalog information, where it is unnecessary to prescribe the location and extent of domains, as other methodologies do. 
This study focuses on obtaining 2D spatial b-values changes across the seismic region. To explore the possible changes in the b-value along the space, we use the TransTessellate2D algorithm that allows us to implement the trans-dimensional inference methodology for 2D cartesian problems with Voronoi cells. The synthetic tests were performed to analyze the spatial resolution of the methodology and the smallest b-value variation that the method can retrieve. This methodology has been successfully implemented in central-northern Chile and California, allowing us to characterize the mechanical behavior on the plate interface of subduction and cortical zones, obtaining a similar solution to previous studies, evidencing the reliability of the Bayesian transdimensional method for capturing robust b-value variations. Our future work includes extending the approach to other 2D dimensions (e.g., time, latitude, longitude, depth). 

How to cite: Morales-Yáñez, C., Benavente, R., Cummins, P., Sambridge, M., and Hawkins, R.: 2D Bayesian transdimensional inversion for b-value variations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14527, https://doi.org/10.5194/egusphere-egu25-14527, 2025.

EGU25-15670 | PICO | ESSI2.10

TerraceM 3.0: Advancing marine terrace mapping using worldwide open satellite altimetry of the ICESat-2 mission. 

Julius Jara-Muñoz, Markus Weiß, Jürgen Mey, Kevin Pedoja, and Daniel Melnick

TerraceM is an open-source software for mapping and analysing marine terraces. One of the primary challenges in accurately mapping marine terraces is the limited availability of digital elevation data with the resolution necessary to capture the subtle and ephemeral morphology of these geomorphic features. Recent advancements in remote sensing, such as NASA's ICESat-2 satellite mission, offer new opportunities to address this limitation. The ICESat-2 was designed to study Earth's polar ice, land canopy, and bare-earth topography using its Advanced Topographic Laser Altimeter System (ATLAS), a laser-based instrument similar to a LiDAR sensor, providing highly accurate surface elevation measurements in the form of geolocated photons along profiles. While the data are not continuous, the mission has completed thousands of orbits, densely covering most of the world's coastal areas with photon profiles, making it possible to achieve highly accurate mapping of marine terraces.

 

The latest version of TerraceM introduces new scripts and graphical user interfaces (GUIs) to efficiently interact with ICESat-2 photon data. These features enable users to select, download, preprocess, and map marine terraces interactively. Preprocessing capabilities include filtering canopy signals and reconstructing nearshore bathymetry, allowing the analysis of both subaerial and submarine terraces. Additionally, the new version of TerraceM supports MATLAB and Python, broadening its accessibility to a wider range of users. TerraceM-3 delivers advanced modelling and mapping functionalities, empowering researchers and students involved in marine terrace studies. By leveraging ICESat-2 data, TerraceM significantly extends our ability to analyse past sea-level changes and understand the interplay between tectonics and climate processes in coastal environments.

How to cite: Jara-Muñoz, J., Weiß, M., Mey, J., Pedoja, K., and Melnick, D.: TerraceM 3.0: Advancing marine terrace mapping using worldwide open satellite altimetry of the ICESat-2 mission., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15670, https://doi.org/10.5194/egusphere-egu25-15670, 2025.

EGU25-17793 | PICO | ESSI2.10

Pilot development of a satellite image-based spatial analysis tool to support rural spatial planning 

Kwan-Young Oh, Kwang-Jae Lee, Jae-Young Chang, Moung-Jin Lee, Gyu-Yeol Chae, and No-Jun Park

This study is about the pilot development of a satellite image-based spatial analysis tool to support rural spatial planning. For the sustainable development and systematic use of space in rural areas, accurate, integrated, and data-driven decision-making is essential. However, the dispersed management of rural-related data, non-standardized formats, and the absence of periodic monitoring systems are acting as obstacles to establishing effective plans. To address this issue, this study applied the following methodology. First, rural-related data stored separately by the central government and local governments were collected and processed into standardized spatial data based on Geographic Information System (GIS).  Second, a satellite image-based facility detection and classification tool was developed for periodic and efficient monitoring of rural facilities. The target facilities were selected as livestock, factories, and solar panels, and the latest deep learning model based on HRNet-OCR architecture was implemented and optimized for the rural environment. For the training and validation data, the mosaic image of the Korean Peninsula (2019~2020) produced by KOMPSAT satellite images was used, which provided a high-resolution spatial resolution of 1m and multiple spectral bands to enable the analysis of various indicator characteristics. Third, to verify the effectiveness of the developed tool, Seosan City, Anseong City, Naju City, and Geochang County in the Republic of Korea were selected as pilot areas. These regions were deemed to represent diverse rural characteristics and facility distributions. Finally, a user-friendly web-based information support tool was developed by integrating processed rural data and satellite image analysis results. The results of this study are expected to be utilized as foundational data for establishing rural spatial plans to support rural spatial restructuring and regeneration, and the developed spatial analysis tool is deemed capable of contributing to the formulation of more efficient and sustainable rural development strategies by providing a data-driven decision support system to rural policymakers.

How to cite: Oh, K.-Y., Lee, K.-J., Chang, J.-Y., Lee, M.-J., Chae, G.-Y., and Park, N.-J.: Pilot development of a satellite image-based spatial analysis tool to support rural spatial planning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17793, https://doi.org/10.5194/egusphere-egu25-17793, 2025.

EGU25-19627 | PICO | ESSI2.10 | Highlight

Applying Statistical and Machine Learning Methods for rapid earthquake alert system in Greece with a new mobile application  

Parthena Paradisopoulou, George Spyrou, Ioanna Karagianni, Angeliki Adamaki, and Konstantinos Leptokaropoulos

Confirming the prompt and accurate notification of earthquakes is vital for mitigating their potential impacts. To achieve this, statistical approaches, including Machine Learning (ML), have become indispensable tools across various scientific fields, particularly in Seismology and seismic data. This research explores the utilization of ML techniques to improve earthquake real time alerts. The case study is Greece and the surrounding region, an area with highest seismic activity throughout the Mediterranean.   

This work is focused on the real time collection and processing of an extensive earthquake dataset to generate earthquake alerts by making phone calls and providing details about the time, magnitude, and epicenter of each seismic event. Previous efforts aimed to extend these alerts beyond the notifications (emails and messages) that analysts at the Seismological Center of AUTH (Aristotle University of Thessaloniki) received during their duty. The goal was to make these alerts accessible to all citizens, communities, civil protection agencies and various authorities (e.g. municipalities, schools, police, etc.). The island of Kefalonia served as a pilot region where this functionality was initially implemented. We then chose to extend the application to all Ionian islands to encompass the entire region.

The new insight here is the development of a mobile application that allows users to define a specific geographical region for receiving notifications-alerts. The AI Service will combine the real time earthquake information in conjunction with the geometry defined by each user in order to classify whether a notification should be sent to that specific user.

As training input data used in the application, we first require a catalog of earthquakes spanning from 2011 to 2025 with M≥3.0, along with demographic data for Greece region provided by the Hellenic Statistical Authority. A radius around each epicenter is calculated by considering the earthquake’s macroseismic Intensity (I), the earthquake’s magnitude (M), earthquake depth, total population and number of households within the calculated radius. The labeled dataset is then used to train a classification model via Azure AutoML. This model identifies significant earthquakes and determines which areas to call in order to provide earthquake alert. Notification messages could be to any subscribed mobile number with the calling voice available in Greek, English, or French. 

How to cite: Paradisopoulou, P., Spyrou, G., Karagianni, I., Adamaki, A., and Leptokaropoulos, K.: Applying Statistical and Machine Learning Methods for rapid earthquake alert system in Greece with a new mobile application , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19627, https://doi.org/10.5194/egusphere-egu25-19627, 2025.

Thermochronology is one of the most versatile tools available to geoscientists to constrain the colling and exhumation history of rocks. In tectonically active mountain belts around the world, it is not unusual to have many hundreds, if not thousands of published ages available from various studies. Although several well-established thermal models allow for a detailed exploration of how cooling or exhumation rates evolved in a limited area or along a transect, integrating large, regional datasets into such models remains a major challenge. Here, we present age2exhume, a thermal model in the form of a Matlab or Python script, which can be used to rapidly obtain a synoptic overview of exhumation rates from large, regional thermochronometric datasets. The model incorporates surface temperature based on a defined lapse rate and a local topographic relief correction that is dependent on the thermochronometric system of interest. Other inputs include sample cooling age, uncertainty, and an initial (unperturbed) geothermal gradient. The model is simplified in that it assumes steady, vertical rock-uplift and unchanging topography when calculating exhumation rates. For this reason, it does not replace more powerful and versatile thermal-kinematic models, but it has the advantage of simple implementation and rapidly calculated results. In our example datasets, we show exhumation rates calculated from 1785 cooling ages from the Himalaya, 1587 cooling ages from New Zealand, and 916 cooling ages from Central Asia (Tian Shan and Pamir). Despite the synoptic nature of the results, they reflect known segmentation patterns and changing exhumation rates in areas that have undergone structural reorganization. These regionally estimated exhumation rates have been used in combination with other datasets to assess regional climatic versus tectonic controls on key aspects of the landscape, including river valley width and modern erosion patterns.

How to cite: Schildgen, T. and van der Beek, P.: Age2exhume - A Matlab/Python script to calculate exhumation rates from thermochronometric ages, with application to the Himalaya, New Zealand, and Central Asia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20472, https://doi.org/10.5194/egusphere-egu25-20472, 2025.

The International Seismological Centre (ISC) combines seismic observations from ~150 agencies in ~100 counties to produce the definitive global earthquake catalogue by combining seismic phase arrivals. As well as seismic phase data, hypocentres and magnitudes the ISC Bulletin includes other earthquake parameters such as moment tensors that are reported by many agencies. This data is freely accessible, searchable and downloadable through the ISC website (www.isc.ac.uk/iscbulletin). The ISC Earthquake Toolbox for MATLAB provides access to this parametric earthquake data via a graphical user interface (GUI) within the MATLAB environment. The GUI replicates the search options of the ISC website and reads this data into MATLAB. Several live scripts are included to demonstrate how to interrogate the ISC Bulletin data. Examples include plotting earthquake aftershock sequences, comparing different magnitude and hypocentre types and authors, as well as plotting moment tensors reported in the ISC Bulletin. The toolbox also enables 3D visualisation of earthquake distributions, 2D and 3D moment tensor plotting, as well as introducing new functionality to plot moment tensors within MATLAB mapping toolbox figures. It is hoped that the ISC Earthquake Toolbox for MATLAB will be used as a teaching tool to explore the wealth of earthquake data available at the ISC, as well as a tool for researchers to build more complex applications upon. The toolbox is publicly available to download via GitHub (github.com/tomgarth/ISC_Earthquake_Toolbox) and MathWorks file exchange (https://uk.mathworks.com/matlabcentral/fileexchange/167786-isc-earthquake-toolbox).

How to cite: Garth, T., Gallacher, R., and Leptokaropoulos, K.: The International Seismological Centre (ISC) Earthquake Toolbox for MATLAB: Interactive Access to Earthquake Observations & Parameters, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20943, https://doi.org/10.5194/egusphere-egu25-20943, 2025.

EGU25-1294 | Posters on site | ESSI2.13

A new sub-chunking strategy for fast netCDF-4 access in local, remote and cloud infrastructures.  

Flavien Gouillon, Cédric Penard, Xavier Delaunay, and Sylvain Herlédan

NetCDF (Network Common Data Form) is a self-describing, portable and platform-independent format for array-oriented scientific data which has become a community standard for sharing measurements and analysis results in the fields of oceanography, meteorology but also in the space domain.

The volume of scientific data is continuously increasing at a very fast rate. Object storage, a new paradigm that appeared with cloud infrastructures, can help with data storage and parallel access issues, but NetCDF may not be able to get the most out of this technology without some tweaks and fine tuning.

The availability of ample network bandwidth within cloud infrastructures allows for the utilization of large amounts of data. Processing data       where the data is located is preferable as it can result in substantial resource savings. But for some use cases downloading data from the cloud is required (e.g. processing also involving confidential data) and results still have to be fetched once processing tasks have been executed on the cloud.

Networks      exhibit significant variations in capacity and quality (ranging from fiber-optic and copper connections to satellite connections with poor reception in degraded conditions on boats, among other scenarios). Therefore, it is crucial for formats and software libraries to be specifically designed to optimize access to      data by minimizing the transfer to only what is strictly necessary.

In this context, a new approach has emerged in the form of a library that indexes the content of netCDF-4 datasets. This indexing enables the retrieval of sub-chunks, which are pieces of data smaller than a chunk, without the need to reformat the existing files. This approach targets access patterns such as time series in netCDF-4 datasets formatted with large chunks.

This report provides a performance assessment of netCDF-4 datasets for varied use cases. This assessment executes these use cases under various conditions, including POSIX and S3 local filesystems, as well as a simulated degraded network connection. The results of this assessment may provide guidance on the most suitable and most efficient library for reading netCDF data in different situations.

How to cite: Gouillon, F., Penard, C., Delaunay, X., and Herlédan, S.: A new sub-chunking strategy for fast netCDF-4 access in local, remote and cloud infrastructures. , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1294, https://doi.org/10.5194/egusphere-egu25-1294, 2025.

EGU25-4155 | Orals | ESSI2.13

How open software, data and platforms are transforming Earth observation data science 

Wolfgang Wagner, Matthias Schramm, Martin Schobben, Christoph Reimer, and Christian Briese

One of the most time-consuming and cumbersome tasks in Earth observation data science is finding, accessing and pre-processing geoscientific data generated by satellites, ground-based networks, and Earth system models. While the much increased availability of free and open Earth observation datasets has made this task easier in principle, scientific standards have evolved according to data availability, now emphasizing research that integrates multiple data sources, analyses longer time series, and covers larger study areas. As a result of this “rebound effect”, scientists and students may find themselves spending even more of their time on data handling and management than in the past. Fortunately, cloud platform services such as Google Earth Engine can save significant time and effort. However, until recently, there were no standardized methods for users to interact with these platforms, meaning that code written for one service could not easily be transferred to another (Schramm et al., 2021). This created a dilemma for many geoscientists: should they use proprietary cloud platforms to save time and resources at the risk of lock-in effects, or rely on publicly-funded collaborative scientific infrastructures, which require more effort for data handling? In this contribution, we argue that this dilemma is about to become obsolete thanks to rapid advancements in open source tools that allow building open, reproducible, and scalable workflows. These tools facilitate access to and integration of data from various platforms and data spaces, paving the way for the “Web of FAIR data and services” as envisioned by the European Open Science Cloud (Burgelman, 2021). We will illustrate this through distributed workflows that connect Austrian infrastructures with European platforms like the Copernicus Data Space Ecosystem and the DestinE Data Lake (Wagner et al., 2023). These workflows can be built using Pangeo-supported software libraries such as Dask, Jupyter, Xarray, or Zarr (Reimer et al., 2023). Beyond advancing scientific research, these workflows are also valuable assets for university education and training. For instance, at TU Wien, Jupyter notebooks are increasingly used in exercises involving Earth observation and climate data, and as templates for student projects and theses. Building on these educational resources, we are working on an Earth Observation Data Science Cookbook to be published on the Project Pythia website, a hub for education and training in the geoscientific Python community.

References

Burgelman (2021) Politics and Open Science: How the European Open Science Cloud Became Reality (the Untold Story). Data Intelligence 3, 5–19. https://doi.org/10.1162/dint_a_00069

Reimer et al. (2023) Multi-cloud processing with Dask: Demonstrating the capabilities of DestinE Data Lake (DEDL), Conference on Big Data from Space (BiDS’23), Vienna, Austria. https://doi.org/0.2760/46796

Schramm et al. (2021) The openEO API–Harmonising the Use of Earth Observation Cloud Services Using Virtual Data Cube Functionalities. Remote Sensing 13, 1125. https://doi.org/10.3390/rs13061125

Wagner et al. (2023) Federating scientific infrastructure and services for cross-domain applications of Earth observation and climate data, Conference on Big Data from Space (BiDS’23), Vienna, Austria. https://doi.org/10.34726/5309

How to cite: Wagner, W., Schramm, M., Schobben, M., Reimer, C., and Briese, C.: How open software, data and platforms are transforming Earth observation data science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4155, https://doi.org/10.5194/egusphere-egu25-4155, 2025.

EGU25-4277 | Posters on site | ESSI2.13

BEACON Binary Format (BBF) - Optimizing data storage and access to large data collections 

Tjerk Krijger, Peter Thijsse, Robin Kooyman, and Dick Schaap

As part of European projects, such as EOSC related Blue-Cloud2026, EOSC-FUTURE and FAIR-EASE, MARIS has developed and demonstrated a software system called BEACON with a unique indexing system that can, on the fly with high performance, extract data subsets based on the user’s request from millions of heterogeneous observational data files. The system returns one single harmonised file as output, regardless of whether the input contains many different data types or dimensions. 

Since in many cases the original data collections that are imported in a BEACON installment contain millions of files (e.g. Euro-Argo, SeaDataNet, ERA5, World Ocean Database), it is hard to achieve fast responses. Next to this, these large collections also require a large storage capacity. To mitigate these issues, we wanted to optimize the internal file format that is used within BEACON. With the aim of reducing the data storage size and speeding up the data transfer, while guaranteeing that the information of the original data files is maintained. As a result, the BEACON software has included a unique file format called the “BEACON Binary Format (BBF)” that meets these requirements. 

The BBF is a binary data format that allows for storing multi-dimensional data as apache arrow arrays with zero deserialization costs. This means that computers can read the data stored on disk, as if it were computer memory, significantly reducing computational access time by eliminating the cost for a computer to translate what’s on disk, to computer memory.

Together with making the entire data format “non-blocking”, which means that all computer cores can access the file at the same time and simultaneously use the jump table to read millions of datasets in parallel. This enables a level of performance which reaches speeds of multiple GB/s, making the hardware the bottleneck instead of the software.

Furthermore, the format takes a unique approach to compressing data by adjusting the way it compresses and decompresses on a per dataset level. This means that every dataset is compressed in a slightly different manner, making it much more effective in terms of size reduction and time to decompress the data which can get close to the effective memory speed of a computer.

It does this while retaining full data integrity. No data is ever lost within this format, nor is any data adjusted. If one were to import a NetCDF file into BBF, one could fully rebuild the original NetCDF file from the BBF file itself. In the presentation the added benefits of using the BBF will be highlighted by comparing and benchmarking it to traditional formats such as NetCDF, CSV, ASCII, etc.

In January 2025, BEACON 1.0.0 was made publicly available as an open-source software, allowing everyone to set-up their own BEACON node to enhance the access to their data, while at the same time being able to reduce the storage size of their entire data collection without losing any information. More technical details, example applications and general information on BEACON can be found on the website https://beacon.maris.nl/.

How to cite: Krijger, T., Thijsse, P., Kooyman, R., and Schaap, D.: BEACON Binary Format (BBF) - Optimizing data storage and access to large data collections, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4277, https://doi.org/10.5194/egusphere-egu25-4277, 2025.

EGU25-5977 | Orals | ESSI2.13

A comparative study of algorithms for lossy compression of 2-d meteorological gridded fields 

Uwe Ehret, Jieyu Chen, and Sebastian Lerch

Meteorological observations (e.g. from weather radar) and the output of meteorological models (e.g. from reanalyses or forecasts) are often stored and used in the form of time series of 2-d spatial gridded fields. With increasing spatial and temporal resolution of these products, and with the transition from providing single deterministic fields to providing ensembles, their size has dramatically increased, which makes use, transfer and archiving a challenge. Efficient compression of such fields - lossy or lossless - is required to solve this problem.

The goal of this work was therefore to apply several lossy compression algorithms for 2d spatial gridded meteorological fields, and to compare them in terms of compression rate and information loss compared to the original fields. We used five years of hourly observations of rainfall and 2m air temperature on a 250 x 400 km region over central Germany on a 1x1 km grid for our analysis.

In particular, we applied block averaging as a simple benchmark method, Principal Component Analysis, Autoencoder Neural Network (Hinton and Salakhutdinov, 2006) and the Ramer-Douglas-Peucker algorithm (Ramer, 1972; Douglas and Peucker, 1973) known from image compression. Each method was applied for various compression levels, expressed as the number of objects of the compressed representation, and then the (dis-)similarity of the original field and the fields reconstructed from the compressed fields was measured by Mean Absolute Error, Mean Square Error, and the Image Quality Index (Wang and Bovik, 2002). First results indicate that even for spatially heterogeneous fields like rainfall, very high compression can be achieved with small error.

 

References

Douglas, D., Peucker, T.: Algorithms for the reduction of the number of points required to represent a digitized line or its caricature. In: The Canadian Cartographer. Bd. 10, Nr. 2, 1973, ISSN 0008-3127, S. 112–122, 1973.

Hinton, G. E., & Salakhutdinov, R. R.: Reducing the dimensionality of data with neural networks. science, 313(5786), 504-507, 2006.

Ramer, U.: An iterative procedure for the polygonal approximation of plane curves, Computer Graphics and Image Processing, 1, 244-256, http://dx.doi.org/10.1016/S0146-664X(72)80017-0, 1972.

Zhou Wang, and A. C. Bovik: A universal image quality index, IEEE Signal Processing Letters, 9, 81-84, 10.1109/97.995823, 2002.

How to cite: Ehret, U., Chen, J., and Lerch, S.: A comparative study of algorithms for lossy compression of 2-d meteorological gridded fields, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5977, https://doi.org/10.5194/egusphere-egu25-5977, 2025.

EGU25-7371 | ECS | Orals | ESSI2.13

Evaluating Advanced Scientific Compressors on Climate Datasets 

Robert Underwood, Jinyang Liu, Kai Zhao, Sheng Di, and Franck Cappello

    As climate and weather scientists strive to increase accuracy and understanding of our world, models of weather and climate have increased in their resolution to square kilometers scale and become more complex increasing their demands for data storage. A recent study SCREAM run at 3.5km resolution produced nearly 4.5TB of data per simulated day, and the recent CMIP6 simulations produced nearly 28PB of data. At the same time, storage and power capacity at facilities conducting climate experiments are not increasing at the same rate as the volume of climate and weather datasets leading to a pressing challenge to reduce data volumes. While some in the weather and climate community have adopted lossless compression, these techniques frequently produce compression ratios on the order of 1.3$\times$, which are insufficient to alleviate storage constraints on facilities. Therefore, additional techniques, such as science-preserving lossy compression that can achieve higher compression ratios, are necessary to overcome these challenges.

    While data compression is an important topic for climate and weather applications, many of the current assessments of the effectiveness of climate and weather datasets do not consider the state of the art in compressor design and instead, asses scientific compressors that are 3-11 years old, substantially behind the state of the art. In this report: 

 

  •  We assess the current state of the art in advanced scientific lossy compressors against the state of the art in quality assessment criteria proposed for the ERA5 dataset to assess the current gaps between needed performance requirements and the capabilities of the current compressors.
  • We present new capabilities that allow us to build an automated, user-friendly, and extensible pipeline for quickly finding compressor configurations that maximize compression ratios while preserving scientific integrity of the data using codes developed as part of the NSF FZ project.
  • We demonstrate a number of capabilities that facilitate use within in the weather and climate community including NetCDF, HDF5, and GRIB file format support; support for innovation via Python, R, and Julia as well as low level languages such as C/C++; and the implementations of commonly used climate quality metrics including dSSIM, and the ability to extend to add new metrics in high-level languages
  • Utilizing this pipeline, We find that with advanced scientific compressors, it is possible to achieve a 6.4x improvement or more in compression ratio over previously evaluated compressors

How to cite: Underwood, R., Liu, J., Zhao, K., Di, S., and Cappello, F.: Evaluating Advanced Scientific Compressors on Climate Datasets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7371, https://doi.org/10.5194/egusphere-egu25-7371, 2025.

EGU25-11118 | Orals | ESSI2.13

Too Big to Handle? Hexagonizing LIDAR and Satellite Data in Geoscience Applications 

Bo Møller Stensgaard, Casper Bramm, Marie Katrine Traun, and Søren Lund Jensen

The exponential growth of LIDAR and satellite data in geoscience presents both opportunities and challenges for users. Traditional data handling methods often struggle with the sheer volume and complexity of these datasets, hindering easy accessibility, efficient analysis and decision-making processes. This presentation introduces the Scandinavian Highlands HEX-Responder platform and database structure, a cutting-edge solution that leverages the power of hexagonal discrete global grid system, Uber H3, and developed processes to revolutionize geospatial data management, fast responsive visualization and analysis.

We will showcase real-world applications, highlighting the platform's potential to accelerate scientific discovery and improve decision-making processes using satellite and remote sensing data.

The platform’s approach offers several advantages over conventional methods:

  • Efficient data organization and retrieval
  • Improved advanced spatial data analyses opportunities
  • Seamless integration of multi-scale and multi-dimensional data without losing information
  • Enhanced, responsive and fast visualisation capabilities

Our ELT (extract, load, transform) and subsequent visualisation procedure can be applied to any big raster data formats. First, the raw raster data is transformed into optimised parquet files through chunked reading and compression based on a low-resolution H3 hexagon cell index (hexagonization), enabling rapid data import to a column-oriented database management system for big data storage, processing and analytics. The H3 cell organisation is preserved in the database through partitioned fetching for visualisation on the platform. This method allows for horizontal scaling and accurate multi-resolution aggregation, preserving data integrity across scales and significantly overcomes typical computational memory limitations.

The platform's capabilities are exemplified by its approach to LIDAR and satellite emissivity data processing using the H3 grid. High-resolution LIDAR data is efficiently gridded and visualized to H3 resolution level 15 hexagons (0.9m2 hexagon cells). The gridding preserves all original pixel raster points while providing aggregated views for seamless zooming.

Another prime example of the capabilities is the handling of NASA’s ASTER Global Emissivity Data (100m resolution). Here, our pipeline transformed 2.1 terabytes of extracted raw CSV-data derived from NASA’s emissivity data into a compressed format based on the H3 index occupying only 593 gigabytes in the database.

This approach not only saves data storage space but also dramatically improves data accessibility and processing speed for the users, allowing users to work in a responsive environment with this massive dataset in ways previously not possible. Each hexagon represents an opportunity to store unlimited amount, types and categories of pre-processed data for more integrative analyses and data insight.

By hexagonizing LIDAR and satellite data, the HEX-Responder platform enables users to explore massive datasets with ease and efficiency in a responsive environment. The integrated procedures allow for detailed information maintenance and retrieval, paving the way for advanced predictive modelling in geoscience applications using earth observation data in a new way.  

How to cite: Stensgaard, B. M., Bramm, C., Traun, M. K., and Jensen, S. L.: Too Big to Handle? Hexagonizing LIDAR and Satellite Data in Geoscience Applications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11118, https://doi.org/10.5194/egusphere-egu25-11118, 2025.

EGU25-12760 | ECS | Posters on site | ESSI2.13

Tree-Based Adaptive Data Reduction Techniques for Scientific Simulation Data 

Niklas Böing, Johannes Holke, Achim Basermann, Gregor Gassner, and Hendrik Fuchs

Large-scale Earth system model simulations produce huge amounts of data. Due to limited I/O bandwidth and available storage space this data often needs to be reduced before written to disk or stored permanently. Error-bounded lossy compression is an effective approach to tackle the trade-off between accuracy and storage space.

We are exploring and discussing lossless as well as error-bounded lossy compression based on tree-based adaptive mesh refinement/coarsening (AMR) techniques. Our lossy compression schemes allow for absolute and relative error bounds. The data reduction methods are closely linked to an underlying (adaptive) mesh which easily permits error regions of different error tolerances and criteria – in particular, we allow nested domains of varying error tolerances specified by the user. Moreover, some of the compressed data structures allow for an incremental decompression in the resolution of the data which may be favorable for transmission and visualization.

We implement these techniques as the open source tool cmc, which is based on the parallel AMR library t8code. The compression tool can be linked to and used by arbitrary simulation applications or executed as a post-processing step. We show different application results of the compression in comparison to current state-of-the-art compression techniques on several benchmark data sets.

How to cite: Böing, N., Holke, J., Basermann, A., Gassner, G., and Fuchs, H.: Tree-Based Adaptive Data Reduction Techniques for Scientific Simulation Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12760, https://doi.org/10.5194/egusphere-egu25-12760, 2025.

EGU25-13394 | ECS | Orals | ESSI2.13

Challenges and perspectives of climate data compression in times of kilometre-scale models and generative machine learning 

Milan Klöwer, Tim Reichelt, Juniper Tyree, Ayoub Fatihi, and Hauke Schulz

Climate data compression urgently needs new standards. The continuously growing exascale mountain of data requires compressors that are widely used and supported, essentially hiding the compression details from many users. With the advent of AI revolutionising scientific computing, we have to set the rules of this game. Minimizing information loss, maximising compression factors, at any resolution, grid and dataset size, for all variables, with chunks and random access, while preserving all statistics and derivatives, at a reasonable speed — are squaring the compression circle. Many promising compressors are hardly used as trust among domain scientists is hard to gain: The large spectrum of research questions and applications using climate data is very difficult to satisfy simultaneously.

Here, we illustrate the motivation behind the newly defined climate data compression benchmark ClimateBenchPress, designed as a quality check in all those dimensions of the problem. Any benchmark will inevitably undersample this space, but we define datasets from atmosphere, ocean, and land as well as evaluation metrics to pass. Results are presented as score cards, highlighting strengths and weaknesses for every compressor.

The bitwise real information content shows a systematic way in case no error bounds are known. In the case of the ERA5 reanalysis, errors are estimated and allow us to categorize many variables into linear, log and beta distributions with values bounded from zero, one or both sides, respectively. This allows us to define error thresholds arising from observation and model errors directly, providing another alternative to the still predominant subjective choices. Most error-bounded compressors come with parameters that can be automatically chosen following this analysis.

Also new data formats are on the horizon: Chunking and hierarchical data structures allow and force us to adapt compressors to spatially or length-scale dependent information densities. Extreme events, maybe counterintuitively, often increase the compressibility through higher uncertainties, but lie on the edge or outside of the training data for machine learned-compressors. This again increases the need for well-tested compressors. Benchmarks like ClimateBenchPress are required to encourage new standards for safe lossy climate data compression.

How to cite: Klöwer, M., Reichelt, T., Tyree, J., Fatihi, A., and Schulz, H.: Challenges and perspectives of climate data compression in times of kilometre-scale models and generative machine learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13394, https://doi.org/10.5194/egusphere-egu25-13394, 2025.

EGU25-13567 | Posters on site | ESSI2.13

Tables as a way to deal with a variety of data formats and APIs in data spaces 

Joan Masó, Marta Olivé, Alba Brobia, Nuria Julia, Nuria Cartell, and Uta Wehn

The Green Deal Data Space is born in the big data paradigm where there is a variety of data formats and data models that are exposed as files or web APIs. As a result, we need to default in simple data structure that is transversal enough to be able to represent most of the more specific data models, formats and API payloads. Many data models present a structure that can be represented as tables.

TAPIS stands for "Tables from APIS". It is a JavaScript code that uses a common data model that is an array of objects with a list of properties that can contain a simple or a complex value. In TAPIS offers a series of operations that use one or more arrays of objects as inputs and produce a new array of objects as an output. There are operations that create the arrays of objects from files or API queries (a.k.a. data import), others that manipulate the objects (e.g. merge two arrays in a single one) and some operations that generate visual representations of the common data structure including tabular, a map, a graph, etc.

TAPIS is limited by its own data model. While many of the data models can be mapped to the common data model, a multidimensional data cube or a data tree cannot be represented in a single table in an efficient way. In the context of the Green Deal Data Space, most of the sensor data, statistical data, geospatial feature based data and administrative data can be considered object based data and can be used in TAPIS. TAPIS is able to connect to Sensor Things API (the sensor protocol selected in AD4GD and CitiObs), S3 buckets (the internal cloud repository used in AD4GD), GeoNetwork (the geospatial metadata catalogue selected in AD4GD and more4nature), and the OGC API features and derivates (the modern web API interfaces standardized by the OGC) but other data inputs will be incorporated, such as Citizen Science data sources and other popular APIs used in the more4nature project. More analytical functionalities are going to be incorporated in the CitiObs project. As part of the AD4GD Green Deal Information Model, there is an operation to associate semantics to each column of a table by linking it to a URI that defines the concept in an external vocabulary (as well as units of measure if appropriate). In order to be compatible with the data space architecture recommended by the International Data Space Association, we are working on supporting the catalogue of the Eclipse Data Connector, and to be able to negotiate a digital contract as a previous step to request access to the relevant data offered in the data space. To do so, we are working on incorporating the data space protocol as part of the TAPIS operations for data import. TAPIS is available as open source at https://github.com/joanma747/TAPIS.

AD4GD, CitiObs and more4nature are Horizon Europe projects co-funded by the European Union, Switzerland and the United Kingdom.

How to cite: Masó, J., Olivé, M., Brobia, A., Julia, N., Cartell, N., and Wehn, U.: Tables as a way to deal with a variety of data formats and APIs in data spaces, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13567, https://doi.org/10.5194/egusphere-egu25-13567, 2025.

Early career scientists rarely have the resources to work with earth observation data at continental to global scale. This is caused by a combination of factors: large scale data analysis often involves teamwork, connecting data scientists, code developers, IT specialists, statisticians and geoscientists. Young researchers are rarely able to coordinate such a team. Meanwhile, all scientists can have relevant ideas or pose powerful research questions that merit investigation. Copernicus Data Space Ecosystem provides a public, free platform for large-scale processing of earth observation data. It combines instant access to all Sentinel satellite imagery with cloud-based processing in the form of API requests and a powerful browser-based viewing interface. This new approach is enabled by storing the data in a different way: uncompressed formats such as JPEG2000, COG or ZARR support subsetting and querying the image rasters without first unzipping the file, thereby allowing direct streaming of only the area and bands that the user requests. Additionally, this means that most calculations and visualization tasks can be carried out on the server side, directly within the request process. The backend tasks of data storage and management are taken care of by the system, while the user can concentrate on the research itself.

Copernicus Data Space Ecosytem supports several API families. OGC API-s directly enable the creation of Open Geospatial Consortium compatible map products such as WMS, WMTS, WFS or WCS services. These can be accessed with GIS software or displayed in web map tools. OData, STAC, and OpenSearch are Catalog API-s, supporting the querying and of datasets in preparation for analysis. Sentinel Hub is an API family that can handle queries, raster operations, and raster-vector integration for deriving statistics. The main advantages of Sentinel Hub API-s are their efficient use and integration with advanced visualization in the Copernicus Browser.

OpenEO is a fully open-source data analysis framework designed specifically to support FAIR principles. It is independent from data formats with its own data cube format, and can be edited using several coding languages. openEO connects to all STAC-compliant repositories, enabling integration between Sentinel data and other sources. Processing tools include many mathematical operations, but also standard machine learning processes. The system is designed with upscaling in mind: the command structure is the same for small and large areas, with storage and asynchronous processing managed by the backend.

Both API families come with a comprehensive scheme of tutorials and documentation to allow step-by-step learning, and an online Jupyter Lab virtual machine facility. Therefore, early-career scientists with a basic understanding of programming can quickly learn to apply their domain knowledge, while creating solutions that are easy to share and replicate.

All in all, Copernicus Data Space Ecosystem is a transformative tool for earth observation, significantly lowering the bar for applying earth observation at large scale in the geosciences.

How to cite: Zlinszky, A. and Milcinski, G.: Copernicus Data Space Ecosystem empowers early-career scientists to do global scale earth observation data analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15282, https://doi.org/10.5194/egusphere-egu25-15282, 2025.

EGU25-15672 | ECS | Posters on site | ESSI2.13

Scaling Down ESS Datasets: Lessons from the EERIE Project on Compression 

Oriol Tinto, Xavier Yepes, and Pierre Antoine Bretonniere

The rapid growth of Earth System Sciences (ESS) datasets, driven by high-resolution numerical modeling, has outpaced storage and data-sharing capabilities. To address these challenges, we investigated lossy compression techniques as part of the EERIE project, aiming to significantly reduce storage demands while maintaining the scientific validity of critical diagnostics.

Our study examined two key diagnostics: Sea Surface Height (SSH) variability and ocean density, essential for understanding climate dynamics. Leveraging tools such as SZ3 and enstools-compression, we achieved data volume reductions by orders of magnitude without compromising the diagnostics' accuracy. Compression-induced differences were found to be negligible compared to the inherent variability between model outputs and observational datasets, underscoring the robustness of these methods.

Additionally, our work highlighted inefficiencies in current workflows, including the prevalent use of double precision in post-processing. We proposed improvements to align data precision with the original model outputs, further optimizing storage and computation. Integrating lossy compression into existing workflows via widely used formats like NetCDF and HDF5 demonstrates a practical path forward for sustainable ESS data management.

This study showcases the transformative potential of lossy compression to make high-resolution datasets more manageable, ensuring they remain accessible and scientifically reliable for stakeholders while significantly reducing resource demands.

How to cite: Tinto, O., Yepes, X., and Bretonniere, P. A.: Scaling Down ESS Datasets: Lessons from the EERIE Project on Compression, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15672, https://doi.org/10.5194/egusphere-egu25-15672, 2025.

EGU25-15864 | Posters on site | ESSI2.13

The Sentinels EOPF Toolkit: Driving Community Adoption of the Zarr data format for Copernicus Sentinel Data 

Sabrina H. Szeto, Julia Wagemann, Emmanuel Mathot, and James Banting

The Standard Archive Format for Europe (SAFE) specification has been the established approach to publishing Copernicus Sentinel data products for over a decade. While SAFE has pushed the ecosystem forward through new ways to search and access the data, it is not ideal for processing large volumes of data using cloud computing. Over the last few years, data standards like STAC and cloud-native data formats like Zarr and COGs have revolutionised how scientific communities work with large-scale geospatial data and are becoming a key component of new data spaces, especially for cloud-based systems.

The ESA Copernicus Earth Observation Processor Framework (EOPF) will be providing access to “live” sample data from the Copernicus Sentinel missions -1, -2 and -3 in the new Zarr data format. This set of reprocessed data allows users to try out accessing and processing data in the new format and experiencing the benefits thereof with their own workflows.

This presentation introduces a community-driven toolkit that facilitates the adoption of the Zarr data format for Copernicus Sentinel data. The creation of this toolkit was driven by several motivating questions: 

  • What common challenges do users face and how can we help them overcome them? 
  • What resources would make it easier for Sentinel data users to use the new Zarr data format? 
  • How can we foster a community of users who will actively contribute to the creation of this toolkit and support each other?

The Sentinels EOPF Toolkit team, comprising Development Seed, SparkGeo and thriveGEO, together with a group of champion users (early-adopters), are creating a set of Jupyter Notebooks and plug-ins that showcase the use of Zarr format Sentinel data for applications across multiple domains. In addition, community engagement activities such as a notebook competition and social media outreach will bring Sentinel users together and spark interaction with the new data format in a creative yet supportive environment. Such community and user adoption efforts are necessary in order to overcome adoption and uptake barriers and to build up trust and excitement to try out new technologies and new developments around data spaces.

In addition to introducing the Sentinels EOPF Toolkit, this presentation will also highlight lessons learned from working closely with users on barriers they face in adopting the new Zarr format and how to address them. 

How to cite: Szeto, S. H., Wagemann, J., Mathot, E., and Banting, J.: The Sentinels EOPF Toolkit: Driving Community Adoption of the Zarr data format for Copernicus Sentinel Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15864, https://doi.org/10.5194/egusphere-egu25-15864, 2025.

EGU25-16791 | ECS | Posters on site | ESSI2.13

Development and performance evaluation of dissolved oxygen climatology in the Northwestern Pacific 

Jae-Ho Lee, Yong Sun Kim, and Sung-Dae Kim

This study developed a monthly regional atlas for dissolved oxygen (DO) with a quarter-degree horizontal resolution and 73 vertical levels over the northwestern Pacific. We used observed profiles of 586,851 and gridded World Ocean Atlas 2023 (WOA23) with 1° resolution by adopting simple kriging horizontal interpolation and vertical stabilizing techniques to produce the new atlas. This approach efficiently mitigates artificial water masses and statistical noise. The new DO climatology provides detailed information along coasts and renders realistic oxygen distribution associated with the current system in the western North Pacific compared to WOA23. A meridional section demonstrates that the newly developed atlas does not yield artificial noise-like spikes frequently observed in WOA23 in the East Sea. This study expects that this new atlas can allow bio-geochemical numerical models to enhance diagnostic and forecasting performance.

How to cite: Lee, J.-H., Kim, Y. S., and Kim, S.-D.: Development and performance evaluation of dissolved oxygen climatology in the Northwestern Pacific, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16791, https://doi.org/10.5194/egusphere-egu25-16791, 2025.

EGU25-17102 | Posters on site | ESSI2.13

Calculation of Gridded Surface Current from Observed Lagrangian Trajectories in the East Sea 

Mi-Jin Jang, Jae-Ho Lee, and Yong Sun Kim

Surface ocean current is crucial for enhancing the safety and efficiency of maritime logistics and transportation, boosting fisheries production and management, and supporting military operations. This study analyzed 25,342 trajectories from NOAA’s Global Drifter Program (1991–2020), 12 from KIOST, and 63 from KHOA (2015–2024). The surface drifters entering the East Sea were extracted, and a five-step quality control process was implemented. Unobserved values were removed, quality control was applied based on drogue lost, abnormally speed or stuck, unrealistic acceleration. To estimated the gridded oceanic current with high-resolution, we removed the Ekman current and tides from the observed velocity and took advantage of a simple kriging approach. The validation against existing datasets confirmed that major ocean currents exhibited similar patterns compared to absolute geostrophic current from the satellite-based altimetry. The constructed dataset is expected to contribute to the accurate identification of surface current movements and the development of realistic models that incorporate regional characteristics based on data assimilation.

How to cite: Jang, M.-J., Lee, J.-H., and Kim, Y. S.: Calculation of Gridded Surface Current from Observed Lagrangian Trajectories in the East Sea, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17102, https://doi.org/10.5194/egusphere-egu25-17102, 2025.

The Copernicus Program is the largest and most successful public space program globally. It provides continuous data across various spectral ranges, with an archive exceeding 84 petabytes and a daily growth of approximately 20 TB, both of which are expected to increase further. The openness of its data has contributed to the widespread use of Earth observation and the development of commercial products utilizing open data in Europe and worldwide. The entire archive, along with cloud-based data processing capabilities, is available free of charge through the Copernicus Data Space Ecosystem initiative and continues to evolve to meet global user standards. 

This paper presents the process of creating the STAC Copernicus Data Space Ecosystem catalog—the largest and most comprehensive STAC catalog in terms of metadata globally. It details the workflow, starting from the development of a metadata model for Sentinel data, through efficient indexing based on the original metadata files accompanying the products, to result validation and backend system ingestion (via database DSN). A particular highlight is that this entire process is executed using a single tool, eometadatatool, initially developed by DLR, further enhanced, and released as open-source software by the CloudFerro team. The eometadatatool facilitates metadata extraction from the original files accompanying Copernicus program products and others (e.g., Landsat, Copernicus Contributing Missions) using a CSV file containing the metadata name, the file in which it occurs, and the path to the key within the file. Since the CDSE repository operates as an S3 resource offering users free access, the tool supports product access via S3 resources by default, configurable through environment variables. All the above characterizes eometadatatool as the most powerful stactool (a high-level command-line tool and Python library for working with STAC) package available, providing both valid STAC items and a method for uploading them to the selected backend. 

The standard specification itself has been influenced by the CDSE catalog development process, which contributed to the evolution of the standard by introducing version 1.1 and updated extensions (storage, eo, proj) that better meet user needs. The paper discusses the most significant modifications, their impact on the catalog’s functionality, and outlines the main differences. 

Particular attention is given to performance optimization due to the substantial data volume and high update frequency. The study examines the configuration and performance testing (using Locust) of the frontend layer (stac-fastapi-pgstac) and backend (pgstac). The stac-fastapi-pgstac implementation was deployed on a scalable Kubernetes cluster and underwent a product hydration process (specific to managing JSON data in pgstac), leveraging Python's native capabilities for this task. The pgstac schema was deployed on a dedicated bare-metal server with a PostgreSQL database, utilizing master-worker replication enabled through appropriate pgstac configuration. Both software tools are open source, and the achieved optimal configurations are documented and will be presented in detail. 

The presented solution empowers the community to fully utilize the new catalog, leverage its functionalities, and access open tools that enable independent construction of STAC catalogs compliant with ESA and community recommendations. 

How to cite: Niemyjski, M. and Musiał, J.: Building the Copernicus Data Space Ecosystem STAC Catalog: Methodologies, Optimizations, and Community Impact, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17171, https://doi.org/10.5194/egusphere-egu25-17171, 2025.

EGU25-17172 | ECS | Orals | ESSI2.13

Neural Embedding Compression for Earth Observation Data – an Ablation Study 

Amelie Koch, Isabelle Wittmann, Carlos Gomez, Rikard Vinge, Michael Marszalek, Conrad Albrecht, and Thomas Brunschwiler

The exponential growth of Earth Observation data presents challenges in storage, transfer, and processing across fields such as climate modeling, disaster response, and agricultural monitoring. Efficient compression algorithms—either lossless or lossy—are critical to reducing storage demands while preserving data utility for specific applications. Conventional methods, such as JPEG and WebP, rely on hand-crafted base functions and are widely used. However, Neural Compression, a data-driven approach leveraging deep neural networks, has demonstrated superior performance by generating embeddings suitable for high levels of entropy encoding, enabling more accurate reconstructions at significantly lower bit rates.

In our prior work, we developed a Neural Compression pipeline utilizing a masked auto-encoder, embedding quantization, and an entropy encoder tailored for satellite imagery [1]. Instead of reconstructing original images, we evaluated the reconstructed embeddings for downstream tasks such as image classification and semantic segmentation. In this study, we conducted an ablation analysis to quantify the contributions of individual pipeline components—encoder, quantizer, and entropy encoder—toward the overall compression rate. Our findings reveal that satellite images achieve higher compression rates compared to ImageNet samples due to their lower entropy. Furthermore, we demonstrate the advantages of learned entropy models over hand-crafted alternatives, achieving better compression rates, particularly for datasets with seasonal or geospatial coherence. Based on these insights, we provide a list of recommendations for optimizing Neural Compression pipelines to enhance their performance and efficiency.

This work was conducted under the Embed2Scale project, supported by the Swiss State Secretariat for Education, Research and Innovation (SERI contract no. 24.00116) and the European Union (Horizon Europe contract no. 101131841).

[1] C. Gomes and T. Brunschwiler, “Neural Embedding Compression for Efficient Multi-Task Earth Observation Modelling,” IGARSS 2024, Athens, Greece, 2024, pp. 8268-8273, doi: 10.1109/IGARSS53475.2024.10642535.

How to cite: Koch, A., Wittmann, I., Gomez, C., Vinge, R., Marszalek, M., Albrecht, C., and Brunschwiler, T.: Neural Embedding Compression for Earth Observation Data – an Ablation Study, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17172, https://doi.org/10.5194/egusphere-egu25-17172, 2025.

EGU25-17326 | Orals | ESSI2.13

The UK EO DataHub - a pathfinder programme to develop a data space for UK industry, public and academic sectors 

Philip Kershaw, Rhys Evans, Fede Moscato, Dave Poulter, Alex Manning, Jen Bulpett, Ed Williamson, John Remedios, Alastair Graham, Daniel Tipping, and Piotr Zaborowski

The EO DataHub is a new national data space which has been under development as part of a two-year pathfinder programme to facilitate the greater exploitation of EO data for UK industry, public sector and academia. The project has been led by the UK National Centre for Earth Observation partnered with public sector bodies, the UK Space Agency, Met Office, Satellite Applications Catapult and National Physical Laboratory and enlisting commercial suppliers for the development and delivery of the software.

The Hub joins a crowded space in this sector as it joins a growing number of similar such platforms. However, as a national platform (with government as an anchor tenant) it is seeking to provide a unique offering as a trusted source of data, integrating curated data products from the science community building on UK strengths in climate research.

The architecture can be considered as a three layer model. At the base layer, different data sources are integrated - both commercial (Airbus and Planet Labs) and academic providers - from the CEDA data archive (https://archive.ceda.ac.uk) hosted on the JASMIN supercomputer (https://jasmin.ac.uk). The data catalogue now includes high and very high resolution SAR and optical products, Sentinel, UK Climate Projections, CMIP (https://wcrp-cmip.org), CORDEX (https://cordex.org) and outputs from EOCIS (https://eocis.org) consisting of a range of satellite-derived climate data products.

The middle layer, the Hub Platform provides services and APIs including federated search which integrates the data from the various providers, image visualisation, a workflow engine, user workspaces and interactive analysis environments. These build on the work of ESA's EOEPCA (https://eoepca.org) and apply open standards from the Open Geospatial Consortium and STAC (https://stacspec.org/) for cataloguing. In providing this suite of services, the goal is to provide a toolkit to facilitate application developers and EO specialists in building new applications and tools to exploit the data. This forms the final layer in the architecture: as part of the programme, three example application scenarios have been funded, each partnered with a target set of users. These include 1) an application taking climate projections and land surface temperature datasets to provide risk assessments for land assets (led by SparkGeo); 2) a land cover application (Spyrosoft) and finally 3), rather than an application in its own right, a project to develop a client toolkit for use with Jupyter Notebooks and a plugin integrating the Hub’s functionality into the open source GIS desktop application QGIS (work led by Oxidian).

Over the course of the programme, running in parallel to the system development, a dedicated study has been undertaken to develop a model for future sustainability of the platform tackling engagement with potential users and cost models. At the beginning, a funding call seeded early pilots to investigate application scenarios that the platform could support. As this initial phase of the Hub completes, work is underway to engage with early adopters and provide training resources for new users.

How to cite: Kershaw, P., Evans, R., Moscato, F., Poulter, D., Manning, A., Bulpett, J., Williamson, E., Remedios, J., Graham, A., Tipping, D., and Zaborowski, P.: The UK EO DataHub - a pathfinder programme to develop a data space for UK industry, public and academic sectors, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17326, https://doi.org/10.5194/egusphere-egu25-17326, 2025.

EGU25-17799 | Posters on site | ESSI2.13

Data Spaces and geodata workflows for environmental protection 

Matthes Rieke, Benjamin Proß, Simon Jikra, Sotiris Aspragkathos, Iasonas Sotiropoulos, Stamatia Rizou, and Lisa Pourcher

The concept of Data Spaces has gained traction in recent years. Major representatives emerged which have the technological maturity as well as support by relevant decision and policy makers (e.g.  the International Data Spaces Association (IDSA) or Gaia-X). These follow different architectural approaches. In this session we want to illustrate the challenges of integrating the Data Space architectures with established concepts of Spatial Data Infrastructure.

During the next 4 years, the ENFORCE project (Empower citizeNs to join Forces with public authORities in proteCting the Environment) is dedicated to fostering sustainable practices and ensuring environmental regulatory compliance by integrating citizen science with innovative technologies. By employing Living Labs and citizen science methodologies, ENFORCE will create innovative tools that bridge the gap between data reporting, monitoring, and policy enforcement. The project integrates data collection (e.g. Copernicus satellite data), analysis, and stakeholder participation to meet these goals. ENFORCE will leverage geospatial intelligence and explainable AI to enhance environmental governance. These tools and strategies will be tested and refined at eight pilot sites in seven countries, supplemented by capacity-building and policy recommendation efforts.

The design and development of a geospatial information infrastructure that supports the envisioned data workflows is a key challenge addressed by ENFORCE. This infrastructure will prioritize the integration of OGC API-driven systems into the Data Space ecosystem, forming a central component of the project’s agenda. Through development of a blueprint architecture for integration, the project will identify gaps and missing components in current systems, aligning with standards such as the FAIR principles and open data. The concepts will be facilitated in an ENFORCE “Tools Plaza”, an innovative platform providing data science and analytical capabilities for environmental compliance workflows.

How to cite: Rieke, M., Proß, B., Jikra, S., Aspragkathos, S., Sotiropoulos, I., Rizou, S., and Pourcher, L.: Data Spaces and geodata workflows for environmental protection, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17799, https://doi.org/10.5194/egusphere-egu25-17799, 2025.

EGU25-19418 | Posters on site | ESSI2.13

Lossy Data Compression Exploration in an Online Laboratory and the Link to HPC Design Decisions 

Karsten Peters-von Gehlen, Juniper Tyree, Sara Faghih-Naini, Peter Dueben, Jannek Squar, and Anna Fuchs

It is apparent that the data amounts expected to be generated by current and upcoming Earth System Science research and operational activities stress the capabilities of HPC and associated data infrastructures. Individual research projects focusing on running global Earth System Models (ESMs) at spatial resolution of 5km or less can easily occupy several petabytes on disk. With multiple of such projects running on a single HPC infrastructure, the challenge of storing the data alone becomes apparent. Further, community-driven activities like model intercomparison projects – which are conducted for both conventional and high-resolution model setups – add to the aforementioned strain on storage systems. Hence, when planning for next-generation HPC systems, the storage requirements of state-of-the-art ESM-centered projects have to be clear so that systems are still fit-for-use 5 years down the road from the initial planning stage.

As computational hardware costs per performance unit (FLOP or Byte) are not decreasing anymore like they have in the past decades, HPC system key figures do not increase substantially anymore from one generation to the next. The mismatch between demands of research and what future systems can offer is therefore clear.

One apparent solution to this problem is to simply reduce the amount of data from ESM simulations stored on a system. Data compression is one candidate to achieve this. Current ESM projects already utilize application-side lossless compression techniques, which help reduce storage space. However, decompression may incur performance penalties, especially when read patterns misalign with the compression block sizes. Lossy compression offers the potential for higher compression rates, without access penalties for data retrieval. However, its suitability is highly content-dependent, raising questions about which lossy compression methods are best suited for specific datasets. On a large scale, applying lossy compression also prompts the consideration of how such data reduction could shape the design of next-generation HPC architectures.

With lossy compression not being very popular in the ESM-community so far, we present a key development of the ongoing ESiWACE3 project: an openly accessible Jupyter-based online laboratory for testing lossy compression techniques on ESM output datasets. This online tool currently comes with a set of notebooks allowing users to objectively evaluate the impact lossy compression has on analyses performed on the compressed compared to the input data. With some compressors promising compression ratios of 10x-1000x, providing such tools to ensure compression quality is essential. The motivation behind the online compression laboratory is to foster the acceptance of lossy compression techniques by conveying first-hand experience and immediate feedback of benefits or drawbacks of applying lossy compression algorithms. 

Going one step further, we illustrate the impacts that applying lossy-compression techniques on ESM data on large-scales can have on the design decisions made for upcoming HPC infrastructures. We illustrate, among others, that increased acceptance and application of lossy compression techniques enables more efficient resource utilization and allows for smarter reinvestment of funds saved from reduced storage demands, potentially leading to the acquisition of smaller systems and thus enabling increased research output per resource used.

How to cite: Peters-von Gehlen, K., Tyree, J., Faghih-Naini, S., Dueben, P., Squar, J., and Fuchs, A.: Lossy Data Compression Exploration in an Online Laboratory and the Link to HPC Design Decisions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19418, https://doi.org/10.5194/egusphere-egu25-19418, 2025.

EGU25-20188 | ECS | Posters on site | ESSI2.13

Creating TROPOMI superobservations for data assimilation and model evaluation 

Pieter Rijsdijk, Henk Eskes, Kazuyuki Miyazaki, Takashi Sekiya, and Sander Houweling

Satellite observations of tropospheric trace gases and aerosols are evolving rapidly. Recently launched instruments provide increasingly higher spatial resolutions with footprint diameters in the range of 2–8 km, with daily global coverage for polar orbiting satellites or hourly observations from geostationary orbit. Often the modelling system has a lower spatial resolution than the satellites used, with a model grid size in the range of 10–100 km. When the resolution mismatch is not properly bridged, the final analysis based on the satellite data may be degraded. Superobservations are averages of individual observations matching the resolution of the model and are functional to reduce the data load on the assimilation system. In this paper, we discuss the construction of superobservations, their kernels and uncertainty estimates. The methodology is applied to nitrogen dioxide tropospheric column measurements of the TROPOMI instrument on the Sentinel-5P satellite. In particular, the construction of realistic uncertainties for the superobservations is non-trivial and crucial to obtaining close to optimal data assimilation results. We present a detailed methodology to account for the representativity error when satellite observations are missing due to e.g. cloudiness. Furthermore, we account for systematic errors in the retrievals leading to error correlations between nearby individual observations contributing to one superobservation. Correlation information is typically missing in the retrieval products where an error estimate is provided for individual observations. The various contributions to the uncertainty are analysed: from the spectral fitting, the estimate of the stratospheric contribution to the column and the air-mass factor. The method is applied to TROPOMI data but can be generalised to other trace gases such as HCHO, CO, SO2 and other instruments such as OMI, GEMS and TEMPO. The superobservations and uncertainties are tested in the ensemble Kalman filter chemical data assimilation system developed by JAMSTEC. These are shown to improve forecasts compared to thinning or compared to assuming fully correlated or uncorrelated uncertainties within the superobservation. The use of realistic superobservations within model comparisons and data assimilation in this way aids the quantification of air pollution distributions, emissions and their impact on climate.

How to cite: Rijsdijk, P., Eskes, H., Miyazaki, K., Sekiya, T., and Houweling, S.: Creating TROPOMI superobservations for data assimilation and model evaluation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20188, https://doi.org/10.5194/egusphere-egu25-20188, 2025.

EGU25-20430 | Orals | ESSI2.13

Compression and Aggregation: a CF data model approach 

David Hassell, Sadie Bartholomew, Bryan Lawrence, and Daniel Westwood

The CF (Climate and Forecast) metadata conventions for netCDF datasets describe means of "compression-by-convention", i.e. methods for compressing and decompressing data according to algorithms that are fully described within the conventions themselves. These algorithms, which can be lossless or lossy, are not applicable to arbitrary data, rather the data have to exhibit certain characteristics to make the compression worthwhile, or even possible.

Aggregation, available in CF-1.13, provides the utility of being able to view, as a single entity, a dataset that has been partitioned across multiple other independent datasets on disk, whilst taking up very little extra space on disk since the aggregation dataset contains no copies of the data in each component dataset. Aggregation can facilitate a range of activities such as data analysis, by avoiding the computational expense of deriving the aggregation at the time of analysis; archive curation, by acting as a metadata-rich archive index; and the post-processing of model simulation outputs, by spanning multiple datasets written at run time that together constitute a more cohesive and useful product. CF aggregation currently has cf-python and xarray implementations.

The conceptual CF data model does not recognise compression nor aggregation, choosing to view all CF datasets as if they were uncompressed and containing all of their own data. As a result, the cf-python data analysis library, that is built exactly on the CF data model, also presents datasets lazily to the user in this manner, without decompressing or re-combining the data in memory until the user actually accesses the data, at which time it occurs automatically. This approach allows the user to interact with their data in an intuitive and efficient manner; and also removes the need for the user to have to assimilate large parts of the CF conventions and having to create their own code for dealing with the compression and aggregation algorithms.

We will introduce compression by ragged arrays (as used by Discrete Sampling Geometry features, such as timeseries and trajectories) and dataset aggregation, with cf-python examples to demonstrate the ease of use that arises from using the CF data model interpretation of the data.

How to cite: Hassell, D., Bartholomew, S., Lawrence, B., and Westwood, D.: Compression and Aggregation: a CF data model approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20430, https://doi.org/10.5194/egusphere-egu25-20430, 2025.

Integrating High-Performance Computing (HPC) and cloud computing in climate sciences is difficult, due to intricate hardware/software, compatibility, performance and reproducibility issues. Here, we address these challenges in a user-friendly way by leveraging the Conda ecosystem and containers.

Containerization allows to match or exceed native performance on HPC while ensuring bit-for-bit reproducibility for deterministic algorithms and similar processor architectures. This approach simplifies deploying climate models across different platforms; for example, CESM 2.2.2 (Community Earth System Model) provides on various clusters throughputs in simulated years per computational day within +/- 1% of bare-metal performance for simulations spanning thousands of processors.

Exclusively using generic Conda packages for MPI (Message Passing Interface) applications was challenging in HPC. Although OpenMPI included UCX (Universal Communication X) and OFI (Open Fabric Interface), it lacked UCC (Unified Collective Communication) and wasn't optimized by default for high-performance networks like InfiniBand, RoCE (Remote Direct Memory Access over Converged Ethernet) and HPE (Hewlett Packard Enterprise) Slingshot-11, often defaulting to TCP/IP (Transmission Control Protocol/Internet Protocol) or failing. 
 
After updating Conda-Forge’s OpenMPI and MPICH feedstocks, we are adding MVAPICH and ParaStationMPI support to PnetCDF, HDF5, NetCDF-C, NetCDF-Fortran and ESMF (Earth System Modeling Framework) libraries critical for modellers, alongside libFabric and openPMIx (Process Management Interface - Exascale). This incidentally exposed ABI (Application Binary Interface) compatibility issues. Now, MPI toolchains featuring major UCX/OFI/PMIx versions ensure consistent operation across different hosts without affecting numerical results. Using the same Conda environment inside a container, and no hardware-specific optimization, preserves bitwise reproducibility. OMB (Ohio State University Micro-Benchmark) tests for latency, bandwidth and other metrics help confirm if optimal performance can be achieved or not. 

Such developments enable climate scientists to focus on addressing scientific questions rather than sorting out software dependencies and technical problems. One can write code on a laptop then effortlessly scale to cloud or supercomputers, and seamlessly run climate simulations somewhere then continue these wherever compute resources are available without worrying about discontinuities. This also releases expensive HPC resources for production instead of wasting them for training, learning, development or testing which can be performed comfortably elsewhere, without job scheduling constraints, in the very same software environment.

Conda has primarily been developed with a focus on compatibility which limits its suitability in highly performance-sensitive applications where locally optimized builds of specific key components are paramount, typically in climate modeling. Additionally, instead of relying on local engineers to install and maintain host software, Conda users can benefit from the work of thousands of open-source contributors who continuously update and test the entire ecosystem.

This strategy fits the session's theme by providing a framework where cloud resources can be utilized for big data without compromising the performance or rigor of HPC environments. Conda and container technologies ought to change how climate scientists approach software management, focusing on ease of use, scalability and reproducibility, thereby potentially altering practices within the field to improve usage of computational resources and leverage community efforts to remain at the forefront.

How to cite: Iaquinta, J., Fouilloux, A., and Ragan-Kelley, B.: Climate Modeling with Conda and Containers to Improve Computational Resource Usage while Achieving Native Performance and Reproducibility, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2605, https://doi.org/10.5194/egusphere-egu25-2605, 2025.

EGU25-6798 | Orals | ESSI2.15

Advancing Geophysical Data Analysis: HEALML for Efficient Sphere-Based Statistics on Pangeo-EOSC 

Jean-Marc Delouis, Erwan Allys, Justus Mangin, Louise Mousset, and Tina Odaka

A significant challenge in data integration and ML methodologies on cloud infrastructures is accurately determining correlated statistics. Initially, aligning data to a consistent pixel grid is essential, motivating the use of Discrete Global Grid Systems (DGGS). In geophysical studies, data reside on a sphere, and approximating with tangent planes can distort results. Our solution is the HEALPix pixelization as our DGGS framework, standardizing data on a common grid for consistent statistical analysis. HEALPix's unique features, such as its iso-latitude layout and uniform pixel areas, enable the use of spin-weighted spherical harmonics in managing vector fields. This enables the accurate calculation of  correlation statistics, such as between velocity and scalar fields on the sphere, while minimizing biases due to spherical approximations. By utilizing the HEALPix framework, known in cosmology, with TensorFlow or PyTorch as backends, we created the: HEALML library. This library facilitates gradient computations of all derived statistics for AI optimization, and has been validated on the Pangeo-EOSC platform. This library parallelizes the computation of localized spherical harmonics and includes features like scattering covariance calculations, allowing the extraction of more complex nonlinear statistics beyond the power spectrum. We compare these results to traditional 2D planar methods, demonstrating the advantages of sphere-based statistics on platforms like Pangeo-EOSC. Furthermore, we demonstrate: HEALML's ability to emulate using a substantially smaller dataset. This demonstration emphasizes the ways in which incorporating spherical statistical methods into Pangeo-EOSC fosters innovative and efficient statistical analysis within geophysical research.

How to cite: Delouis, J.-M., Allys, E., Mangin, J., Mousset, L., and Odaka, T.: Advancing Geophysical Data Analysis: HEALML for Efficient Sphere-Based Statistics on Pangeo-EOSC, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6798, https://doi.org/10.5194/egusphere-egu25-6798, 2025.

EGU25-8127 | ECS | Posters on site | ESSI2.15

Seamless Upscaling Research from Cloud to HPC using eWaterCycle 

Mark Melotto, Rolf Hut, and Bart Schilperoort

The eWaterCycle platform provides hydrologists with a platform that allows them to work with each other's models and data without having to become a computer scientist in the process. The eWaterCycle platform supports existing hydrological models and makes them available for scientists using the BMI model interface as a communication layer. Models run in containers for reproducibility and dependency control. Popular hydrological models are readily available (PCRGLobWB, WFLOW, HBV, etc.). Scientists develop their analyses or experiments in the widely known JupyterHub environment. 

While in theory eWaterCycle can be installed and run on any hardware, in practice most users interact with it on the SURF Research Cloud, a cloud computing infrastructure available to the Dutch academic ecosystem. Until recently upscaling from Cloud to HPC infrastructure for larger model runs required extensive knowledge of the HPC system. Here we will present our work on building a seamless workflow that allows scientists to upscale their cloud based work to the Snellius supercomputer and the Spider grid computer without having to worry about technical issues like mounting points for (large) datasets and container engines.

Our workflow opens up the possibility for more scientists to benefit from HPC and Grid resources while focussing on their domain science. We present the workflow in such a format that it should be easily portable to other hybrid cloud - HPC infrastructures, including the DestinE systems.

How to cite: Melotto, M., Hut, R., and Schilperoort, B.: Seamless Upscaling Research from Cloud to HPC using eWaterCycle, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8127, https://doi.org/10.5194/egusphere-egu25-8127, 2025.

We present recent progress around the EERIE cloud data server (https://eerie.cloud.dkrz.de) and its software stack “cloudify”. The EERIE cloud provides efficient open access to prominent climate datasets stored on disk at the German Climate Computing Center (DKRZ).

A new kerchunk-plugin enables data access to raw model output as-is to enable verifiable data transfer with better latency. STAC (Spatio Temporal Assets Catalog) catalogs are deployed and displayed through the EERIE cloud to make the provided DKRZ datasets findable and accessible. Two in-browser apps can be started, pre-configured for each dataset, by just clicking buttons: (1) the data visualization app “gridlook” as well as a (2) jupyterlite for interactive analysis and monitoring. 

We leverage the python package xpublish, a plugin for Pangeo's central analysis package Xarray. Its main feature is to provide ESM output by mapping any input data to virtual zarr datasets. Users can retrieve these datasets as if they were cloud-native and cloud-optimized.

How to cite: Wachsmann, F.: The EERIE cloud: Apps and Catalogs for Cloudified Earth System Model Output, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8591, https://doi.org/10.5194/egusphere-egu25-8591, 2025.

EGU25-8754 | ECS | Orals | ESSI2.15

Leveraging Cloud, Earth Observation and In-Situ Sensors for Agri-Environmental Monitoring and Policy Decision-Making  

Georgios Charvalis, Panagiota Louka, Vassileios Gkoles,  Thanasis Manos, Nikos Kalatzis, Dionysios Solomos, Anastasios Trypitsidis, and Odysseas Sekkas

Cloud infrastructures play a significant role in delivering secure, scalable and efficient data processing for Earth Observation (EO) and agricultural management applications. As part of the ScaleAgData project, we present a hierarchical Agri-Environmental Monitoring Tool running on a private cloud infrastructure. The system combines data from EO, in-situ sensors and farm management information systems (FMIS), including parcel calendars, to provide farmers and policymakers multi-scale insights.  

The solution is cloud-based and designed with an underlying architecture that ensures both scalability and interoperability, leveraging OGC-compliant data formats where applicable. EO and in-situ data streams can be processed and analyzed efficiently with the help of containerized apps and microservices to facilitate modular development and simplify deployment. By using a web-based dashboard with hierarchical design, stakeholders can navigate from overviews at the municipal level to individual parcels. Aggregated summaries that comply with Common Agricultural Policy (CAP) criteria are useful to policymakers and farmers can get comprehensive parcel-level metrics to optimize irrigation, pesticide use and other agro-related activities.  

Specifically, the tool combines EO data to derive vegetation indices (e.g., NDVI, EVI) and other parameters requiring advanced processing for crop type classification. Furthermore, these datasets are enriched with in-situ sensor measurements (e.g. soil moisture, weather data) and farm logs managed within FMIS (irrigation schedule, pesticide usage). Parcel-level data (L1) is processed to generate statistics, which are then calibrated with nearby parcels data with similar properties and crop type(L2), serving as control level, and finally extrapolated to the municipal level (L3) using spatial averaging techniques  to provide indicators related to irrigation water, pesticide, fertilizer usage, etc.  Farm calendars stored within FMIS provide a reliable source of ground-truth data, enhancing the tool’s ability to validate aggregated metrics. The aggregation at L2 and L3 allows for the identification of regional trends and patterns in agricultural practices, empowering policymakers and stakeholders to implement targeted interventions at both levels, thereby promoting sustainable agriculture.   

This work showcases the potential of private cloud infrastructures to enhance agri-environmental monitoring by processing and integrating heterogeneous data streams (EO, in-situ sensors and farm log data) into a unified system. The system is being applied in diverse agricultural regions of Greece (Crete, Thessaly, Macedonia) with ongoing validation efforts aimed at refining its accuracy and adaptability. Future work includes the integration of cloud-based machine learning models and EO-derived evapotranspiration data to enhance the efficiency of extrapolating parcel-level (L1) and regional (L2) metrics into policy-level indicators (L3). Additionally, alternative aggregation methods, such as model-based approaches, spatial regression, and interpolation techniques like Kriging, will be tested to improve the accuracy and reliability of aggregated insights. 

How to cite: Charvalis, G., Louka, P., Gkoles, V., Manos,  ., Kalatzis, N., Solomos, D., Trypitsidis, A., and Sekkas, O.: Leveraging Cloud, Earth Observation and In-Situ Sensors for Agri-Environmental Monitoring and Policy Decision-Making , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8754, https://doi.org/10.5194/egusphere-egu25-8754, 2025.

EGU25-9432 | Posters on site | ESSI2.15

Co-Creating Cloud-Based Tools for Urban Climate-Resilience: The CLIMRES Project 

Claudio Pisa, Marica Antonacci, Vasileios Baousis, Sotirios Aspragkathos, Iasonas Sotiropoulos, and Stamatia Rizou

Europe faces a growing frequency of extreme weather events, from heatwaves and floods to wildfires and earthquakes, increasingly threatening urban environments. Unusually warm winters are becoming progressively common, destabilizing ecosystems and altering traditional weather dynamics. 

Addressing these crucial changes, the CLIMRES project aims to foster a “Leadership for Climate-Resilient Buildings” by identifying and categorizing vulnerabilities within the built environment and assessing their effects within urban systems. This effort integrates diverse data sources, including Copernicus services, IoT networks, and municipal datasets, and considers hazard warnings and weather forecasts. Moreover, a liaison with the Destination Earth initiative enhances the project with the capacity to leverage extreme weather predictions and future climate models. 

CLIMRES aims to deliver vulnerability assessment and impact evaluation methodologies, along with a “hub of measures” inventory for cost-effective building design and materials against climate risks, as well as decision support tools, to aid building owners, policymakers and stakeholders in planning effective interventions and to address vulnerabilities, targeting three levels of decision making at strategic, tactical and operational levels. The project deploys cloud technologies like OpenStack and Kubernetes to host an interoperable platform for vulnerability analysis, data harmonization, and decision-making. Its solutions will be tested and validated on 3 Large Scale Pilots in Spain, Greece, Italy, and Slovenia, addressing hazards such as heatwaves, flooding, fires, and earthquakes. A multi-hazard replication pilot in France will further evaluate the scalability and versatility of these approaches across diverse contexts. 

Insights from these pilots will feed into a replication roadmap and a capacity-building program designed to train future leaders in climate-resilient urban development. By fostering co-creation with local stakeholders and communities, CLIMRES ensures its innovative solutions are practical, cost-effective, and replicable, targeting Technology Readiness Levels (TRL) 6-8. 

CLIMRES aims to bridge innovation with actionable solutions, equipping building owners, policymakers, and communities with the tools needed to enhance urban climate resilience. This presentation highlights the project’s interdisciplinary approach, outputs and technological underpinnings, offering insights into scalable solutions for climate adaptation in urban settings. 

How to cite: Pisa, C., Antonacci, M., Baousis, V., Aspragkathos, S., Sotiropoulos, I., and Rizou, S.: Co-Creating Cloud-Based Tools for Urban Climate-Resilience: The CLIMRES Project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9432, https://doi.org/10.5194/egusphere-egu25-9432, 2025.

EGU25-10683 | Orals | ESSI2.15

Copernicus data and services uptake with EO4EU platform: an AI-augmented ecosystem for Earth Observation data accessibility and exploitation. 

Federico Fornari, Vasileios Baousis, Mohanad Albughdadi, Marica Antonacci, Tolga Kaprol, Claudio Pisa, Charalampos Andreou, Kakia Panagidi, and Stathes Hadjiefthymiades

The Copernicus program has fostered Earth Observation (EO) and Earth Modeling by offering extensive data and services to European Citizens. Sentinel satellites’ data is accessible  through platforms like the Copernicus Open Access Hub and the Copernicus Data Space Ecosystem, which provide a wide range of information on land, ocean and atmospheric conditions. Complementing these resources, six specialized Copernicus services deliver data in domains such as the atmosphere, marine environment, land monitoring, climate change, security and emergency response. To streamline access and usability, cloud-based Copernicus Data and Information Access Services (DIAS) offer centralised platforms equipped with cloud infrastructure and processing tools. Building on these efforts, the Copernicus Data Space Ecosystem (https://dataspace.copernicus.eu/) enhances existing DIAS services with advanced functionalities like improved search capabilities, virtualizations and APIs. Meanwhile, the Destination Earth (DestinE) initiative led by ECMWF, EUMETSAT and ESA, aims to develop high-precision digital Earth models - or digital twins - that simulate natural and human activities. These models mainly focus on weather-induced extremes and climate change adaptation, generating valuable Earth Modeling data. Furthermore, European Data Spaces integrate datasets across diverse domains, including agriculture, health, energy, and environmental monitoring, creating opportunities to combine these resources with Copernicus and DestinE data through advanced technologies like artificial intelligence (AI) and machine learning (ML). This integration paves the way for innovative solutions and public-facing products and services. Despite the volume and richness of Copernicus and related EO data, its accessibility remains limited, with most users being experts or scientists. For broader industry adoption and the development of impactful applications that benefit society and the enviroment, significant barriers must be addressed. EO data is often fragmented, complex, and difficult to process, requiring domain expertise for tasks such as data discovery, pre-processing, storage, and conversion into formats suitable for analytics and Geographic Information Systems (GIS).

The EO4EU platform (https://www.eo4eu.eu/), showcased in this presentation, introduces a multi-cloud ecosystem designed for holistic management of EO data. Its primary objective is to bridge the gap between domain experts and end users, leveraging technological advancements to broaden the adoption of EO data across diverse markets. By enhancing the accessibility and usability of EO data, EO4EU supports market growth through advanced data modeling, dynamic annotation, and state-of-the-art processing, powered by European cloud infrastructures such as WEkEO/DIAS and CINECA. EO4EU provides a suite of innovative tools and methodologies to assist a wide range of users, from professionals and domain experts to general citizens, in benefiting from EO data. Its key features include:

  • Knowledge Graph-based Decision Making: Facilitates insightful feature extraction from diverse repositories, enabling a more comprehensive understanding of datasets.
  • AI/ML Marketplace: A centralized hub for AI & ML models, algorithms, techniques, and metadata.
  • Big Data Processing Engines: Optimized for cloud environments to efficiently manage large-scale datasets.
  • User-friendly Interfaces: GUI, CLI, APIs, and immersive VR experiences, targeting both technical and non-technical users.
  • Workflow Engine: Simplifies the definition and execution of recurring tasks for EO data retrieval and processing.

How to cite: Fornari, F., Baousis, V., Albughdadi, M., Antonacci, M., Kaprol, T., Pisa, C., Andreou, C., Panagidi, K., and Hadjiefthymiades, S.: Copernicus data and services uptake with EO4EU platform: an AI-augmented ecosystem for Earth Observation data accessibility and exploitation., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10683, https://doi.org/10.5194/egusphere-egu25-10683, 2025.

EGU25-10977 | Posters on site | ESSI2.15

Cloud-Powered Earth Observation Tools for Urban Resilience: The BUILDSPACE Project 

Marica Antonacci, Vasileios Baousis, Claudio Pisa, Stamatia Rizou, and Iasonas Sotiropoulos

The BUILDSPACE project harnesses the transformative potential of cloud computing to evolve urban development and resilience practices. By integrating advanced Earth Observation (EO) data with state-of-the-art satellite and cloud technologies, BUILDSPACE addresses critical urban challenges, including climate adaptation, energy efficiency, and disaster resilience, while contributing to the European Green Deal’s objectives of sustainability and carbon neutrality. 

Central to BUILDSPACE are five innovative services designed to support urban decision-making. At the building scale, the project facilitates the generation and visualization of detailed digital twins through interactive displays, virtual reality (VR), and augmented reality (AR) interfaces. These digital twins enable precise simulations for energy optimization, operational efficiency, and climate impact assessment. At the city scale, BUILDSPACE provides tools to address climate scenarios, such as urban heat islands and flooding, empowering municipalities and urban planners with actionable insights through interactive, map-based platforms. 

The project’s technical foundation lies in a robust, cloud-native architecture built on Kubernetes and OpenStack, combined with a DevOps methodology to streamline both infrastructure services and application deployment. Kubernetes orchestrates containerised workloads, enabling efficient automated deployment, scaling and management of applications, while OpenStack provides a flexible infrastructure for managing compute, storage, and networking resources. Through the DevOps approach, we ensure continuous integration and delivery (CI/CD), fostering rapid development cycles and operational agility. By adopting open-source cloud platforms, the project ensures interoperability, reproducibility and automation across diverse environments, driving consistency and efficiency throughout the lifecycle of both infrastructure and applications. 

The project’s services are being validated across four European cities representing diverse climatic conditions, namely Warsaw, Riga, Piraeus and Ljubljana. These validations focus on two scenarios: construction companies monitoring building processes with advanced digital tools, and municipalities analysing the impacts of climate change on urban infrastructure. 

By advancing from TRL 5-6 to TRL 7-8, BUILDSPACE aims to deliver market-ready solutions that align with the European GNSS and Copernicus initiatives and to synchronise with the advances, concerning Digital Twin technologies and data federation mechanisms, of the Destination Earth initiative, while paving the way for a broader adoption of cloud technologies in EO-based urban resilience applications. 

How to cite: Antonacci, M., Baousis, V., Pisa, C., Rizou, S., and Sotiropoulos, I.: Cloud-Powered Earth Observation Tools for Urban Resilience: The BUILDSPACE Project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10977, https://doi.org/10.5194/egusphere-egu25-10977, 2025.

EGU25-11810 | Orals | ESSI2.15

DeployAI Earth Observation Services: Enabling Environmental Insights on the European AI-on-Demand Platform 

Antonis Troumpoukis, Mohanad Albughdadi, Martin Welß, Vasileios Baousis, and Iraklis Klampanos

The DeployAI project [1] designs and delivers a fully operational European AI-on-Demand Platform (AIoDP) to empower the European industry with access to cutting-edge AI technology, and to promote trustworthy, ethical, and transparent European AI solutions, with a focus on SMEs and the public sector. To achieve this, the platform enables the development and deployment of AI solutions through the following core solutions: (i) AI Builder [2], which allows the assembling of reusable AI modules into AI pipelines; (ii) seamless access to Cloud and HPC infrastructures (e.g., MeluXina and LUMI); (iii) a marketplace for the listing and distribution of ready-to-use AI products; (iv) an expansive and growing library of diverse AI-driven use cases.

As part of its domain-driven solutions, AIoDP seeks to empower Environmental Scientists, AI Engineers, Developers, Researchers, and SMEs via the DeployAI Earth Observation Services. These services will accelerate the development of AI-driven environmental applications, by providing pre-trained models that simplify satellite imagery processing, land usage classification, and image segmentation. Key models available as modules within the DeployAI’s AI Builder include:

  • Leaf Area Index (LAI) Model: Enables precise monitoring of vegetation health and ecological dynamics by calculating leaf area per unit ground [3]. 
  • Object Detection Model: Identifies specific objects in high-resolution satellite images, supporting applications such as  infrastructure monitoring, pollution tracking, and deforestation assessment [4].
  • Segment Anything Model (SAM): Simplifies analysis across diverse environmental applications through the capabilities of SAM that allows flexible, prompt-based image segmentation for new datasets, with zero-shot and few-shot learning [5].

These models, along with the broader functionalities of AI Builder, enable users to create custom AI pipelines that address their specific environmental challenges in several environmental areas, including vegetation health monitoring, water balance analysis, climate modeling, urban planning, traffic management, pollution monitoring, and infrastructure maintenance. Users can leverage the visual pipeline editor to easily assemble pipelines from reusable AI modules without needing to write code. Once created, these pipelines can be deployed as AI applications on various execution environments. DeployAI facilitates seamless transitions between these environments by providing connectors to a host of target infrastructures, including Cloud platforms and HPC systems. This empowers users to leverage the most suitable computational resources for their specific needs.

By providing a user-friendly platform with access to cutting-edge AI technology and Cloud/HPC resources, DeployAI empowers users to address critical environmental challenges and unlock new possibilities for sustainable development.

[1] https://deployaiproject.eu
[2] https://gitlab.eclipse.org/eclipse/graphene
[3] https://github.com/DeployAI-Environmental-Services/depai-lai
[4] https://github.com/DeployAI-Environmental-Services/depai-yolov8-obb
[5] https://github.com/DeployAI-Environmental-Services/depai-sam-interactive

This work has received funding from the European Union’s Digital Europe Programme (DIGITAL) under grant agreement No 101146490.

How to cite: Troumpoukis, A., Albughdadi, M., Welß, M., Baousis, V., and Klampanos, I.: DeployAI Earth Observation Services: Enabling Environmental Insights on the European AI-on-Demand Platform, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11810, https://doi.org/10.5194/egusphere-egu25-11810, 2025.

EGU25-12070 | ECS | Orals | ESSI2.15

Performance Benchmarking and Energy monitoring for Climate Modelling 

Sergi Palomas, Mario Acosta, Gladys Utrera, Okke Lennart, Daniel Beltran, Miguel Castrillo, Niclas Schroeter, and Ralf Mueller

The computational intensity of climate models makes them among the most energy-demanding applications in High-Performance Computing (HPC), resulting in significant computational costs and carbon emissions. Addressing the dual challenge of improving climate predictions —by running higher resolution, more accurate and complex models— and ensuring sustainability requires innovative tools to evaluate both computational efficiency and energy consumption across diverse HPC architectures. To address this, and in the context of the Center of Excellence in Simulation of Weather and Climate in Europe (ESiWACE), we have extended the High-Performance Climate and Weather Benchmark (HPCW) framework to incorporate a standardised set of Climate Performance Metrics for Intercomparison Projects (CPMIPs) and energy consumption monitoring.

HPCW, originally designed to maintain a set of relevant and realistic, near-operational weather forecast workloads to benchmark HPC sites, can provide insights beyond generic benchmarks like High-Performance Linpack (HPL) or High-Performance Conjugate Gradients (HPCG) by focusing on domain-specific workloads.

The inclusion of CPMIPs into HPCW brings a widely accepted set of metrics specifically tailored to the particularities of climate workflows. These metrics, already recognized by the scientific community, are key to better understanding climate model performance and allow us to keep the results from the framework relevant for research and operational runs, as well as improving our capacity for multi-model multi-platform performance comparisons.

By integrating energy monitoring, HPCW enables users to evaluate how critical computational kernels in climate models perform in terms of energy consumption. Our review of energy profiling tools across EuroHPC pre-exascale systems, including MareNostrum 5, LUMI, and Leonardo, highlights a fragmented landscape. Current tools offer varying granularity and portability, but limitations such as system configurations, administrative restrictions, and hardware compatibility often hinder their application. Low-level interfaces like Running Average Power Limit (RAPL) and Performance Application Programming Interface (PAPI) counters offer precise energy measurements but are constrained by accessibility issues.

These advancements aim to improve the allocation of climate experiments, such as those conducted for the Intergovernmental Panel on Climate Change (IPCC) in Coupled Model Intercomparison Projects (CMIPs), to the most suitable HPC resources, while also identifying architectural bottlenecks before running production experiments. Additionally, by enhancing energy consumption quantification, this work contributes to ongoing efforts to measure and reduce the carbon footprint of the climate research community. Furthermore, these analyses are expected to be particularly valuable for climate researchers, especially in the context of upcoming large-scale initiatives like CMIP7, enabling them to make informed resource requests and facilitate robust multi-platform comparisons of climate model performance which were not possible in the past. We anticipate that HPC vendors can also benefit from the outcomes of our work in optimising the systems for climate modelling workloads. By combining performance and energy metrics within a unified framework, we provide critical insights that align computational advancements with sustainability goals, ensuring efficient and environmentally conscious use of HPC resources for climate research.

How to cite: Palomas, S., Acosta, M., Utrera, G., Lennart, O., Beltran, D., Castrillo, M., Schroeter, N., and Mueller, R.: Performance Benchmarking and Energy monitoring for Climate Modelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12070, https://doi.org/10.5194/egusphere-egu25-12070, 2025.

EGU25-12918 | ECS | Orals | ESSI2.15

Enhancing Pangeo-Fish with HEALPix Convolution: Impact Evaluation and Benefits 

Etienne Cap, Tina Odaka, Jean-Marc Delouis, Justus Magin, and Mathieu Woillez

The Pangeo-Fish project processes biologging data to analyze fish movement and migration patterns.  While SciPy’s convolution methods are robust, they are not optimized for handling spherical datasets inherent to Earth system science. To address this limitation, we propose the integration of HEALPix convolution, a method designed for spherical operations, into Pangeo-Fish.

HEALPix convolution offers distinct advantages for geophysical data analysis, particularly when dealing with spherical datasets in Earth system science. It uses the HEALPix pixelization as a core Discrete Global Grid System (DGGS), which ensures equally-sized pixels globally, removing distortions common in flat projections. This consistency is crucial for maintaining the physical relevance of convolutions across locations. Additionally, HEALPix’s dyadic property supports flexible, multiscale resolution adjustments, allowing for downscaling while preserving accuracy. Such scalability is essential for studying oceanic environments where areas of interest, like coastal zones and basins, are often resolution-dependent.

Our approach evaluates the performance of HEALPix convolution in comparison to traditional SciPy methods, focusing on its ability to enhance the accuracy of habitat mapping and migration pathway modeling for fish. 

This integration is particularly relevant within the Global Fish Tracking System (GFTS), which operates under the European Union’s Destination Earth (DestinE) initiative. GFTS utilizes datasets from Copernicus Marine Services and the European Tracking Network (ETN) to model fish habitats, spawning grounds, and migration swimways. HEALPix convolution strengthens the pangeo-fish’s capacity for studying Species such as tuna and eel that exhibit large-scale, transoceanic migrations.    

In conclusion, this work highlights the transformative potential of HEALPix convolution in spherical data processing. By integrating this innovative method, Pangeo-Fish can provide more accurate, scalable, and actionable insights into fish behaviors and habitats, contributing to sustainable management practices and conservation strategies globally.

 

How to cite: Cap, E., Odaka, T., Delouis, J.-M., Magin, J., and Woillez, M.: Enhancing Pangeo-Fish with HEALPix Convolution: Impact Evaluation and Benefits, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12918, https://doi.org/10.5194/egusphere-egu25-12918, 2025.

EGU25-13725 | Posters on site | ESSI2.15

Cloud-based platform for the management of hydrogeological risks in the Po Basin  

Marco Zazzeri and the PARACELSO team

In recent years, technological advances in the use of geospatial data (such as satellite images, anthropogenic and/or environmental raster and vector open data, etc.) for hydrogeological risk assessment, combined with advanced analysis techniques (e.g., machine learning), have become increasingly valuable. These technologies can be utilized by local and national authorities for land planning and emergency management to better understand the dynamics associated with climate change. This understanding can help guide actions aimed at safeguarding not only environmental resources but also socio-economic assets and citizens’ lives.

In pursuit of this goal, a partnership has been established between the Po River Basin District Authority (AdBPo), the Italian Space Agency (ASI), and academic and research institutions such as the University of Bologna (UNIBO), the University of Modena and Reggio Emilia (UNIMORE), the University of Padova (UNIPD), and the Institute of Environmental Geology and Geoengineering of the National Research Council of Italy (CNR-IGAG). The aim is to implement a downstream service for monitoring landscape evolution related to fluvial systems (geomorphological classification), and slope dynamics (including landslides and rock glaciers) and to quantitatively evaluate the exposed assets.

The PARACELSO project (Predictive Analysis, MonitoRing, and mAnagement of Climate change Effects Leveraging Satellite Observations) aims to develop a modular and interoperable cloud-based platform that supports the analysis of natural phenomena (such as fluvial hydrodynamics, landslides, and rock glaciers) using satellite images provided by:

  • DIAS platforms deployed by the Copernicus Programme (e.g., Sentinel 1-2);
  • ASI missions such as CosmoSkyMed, PRISMA, and SAOCOM.

Furthermore, a methodology integrating Earth Observation and geospatial data analysis has been implemented using open-source libraries.

To facilitate this, the MarghERita supercomputer, named in honor of the scientist Margherita Hack, has been made available by the Emilia-Romagna region. It is used both to store the downloaded satellite images and to run the algorithms developed in the project for studying the temporal evolution of river and slope systems. Finally, it enables the sharing and visualization of processed data.

The project has received funding from ASI through the “I4DP_PA (Innovation for Downstream Preparation for Public Administrations)” Call for Ideas.

How to cite: Zazzeri, M. and the PARACELSO team: Cloud-based platform for the management of hydrogeological risks in the Po Basin , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13725, https://doi.org/10.5194/egusphere-egu25-13725, 2025.

EGU25-13873 | Orals | ESSI2.15

UXarray: Extending Xarray for Enhanced Support of Unstructured Grids 

John Clyne, Hongyu Chen, Philip Chmielowiec, Orhan Eroglu, Cecile Hannay, Robert Jacob, Rajeev Jain, Brian Medeiros, Paul Ullrich, and Colin Zarzycki

Over the past decade, weather and climate models have rapidly adopted unstructured meshes to better leverage high-performance computing systems and approach kilometer-scale resolutions. Output from this new generation of models presents many challenges for their subsequent analysis, largely due to a lack of community tools supporting unstructured grid data. Last year, we introduced UXarray, a class extension of Xarray that provides native support for unstructured meshes. UXarray readily runs in a Jupyter Notebook and offers parallelized execution through its compatibility with Dask, demonstrating its flexibility as both a tool for lightweight exploration and communication, and for supporting intensive calculations applied to vast data volumes. Over the past year, UXarray has matured significantly and is now capable of supporting many real-world analysis workflows applied to outputs from a growing number of high-resolution models and dynamical cores, including ICOsahedral Non-hydrostatic (ICON) atmosphere model, the Finite-Element/volumE Sea ice-Ocean Model (FESOM), NSF NCAR’s Model for Prediction Across Scales (MPAS), and the U.S. DOE’s Energy Exascale Earth System Model (E3SM). This presentation will provide an overview of the UXarray’s current capabilities, which include extensive support for plotting and many foundational analysis operators; demonstrate examples in Jupyter Notebooks; present plans for the future;  and discuss ways for Pangeo and the broader earth system science community to help guide new developments. 

How to cite: Clyne, J., Chen, H., Chmielowiec, P., Eroglu, O., Hannay, C., Jacob, R., Jain, R., Medeiros, B., Ullrich, P., and Zarzycki, C.: UXarray: Extending Xarray for Enhanced Support of Unstructured Grids, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13873, https://doi.org/10.5194/egusphere-egu25-13873, 2025.

EGU25-14306 | ECS | Orals | ESSI2.15

Navigating New Grids: Evaluating DGGS Configurations for Marine Spatial Analysis 

Kayziel Martinez, Alexander Kmoch, Lőrinc Mészáros, Andrew Nelson, and Evelyn Uuemaa

Accurate and efficient spatial analysis is crucial for the mapping and sustainable management of marine environments, where large-scale and diverse datasets present significant analytical challenges. Traditional latitude-longitude methods, while widely used, often encounter limitations in data integration and handling distortion caused by Earth’s curvature. Discrete Global Grid Systems (DGGS) have emerged as a promising solution, offering a hierarchical, global, and equal-area framework for geospatial analysis. Despite their potential, the performance in marine spatial analysis remains underexplored.

This study evaluates the impact and suitability of DGGS-based spatial analysis by comparing its performance with the traditional latitude-longitude approaches. Using marine datasets representing point and raster data formats, the workflow begins with quantization, converting the data into DGGS cells.The implementation utilizes open-source Python tools from the Pangeo ecosystem, including xarray-xdggrid, to enable seamless integration and efficient analysis of large geospatial datasets. Three DGGS configurations – ISEA7H, HEALPIX, and ISEA3H are compared alongside traditional latitude-longitude grid for computational efficiency (processing time and memory usage) and their ability to preserve spatial patterns. Spatial analysis methods include density estimation, nearest neighbor evaluation, and clustering for point data, as well as zonal statistics, spatial autocorrelation, and resampling for raster data.

To further illustrate the application of DGGS-based methods, the study includes a case study on estuary characterization. This characterization relies on spatial analysis methods, integrating physical oceanographic parameters from Delft3D-FM, biogeochemical and optical data products, and in-situ point measurements from the Copernicus Marine Environment Monitoring Service (CMEMS). Representing these diverse datasets within the DGGS framework highlights its ability to manage varying data types and scales, offering insights into estuarine environments and demonstrating its scalability for addressing complex marine spatial challenges.

Results indicate that DGGS frameworks deliver comparable computational performance while offering consistent spatial representation. Configuration-specific trade-offs influence their effectiveness, emphasizing the importance of aligning DGGS configurations with specific analytical tasks and applications. Findings suggest that DGGS-based methods offer a promising alternative to traditional analysis techniques, providing greater flexibility in adapting to datasets, scale, and resolution. This contributes to more efficient mapping, sustainable marine environmental management, and advancing geospatial applications through open-source tools from the Pangeo ecosystem.

How to cite: Martinez, K., Kmoch, A., Mészáros, L., Nelson, A., and Uuemaa, E.: Navigating New Grids: Evaluating DGGS Configurations for Marine Spatial Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14306, https://doi.org/10.5194/egusphere-egu25-14306, 2025.

EGU25-14400 | ECS | Orals | ESSI2.15

A community oriented approach to enabling open science with Earth science data at scale 

Max Jones, Aimee Barciauskas, Jonas Sølvsteen, Brian Freitag, Yuvi Panda, Kyle Barron, Julia Signell, Alex Mandel, Chuck Daniels, Nathan Zimmerman, Sean Harkins, Henry Rodman, Zac Deziel, Slesa Adhikari, Anthony Boyd, Alexandra Kirk, David Bitner, and Vincent Sarago

To enable wider participation in open science with geospatial data at scale, we need to reduce the effort and custom approaches required for setting up scalable scientific data analysis environments and computing workflows. We have made great strides in this pursuit by evolving and promoting community-developed open source frameworks, tools, and libraries for cloud-native data access and analysis, making them the default for scientists on the public cloud and local systems.

Many of our achievements have been supported by the NASA Visualization, Exploration, and Data Analysis (VEDA) project which seeks to proliferate cloud-native approaches for open science on Earth science data from NASA’s rich archives and many other providers. Our presentation highlights how we have engaged with communities like Pangeo, OpenScapes, Earth Science Information Partners, and the Cloud Native Geospatial Forum to build joint initiatives, target development, and ensure uptake of new solutions. We present key results from working groups, community showcases, and hackdays and hackweeks organized by VEDA team members, as well as specific contributions to the open source ecosystem, including the eoAPI platform for quickly and easily deploying an open-source Earth Observation stack, JupyterHub fancy profiles (with BinderHub) for seamless environment building, and Lonboard for fast, interactive vector visualization.

How to cite: Jones, M., Barciauskas, A., Sølvsteen, J., Freitag, B., Panda, Y., Barron, K., Signell, J., Mandel, A., Daniels, C., Zimmerman, N., Harkins, S., Rodman, H., Deziel, Z., Adhikari, S., Boyd, A., Kirk, A., Bitner, D., and Sarago, V.: A community oriented approach to enabling open science with Earth science data at scale, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14400, https://doi.org/10.5194/egusphere-egu25-14400, 2025.

EGU25-14610 | Orals | ESSI2.15

Weather Data Streaming with Kerchunk: Strengthening Early Warning Systems  

Nishadh Kalladath, Masilin Gudoshava, Shruti Nath, Jason Kinyua, Fenwick Cooper, Hannah Kimani, David Koros, Christine Maswi, Zacharia Mwai, Asaminew Teshome, Samrawit Abebe, Isaac Obai, Jesse Mason, Ahmed Amdihun, and Tim Palmer

The Ensemble Prediction System (EPS) provided by global weather forecast centres generates vast amounts of data that is crucial for early warnings of extreme weather and climate. However, regional and national meteorological services often face challenges in processing this data efficiently, particularly during regional downscaling and post-processing. Conventional methods of downloading and storing GRIB-format data have become increasingly inefficient and unsustainable. The Strengthening Early Warning Systems for Anticipatory Actions (SEWAA) project aims to address these challenges by exploring the use of cloud native operations and GenAI-cGAN driven post-processing systems.   

Kerchunk provides a groundbreaking solution for real-time weather data streaming, catering to the transition towards open and free to use cloud-based object storage from global weather forecasting centres. Kerchunk, in conjunction with GRIB index files, enables efficient, real-time access to weather data, fostering more sustainable workflows in weather and climate services, thus strengthening early warning systems.  

This study developed a workflow for streaming forecast data using Kerchunk with two primary objectives:  

1. Using GRIB index files to reduce redundant readings and generate Kerchunk reference files.  

2. Through streaming-like access, convert the reference files into virtual Zarr datasets and utilise Dask compute for scalable data handling   

The methodology utilised recent improvements in the Kerchunk library that integrate GRIB scanning with its index files. This allowed the system to sample subsets of the GRIB corpus instead of processing entire Forecast Model Run Collections (FMRC), significantly optimising performance.  

The workflow was implemented using cloud-based compute operations via Coiled python library and its service on the Google Cloud Platform. Dask cluster, managed through Coiled, enabled the creation of Zarr virtual datasets for analysis and visualisation. This streaming approach efficiently loads weather data into memory on demand, avoiding unnecessary data downloads and duplication.   

We validated the solution with NOAA GFS/GEFS datasets stored in AWS S3 bucket as open datasets. The optimised workflow demonstrated remarkable efficiency, requiring only <5% of the original GRIB data to be read, with the rest replaced by index files as input for reference file creation. This is followed by the step of Kerchunk reference files to virtual Zarr conversion by Dask clusters to process on a regional scale, such as East Africa’s in minutes supporting near real-time applications across spatial and temporal scales.  

This approach significantly enhances post processing workflows for EPS weather forecast, bolstering early warning systems and anticipatory action. Future work will focus on using the method to scaling training datasets and improving the cost efficiency of cGAN training to advance operational early warning systems. This innovative solution directly addresses the challenges faced by meteorological services in processing massive weather datasets, providing a scalable, cost-effective, development foundation for applying GenAI based post-processing and improving early warning systems. 

How to cite: Kalladath, N., Gudoshava, M., Nath, S., Kinyua, J., Cooper, F., Kimani, H., Koros, D., Maswi, C., Mwai, Z., Teshome, A., Abebe, S., Obai, I., Mason, J., Amdihun, A., and Palmer, T.: Weather Data Streaming with Kerchunk: Strengthening Early Warning Systems , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14610, https://doi.org/10.5194/egusphere-egu25-14610, 2025.

EGU25-15406 | Orals | ESSI2.15

Platform Engineering for Earth Observation: A Unified Approach to HPC and Cloud Systems 

Armagan Karatosun and Vasileios Baousis

The growing volume of Earth Observation (EO) and Earth modeling data makes it increasingly impractical to download and analyze it locally. Furthermore, as cloud-native data formats and AI/ML-driven models gain popularity, the community requires powerful computing and storage solutions to efficiently process and analyze EO data. High-performance computing (HPC) and cloud infrastructures can help accomplish this, but both bring significant challenges in maintaining those resources, putting additional workloads on the scientists and developers.

In this paper, we will present our solution, which uses cloud-native technologies and a “Control Plane” approach to seamlessly interact with HPC scheduling endpoints like SLURM and PBS, as well as cloud infrastructure resources, allowing HPC jobs to be submitted and monitored directly from a Kubernetes-based infrastructure. In contrast to traditional IT architecture, Platform Engineering is concerned with lowering operational complexity by introducing control planes to provide self-service capabilities. By abstracting away the complexities of the underlying infrastructure, this method gives teams a customized, scalable, and dependable environment to suit their unique requirements. We will thoroughly analyze existing technologies, including their methodologies, strengths, limits, and potential as universal solutions. Furthermore, we will assess their adaptation to various cloud and HPC infrastructures, providing insights into their suitability for larger applications. 

We will conclude our discussion with practical examples showing how the technical benefits of these two computing paradigms, combined with the Platform Engineering approach, may be effectively used in real-world EO data processing scenarios.

How to cite: Karatosun, A. and Baousis, V.: Platform Engineering for Earth Observation: A Unified Approach to HPC and Cloud Systems, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15406, https://doi.org/10.5194/egusphere-egu25-15406, 2025.

High-resolution regional climate model datasets, such as those produced within the Coordinated Regional Downscaling Experiment (CORDEX) framework, are critical for understanding climate change impacts at local and regional scales. These datasets, with their high spatial and temporal resolution, provide detailed insights into region-specific climate phenomena, including urban heat islands, mountainous climates, and extreme weather events. However, their accessibility and usability are often constrained by technical challenges such as fragmented data storage, inconsistent formats, and limited interoperability.

To address these barriers, we are developing the Climate Service Database (CSD) - a centralized data warehouse designed to streamline the temporal and spatial aggregation of CORDEX datasets for climate service applications. The CSD ingests raw CORDEX datasets and applies automated extraction, transformation, and loading (ETL) workflows to produce analysis-ready datasets tailored to user needs. By leveraging cloud-based infrastructure and adhering to Climate and Forecast (CF) conventions, the CSD ensures consistent, interoperable data products that are optimized for scalable access and analysis.

A core functionality of the CSD is its ability to aggregate datasets at multiple spatial and temporal scales, ranging from daily extremes to decadal averages, and across diverse spatial resolutions (e.g., countries, administrative regions, or watersheds). This capability enables the generation of climate indicators (e.g., hot summer days, heavy precipitation events) that are directly relevant for local decision-making and impact assessments. By providing data in cloud-optimized, analysis-ready formats (ARCO) and offering Software as a Service (SaaS), the CSD significantly lowers the technical barriers for researchers, businesses, and policymakers seeking to access user-tailored climate service datasets.

By centralizing and optimizing the processing of regional climate model datasets, the CSD fosters collaboration across research institutions, public agencies, and climate-tech startups. It enables users to efficiently access consistent and up-to-date data while eliminating the redundancies of localized data storage and processing. This approach also opens new opportunities for applying AI-driven analytics and machine learning models to CORDEX data, paving the way for innovative climate services and applications.

Through its focus on regional climate model datasets, the CSD exemplifies how modern data infrastructures can enhance the usability of high-resolution climate data, empowering stakeholders to develop robust, data-driven adaptation and mitigation strategies in response to the challenges of climate change.

How to cite: Buntemeyer, L.: Advancing Regional Climate Data Accessibility through a Cloud-native Climate Service Database, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16230, https://doi.org/10.5194/egusphere-egu25-16230, 2025.

EGU25-17137 | Orals | ESSI2.15

The Sentinels EOPF Toolkit: Community Notebooks and Plug-ins for using Copernicus Sentinel Data in Zarr format 

Dr. Julia Wagemann, Sabrina Szeto, Emmanuel Mathot, and James Banting

Zarr is a key component of the Pangeo ecosystem and instrumental for effectively accessing and processing multi-dimensional Earth data in cloud-based systems. More and more leading satellite data providers are exploring the transition of their data archives to a cloud environment. 

As part of the ESA Copernicus Earth Observation Processor Framework (EOPF), ESA is in the process of providing access to “live” sample data from the Copernicus Sentinel missions -1, -2 and -3 in the new Zarr data format. This set of reprocessed data allows users to try out accessing and processing data in the new format and experiencing the benefits thereof with their own workflows.

To help Sentinel data users to experience and adopt the new data format, a set of resources called the Sentinels EOPF Toolkit is being developed. Development Seed, SparkGeo and thriveGEO, together with a group of champion users (early-adopters), are creating a set of Jupyter Notebooks, plug-ins and libraries that showcase the use of Sentinel data in Zarr for applications across multiple domains for different user communities, including users of Python, Julia, R and QGIS.

This presentation will give a demo of the first set of notebooks and plugins of the Sentinels EOPF toolkit that were developed and that facilitate the adoption of the Zarr data format for Copernicus Sentinel data users. Additionally, we will give an overview of toolkit developments and community activities that are planned throughout the project period.

How to cite: Wagemann, Dr. J., Szeto, S., Mathot, E., and Banting, J.: The Sentinels EOPF Toolkit: Community Notebooks and Plug-ins for using Copernicus Sentinel Data in Zarr format, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17137, https://doi.org/10.5194/egusphere-egu25-17137, 2025.

EGU25-18285 | ECS | Posters on site | ESSI2.15

Regridding Satellite and Model Data to DGGS (HEALPix) Using the Pangeo Ecosystem 

Justus Magin, Jean-Marc Delouis, Lionel Zawadski, Julien Petiton, Max Jones, and Tina Odaka

Regridding data from diverse sources, such as satellite observations and numerical models, is a critical task in Earth system sciences. Proper interpolation methods are essential to ensure data fidelity when combining or comparing datasets on different grids. This becomes especially relevant in the context of emerging grid systems like Discrete Global Grid Systems (DGGS), specifically HEALPix.

DGGS are spatial reference systems designed to partition the Earth’s surface into a hierarchy of equal-area cells. Unlike traditional latitude-longitude grids, DGGS uses tessellations, such as hexagons, to represent the Earth’s curved surface with minimal distortion. This grid system is particularly suited for handling global-scale geospatial data by providing uniform coverage and resolution, enabling efficient storage, processing, and analysis.

HEALPix (Hierarchical Equal Area isoLatitude Pixelation) is a specific implementation of DGGS widely used in astronomy and Earth sciences. HEALPix divides the sphere into equal-area cells following an iso-latitude structure, making it computationally efficient for operations such as spherical harmonics and multi-resolution analysis. Originally developed for astrophysical applications, it has become increasingly popular in the Earth sciences for representing satellite data, model outputs, and other geospatial datasets in a way that preserves area integrity and facilitates seamless multi-resolution data integration.

By leveraging these grid systems, particularly HEALPix, we can achieve a more accurate and efficient representation of geospatial data.

The Pangeo ecosystem includes an array of powerful regridding tools, each tailored to specific grid types and applications. However, navigating this ecosystem to identify the most suitable tool and workflow can be challenging.

In this presentation, we will show an overview of regridding solutions within Pangeo, highlighting their capabilities and limitations, as well as  their application. We will also demonstrate a practical regridding workflow using model outputs or simulated satellite data such as the Odysea dataset (Aviso+ Altimetry. (n.d.). Simulated Level-2 Odysea Dataset. Retrieved from https://www.aviso.altimetry.fr/en/data/products/value-added-products/simulated-level-2-odysea-dataset.html on January 14, 2025), to the HEALPix grid. This workflow will make use of recent advances in technology to make it reproducible to make it efficient and reproducible, such as virtualizarr for fast metadata access and dask for scalable operations, with the output saved as chunked zarr files for seamless integration with downstream analysis.

How to cite: Magin, J., Delouis, J.-M., Zawadski, L., Petiton, J., Jones, M., and Odaka, T.: Regridding Satellite and Model Data to DGGS (HEALPix) Using the Pangeo Ecosystem, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18285, https://doi.org/10.5194/egusphere-egu25-18285, 2025.

EGU25-18336 | Posters on site | ESSI2.15

Advancing Earth System Science through collaboration: An overview of ECMWF Special Projects 

Milana Vuckovic and Becky Hemingway

ECMWF has been providing resources on its operational high-performance computing (HPC) and cloud facilities (European Weather Cloud) to researchers and institutions through the Special Projects framework. This framework has been established almost 50 years ago as part of the creation of ECMWF. ECMWF's HPC facility is specifically designed to support both operational time-critical production of global weather forecasts and typical research workflows, therefore through Special Projects, researchers can get access not only to a top high-performance computing and cloud facility and one of the largest meteorological archives in the world, but also full user support.
Special Projects are defined as experiments or investigations of a scientific or technical nature, undertaken by one or more ECMWF Member States, likely to be of interest to general scientific community. The main aim of this initiative is to facilitate collaboration, enabling the development of innovative methodologies and tools for numerical weather prediction, climate and environmental modelling, and other disciplines within Earth System Sciences. All Special Project applications undergo a review process by ECMWF and its Scientific Advisory Committee (SAC), as well as ECMWF Member State's meteorological services and are ranked primarily by their scientific quality.
This poster will describe the Special Projects framework and showcase three recent Special Projects that illustrate collaborative nature of the initiative using ECMWF's HPC and European Weather Cloud facilities, including validating ICON model on ECMWF systems, the development of next-generation European Earth System Model (EC-EARTH4) and mapping the yet uncharted continuum of cyclone dynamics for the Euro Atlantic domain.
Through these examples, the poster will demonstrate how ECMWF Special Projects foster international collaboration, resource sharing, and innovation, enabling advancement in Earth System Science. 

How to cite: Vuckovic, M. and Hemingway, B.: Advancing Earth System Science through collaboration: An overview of ECMWF Special Projects, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18336, https://doi.org/10.5194/egusphere-egu25-18336, 2025.

EGU25-20590 | Posters on site | ESSI2.15

Dhemeter: Data Hub for Environmental and METEorological Resources 

Cédric Pénard, Nathan Amsellem, Boris Gratadoux, Bastien Barthet, Jean Christophe Pere, Johannes Staufer, Laure Chaumat, and Alexia Mondot

Dhemeter is a weather and environmental data aggregator. It is developed using a microservices architecture to handle a wide variety of data from various providers, such as NOAA, ECMWF, Eumetsat, Météo-France, DWD, and Copernicus. The implementation of aggregation, concatenation, and consistency functionalities has been successfully executed for meteorological data. This versatile tool accommodates numerical model data, in-situ observations, remote sensing data, and reanalyses, allowing for online data retrieval from multiple sources.

Key features of the aggregator include:

  • Concatenation of Multiple Data Sources: Users can combine data according to selected categories such as Observations, Forecasts, and Reanalyses.
  • Standardization of Physical Data: This involves spatial and temporal interpolation as well as geographical selections to ensure uniformity.
  • Storage of Resulting Data Structures: The data is stored in a pivot format that facilitates access and distribution of scientific data, specifically in the NetCDF format.

The microservices architecture of the aggregator allows for the extensibility of the offered data catalog, and an API is available for users to make direct queries to chosen data sources.

In the short to medium term, the goal is to enhance the tool further, evolving it into a comprehensive data distribution and aggregation system that centralizes and simplifies access to various types of data, including meteorological, oceanographic, and air quality data.

Dhemeter focuses on ease of use, extensibility, scalability, and customization, offering users capabilities for data fusion and harmonization.

How to cite: Pénard, C., Amsellem, N., Gratadoux, B., Barthet, B., Pere, J. C., Staufer, J., Chaumat, L., and Mondot, A.: Dhemeter: Data Hub for Environmental and METEorological Resources, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20590, https://doi.org/10.5194/egusphere-egu25-20590, 2025.

EGU25-20676 | Orals | ESSI2.15

Integrated geospatial Python libraries for efficient analysis of modern elevation measurements 

Scott Henderson, David Shean, Jack Hayes, and Shashank Bhushan

NASA established the Surface Topography and Vegetation (STV) Incubation program to develop and mature the next-generation measurement approaches to precisely map Earth’s changing surface and overlying vegetation structure, and prepare for a dedicated satellite mission within the next decade. Over the past two decades, large archives of 3D surface elevation measurements by airborne and satellite instruments including LiDAR, altimeters, Synthetic Aperture Radar, and stereo optical imagery have been systematically collected, though not always in a coordinated way. Yet, many of these datasets are fortuitously acquired over the same location within a short temporal window (e.g., <1-14 days) and many are now publicly available and hosted on the cloud. In theory, this is a great opportunity to synthesize myriad elevation measurements for STV researchers, but in practice merging these datasets accurately for scientific analysis requires dealing with numerous data formats, complex 4D coordinate reference systems, and securing access to significant computational resources.

We are developing an open-source Python library to identify, curate, and efficiently process coincident elevation measurements spanning the last several decades. This work would not be possible without well-integrated geospatial libraries (e.g. Geopandas, Xarray, Dask), as well as emerging cloud-native data and metadata formats such as Cloud-Optimized Geotiff and STAC-GeoParquet. We will describe our work to-date and reflect on the process of collaborative development across libraries, on our increasing reliance on Cloud resources, and current and future research directions.

How to cite: Henderson, S., Shean, D., Hayes, J., and Bhushan, S.: Integrated geospatial Python libraries for efficient analysis of modern elevation measurements, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20676, https://doi.org/10.5194/egusphere-egu25-20676, 2025.

EGU25-21202 | Orals | ESSI2.15

From SAFE to Zarr: The EOPF Sample Service Initiative 

Christian Briese, Christoph Reimer, Christian Briese, Christoph Reck, Dimitrios Papadakis, Michele Claus, Gunnar Brandt, Anne Fouilloux, and Tina Odaka

Over the past decade, the operational Copernicus Sentinels Data Processors have generated vast amounts of Earth observation data, supporting various scientific and commercial applications. However, the current format used by ESA to provide Copernicus data, known as SAFE (Standard Archive Format for Europe), has become outdated. To address this, ESA has initiated the transition to a new Zarr-based data format. The Earth Observation Processing Framework (EOPF) Sample Service is ESA’s official initiative to support this transition by providing early access to the new format for users. This shift is essential for creating a cloud-native and interoperable solution that enhances data accessibility and integration with modern processing frameworks. The primary goal is to standardize data formats across Sentinel missions, enable scalable processing on cloud platforms, and ensure compatibility with contemporary data science tools. This initiative is crucial for minimizing disruption and ensuring continuity for users, applications, and services built around existing data formats.

The EOPF Sample Service comprises several key components. The EOPF Core Platform re-formats ingested SAFE data products into the new cloud-optimized EOPF Zarr data products and provides data access via STAC API and S3 API. To ensure timely conversion, the platform utilizes Argo Events and the Copernicus Data Space Ecosystem's subscription service. This platform is maintained by experts from EODC and DLR. The EOPF User Platform offers additional user services, including JupyterHub (BinderHub), Dask, and a STAC Browser, which are essential for supporting user adoption by lowering the entrance barrier to cloud applications and data discovery capabilities. The service is designed to make use of advanced technologies such as Kubernetes for container orchestration and Dask for parallel computing. User and identity management is achieved in cooperation with the Copernicus Data Space Ecosystem.

User adoption is further facilitated through Jupyter Notebooks designed by experts within the consortium, including members from the Pangeo community. These notebooks showcase the use of the new format within the community and are continuously improved by incorporating user feedback. In addition, enhancements are made to widely-used software tools like GDAL to support the new format, with practical demonstrations available through Jupyter Notebooks. The consortium selected by ESA to carry out this implementation includes experts from Brockmann Consult, DLR, Ifremer, EURAC, Evenflow, Simula, and EODC, each contributing their specialized knowledge in Earth observation, data management, and user engagement.

This contribution aims to present the EOPF Sample Service initiative and the current status of its implementation. The first Jupyter Notebooks demonstrating the new format will also be showcased, providing users with an intuitive and user-friendly interface for accessing and processing sample data in the new EOPF format.

How to cite: Briese, C., Reimer, C., Briese, C., Reck, C., Papadakis, D., Claus, M., Brandt, G., Fouilloux, A., and Odaka, T.: From SAFE to Zarr: The EOPF Sample Service Initiative, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21202, https://doi.org/10.5194/egusphere-egu25-21202, 2025.

EGU25-21279 | Orals | ESSI2.15

Advancing Cloud-Native Data Analysis and Publishing with Pangeo Tools in EarthCODE 

Deyan Samardzhiev, Anne Fouilloux, Tina Odaka, and Benjamin Ragan-Kelley

EarthCODE (Earth Science Collaborative Open Development Environment) is a platform that leverages cloud-native tools to empower Earth system researchers in accessing, analyzing, and sharing data across distributed infrastructures, such as the Copernicus Data Space Ecosystem and Deep Earth System Data Laboratory (DeepESDL). By integrating Pangeo ecosystem tools—including Xarray, Dask, and Jupyter—EarthCODE supports scalable, FAIR-aligned workflows tailored to the challenges of Earth system science.

EarthCODE streamlines cloud-based data analysis and publishing by enabling collaborative research through interoperable workflows for analyzing complex datasets, including satellite observations, climate models, and in-situ measurements. Researchers can publish their analyses and workflows as reusable, executable resources in EarthCODE’s science catalog, fostering alignment with open science principles.

Through its integration of Pangeo tools, EarthCODE offers an intuitive environment for reproducibility, scalability, and collaboration, bridging the gap between data analysis and actionable insights. This presentation will demonstrate EarthCODE’s capabilities, including live, executable Jupyter notebooks that highlight its potential for sharing workflows and engaging diverse user groups. EarthCODE exemplifies the transformative power of cloud-native research, promoting open science and advancing the accessibility of Earth system data.

How to cite: Samardzhiev, D., Fouilloux, A., Odaka, T., and Ragan-Kelley, B.: Advancing Cloud-Native Data Analysis and Publishing with Pangeo Tools in EarthCODE, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21279, https://doi.org/10.5194/egusphere-egu25-21279, 2025.

EGU25-21603 | ECS | Orals | ESSI2.15 | Highlight

The Pangeo Ecosystem Supporting Climate Change Adaptation: The FAIR2Adapt RiOMar Case Study 

Even Moa Myklebust, Ola Formo Kihle, and Justus Magin

The RiOMar (River dominated Ocean Margins) case study, part of the FAIR2Adapt (FAIR to Adapt to Climate Change) project (EU funded project grant agreement No 101188256), focuses on supporting science-based climate change adaptation strategies for coastal water quality and marine ecosystem management. The case study uses large environmental datasets, such as sea temperature, salinity, and other marine parameters, to assess and model the impacts of climate change on coastal ecosystems. As part of the FAIR2Adapt project, which aims to enhance the shareability, accessibility, interoperability, and reusability of environmental data through the FAIR (Findable, Accessible, Interoperable, and Reusable) principles, the RiOMar case study emphasizes the use of cutting-edge data processing and analysis methods to support adaptive strategies for climate resilience.

In this presentation, we present our approach to reading the RiOMar large environmental datasets in netCDF format, creating VirtualZarr archives for efficient data handling, transforming them into a Discrete Global Grid System (DGGS) using the Healpix grid.Leveraging the Pangeo ecosystem, we use tools such as Kerchunk to create simpler access to multiple data sources, parallelize dataset processing using Dask or Cube, enabling scalable analysis of these complex, multi-dimensional data. We will show a comparison of performance between traditional cube-based approaches and Dask, highlighting the advantages of parallelized processing. Furthermore, we will showcase how to interactively visualize these datasets using tools like XDGGs and Lonboard, facilitating seamless exploration and analysis of the underlying environmental patterns. This work underscores the potential of open-source tools, scalable computing techniques, and the Pangeo ecosystem to enhance the accessibility and usability of large geospatial datasets in climate adaptation research.

How to cite: Moa Myklebust, E., Formo Kihle, O., and Magin, J.: The Pangeo Ecosystem Supporting Climate Change Adaptation: The FAIR2Adapt RiOMar Case Study, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21603, https://doi.org/10.5194/egusphere-egu25-21603, 2025.

Let us focus on a specific question that may has an ability to build an efficient method toward extracting significantly major ingredients of pre-active events going ahead of significant seismic activities. What is the common point at the state spaces of significant earthquakes of Türkiye in 1999 and 2023? The answer comes from some live but non-instrumented observations, those are devised privately. Those observations are related to both waveguide and cavity effects of natural and/or manmade significant structures replaced in both underground and/or atmosphere. The effects are studied on the electromagnetic wave propagation at significant pre-seismic activities of both circularly cylindrical wave guide and cavity structures meshed in underground and/or atmosphere by considering the extended wave equations in irregularly deviating environs1. Those structures have excessive dimensions as in subway tunnels2 and/or layered guiding pathways in atmosphere3.

The answer comes from two big tunnels excavated before abovesaid two earthquakes of Türkiye. First is Mount Bolu Tunnel, that is almost finished in 2007 and begun in 1993 and second is New Mount Zigana Tunnel, that is finished in 2023 and begun in 2016. Why? First of all, both tunnels are into mounts area of Northern Anatolia. The reason is related to the changing character of seismic activities after 5.9 R (included) magnitude that converts the seismic activities to electromagnetic activities majorantly4.

There is one more tunnel process that still continues for constructions: Between Bahce (37° 12′ 0″ N, 36° 35′ 0″ E) and Nurdagi (37° 10′ 44″ N, 36° 44′ 23″ E) districts of Gaziantep Province, Türkiye. This tunnel construction may have a potential on future seismic activities as two tunnel constructions said in previous paragraph.

The cavities and tunnels behave as layered guiding pathways for propagating waves either homogeneous and/or inhomogeneous fillings; therefore, the activities of waveforms may propagate along long distances under the Earth; i.e., between NAF and SAF by suitable transmissions, propagations, and guiding of waves. The majorant contributions come through Casimir and Casimir-like activities from the boundary interfaces between different materials with specific conditions under stochastic processes. The propagating waves create similar effects among transmitters and receivers through atmosphere layers. Author calls transmission effect by the cavity tunneling and layered guiding pathways these effects.

Those circumstances are studied in above paragraphs by considering the state space formulation of equivalent electrical circuits models through the possible mechanical circuits into the Earth.

The equivalent circuit model governs the significant Seismic Activities, sSAs, by the interactions among source and sink structures available in the distributed networks of equivalent circuits. New constructions have the ability to trigger and produce sSAs close to both specific domains of sSAs and their neighbor domains even if they never generated sSAs in past, of tunnel projects in paragraph 3 and similar ones. Temporal intervals may not coincide with the time spans of excavations of sSAs processes and their triggering effects may either decrease, mostly and/or increase, asymptotically as depending to coupling activities in environ.

 

1https://doi.org/10.1109/APS.1996.549734

2https://doi.org/10.5194/egusphere-egu2020-22589.

3 https://doi.org/10.1109/RAST.2003.1303999.

4 https://doi.org/10.5194/egusphere-egu2020-21121.

How to cite: Sengor, T.: The Cavity Tunneling and Layered Guiding Pathways in Significant Seismic Activities: Pre-fingerprints in Significant Earthquakes of Türkiye in 1999 and 2023, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-988, https://doi.org/10.5194/egusphere-egu25-988, 2025.

There has long been research on the phenomenon of abnormal microwave radiation emitted from the Earth's surface before a major earthquake. However, the enhanced microwave radiation received by satellite sensors is affected by a combination of factors such as surface vegetation, soil moisture, land surface temperature, and atmospheric environment. So far, it has been difficult to remove non-seismic interference through quantitative physical modeling, leaving only the earthquake-related additional components for earthquake precursor analysis and short-term earthquake prediction. To tackle with this, we developed a knowledge-guided deep learning model that leverages a large amount of remote sensing observation data for training, incorporating prior knowledge of earthquake anomaly analysis. In the modelling process, a large amount of multi-source data, such as surface microwave brightness temperature (MBT), land surface temperature (LST), surface vegetation index, soil moisture index, atmospheric water vapor content, cloud cover, land cover type, digital elevation model (DEM), and geological type, were collected, and a regression model between multiple factors and surface MBT were firstly established through deep learning methods. In the same way, another regression model was developed between non-temperature parameters and LST by using historical records. During the seismic window of one month before and after the target earthquake, the LST was obtained by using non-temperature data through the second regression model, and then was substituted it into the first regression model to get the MBT value that does not include the additional effects of earthquakes. Eventually, we can obtain the additional MBT value due to seismic activity by calculating the difference with the actual observation, which represents the earthquake-related MBT anomaly. Since the deep learning-based modeling is based on long time series data and the output results of the model already include the contribution of multiple factors on the surface to the MBT, the differential results are mainly affected by the additional impacts of the earthquake, so they can be considered 'pollution-free'. In other words, there is no need to use additional auxiliary data to discriminate and separate the non-seismic disturbances. For a specific target area, such as the Tibetan Plateau, after establishing a model based on historical data using the aforementioned method, we can obtain real-time earthquake MBT variations as the input data is continuously updated. This can be used to analyze and identify potential earthquake precursors, and consequently, for short-term earthquake prediction.

How to cite: Qi, Y., Mao, W., Wu, L., and Huang, B.: A Knowledge-Guided Deep Learning Model for Extracting Pollution-free Seismic Microwave Brightness Temperature Anomalies, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1408, https://doi.org/10.5194/egusphere-egu25-1408, 2025.

EGU25-2283 | Orals | NH4.4

On the Ionosphere-Atmosphere-Lithosphere coupling during theNovember 9, 2022 Italian Earthquake 

Mirko Piersanti, Giulia D'Angelo, Dario Recchiuti, Fabio Lepreti, Paola Cusano, Enza De Lauro, Vincenzo Carbone, Pietro Ubertini, and Mariarosaria Falanga

In the last decades, the scientific community has been focused on searching earthquake signatures in the Earth's atmosphere, ionosphere, and magnetosphere. This work investigates an offshore Mw 5.5 earthquake that struck off the Marche region's coast (Italy) on November 9, 2022, with a focus on the potential coupling between the Earth's lithosphere, atmosphere, and magnetosphere triggered by the seismic event. Analysis of atmospheric temperature data from ERA5 reveals a significant increase in potential energy (Ep) at the earthquake's epicenter, consistent with the generation of Atmospheric Gravity Waves (AGWs). This finding is further corroborated by the MILC analytical model, which accurately simulates the observed Ep trends (within 5%), supporting the theory of Lithosphere-Atmosphere-Ionosphere-Magnetosphere Coupling. The study also examines the vertical Total Electron Content (vTEC) and finds notable fluctuations at the epicenter, exhibiting periodicities (7-12 minutes) characteristic of AGWs and traveling ionospheric disturbances. The correlation between ERA5 observations and MILC model predictions, particularly in temperature deviations and Ep distributions, strengthens the hypothesis that earthquake-generated AGWs impacted atmospheric conditions at high altitudes, leading to observable ionospheric perturbations. This research contributes to a deeper understanding of Lithosphere-Atmosphere-Ionosphere-Magnetosphere Coupling mechanisms and the potential for developing reliable earthquake prediction tools.

How to cite: Piersanti, M., D'Angelo, G., Recchiuti, D., Lepreti, F., Cusano, P., De Lauro, E., Carbone, V., Ubertini, P., and Falanga, M.: On the Ionosphere-Atmosphere-Lithosphere coupling during theNovember 9, 2022 Italian Earthquake, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2283, https://doi.org/10.5194/egusphere-egu25-2283, 2025.

We discuss the potential impact of the Geospace environment on the significant earthquake preparation processes. In this work, we investigate the response of major seismic activity to geomagnetic storms with a joint analysis of solar wind, geomagnetic field, and earthquake catalog. As a test case, we processed the seven strongest earthquakes in Italy for the period  1980 - 2016:  Amatrice M6.2 of Aug 24, 2016; Visso M6.1 of 26 Oct 2016; Norcia M6.6 of 30 Oct 2016; Emilia-Romagnia M6 of May 20, 2012;  L’Aquila M6.3 of Apr 6, 2009;  Foligno M6 of Sep 26,1997  and  Irpina of M6.9 of 23 Nov 1980. All of the seismic events were preceded by geomagnetic storms, which satisfied a given criterion: at the time of geomagnetic storm onset, the high-latitude part of the longitudinal region, where in the future an earthquake occur, was located under the polar cusp, where the solar wind plasma would directly access the Earth’s environment [Ouzounov and Khachikyan, 2024]. The number of preceded storms varied for different earthquakes from two to five. This results in different time delays between the day of the magnetic storm onset and the day of earthquake occurrence; it ranges between 9-80 days. Because of the existing delay between a shocked solar wind arrival and earthquake occurrence up to some months, this may suggest that solar wind energy does not trigger earthquakes immediately (as it is believed at present); instead, it may accelerate the processes of lithosphere dynamics, such as fluid and gas upwelling, which are active participants in tectonic earthquakes. For comparison, we present the results of the same analysis applied to other territories of the Mediterranean region: the Anatolian Plate (Turkey) and Crete Island (Greece), which look strikingly similar.

 

How to cite: Ouzounov, D. and Khachikyan, G.: The impact of the geospace environment on earthquake preparation processes. Case studies for M>6 in Italy for 1980-2016, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3686, https://doi.org/10.5194/egusphere-egu25-3686, 2025.

EGU25-5269 | Orals | NH4.4

Similarities and differences of the preparation of three (M≈6) earthquake doublets around the Arabian Plate 

Essam Ghamry, Dedalo Marchetti, and Mohamed Metwaly

In this study, we compared the results of multiparametric and multilayers investigations of three doublet earthquakes that occurred around the Arabian plate (M6.2 + M6.0 on 18 August 2014 close to Dehloran, Iran; M6.0 + M6.0 occurred on 15 July 2018 offshore Kilmia, Yemen and M6.0 + M6.0 occurred on 1 July 2022 close to Bandar-e Lengeh). We applied identical methods to the same dataset for all three cases. In particular, we investigated lithospheric, atmospheric, and ionospheric data six months before the three events. The lithosphere was investigated by calculating the cumulative Benioff strain with the USGS earthquake catalogue. Several atmospheric parameters (aerosol, SO2, CO, surface air temperature, surface latent heat flux humidity, and dimethyl sulphide) have been monitored using the homogeneous data from the MERRA-2 climatological archive. We used the three-satellite Swarm constellation for magnetic data, analysing the residuals after removing a geomagnetic model. All the cases present some patterns of anomalies, and when comparing them, we noticed some similarities but also differences. We pointed out that the released energy by the three events is very similar and occurred around the same plate. Still, they involved two different tectonic contexts (compressional on the Iranian side and extensional and transcurrent on the African Plate border). For the above reasons, their comparison is very interesting. Some similarities seem to be explainable in the tectonic context, and some are caused by the ocean's influence at the epicentre location. However, we also identified some differences that still require further investigation and comparison with other case studies.

Finally, this work can be considered a preliminary test of an extensive investigation and systematical search of LAIC patterns before the earthquake occurrences and the study of the possible influence of focal mechanism, location, geological factors, and other constraints.

 

References :

Ghamry Essam; Marchetti Dedalo; Metwaly Mohamed. Geophysical Coupling Before Three Earthquake Doublets Around the Arabian Plate. Atmosphere 2024, 15, 1318. https://doi.org/10.3390/atmos15111318

 

How to cite: Ghamry, E., Marchetti, D., and Metwaly, M.: Similarities and differences of the preparation of three (M≈6) earthquake doublets around the Arabian Plate, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5269, https://doi.org/10.5194/egusphere-egu25-5269, 2025.

EGU25-5321 | Posters on site | NH4.4

Novel experimental design for the study of seismic processes based on the stick-slip mechanism. 

Alejandro Ramírez-Rojas, Luciano Telesca, and Elsa Leticia Flores-Márquez

Seismicity is the result of the interaction between tectonic plates in relative motion where the underlying mechanism of earthquake generation in seismic subduction areas is stick-slip. In reality, seismicity is a complex phenomenon as it involves processes that take place from within the Earth. A thorough understanding of seismicity requires theoretical and experimental approaches. The dynamics in subduction zones occur when two tectonic plates, one on top of the other, are in relative motion where the plate below is in motion due to convective processes within the Earth. Due to the roughness of both surfaces, the underlying mechanism that gives rise to seismicity is stick-slip. In this work, an experimental stick-slip model is proposed, which simulates the relative motion of two rough surfaces by the interaction of two blocks covered by sandpaper with a certain degree of roughness. In this experimental model, the interaction between rough surfaces (sandpaper), with a relative motion in opposite directions to each other, produces stick-slip events (synthetic seismicity), which mimic real seismicity. Here we present the first analyses of synthetic seismicity by calculating the Gutenberg-Richter law, temporal correlations and characterization in terms of organization and order from the Fisher-Shannon method for each synthetic catalogue.

How to cite: Ramírez-Rojas, A., Telesca, L., and Flores-Márquez, E. L.: Novel experimental design for the study of seismic processes based on the stick-slip mechanism., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5321, https://doi.org/10.5194/egusphere-egu25-5321, 2025.

EGU25-5493 | Orals | NH4.4

Toward Real-Time Forecasting of Earthquake Occurrence and Ground-Shaking Intensity Using ETAS and GMM: Insights from Recent Large Earthquakes in Taiwan 

Ming-Che Hsieh, Chung-Han Chan, Kuo-Fong Ma, Yin-Tung Yen, Chun-Te Chen, Da-Yi Chen, and Yi-Wun Liao

Earthquake forecasting, combined with precise ground-shaking estimations, plays a pivotal role in safeguarding public safety, fortifying infrastructure, and bolstering the preparedness of emergency services. This study introduces a comprehensive workflow that integrates the epidemic-type aftershock sequence (ETAS) model with a preselected ground-motion model (GMM), facilitating accurate short-term forecasting of ground-shaking intensity (GSI), which is crucial for adequate earthquake warning for earthquake-prone regions like Taiwan. First, an analysis was conducted on a Taiwanese earthquake catalog from 1994 to 2022 to optimize the ETAS parameters. The dataset used in this analysis allowed for the further calculation of total, background, and clustering seismicity rates, which are crucial for understanding spatiotemporal earthquake occurrence. Subsequently, short-term earthquake activity simulations were performed using these up-to-date seismicity rates to generate synthetic catalogs. The ground-shaking impact on the target sites from each synthetic catalog was assessed by determining the maximum intensity using a selected GMM. This simulation process was repeated to enhance the reliability of the forecasts. Through this process, a probability distribution was created, serving as a robust forecasting for GSI at sites. The performance of the forecasting model was validated through an example of the Taitung, Taiwan earthquake sequence in September 2022, showing its effectiveness in forecasting earthquake activity and site-specific GSI. The other example is the Hualien, Taiwan earthquake sequence from April 2024, which serves as an excellent demonstration of a workflow designed to provide real-time aftershock forecasting following an M7.2 event. The proposed forecasting model can quickly deliver short-term seismic hazard curves and warning messages, facilitating timely decision-making.

How to cite: Hsieh, M.-C., Chan, C.-H., Ma, K.-F., Yen, Y.-T., Chen, C.-T., Chen, D.-Y., and Liao, Y.-W.: Toward Real-Time Forecasting of Earthquake Occurrence and Ground-Shaking Intensity Using ETAS and GMM: Insights from Recent Large Earthquakes in Taiwan, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5493, https://doi.org/10.5194/egusphere-egu25-5493, 2025.

EGU25-8052 | Orals | NH4.4

Multiparameter observations of Lithosphere–Atmosphere–Ionosphere pre-seismic anomalies: Insights from the 2022 M6.8 Chihshang earthquake in southeastern Taiwan 

Ching-Chou Fu, Hao Kuo-Chen, Chung-Hsiang Mu, Hau-Kun Jhuang, Lou-Chuang Lee, Vivek Walia, and Tsung-Che Tsai

This study conducted a systematic analysis of the 2022 Chihshang earthquake sequence in eastern Taiwan, integrating multidimensional observational parameters related to the lithosphere, atmosphere, and ionosphere. High-resolution data from the MAGIC (Multidimensional Active fault of Geo-Inclusive observatory - Chihshang) at the Chihshang fault area provided a comprehensive and diverse dataset. The analysis revealed significant pre-earthquake anomalies across various parameters. These include a marked increase in soil radon concentration one month prior to the earthquake, concurrent anomalies in hydrogeochemical parameters (e.g., elevated groundwater temperature, reduced pH, and decreased chloride ion concentration), and active foreshock activity detected by a dense microseismic network starting mid-August, suggesting the development of microfractures within the lithosphere. Additionally, persistent OLR (Outgoing Longwave Radiation) anomalies, indicating hotspots near the epicenter, were observed from September 5 to 7. Pre-earthquake signals in TEC (Total Electron Content) were identified between August 20 and September 13 in two independent datasets, GIM-TEC and CWA-TEC.

Post-earthquake observations revealed a significant increase in CO2 flux in the region, likely attributable to the release of deep-seated gas sources or enhanced permeability of the fault system. These combined observations suggest that all anomalies can be classified as short-term precursors, which can be interpreted within the theoretical framework of lithosphere-atmosphere-ionosphere coupling (LAIC). The findings also contribute to a deeper understanding of the earthquake preparation process. This study underscores the critical importance of real-time integration of multi-parameter observations, offering new insights and improvements for seismic hazard assessment and advancing the predictive capability of earthquake precursors.

How to cite: Fu, C.-C., Kuo-Chen, H., Mu, C.-H., Jhuang, H.-K., Lee, L.-C., Walia, V., and Tsai, T.-C.: Multiparameter observations of Lithosphere–Atmosphere–Ionosphere pre-seismic anomalies: Insights from the 2022 M6.8 Chihshang earthquake in southeastern Taiwan, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8052, https://doi.org/10.5194/egusphere-egu25-8052, 2025.

EGU25-8652 | ECS | Posters on site | NH4.4

Recent achievements on the application of Robust Satellite Techniques to the short-term seismic hazard forecast 

Roberto Colonna, Carolina Filizzola, Nicola Genzano, Mariano Lisi, Iacopo Mancusi, Carla Pietrapertosa, and Valerio Tramutoli

Robust Satellite Techniques applied to long-term satellite TIR (Thermal InfraRed) radiances have
been, since more than 25 years, employed to identify those anomalies (in the spatial/temporal
domain) possibly associated to the occurrence of major earthquakes.
The results until now achieved by processing multi-annual (more than 10 years) time series of TIR
satellite images collected in different continents and seismic regimes, showed that more than 67%
of all identified (space-time persistent) anomalies occur in the pre-fixed space-time window around
the occurrence time and location of earthquakes (M≥4), with a false positive rate smaller than 33%.
Moreover, Molchan error diagram analysis gave a clear indication of non-casualty of such a
correlation, in comparison with the random guess function.
After the most comprehensive test performed over Greece, Italy, Turkey and Japan, here, we will
critically discuss the preliminary results achieved over California by applying RST analyses to
long-term series of GOES-17 radiances.

How to cite: Colonna, R., Filizzola, C., Genzano, N., Lisi, M., Mancusi, I., Pietrapertosa, C., and Tramutoli, V.: Recent achievements on the application of Robust Satellite Techniques to the short-term seismic hazard forecast, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8652, https://doi.org/10.5194/egusphere-egu25-8652, 2025.

EGU25-8809 | Orals | NH4.4

Noise reductions of VLF signals and excitation/attenuation of waves with small wave periods before earthquakes 

Giovanni Nico, Aleksandra Nina, Pierfrancesco Biagi, Hans Ulrich Eichelberger, Mohammed Y. Boudjada, and Luka Č. Popović

Various types of changes in the characteristics of very low frequency (VLF) signals before earthquakes have been presented during the past few decades. Most of these changes have been observed on data with time sampling of the order of a few tenths of a second or of the order of minutes. Improvements in this sampling in recent years have indicated three new types of changes whose onsets have been observed a few minutes or tens of minutes before the earthquake. These changes manifest themselves as reductions in the VLF signal amplitude and phase noises, and excitation and attenuation of waves with small wave periods in both of these signal characteristics [1-5].

In this work, we present these changes and list the parameters in the time and frequency domains that are significant for statistical analyses. A central issue is the relationship of the changes with the characteristics of earthquakes, the observed signals, and their spread in the surrounding area. The presented analyses were conducted on data recorded by a VLF receiver in Belgrade, Serbia.

 

References:

[1] A. Nina, S. Pulinets, P.F. Biagi, G. Nico, S.T. Mitrović, M. Radovanović, L.Č. Popović, “Variation in natural short-period ionospheric noise, and acoustic and gravity waves revealed by the amplitude analysis of a VLF radio signal on the occasion of the Kraljevo earthquake (Mw = 5.4)”, Science of The Total Environment, 710, 136406, 2020.

[2] A. Nina, P. F. Biagi, S. T. Mitrović, S. Pulinets, G. Nico, M. Radovanović,  L. Č. Popović, “Reduction of the VLF signal phase noise before earthquakes”, Atmosphere 12 (4), 444, 2021.

[3] A. Nina, P. F. Biagi, S. A. Pulinets, G. Nico, S. T. Mitrović, V. M. Čadež, M. Radovanović, M. Urošev,  L. Č. Popović, “Variation in the VLF signal noise amplitude during the period of intense seismic activity in Central Italy from 25 October to 3 November 2016”, Frontiers in Environmental Science, 10, 10:1005575, 2022.

[4] A. Nina, “Analysis of VLF Signal Noise Changes in the Time Domain and Excitations/Attenuations of Short-Period Waves in the Frequency Domain as Potential Earthquake Precursors”, Remote Sensing, 16(2), 397, (2024)

[5] A. Nina “VLF Signal Noise Reduction during Intense Seismic Activity: First Study of Wave Excitations and Attenuations in the VLF Signal Amplitude”, Remote Sensing, 16(8), 1330, 2024.

 

How to cite: Nico, G., Nina, A., Biagi, P., Eichelberger, H. U., Boudjada, M. Y., and Popović, L. Č.: Noise reductions of VLF signals and excitation/attenuation of waves with small wave periods before earthquakes, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8809, https://doi.org/10.5194/egusphere-egu25-8809, 2025.

A critical review of geoelectrical monitoring activities carried out in seismically active areas is presented and discussed. The electrical resistivity of rocks is one of the geophysical parameters of greatest interest in the study of possible seismic precursors, and it is strongly influenced by the presence of highly fractured zones with high permeability and fluid levels. The analysis in this study was based on results obtained over the last 50 years in seismic zones in China, Japan, the USA and Russia. These previous works made it possible to classify the different monitoring strategies, to analyze the theoretical models for interpreting possible correlations between anomalies in resistivity signals and local seismicity, and to identify the main scientific and technological gaps. In addition, much attention is given to some recent work on the study of correlations between focal mechanisms and the shapes of anomalous patterns in resistivity time series, and to the new possibilities offered by the AI-based methods for geophysical data processing. Finally, new strategies and activities for investigating the spatial and temporal dynamics of the electrical resistivity changes in seismically active areas were identified.

How to cite: Lapenna, V.: Detecting DC Electrical Resistivity Changes in Seismic Active Areas: State-of-the-Art and Future Directions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9250, https://doi.org/10.5194/egusphere-egu25-9250, 2025.

EGU25-9938 | ECS | Posters on site | NH4.4

High resolution tsunami inundation maps: towards multi-hazard risk analysis. 

Hany M. Hassan and Antonella Peresan

Multi-hazard disaster risk analyses in coastal areas requires the integration of data and models concerning hazard, exposure and vulnerability data and models, all developed with high spatial resolution. Indeed, accurate high-resolution models and data are essential for properly assessing the impact of specific hazards that threaten coastal areas, such as tsunamis, floods, landslides, and coastal erosion. Nevertheless, this level of detail remains unachieved for many coastal hazards in various locations. Consequently, critical fine-scale differences in localized risk assessment are overlooked, leading to potential underestimations or overestimations of the actual risk to coastal communities. It is vital to address this gap in order to enhance the accuracy and reliability of risk assessments.

A key step in tsunami hazard and risk assessment involves the development of inundation maps, specifically maps describing inundated areas and related depths. To date, such maps are not yet available at proper resolution for the coastal areas of the Friuli-Venezia-Giulia Region (FVG). Accordingly, this study aims to enhance the characterization of tsunami hazard in the Northern Adriatic by developing detailed inundation maps and possibly addressing the identified research gaps. Leveraging on accurate and high resolution bathymetry and topographic data is crucial for reliable tsunami modelling for the FVG coastal areas. To this purpose, bathymetry and topographic data are refined and are used, along with existing databases of tsunamigenic earthquake sources, for modelling tsunami waves propagation and inundation by means of the NAMI DANCE code (e.g. Yalciner et al. 2014, Mediterranean Sea Oceanography and references therein).

Existing datasets from open access and local data sources are collected and then refined, particularly addressing inaccuracies in lagoon bathymetry. This involves incorporating high-resolution data and considering small-scale coastal features that can significantly impact tsunami inundation. Multiple bathymetry and topography datasets are used to develop high resolution refined data at 25 meters, and 10 meters resolution. The database of co-seismic seafloor displacement for all individual scenarios, developed based upon the DISS-3.3.0 database, is adopted to carry out a reappraisal of tsunami wave amplitude maps (Peresan & Hassan, MEGR 2024 and references therein) and to estimate realistic tsunami inundation maps. Additionally, tsunami sources caused by local earthquakes relevant to the FVG region are investigated, providing local scale maps of wave amplitudes and inundation estimates; this involves using appropriate fault rupture realisations for local tsunami scenarios (ITCS100&101), as specified in the DISS-3.3.0 database.

The outcomes from this study provide the basis for multi-scenario tsunami hazard assessment, contributing to the development of high-resolution and comprehensive tsunami hazard maps for the Northern Adriatic coasts. Moreover, along with high-resolution exposure maps, they contribute improving precision and accuracy of related risk assessment, and hence are an important step in preparedness, response, and prevention efforts in the framework of disaster risk management.

This research is a contribution to the RETURN Extended Partnership (European Union Next-Generation EU—National Recovery and Resilience Plan—NRRP, Mission 4, Component 2, Investment 1.3—D.D. 1243 2/8/2022, PE0000005).

How to cite: Hassan, H. M. and Peresan, A.: High resolution tsunami inundation maps: towards multi-hazard risk analysis., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9938, https://doi.org/10.5194/egusphere-egu25-9938, 2025.

EGU25-10351 | Posters on site | NH4.4

Investigation of VLF/LF electromagnetic wave propagation as recorded by the receivers of the INFREP network 

Iren-Adelina Moldovan, Victorin Emilian Toader, Hans Ulrich Eichelberger, Pier Francesco Biagi, Mohammed Boudjada, Mihai Anghel, Liviu Marius Manea, Andrei Mihai, and Bogdan Antonescu

In recent decades, significant efforts have been devoted to understanding and interpreting the link between ionospheric perturbations and natural or anthropogenic phenomena, such as seismic activity, electrical or geomagnetic storms, and unidentified radio emissions. This is achieved through various methods among which is also the study of electromagnetic (EM) wave propagation in the very low frequency (VLF, 3–30 kHz) and low frequency (LF, 30–300 kHz) bands. These bands enable long-distance communication, navigation, and military applications, including submarine contact, AM broadcasting, lightning detection, and weather systems. Due to their long wavelengths, VLF and LF waves exhibit unique propagation characteristics. VLF waves propagate globally by using Earth-ionosphere waveguides, reflecting off the D and E layers as skywaves, and are influenced by solar and atmospheric conditions. LF waves primarily rely on ground waves for extensive coverage, although they can also utilize ionospheric reflection (skywaves) for longer-distance communication.

This paper introduces fundamental concepts related to VLF/LF electromagnetic wave emission, propagation, reception, and the perturbing factors that affect them. Additionally, it presents key findings from the European INFREP Receivers Network, which studies seismo-ionospheric anomalies linked to earthquake activity. Established in 2009, the INFREP network monitors VLF/LF signals from transmitters across Europe and neighboring regions. The network currently comprises 10 receivers, built by Elettronika (Italy), and operates at a sampling rate of one sample per minute. The Romanian segment of INFREP includes two receivers, operational since 2009 and 2017, with only brief interruptions, notably during the pandemic when travel restrictions hindered access to the observatories.

The paper discusses the current state of the INFREP network and outlines methods for providing near real-time data access. It highlights advancements in real-time electromagnetic data transmission, archiving, and the use of 2D and 3D online signal visualization and processing techniques. Data access is available through the INFREP headquarters in Graz, Austria (https://infrep.iwf.oeaw.ac.at/data-access/) and the National Institute for Earth Physics in Romania (https://mg.infp.ro/d/ch-aqZXIz/vlf-lf-radio-data?orgId=1&from=now-6M&to=now). The paper also shares findings from the detection of potential ionospheric anomalies in EM signals preceding large earthquakes that occurred between 2012 and 2024. All anomalies are analyzed in correlation with space weather events and extreme meteorological phenomena.

This paper was carried out within Nucleu Program SOL4RISC, supported by MCI, project no PN23360201, and PNRR- DTEClimate Project nr. 760008/31.12.2023, Component Project Reactive, supported by Romania - National Recovery and Resilience Plan

 

How to cite: Moldovan, I.-A., Toader, V. E., Eichelberger, H. U., Biagi, P. F., Boudjada, M., Anghel, M., Manea, L. M., Mihai, A., and Antonescu, B.: Investigation of VLF/LF electromagnetic wave propagation as recorded by the receivers of the INFREP network, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10351, https://doi.org/10.5194/egusphere-egu25-10351, 2025.

EGU25-10353 | ECS | Orals | NH4.4

High-resolution exposure models for coastal cities in Northern Adriatic for multi-risk analysis 

Hazem Badreldin, Chiara Scaini, Hany M Hassan, and Antonella Peresan

Multi-hazard disaster risk reduction and mitigation require high-resolution exposure models that grasp the characteristics of assets at the local scale. High-resolution exposure models may allow improving precision/accuracy of risk and damage assessments, especially for hazards which are characterised by high spatial variability or may be influenced by the presence of the assets, such as tsunami or flooding. We propose a methodology for developing a high-resolution population and residential buildings exposure models, to be used for multi-hazard risk reduction purposes at the local scale.  This method has been tested and validated for a selected coastal area in the upper Adriatic, exposed to multiple hazards including earthquakes, tsunamis, meteorological events and coastal erosion. For the development of the population exposure model, a high-resolution population density data, collected at global scale, is combined with the national population census data, leveraging  both on the accuracy of the national census and on the resolution of the global data. Also, the building census data is complemented with exposure indicators extracted from digital building footprints from the Carta Tecnica Regionale Numerica (CTRN),  which is missing in census data, such as average built area, total built area, replacement cost, height and plan regularity. The final exposure layers are assembled at two resolutions: 100 meters and 30 meters, with information also provided at the census unit level. We discuss the development and use of these layers for multi-risk assessment and their potential combination with artificial intelligence. 

This research is a contribution to the projects: RETURN Extended Partnership (European Union Next-Generation EU—National Recovery and Resilience Plan—NRRP, Mission 4, Component 2, Investment 1.3—D.D. 1243 2/8/2022, PE0000005); PRIN-PNRR project SMILE: Statistical Machine Learning for Exposure development, funded by the European Union- Next Generation EU, Mission 4 Component 1 (CUP F53D23010780001). 

How to cite: Badreldin, H., Scaini, C., M Hassan, H., and Peresan, A.: High-resolution exposure models for coastal cities in Northern Adriatic for multi-risk analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10353, https://doi.org/10.5194/egusphere-egu25-10353, 2025.

Japan is frequently hit by major earthquakes, such as the 2011 off the Pacific coast of Tohoku Earthquake and the 2024 Noto Peninsula Earthquake, which cause enormous human and economic losses. Short-term forecast of earthquakes is effective for mitigating such damage, but this has not been achieved to date. On the other hand, there have been reports of electromagnetic phenomena preceding major earthquakes in various frequency bands, including precursor phenomena in the VLF/LF band (3-300 kHz). In this study, we investigated earthquake-related VLF/LF signals, which has strong electromagnetic emissions due to lightning activity, and it is important to discriminate the VLF/LF signals from those due to lightning activity. In this study, two approaches were attempted: (1) development of a source localization method using VLF/LF broadband interferometry and (2) removal of signals caused by lightning discharges using machine learning.
The first approach is expected to spatially discriminate between VLF/LF signals related to earthquakes (which are located near the epicenter and do not move) and signals related to lightning activity (which move with fronts and thunderclouds). The second is to utilize machine learning technology, which has been rapidly developed in recent years, for detection and removal of lightning discharge signals. For example, Wu et al. at Gifu University have succeeded in classifying lightning discharge waveforms in the thunderstorm activity process with an accuracy of approximately 99% using a machine learning technique called Random Forest. In this study, machine learning is expected to efficiently discriminate and eliminate known lightning discharge signals from a large amount of observation data with high accuracy, and analyze the remaining unknown signals to efficiently investigate the relationship between lightning and earthquakes. In this paper, we will describe the specific methods and results of the above two approaches.

How to cite: Hattori, K., Ota, Y., Yoshino, C., and Imazumi, N.: Construction of a VLF/LF band interferometer using a capacitive circular flat-plane antenna and discrimination and identification of observed VLF/LF band signals by machine learning: Preliminary results, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10447, https://doi.org/10.5194/egusphere-egu25-10447, 2025.

EGU25-13142 | Orals | NH4.4

Sub-ionospheric VLF/LF waveguide electric field investigation from Mw≥5.0 earthquake events with multiple receivers in Europe 

Hans U. Eichelberger, Mohammed Y. Boudjada, Aleksandra Nina, Bruno P. Besser, Daniel Wolbang, Maria Solovieva, Pier F. Biagi, Patrick H. M. Galopeau, Christoph Schirninger, Iren-Adelina Moldovan, Giovanni Nico, Manfred Stachel, Özer Aydogar, Cosima Muck, Josef Wilfinger, and Irmgard Jernej

Electric field amplitude and phase measurements between narrowband VLF/LF transmitters and receivers in the sub-ionospheric waveguide are affected and altered by man-made and natural sources (Nina 2024; Boudjada et al., 2024a,b). In this study we investigate Mw≥5.0 earthquakes (EQs) which occurred in Europe during the year 2024 based on data from the INFREP receiver network (Biagi et al., 2019; Moldovan et al., 2015; Galopeau et al., 2023). In the selected Mediterranean area with geographical longitude [-10°E, 40°E] and latitude [30°N, 50°N] the United States Geological Survey EQ catalog (USGS, 2025) provides 20 events with Mw≥5.0. For these EQs we apply the night-time amplitude method and consider variations in the terminator times (Hayakawa et al., 2010). The main radio links that cross the EQ prone areas are from transmitters localized in the southern part of Europe, including TBB (26.70 kHz, Bafa, Turkey), ITS (45.90 kHz, Niscemi, Sicily, Italy), and ICV (20.27 kHz, Tavolara, Italy). 

We find statistically significant electric field anomalies for various VLF/LF paths, particularly for events with higher magnitudes. The continuous VLF/LF electric field amplitude and phase datasets can be important parameters for real-time observations and services to assess seismic hazards and disturbing physical phenomena within the waveguide.

References:

Biagi, P.F., et al., The INFREP network: Present situation and recent results, OJER, 8, 101-115, 2019. https://doi.org/10.4236/ojer.2019.82007

Boudjada, M.Y., et al., Unusual sunrise and sunset terminator variations in the behavior of sub-ionospheric VLF phase and amplitude signals prior to the Mw7.8 Turkey Syria earthquake of 6 February 2023, Remote Sens., 16, 4448, 2024. https://doi.org/10.3390/rs16234448

Boudjada, M.Y., et al., Analysis of pre-seismic ionospheric disturbances prior to 2020 Croatian earthquakes, Remote Sens., 16, 529, 2024. https://doi.org/10.3390/rs16030529

Galopeau, P.H.M., et al., A VLF/LF facility network for preseismic electromagnetic investigations, Geosci. Instrum. Method. Data Syst., 12, 231–237, 2023. https://doi.org/10.5194/gi-12-231-2023

Hayakawa, M., et al., A statistical study on the correlation between lower ionospheric perturbations as seen by subionospheric VLF/LF propagation and earthquakes, JGR Space Physics, 115(A9), 09305, 2010. https://doi.org/10.1029/2009JA015143

Moldovan, I.A., et al., The development of the Romanian VLF/LF monitoring system as part of the International Network for Frontier Research on Earthquake Precursors (INFREP), Romanian Journal of Physics, 60 (7-8), 1203-1217, 2015. Bibcode: 2015RoJPh..60.1203M https://rjp.nipne.ro/2015_60_7-8/RomJPhys.60.p1203.pdf

Nina, A., VLF signal noise reduction during intense seismic activity: First study of wave excitations and attenuations in the VLF signal amplitude, Remote Sens., 16, 1330, 2024. https://doi.org/10.3390/rs16081330

USGS, United States Geological Survey earthquake catalog, https://www.usgs.gov/programs/earthquake-hazards, as of Jan 2025.

How to cite: Eichelberger, H. U., Boudjada, M. Y., Nina, A., Besser, B. P., Wolbang, D., Solovieva, M., Biagi, P. F., Galopeau, P. H. M., Schirninger, C., Moldovan, I.-A., Nico, G., Stachel, M., Aydogar, Ö., Muck, C., Wilfinger, J., and Jernej, I.: Sub-ionospheric VLF/LF waveguide electric field investigation from Mw≥5.0 earthquake events with multiple receivers in Europe, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13142, https://doi.org/10.5194/egusphere-egu25-13142, 2025.

EGU25-13210 | Orals | NH4.4

Comparative multifractal study of seismicity in two seismic zones of Türkiye in the period from 2010 to 2024. 

Elsa Leticia Flores-Marquez, Alejandro Ramirez Rojas, and Jennifer Pérez-Oregon

Intense earthquakes have been natural phenomena that produce enormous disasters, mainly in large urban areas, due to the intense energy released in a very short period. Earthquakes are inevitable natural phenomena, and up to now, they cannot be predicted. On February 6, 2023, a M 7.8 earthquake occurred in southern Türkiye, near the northern border of Syria. This earthquake was followed by a M 7.5 earthquake to the north. The relative motions of three major tectonic plates (Arabian, Eurasian, and African) and one smaller tectonic block (Anatolian) are responsible for the seismicity in Türkiye. Recently, Onur investigated the aftershock distribution and its relation to energy release on the faults and Coulomb stress change areas, his study allowed the relocation of two-catastrophic earthquakes. In the present work we analyze the behavior of multifractality and its complexity parameters calculated from the catalog of seismic magnitudes during a period of 14 years monitored within two regions of Türkiye: the first one (west) between (35-42) Latitude, (25-34) Longitude and the second one (East) between (35-42) Latitude and (34-42) Longitude, being this area where the doublet occurred. Our results show differences in both multifractality and its complexity measures between the two regions. These findings may be indicators of expected seismicity in each region.

 

How to cite: Flores-Marquez, E. L., Ramirez Rojas, A., and Pérez-Oregon, J.: Comparative multifractal study of seismicity in two seismic zones of Türkiye in the period from 2010 to 2024., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13210, https://doi.org/10.5194/egusphere-egu25-13210, 2025.

EGU25-14706 | Orals | NH4.4

Design of the PRELUDE CubeSat for investigating ionospheric D-region earthquake precursor 

Masashi Kamogawa, Masashiko Yamazaki, and Nagisa Sone

Despite advances in satellite remote sensing, predicting large earthquakes, remains a significant challenge due to the unpredictable nature of these events. To address this challenge, our study, building upon the achievements of the French DEMETER satellite, focuses on atmospheric and space electrical variations as potential indicators of ionospheric D-region precursors to earthquakes. This approach is expected to contribute to the enhancement of short-term prediction capabilities. For this purpose, we would like to introduce our CubeSat PRELUDE (Precursory electric field observation CubeSat Demonstrator), a tiny satellite dedicated to the earthquake precursor detection and elucidated the physical mechanism. PRELUDE is scheduled for launch in JFY2025 as part of JAXA’s Innovative Satellite Technology Demonstration Program. This study presents the results of the system design, development, and mission planning of the PRELUDE, aimed at clarifying the physical mechanisms behind the statistically significant earthquake precursor ionospheric phenomena. PRELUDE is a 6U CubeSat specialized in VLF electromagnetic wave intensity observation, weighing 8 kg. To achieve miniaturization, it incorporates a drive recording function to downlink only the data surrounding the EQ epicenter to ground stations, reducing data storage and transmission requirements. Additionally, it hybridizes the Langmuir and electric field probes, typically found on satellites weighing over 100 kg like DEMETER, into a compact design suitable for CubeSats weighing just a few kilograms. The hybrid sensor unit extends booms bidirectionally by 1.5 m from the satellite using a folding extension mechanism, In this presentation, we show the satellite design requirements for elucidating the mechanism of earthquake precursor phenomena.

How to cite: Kamogawa, M., Yamazaki, M., and Sone, N.: Design of the PRELUDE CubeSat for investigating ionospheric D-region earthquake precursor, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14706, https://doi.org/10.5194/egusphere-egu25-14706, 2025.

EGU25-14734 | Posters on site | NH4.4

Rapid prediction method of earthquake damage to masonry structures based on machine learning 

Lingxin Zhang, Yan Liu, Li Liu, and Baijie Zhu

Masonry structures are one of the most vulnerable to severe and extensive damage in terms of previous earthquakes. It is significant to quickly evaluate the seismic damage levels of masonry structures, to reduce casualties and economic losses caused by earthquakes. However, traditional methods based on manual judgment or finite element simulations tend to be relatively slower . In this paper, a machine learning-based rapid prediction method was proposed for assessing the seismic damage of masonry structures. By analysis of building data from several cities and combining ground motion with structural characteristics, 11 impact factors were identified as input variables. The LM-BP neural network model was developed by a backpropagation (BP) neural network with strong nonlinear modeling capabilities, and by the Levenberg-Marquardt (LM) algorithm. The accuracy and stability of the model were verified by comparing the predicted values with actual earthquake examples. The results show that the selected seismic damage impact factors can accurately reflect the structural damage level. By comparing methods using parameters on either the structure or ground motion, the predictive accuracy of the proposed method is significantly enhanced. It provides a basis for post-earthquake structural safety assessments and disaster prevention and mitigation work.

How to cite: Zhang, L., Liu, Y., Liu, L., and Zhu, B.: Rapid prediction method of earthquake damage to masonry structures based on machine learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14734, https://doi.org/10.5194/egusphere-egu25-14734, 2025.

EGU25-16143 | ECS | Orals | NH4.4

Machine Learning based EStimator for ground shaking maps workflow applied to New Zealand 

Rut Blanco Prieto, Marisol Monterrubio Velasco, Brendon Bradley, Claudio Schill, and Josep de la Puente

Earthquakes are among the most frequent yet unpredictable natural hazards, posing substantial risk to human safety and infrastructure globally, particularly, when large-magnitude earthquakes occur. This highlights the urgent need to develop innovative and alternative methodologies for rapidly assessing the intensity of ground shaking following an earthquake.

This study explores the application of the Machine Learning Estimator for Ground Shaking Maps (MLESmap) methodology in New Zealand, a region characterized by  high seismic activity.

MLESmap utilizes extensive datasets of high-fidelity, physics-based seismic scenarios to rapidly estimate ground-shaking intensity in near real-time following an earthquake. This methodology has demonstrated evaluation times similar to those of empirical ground motion models, while offering superior predictive accuracy in the two previously tested regions: the Los Angeles basin and the South Iceland Seismic Zone (SISZ).

To adapt MLESmap for New Zealand’s seismicity, seismic simulations tailored to the unique geological and tectonic context of the region are implemented. Specifically, we use the dataset generated by CyberShake NZ, a probabilistic seismic hazard analysis (PSHA) software developed by the University of Canterbury. Using this software, a total of 11,362 finite-fault rupture simulations were performed across the region and seismic hazard results were calculated on a grid of 27,481 synthetic seismic stations. A ‘forward’ simulation approach was adopted due to the large number of output locations relative to rupture locations, the optimisation of the grid for each rupture and the intention to include plasticity.

The expected results aim to demonstrate the applicability of MLESmap to New Zealand, providing ML-based tools for rapid response actions. This study also takes the first steps in applying cascading effects to MLESmap, in order to improve the overall risk assessment and to advance prevention efforts through innovative and multidisciplinary methodologies.

 

 

©2023 ChEESE-2P Funded by the European Union. This work has received funding from the European High Performance Computing Joint Undertaking (JU) and Spain, Italy, Iceland, Germany, Norway, France, Finland and Croatia under grant agreement No 101093038.

How to cite: Blanco Prieto, R., Monterrubio Velasco, M., Bradley, B., Schill, C., and de la Puente, J.: Machine Learning based EStimator for ground shaking maps workflow applied to New Zealand, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16143, https://doi.org/10.5194/egusphere-egu25-16143, 2025.

The region near the India-Eurasia plate boundary has a long history of large earthquakes. Over the past century, more than 50 earthquakes with magnitudes of 7 or greater have occurred within 500 km of the Indo-Eurasian collision zone. These include the 2015 M7.8 Nepal earthquake, the 1934 M8.0 Bihar-Nepal earthquake, the 1950 M8.6 Assam earthquake, and the 1905 M7.9 Kangra earthquake. The January 7, 2025, M7.1 earthquake in the southern Tibetan Plateau further underscores the seismic significance of this region. This study examines the temporal variation in seismicity within the Indo-Eurasian collision zone and its adjacent areas by utilizing historical records and instrumentally recorded earthquake data from 1900 to 2024. Based on seismic behaviour, clustering of events, and tectonic structures, the collision zone is divided into 26 distinct seismic zones. The temporal variation in seismicity for each zone is analyzed, and a susceptibility index, ESI6, is calculated. This index considers the return period of earthquakes with Mw ≥ 6 and the time elapsed since the last Mw ≥ 6 earthquake in each zone. The ESI6 represents the number of pending Mw ≥ 6 earthquakes in each seismic zone. Ten zones with high ESI6 values (>2.5) have been identified; these zones were seismically active in the past but have remained without major earthquakes for the last three decades. To mitigate potential losses and raise awareness, it is critical to implement GPS monitoring of plate movements, satellite-based deformation monitoring, and seismic health assessments of crucial infrastructure in these silent zones.

How to cite: Kumar, S.: Spatio Temporal Analysis of Earthquake Potential in the Indo-Eurasian Collision Zone: Identifying Future Seismic Hotspots Using the Earthquake Susceptibility Index, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16591, https://doi.org/10.5194/egusphere-egu25-16591, 2025.

EGU25-17662 | Orals | NH4.4

Automated Site Effects Mapping in Mayotte Using Airborne Electromagnetic Data and Machine Learning 

Cécile Gracianne, Hugo Breuillard, Célia Mato, Pierre-Alexandre Reninger, Agathe Roullé, Anne Raingeard, and Roxanne Rusch

Recent seismic hazard assessments in Mayotte have highlighted the island's significant exposure to site effects during earthquakes. These effects are closely linked to its complex geological setting, characterized by altered volcanic formations whose heterogeneous geometry leads to strong spatial variations in ground motion. In response to governmental requests, a site effects map is being developed to raise public awareness and support risk-informed urban planning.

A novel methodology for site effects mapping has recently been developed at BRGM, integrating airborne electromagnetic (AEM) data with borehole logs, geological maps, and seismic data (MASW and H/V measurements). This approach was tested on three test sites covering 12 km² of Mayotte surface, and it has demonstrated its potential in imaging the geological interfaces responsible for site effects. However, the current methodology relies on expert-driven data interpretation, making its large-scale application highly labour-intensive and costly. To overcome this limitation, partial automation of the data processing is required in order to handle larger datasets efficiently.

Machine learning techniques offer a promising solution to address this challenge. The test sites provided a unique training dataset, associating resistivity profiles derived from AEM data with the position of geological interfaces responsible for site effects within the soil column. These interface locations were determined through the integration and interpretation of all available geological and geophysical data, including MASW, H/V measurements, and borehole logs. Using this dataset, we trained various models, including Random Forest and Convolutional Neural Networks (CNN), to predict the localization of geological interfaces responsible for site effects based on AEM data.

Preliminary results indicate that the CNN model shows good performances on this task. Nevertheless, further improvements require the expansion of training datasets, underscoring the significant investment needed to generalize this approach to other regions. Future research will focus on refining predictive models and optimizing data acquisition to support large-scale implementation.

How to cite: Gracianne, C., Breuillard, H., Mato, C., Reninger, P.-A., Roullé, A., Raingeard, A., and Rusch, R.: Automated Site Effects Mapping in Mayotte Using Airborne Electromagnetic Data and Machine Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17662, https://doi.org/10.5194/egusphere-egu25-17662, 2025.

Finding a sustainable solution to disaster risk mitigation needs to consider different aspects of the disaster’s impact along with social, economic, and physical characteristics of the region. In this regard, a desirable solution for disaster risk mitigation for a region is the one tailored to the local characteristics. These local characteristics not only help measure the different aspects of a disaster impact but also portray existing pressing issues as priorities. While the former can be modeled using risk and resilience assessment models, the latter can be measured from experts’ points of view. Ultimately, the combination of the expert’s perception on important issues and the output of risk and resilience assessment models can be used to evaluate the optimality of each disaster risk mitigation solution.

In this research, a Multi-Criteria Decision Analysis (MCDA) framework is developed to provide an evaluation of each disaster risk mitigation. The developed framework is designed to be able to run on the action-outcome results from risk and resilience assessment models and the cardinal ranking of the decision criteria, representing decision-makers’ expert opinion on the priorities in mitigating and managing disaster risk. The developed MCDA framework is very practical as it can run on action-outcome results, and these results are accessible from a large variety of risk and resilience assessment models. Furthermore, the developed MCDA framework takes into account the uncertainty in the risk and resilience assessment models. In compatibility with running on minimal available information, the MCDA’s decision model is simplified to one layer with a single layer of the decision criteria.

Additionally, as the number of competing mitigation solutions might increase rapidly in practice, the MCDA framework is developed to handle a huge number of alternatives more efficiently and with relatively limited computational resources. The MCDA framework is developed based on the CAR method of eliciting the preferences among mitigation alternatives. The final results evaluate the competing disaster risk mitigation solution based on available data (as processed by risk and resilience assessment models) and the expert’s opinion on important issues and their preferences on the important aspects of disaster impact. As such, the final results provide an estimation of the expert’s belief on the desirability of each of the disaster risk mitigation solutions.

This MCDA framework is developed as part of the Horizon Europe project MEDiate (Multi-hazard and risk-informed system for Enhanced local and regional Disaster risk management). This project is dedicated to creating a decision-support system (DSS) for disaster risk management that not only takes into account the complexities of multiple interacting natural hazards but also tailors the final solution to the characteristics, priorities, and concerns of the local communities and decision-makers. The MEDiate framework is implemented on four different testbeds (Oslo (Norway), Nice (France), Essex (UK), and Múlaþing (Iceland)), each of which has a different multi-hazard pair and different socio-economic characteristics. The deployment of the developed MCDA framework on different natural hazards and socio-economic characteristics shows its flexible practicality.

How to cite: Yeganegi, M. R., Komendantova, N., and Danielson, M.: Measuring the experts’ perception about the suitability of natural disaster risk mitigation solutions using minimal risk assessment information, a Multi-Criteria Decision Analysis approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17936, https://doi.org/10.5194/egusphere-egu25-17936, 2025.

EGU25-18225 | Posters on site | NH4.4

A web platform for crowdsourced collection, processing, and visualization of exposure data on buildings 

Maria Teresa Artese, Elisa Varini, Isabella Gagliardi, Gianluigi Ciocca, Flavio Piccoli, Claudio Rota, Matteo Del Soldato, Silvia Bianchini, Chiara Scaini, Antonella Peresan, and Piero Brondi

The ultimate objective of our research is to explore the potential of Machine Learning in the dynamic creation of up-to-date exposure layers for buildings. This effort involves integrating remote sensing images, ancillary data such as national census information, and crowdsourced data collected by trained citizens. The crowdsourcing activity builds on a previous successful initiative developed within the CEDAS (building CEnsus for seismic Damage Assessment) project, which engaged high school students from North-East Italy in collecting data on buildings that were either unavailable from conventional exposure data sources or not easily retrievable via remote sensing techniques (Scaini et al., 2022).

To this end, we are developing a complex multimedia information system via web platform designed to collect, process, store, and distribute information to different knowledge users (policymakers, territorial planners, citizens) with targeted visualization strategies. The crowdsourcing initiatives are taking place in selected municipalities of Tuscany and Friuli regions (Italy), exposed to different natural hazards, such as earthquakes, tsunamis and landslides.  An online questionnaire has been created to assist the user in building data collection and minimize input errors. Simultaneously, building data, along with their photos, are stored in a structured database for research purposes.  For instance, building data and images are used as learning set to train a machine learning algorithm to identify specific features such as roof type, number of floors, and the presence of a basement. These algorithms can then be included in the online questionnaire to facilitate further data collection by automatically suggesting features associated to the buildings. A dedicated visualization tool is being developed on the web platform to showcase the effectiveness of this method in recognition of building features. We will demonstrate the data visualization tools developed on the web platform so far, highlighting the key features of the available exposure databases. The web platform is designed to provide an easy-to-use tool for communicating with various knowledge users, while also enhancing disaster awareness and preparedness, which is attained exploring and collecting data on the built environment.

This study is a contribution to the ongoing PRIN 2022 PNRR project SMILE “Statistical Machine Learning for Exposure development” (code P202247PK9, CUP B53D23029430001) within the European Union-NextGenerationEU program.

How to cite: Artese, M. T., Varini, E., Gagliardi, I., Ciocca, G., Piccoli, F., Rota, C., Del Soldato, M., Bianchini, S., Scaini, C., Peresan, A., and Brondi, P.: A web platform for crowdsourced collection, processing, and visualization of exposure data on buildings, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18225, https://doi.org/10.5194/egusphere-egu25-18225, 2025.

EGU25-21907 | Orals | NH4.4

Seismo-electromagnetism: observations and mechanisms 

Qinghua Huang

Seismogenic mechanism of strong earthquakes plays a fundamental role in disaster prevention. Electromagnetic methods, which are sensitive to fluid, have been widely adopted in the study on seismogenic structure and earthquake physics. Due to the increasing environmental disturbances and limited understanding on electromagnetic anomalies, electromagnetic data cannot fully show their potential values in disaster prevention. We propose an integrated work on seismogenic structure, identification of electromagnetic disturbances, and mechanism of seismo-electromagnetic anomalies. Based on the tests of synthetic and field data, we demonstrate that the multiple electromagnetic methods can reveal the feature of the multi-scaled seismogenic structure. With the developments of the new methodology based on deep learning and the seismo-electromagnetic coupling model, one can investigate the spatio-temporal characteristics of electromagnetic anomalies and their possible relationship with earthquakes. This study may contribute to the study on earthquake forecast and disaster prevention.

How to cite: Huang, Q.: Seismo-electromagnetism: observations and mechanisms, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21907, https://doi.org/10.5194/egusphere-egu25-21907, 2025.

ESSI3 – Open Science Informatics for Earth and Space Sciences

EGU25-240 | ECS | Posters on site | ESSI3.1

Leveraging Machine Learning and Remote Sensing for Water Quality Analysis in Lake Ranco, Southern Chile 

Lien Rodríguez-López, Lisandra Bravo Alvarez, Iongel Duran Llacer, David Ruíz-Guirola, Samuel Montejo-Sánchez, Rebeca Martínez-Retureta, Luc Bourel, Frederic Frappart, and Roberto Urrutia

This study examines the dynamics of limnological parameters of a South American lake located in southern Chile with the objective of predicting chlorophyll-a levels, which are a key indicator of algal biomass and water quality, by integrating combined remote sensing and machine learning techniques. Employing four advanced machine learning models, the research focuses on the estimation of chlorophyll-a concentrations at three sampling stations within Lake Ranco. The data span from 1987 to 2020 and are used in three different cases: using only in situ data (Case 1), using in situ and meteorological data (Case 2), using in situ, and meteorological and satellite data from Landsat and Sentinel missions (Case 3). In all cases, each machine learning model shows robust performance, with promising results in predicting chlorophyll-a concentrations. Among these models, LSTM stands out as the most effective, with the best metrics in the estimation, the best performance was Case 1, with R2 = 0.89, an RSME of 0.32 μg/L, an MAE 1.25 μg/L and an MSE 0.25 (μg/L)2, consistently outperforming the others according to the static metrics used for validation. This finding underscores the effectiveness of LSTM in capturing the complex temporal relationships inherent in the dataset. However, increasing the dataset in Case 3 shows a better performance of TCNs (R2 = 0.96; MSE = 0.33 (μg/L)2; RMSE = 0.13 μg/L; and MAE = 0.06 μg/L). The successful application of machine learning algorithms emphasizes their potential to elucidate the dynamics of algal biomass in Lake Ranco, located in the southern region of Chile. These results not only contribute to a deeper understanding of the lake ecosystem but also highlight the utility of advanced computational techniques in environmental research and management.

How to cite: Rodríguez-López, L., Bravo Alvarez, L., Duran Llacer, I., Ruíz-Guirola, D., Montejo-Sánchez, S., Martínez-Retureta, R., Bourel, L., Frappart, F., and Urrutia, R.: Leveraging Machine Learning and Remote Sensing for Water Quality Analysis in Lake Ranco, Southern Chile, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-240, https://doi.org/10.5194/egusphere-egu25-240, 2025.

EGU25-2206 | Orals | ESSI3.1

Integrated and Open geohazard monitoring data in Aotearoa New Zealand: developing an interoperable data service for GeoNet’s time series datasets.   

Elisabetta D'Anastasio, Jonathan B. Hanson, Steve Sherburn, Joshua Groom, and Mark Rattenbury

The GeoNet programme at GNS Science Te Pū Ao (GNS) is the primary agency responsible for collecting, managing, and delivering geohazard data in Aotearoa New Zealand, enabling the monitoring of volcanoes, earthquakes, landslides, and tsunamis. The programme oversees a multi-parametric sensor network along with a diverse array of instrumentation and methodologies to provide both raw and analysis-ready data to its end users. Since its inception in 2002, an "open by default" policy has been the guiding principle of this research and monitoring data infrastructure. 

To enhance the interoperability, accessibility, and usability of GeoNet's data, and in alignment with FAIR data principles, we developed an in-house interdisciplinary solution (Tilde) for storing and accessing time-series datasets managed by the programme. Operating successfully for two years, Tilde has improved the interoperability, usability, and FAIR-ness of GeoNet data. In this presentation, we will outline how Tilde has achieved these improvements, discuss challenges and unresolved questions within the geophysical community, and explore potential future directions for leveraging this open data platform to address CARE principles and indigenous data governance in Aotearoa New Zealand. 

How to cite: D'Anastasio, E., Hanson, J. B., Sherburn, S., Groom, J., and Rattenbury, M.: Integrated and Open geohazard monitoring data in Aotearoa New Zealand: developing an interoperable data service for GeoNet’s time series datasets.  , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2206, https://doi.org/10.5194/egusphere-egu25-2206, 2025.

EGU25-2826 | Orals | ESSI3.1

 Datacubes as enablers for land take quantification in LandSupport Regions 

Giuliano Langella, Piero Manna, and Florindo Antonio Mileti

Land take, a significant driver of land degradation, poses challenges for sustainable land management, particularly in regions under high anthropogenic pressure. Addressing these challenges necessitates robust, data-driven approaches to monitor, quantify, and mitigate land take. This contribution explores the integration of datacube technology within the LandSupport Regions platform, leveraging advances from the European LandSupport project and its extension under the Italian GeoSciences-IR project.

Raster datacubes, structured as multidimensional arrays, enable efficient management and analysis of large-scale spatio-temporal datasets, overcoming traditional file-based limitations. The LandSupport Regions platform utilizes a datacube-based Spatial Decision Support System (S-DSS) to enhance the monitoring of land consumption, land cover, and land use at multiple scales—from municipal to national levels. The system integrates heterogeneous datasets, including satellite imagery (e.g., Copernicus), regional land use maps, and environmental indicators (such as high resolution and multi-temporal imperviousness maps), within a common infrastructure, adhering to the FAIR principles.

A key focus is on land take quantification, supported by high-resolution datacubes capable of tracking land-use changes over time. The platform offers decision-makers a suite of tools for generating actionable indicators, assessing compliance with land-use policies, and proposing mitigation strategies aligned with zero net land take objectives. Moreover, the system’s interoperability and open-access characteristics allow integration of user-defined data and models, fostering innovation and scalability.

The platform’s capabilities are demonstrated through use cases in Italy, where local administrations leverage datacube analytics to refine urban and regional planning. These use cases underscore the role of datacubes in delivering accurate, timely insights for sustainable land management. By aligning regional initiatives with European Green Deal objectives, the LandSupport Regions platform – produced under the GeoSciences-IR project – exemplifies the potential of datacube-enabled S-DSSs to advance environmental governance.

How to cite: Langella, G., Manna, P., and Mileti, F. A.:  Datacubes as enablers for land take quantification in LandSupport Regions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2826, https://doi.org/10.5194/egusphere-egu25-2826, 2025.

EGU25-3362 | Orals | ESSI3.1

Mindat: A crowd-sourced and expert-curated open data ecosystem for mineralogy 

Xiaogang Ma, Jiyin Zhang, and Jolyon Ralph

Over the past three years, we have successfully launched an open data service for Mindat, one of the largest databases focused on mineral species and their global distributions. Our achievements include: 1) a comprehensive review of existing data records, covering the list of data subjects, their characteristics, and inherent biases, 2) the establishment of an open data API (application programming interface) alongside Python and R packages to integrate the API into workflow platforms, and 3) fostering community collaboration on data standards and best practices for open data, such as mineral nomenclature, rock classification, and technical frameworks for applying the FAIR (findable, accessible, interoperable, and reusable) principles. Mindat is both crowd-sourced and expert-curated, and for the past decades it has been proven to be an effective approach to engage both data contributors and users. Mindat has been popular amongst geoscience professionals and the public alike. Through our open data initiatives and community engagement, we have also gathered valuable insights to guide future developments of Mindat open data. In this presentation, we will highlight the current open data capabilities, provide an overview of the review of Mindat's data records, and share our vision for leveraging advanced artificial intelligence technologies to expand and enhance Mindat in the future.

How to cite: Ma, X., Zhang, J., and Ralph, J.: Mindat: A crowd-sourced and expert-curated open data ecosystem for mineralogy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3362, https://doi.org/10.5194/egusphere-egu25-3362, 2025.

EGU25-4337 | Orals | ESSI3.1

BEACON - Accelerating access to multidisciplinary data with Relative Optimized Chunking technology 

Robin Kooyman, Peter Thijsse, Dick Schaap, and Tjerk Krijger

Achieving fast access to analysis-ready data from a large number of multidisciplinary data resources is key for contributing to many of the nowadays societal and scientific challenges via Digital Twins of the Oceans or virtual research environments. However, achieving this kind of performance is a major challenge as original data is often organised in millions of (observation) files which makes it hard to achieve fast responses. Next to this, data from different domains are stored in a large variety of data infrastructures, each with their own data-access mechanisms, which causes researchers to spend much time on trying to access relevant data. In a perfect world, users should be able to retrieve analysis-ready data in a uniform way from different data infrastructures following their selection criteria, including for example spatial or temporal boundaries, parameter types, depth ranges and other filters. 

Therefore, as part of several European projects, MARIS has developed a software system called BEACON with a unique indexing and dynamic chunking system that can, on the fly with high performance, extract specific data based on the user’s request from millions of (observational) data files, containing multiple parameters in diverse units. The system returns one single harmonised file as output, regardless of whether the input contains many different data types or dimensions. In January 2025, BEACON 1.0.0 was made publicly available as an open-source software, allowing everyone to set-up their own BEACON node to enhance the access to their data or use existing BEACON nodes from well-known data infrastructures such as Euro-Argo or the World Ocean Database for fast and easy access to harmonized data subsets. More technical details, example applications and general information on BEACON can be found on the website https://beacon.maris.nl/.

The presentation would focus on one of the core features of BEACON called “Relative Optimized Chunking (ROC)”, which is a unique dynamic chunking technology that has been developed specifically to make the data retrieval as fast as possible. This optimized way of chunking reduces the number of chunks BEACON has to search through when a data request has been made. This is done by applying variable sized chunking on multiple levels at the same time such as geo-location, depth and time, which means that data that is relatively close to each other is chunked accordingly. This enhances the speed because it allows BEACON to traverse the millions of datasets using its index with much more precision by not only finding the relevant datasets, but also the exact data blocks containing the relevant data.

The demonstration will involve the use of an existing BEACON node in the field of marine science to access data subsets via its REST API and demonstrate its performance. This will be done in a Jupyter Notebook by querying data via a JSON request to the BEACON system. By going through the Notebook, it will be explained how the BEACON system can be accessed and used by developers including the most recent developments.

How to cite: Kooyman, R., Thijsse, P., Schaap, D., and Krijger, T.: BEACON - Accelerating access to multidisciplinary data with Relative Optimized Chunking technology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4337, https://doi.org/10.5194/egusphere-egu25-4337, 2025.

EGU25-4843 | Posters on site | ESSI3.1

Aviation Safety Datacubes 

Peter Baumann, Colin Price, Vlad Merticariu, Bang Pham Huu, and Dimitar Misev

Air traffic today is immense, with large numbers of humans and goods transported routinely, but also in search and rescue missions, military contexts, and hobby piloting, etc. Still, aviation today is safer than it ever has been, thanks to advanced technology and procedures which are continuously revisited and, where necessary, improved. Of central importance for planning and conducting flights is the atmospheric condition the aircraft is flying in, represented by various relevant weather parameters. Hence, these are continuously monitored.

In the Cube4EnvSec project, a federated datacube demonstrator has been established which illustrates ad-hoc assessment of atmospheric conditions relevant for aircraft. Data stem from two sources, DWD and WWLLN. 

From German Weather Service (Deutscher Wetterdienst, DWD), WAWFOR (World Aviation Weather Forecast) data are obtained, a digital aviation meteorological dataset based on the ICON6_Nest model in support of air traffic management based on geostationary weather satellites. Components currently used are wind speed, icing parameters, Cumulonimbus tops, temperature, tropopause, turbulence, lightning, precipitation radar, volcanic ash, and dust. Updates are provided every 6 hours, temporal resolution is 1 hour with a forecast window of currently 78 hours. The update batches are harvested from DWD and merged into the respective datacubes, extending it by 6 hours further into the future. The 6 hours not overwritten by the new forecast are retained and create a growing "long tail" of historical weather data, currently about 17,000 timeslices. Some datacubes are 3D x/y/t, most however are 4D x/y/h/t with a spatial resolution of 0.0625° x 0.0625° (approximately 6.5km x 6.5km), altitude between ground and 18,000 feet (FL180).

Lightning data are obtained from the World Wide Lightning Location Network (WWLLN) by the Colin Price research group at Tel Aviv University and aggregated into a 3D x/y/t timeseries of lightning strikes observed. Spatial resolution is 0.1°, temporal resolution is 1 hour.

Altogether, the datacubes have a footprint of currently about 20 TB. APIs offered by the Aviation Safety service include the adopted standards WMS, WMTS, WCS, and WCPS, and additionally the OGC drafts OAPI-Coverages and GeoDataCube. Any client conforming to these APIs can be utilized; in the demonstration the rasdaman dashboard will be used which is configurable for manifold datacube interaction techniques (see Figure at bottom).

The demonstration presented includes the following steps:

  • general overview of the datacubes offered by the service;
  • visualization of the combined forecast/history weather datacubes;
  • information relevant for pilot flight planning: weather hazards overview; severe weather conditions along historic routes;
  • same for ad-hoc chosen flight paths, with 4D corridor cutout;
  • various analytical queries related to flight weather conditions.

Most parts of this demo are publicly accessible under https://cube4envsec.org/aviation-dashboard . Any standard Web browser can access it without any plugin etc. to be installed.

Acknowledgement
Cube4EnvSec has received funding by the NATO Science for Peace and Security (SPS) program.

 

 

 


Fig.: Aviation Safety datacube dashboard

 

How to cite: Baumann, P., Price, C., Merticariu, V., Pham Huu, B., and Misev, D.: Aviation Safety Datacubes, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4843, https://doi.org/10.5194/egusphere-egu25-4843, 2025.

EGU25-6020 | Orals | ESSI3.1

AI and Datacubes - a Happy Marriage? 

Dimitar Mishev and Peter Baumann


Datacubes are an accepted cornerstone for Analysis-Ready Data (ARD). One analysis technique of skyrocketing importance today is AI, and this begs the question: how to generalize evaluation of pre-trained models on datacubes?

From a theoretial viewpoint, the connection is immediate: datacubes mathematically resemble tensors, and EO models evaluate tensors. In practice, though, the situation is less straightforward as our experiments with different models have shown. A main issue is the variety and the lack of standardized interfaces of ML models: different input data are processed, data need model-specific preprocessing, and several more. In our research towards offering ML-on-datacubes as a commodity in a federated datacube infrastructure we have collected challenges and methods for presentation.

In our demo, we present AI-Cubes as an enabling concept uniting AI and datacubes. The demos will approach the theme from two sides:

- AI support for datacube query writing: We have trained a chatbot to explain and assist with datacube queries in the OGC/ISO/INSPIRE WCPS standard. This can act as a productiity-enhancing tool for both expert and non-expert users. We demonstrate live how specific questions get answered, such as phrasing NDVI on Sentinel-2 data.

- AI model evaluation on datacubes: particularly attractive is that datacubes allow simple navigation to any area, any time, and even from federated services. This we demonstrate live.

We also highlight challenges coming with this simple data access: models do not convey the same performance anywhere, anytime. This has led to new research on "model fencing", ie: attempting to restrict model application to situations where they exhibit sufficient accuracy. We present first ideas of this research.

Altogether, we cast light on the combination of datacubes and AI from a service and infrastructure perspective. 

How to cite: Mishev, D. and Baumann, P.: AI and Datacubes - a Happy Marriage?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6020, https://doi.org/10.5194/egusphere-egu25-6020, 2025.

EGU25-7038 | Posters on site | ESSI3.1

Establishing Institutional Workflows to Engage Stakeholder Groups in PID Metadata Maintenance 

Emanuel Soeding, Dorothee Kottmeier, Andrea Poersch, Stanislav Malinovschii, and Sören Lorenz

At the Helmholtz Association, we strive to establish a well-structured, harmonized data space that seamlessly connects information across distributed data infrastructures. Achieving this goal requires the standardization of dataset descriptions through consistent metadata practices, such as leveraging persistent identifier (PID) metadata, to ensure interoperability and machine actionability.

While developing concepts to harmonize PID metadata is a crucial first step in creating a unified data space, it is not sufficient on its own. The practical application of PIDs to facilitate the compilation of rich, relevant metadata for datasets necessitates knowledge, training, support, and cooperation among diverse stakeholder groups, each responsible for different aspects of the information lifecycle.

For example, ORCID is a PID system designed to identify individuals contributing to research. Traditionally, this has primarily applied to scientists publishing journal articles. However, in the context of research data, other stakeholders also play vital roles. These include technicians operating instrumentation, data management personnel curating research data and repositories, and administrative staff maintaining institutional data relevant to research. Currently, these stakeholders are often unaware of their potential roles in data management, and the information they collect is typically not harmonized. To address this, workflows must be implemented to manage, structure, and connect the information they produce to research data where appropriate. In the case of ORCID, these workflows should begin at the earliest stages of the research process, such as during employee onboarding.

PIDINST, a PID system introduced by an RDA working group, provides a simple metadata schema to collect essential information about instruments and registers them with unique IDs. These IDs are invaluable for identifying measurements conducted with the same or similar devices. Therefore, we strongly recommend the adoption of PIDInst within the Helmholtz Association. For PIDINST, successful implementation would involve integrating the workflow into existing processes, starting with the acquisition of an instrument or sensor at the research center. Relevant information would then be passed to technicians responsible for maintaining up-to-date records. For researchers, PIDINST provides reliable identification of devices used in scientific processes.

In this presentation, we highlight critical positions within the centers where minor adjustments to established workflows can significantly support the registration of specific PIDs and the engagement of stakeholder groups. We also explore strategies for implementing these changes across the Helmholtz Association. Furthermore, we assign clear responsibilities for metadata maintenance to appropriate stakeholders. The conclusions drawn from this process aim to redefine roles and responsibilities within our organization, fostering a more integrated and effective approach to data management.

How to cite: Soeding, E., Kottmeier, D., Poersch, A., Malinovschii, S., and Lorenz, S.: Establishing Institutional Workflows to Engage Stakeholder Groups in PID Metadata Maintenance, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7038, https://doi.org/10.5194/egusphere-egu25-7038, 2025.

The World Data System (WDS), a member of the International Science Council, serves a membership of over 150 trusted data repositories and related organizations. It builds on the data sharing legacy of World Data Centers that were initiated seven decades ago. Governed by a Scientific Committee, the WDS consists of an International Program Office (WDS-IPO) based in Oak Ridge, Tennessee, USA, and an International Technology Office (WDS-ITO) based in Victoria, BC, Canada. The WDS mission is to enhance the capabilities, impact, and sustainability of our member data repositories and data services. In this presentation, we outline the 2025 to 2027 Action Plan objectives, highlighting activities and collaborations that are underway or planned to progress open science, integrated data infrastructures and FAIR/CARE/TRUST Principles. 

How to cite: Jenkyns, R.: Advancing Open Science through Trusted Data Repository Intersections at the World Data System, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7551, https://doi.org/10.5194/egusphere-egu25-7551, 2025.

EGU25-7959 | Orals | ESSI3.1

Breaking down data sharing barriers and uplifting FAIR for climate data at scale 

Clare Richards, Kelsey Druken, Romain Beucher, and Felicity Chun

Developing climate models often requires the ability to access and share extremely large datasets (spanning tens to hundreds of terabytes) that are discoverable and optimised for high-performance computing (HPC) applications. This is a major challenge, as researchers frequently lack the storage resources and specialised support needed to ensure efficient data management and sharing practices across the full data life-cycle. The challenges of sharing data are evident even when dealing with curated datasets that are prepared for broad access, citation, and reuse. However, these challenges are amplified during the rapid and iterative stages of model development and prototyping. At this point, multiple versions of datasets must be shared and evaluated by a wide range of experts before the data is finalised and curated for public use. This iterative process requires robust infrastructure and coordination to avoid bottlenecks that can hinder progress.

To help overcome these barriers, Australia’s Climate Simulator (ACCESS-NRI) provides a dedicated merit allocation scheme for compute and storage resources. This includes 3PB of storage for datasets that are for use by community members to undertake scientific research, support model development and/or will be shared for reuse. Experience has shown, that if the usage of these storage resources is not managed then the data can quickly go from being an asset to a liability.  Therefore, to maximise the value of both the data and investment in storage, ACCESS-NRI has developed an approach for sharing datasets that is designed to support science and innovation while enhancing the current practices for making data more accessible and usable in accordance with the FAIR and CARE principles.

We will present the motivating use cases and show how this approach supports the model development cycle while making data and the science it underpins more transparent, open and accessible. This approach encourages data generators to transition their datasets from unmanaged, undocumented spaces into managed environments where curation and oversight are aligned with the data’s intended purpose and use. It acknowledges that supporting FAIR principles does not always require full curation to the standards of a long-term publication. Instead, it focuses on reducing barriers to data sharing by promoting active data management practices. These practices enhance discoverability, trust, and reliability, ensuring that shared data is fit for purpose without imposing unnecessary burdens.

ACCESS-NRI is a national research infrastructure (NRI) established to support the Australian Community Climate and Earth System Simulator, or ACCESS. The ACCESS suite of software and data outputs are essential tools used to simulate past and future climate, weather and Earth systems and to support research and decision making within Australia.

ACCESS-NRI's mission is to build an open collaborative infrastructure that will accelerate research in Earth system, climate and weather modelling as well as enable new research not currently possible. The facility brings together skills in software development, high-performance computing, data management and analysis to enhance the ACCESS modelling framework, making it easier to use and lowering the barrier for scientific innovation.

How to cite: Richards, C., Druken, K., Beucher, R., and Chun, F.: Breaking down data sharing barriers and uplifting FAIR for climate data at scale, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7959, https://doi.org/10.5194/egusphere-egu25-7959, 2025.

EGU25-9104 | Orals | ESSI3.1

Using Data Cubes to Investigate Links Between Lightning and Civil Aviation 

Colin Price, Aviv Shay, and Peter Baumann

Lightning is a hazard for many sectors and industries, including the power utility sector, wind turbines, forest management, and civil aviation.  Commercial aircraft are struck by lightning approximately once every year, but most airlines try to avoid thunderstorms if possible by rerouting around these turbulent and electrified storms.  However, such diversions can delay flights, add costs to fuel demands, while increasing greenhouse gas emissions for the aircraft company.  In this study using data cubes we have combined lightning data from the World Wide Lightning Location Network (WWLLN) together with civil aviation flight data from FlightRadar24 to better understand the risks of lightning to civil aviation.  Combining historic lightning and aviation data we can address questions about risks to aircraft from thunderstorms, the frequency of close encounters with thunderstorms, and the frequency of rerouted flights due to thunderstorm activities.  The emerging concept of Analysis-Ready Data (ARD) attempts to find concepts and methods towards services operating on homogenized data. For spatio-temporal data, datacubes are an accepted cornerstone for ARD providing Big Geo Data easier for users and applications, ready for analysis, visualization, fusion, etc. As part of the Cube4EnvSec NATO Science for Peace and Security (SPS) project we will present live demos of our data cube tools and services related to lightning risks for civil aviation over Europe.  Derived analytics from the datacube will also be presented.

How to cite: Price, C., Shay, A., and Baumann, P.: Using Data Cubes to Investigate Links Between Lightning and Civil Aviation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9104, https://doi.org/10.5194/egusphere-egu25-9104, 2025.

EGU25-10578 | Orals | ESSI3.1

DynAWI Extreme Weather Toolbox: an online platform for agricultural risk assessment and decision support 

Arno de Kock, Timm Waldau, Pedro Batista, Peter Baumann, Thorsten Behrens, Peter Fiener, Jens Foeller, Markus Moeller, Ingrid Noehles, Karsten Schmidt, and Burkhard Golla

The DynAWI Extreme Weather Toolbox represents an innovative approach to addressing climate-related challenges in agriculture. This publicly accessible web application offers three primary functions: a historical agricultural weather indicator atlas, a dynamic configurator for calculating user-specified weather indexes, and a forecast model for predicting reduced yields or complete crop failure due to weather extremes. The web application can perform real-time analyses based on multi-dimensional spatio-temporal data.

The technical implementation is based on a client-server architecture, utilizing a scalable geodata infrastructure and an array database management system rasdaman, enabling efficient processing of multidimensional geodata. The system allows real-time analysis of extreme weather events, such as droughts, heatwaves, and heavy rainfall, dating back to 1995. The toolbox aims to provide stakeholders—from farmers to policymakers—with a comprehensive platform for weather-related risk assessment and decision support in agriculture.

In a live demonstration, we will showcase the platform's key features, emphasizing its interactive capabilities and extensive parameter customization options.

How to cite: de Kock, A., Waldau, T., Batista, P., Baumann, P., Behrens, T., Fiener, P., Foeller, J., Moeller, M., Noehles, I., Schmidt, K., and Golla, B.: DynAWI Extreme Weather Toolbox: an online platform for agricultural risk assessment and decision support, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10578, https://doi.org/10.5194/egusphere-egu25-10578, 2025.

EGU25-10583 | Orals | ESSI3.1

EOEPCA+: a method for an open-sourced EO Exploitation Platform Common Architecture 

Richard Conway, James Hinton, Chandra Taposeea, Claudio Iacopino, Salvatore Pinto, and Simon Hunter

The ‘Exploitation Platform’ concept derives from the need to access and process an ever-growing volume of data. Many web-based platforms have emerged - offering access to a wealth of satellite Earth Observation (EO) data, increasingly collocated with cloud computing resources and applications for exploiting the data. Rather than downloads, the exploitation platform offers a cloud environment with EO data access and associated compute and tools facilitating the analysis and processing of large data volumes. Users benefit from the scalability & performance of the cloud infrastructure, the added-value services offered by the platform – and avoid the need to maintain their own hardware. Data hosted in the cloud infrastructure reaches a wider audience and Infrastructure Providers gain an increased cloud user base.

Users are beginning to appreciate the advantages of exploitation platforms. However, the market now offers a plethora of platforms with various added value services and data access capabilities. This ever-increasing offer is intimidating and confusing for most users, often facing challenges such as inconsistent interfaces, proprietary software and limited interoperability. To fully exploit the potential of these complementary platform resources, interoperation amongst the platforms is needed, such that users of one platform may consume the services of another directly platform-to-platform.

EOEPCA (EO Exploitation Platform Common Architecture) is a European Space Agency (ESA) funded project with the goal to define and agree a re-usable exploitation platform architecture using standard interfaces to encourage interoperation and federation between operational exploitation platforms - facilitating easier access and more efficient exploitation of the rapidly growing body of EO and other data. Interoperability through open standards is a key guiding force for the Common Architecture. EOEPCA adheres to standards from organisations such as Open Geospatial Consortium (OGC) and follows best practices in data management, including implementation of OGC Web Services and emerging OGC API specifications for features, coverages and processes. Platform developers are more likely to invest their efforts in standard implementations that have wide usage; off-the-shelf clients and software are more likely to be found for standards-based solutions.

The EOEPCA system architecture is designed to meet defined use cases for various user levels(expert application developers to data analysts and end users). The architecture is defined as a set of Building Blocks (BBs), exposing well-defined open-standard interfaces. These include Identity and Access Management, Resource Discovery, Data Access, Processing Workflows, Data Cube Access, Machine Learning Operations, and more. Each of these BBs are containerized for Kubernetes deployment, providing an infrastructure-agnostic deployment target.

The exploitation platform is conceived as a ‘virtual work environment’,  withusers accessing data, developing algorithms, conducting analysis and sharing value-adding outcomes. The EOEPCA architecture facilitates this through a Workspace BB, providing collaboration environments for groups of users, including dedicated storage and services for analysis, processing and publishing of added-value data and applications. This is supported by an Application Hub BB, providing interactive web-tooling for analysis, algorithm development, data exploitation and providing a web dashboard capability, whereadded-value outcomes are showcased.

Our presentation will highlight the generalised architecture, standards, best practice and open source software components available.

How to cite: Conway, R., Hinton, J., Taposeea, C., Iacopino, C., Pinto, S., and Hunter, S.: EOEPCA+: a method for an open-sourced EO Exploitation Platform Common Architecture, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10583, https://doi.org/10.5194/egusphere-egu25-10583, 2025.

EGU25-10638 | Orals | ESSI3.1

Climate Science Meets Data Spaces: FAIR Digital Objects as a Gateway to Interdisciplinary Science 

Ivonne Anders, Beate Krüss, Marco Kulüke, Karsten Peters-von Gehlen, Hannes Thiemann, and Heinrich Widmann

In recent years, the concept of data spaces has gained prominence, particularly in industry, as a framework for organizing and sharing data across business ecosystems and institutional and disciplinary boundaries. While the term itself is not yet widely adopted in the scientific community , it can be directly applied to research. Data spaces provide a  structured environment for integrating data sets from diversedisciplines, methods or fieldsand making themaccessible for collaboration and analysis. Climate and climate impact research, which relies on data from different fields such as meteorology, hydrology or socio-economics, is in a unique position to benefit from the application from this approach.

In line with the principles of open science, researchers are increasingly adopting frameworks that promote transparency, accessibility and reproducibility. FAIR Digital Objects (FDOs) offer effective means of achieving these goals while also enabling interactions between different data spaces. As standardized, interoperable, and machine-readable entities, FDOs link data, metadata and software, simplifying integration and promoting reuse across disciplines.

Using an example from climate research, we demonstrate how climate model data from an institutional data space, observational data from field campaigns, and satellite data (e.g., from the Destination Earth Data Lake) can be combined. By employing STAC (Spatio Temporal Asset Catalog) catalogs defined as FAIR Digital Objects facilitating the European Open Science Cloud (EOSC) Data Type Registry, we address a specific interdisciplinary research question. This approach not only illustrates the practical application of FDOs but also highlights how they can provide a robust framework for tackling larger and more complex scientific challenges by streamlining workflows and enabling collaboration across disciplinary and institutional boundaries.

How to cite: Anders, I., Krüss, B., Kulüke, M., Peters-von Gehlen, K., Thiemann, H., and Widmann, H.: Climate Science Meets Data Spaces: FAIR Digital Objects as a Gateway to Interdisciplinary Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10638, https://doi.org/10.5194/egusphere-egu25-10638, 2025.

EGU25-15095 | Orals | ESSI3.1

Advancing Open, FAIR, and Responsible Science through the International Generic Sample Number 

Rorie Edmunds, Jens Klump, Kirsten Elger, Lesley Wyborn, Kerstin Lehnert, Lindsay Powers, and Fabian Kohlmann

Research to address global environmental and societal challenges increasingly depends on the availability of large-scale, multidisciplinary datasets, making the need for robust systems that ensure data discoverability, accessibility, and interoperability evermore critical. However, having the data is not enough, one also needs to know about—and understand the connections among—related outputs and entities that support the veracity and reproducibility of the research.

The International Generic Sample Number (IGSN) is a persistent identifier (PID) for material samples arising from any research discipline. Originally developed in the Earth Sciences, the IGSN provides a vital component in solving the abovementioned challenges, enabling seamless integration of sample data across diverse platforms, disciplines, and organizational and geographic boundaries. By uniquely and permanently linking samples to their descriptions (provided as structured metadata), analytical results, and associated publications, IGSNs facilitate transparency, traceability, and reusability of material samples in line with the FAIR and CARE Principles. This is underpinned by the proven interoperability of the IGSN with the scientific communication infrastructure, which also enables citations of samples in the literature to be automatically captured.

This presentation will showcase the collaborative efforts of the IGSN Organization (IGSN e.V.) and DataCite to establish a resilient, cross-disciplinary, globally harmonized PID system for material samples. Use cases will illustrate how IGSNs enhance research workflows, enabling researchers to be more effective and attributed. We will also discuss governance, technical standards, and best practices that promote trust in the IGSN-DataCite partnership and scalability of sample PID adoption, aligning with UNESCO’s Open Science recommendations.

How to cite: Edmunds, R., Klump, J., Elger, K., Wyborn, L., Lehnert, K., Powers, L., and Kohlmann, F.: Advancing Open, FAIR, and Responsible Science through the International Generic Sample Number, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15095, https://doi.org/10.5194/egusphere-egu25-15095, 2025.

EGU25-16668 | Posters on site | ESSI3.1

The Visual Drill Core Library: A Tool for Improving Access to Samples from the Natural Science Institute of Iceland 

María Helga Guðmundsdóttir, Kjartan Birgisson, Hrafnkell Hannesson, Kristján Jónasson, Anette Th. Meier, Birgir Vilhelm Óskarsson, and Björn Darri Sigurðsson

The Drill Core Library (DCL) of the Natural Science Institute of Iceland is Iceland’s national repository for drill cores and cuttings. As such, the DCL is responsible for preserving these geological samples and making them available to the scientific community. The library comprises an estimated 100 km of core and a 470 km equivalent of cuttings from over 4,000 boreholes, as well as a growing database of analytical results. The collection spans Iceland’s range of diverse geological environments and houses core from significant research projects including the SUSTAIN drilling project in Surtsey, sponsored in part by the International Continental Scientific Drilling Program, and the Iceland Research Drilling Project. The DCL’s drill cores and cuttings are available for study and sampling for research purposes, and DCL staff are available for consultation and assistance in identifying and collecting suitable samples. The DCL’s on-site facilities are maintained in collaboration with the University of Iceland’s Research Centre in Breiðdalsvík, East Iceland.

An emphasis has been placed on developing digital infrastructure to improve access to the collections for the scientific community. To facilitate sample identification, an online map-based interface and WFS service have been created where the collection can be examined and contextualized with geological data. The database of the DCL has also been partly integrated into the European Plate Observing System (EPOS), a collaborative initiative enabling FAIR (Findable, Accessible, Interoperable, and Reusable) and open access to geoscientific data from across Europe.

The latest advance in digital access is the ongoing population of the DCL database with core photographs. These are linked directly to the WFS and map viewer, forming a “visual library” that enables direct examination of the library collections, thereby facilitating identification of sampling targets by researchers around the world. At present, 16% of the drill core collection has already been photographed, with 50% set as a target for the end of 2025. Further development of the interface will be carried out in consultation with users of the DCL collections, and cores of interest to researchers are prioritized for photography.

How to cite: Guðmundsdóttir, M. H., Birgisson, K., Hannesson, H., Jónasson, K., Meier, A. Th., Óskarsson, B. V., and Sigurðsson, B. D.: The Visual Drill Core Library: A Tool for Improving Access to Samples from the Natural Science Institute of Iceland, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16668, https://doi.org/10.5194/egusphere-egu25-16668, 2025.

EGU25-16708 | ECS | Orals | ESSI3.1

Scalable and Interoperable Datacube Framework for Advanced Geospatial Data Analysis 

Chen-Yu Hao, Jo-Yu Chang, I-Liang Shih, and Ya-Chu Change

This study introduces a scalable and integrated datacube framework for efficient geospatial data processing and analysis. Leveraging the advanced cloud infrastructure of the National Center for High-Performance Computing (NCHC), the framework combines the openEO API and OGC services to address challenges in managing multidimensional datasets. By ensuring interoperability, security, and high-performance computing, the framework provides a reliable solution for researchers and practitioners to tackle complex geospatial challenges.

Framework Architecture

The framework architecture integrates advanced tools and services, focusing primarily on the openEO API and OGC standard services (e.g., Web Coverage Service and Web Coverage Processing Service). The openEO API provides a unified interface supporting multiple programming languages, allowing users to design and execute customized workflows and enabling batch processing.

openEO integration
The openEO API plays a central role in the framework, performing the following functions:

  • Unified Data Access and Processing Interface: openEO offers a standardized access and processing layer for Earth observation data, abstracting underlying complexities and enabling users to uniformly access multidimensional data from various sources, such as satellite imagery and terrain datasets.
  • Process Graphs and User-Defined Processes: openEO supports User-Defined Processes and Process Graphs, enabling users to create tailored data processing pipelines based on specific analytical requirements. This is particularly valuable for advanced analyses like temporal change detection or spatial statistics.
  • Seamless Integration with OGC Services: openEO works seamlessly with OGC services (e.g., WCS and WCPS) in the framework, enhancing its ability to handle multi-source data. While openEO provides high-level data access and analytical capabilities, OGC services ensure interoperability and standardization at the data layer.

API Proxy Architecture Design

The API proxy is a critical component of the framework, bridging the openEO API and the backend infrastructure to ensure efficient, secure, and stable interactions between users and the system. Its main functions include authentication, authorization management, traffic control, and caching. With the API proxy, openEO can provide a simplified user experience while ensuring optimal utilization of backend data and resources.

Application Scenarios

1. Terrain Analysis
By transforming digital terrain models (DTMs) into multidimensional structures, the framework significantly improves the processing speed and accuracy of large-scale datasets. openEO’s role includes providing a unified interface for data access, enabling users to quickly retrieve and process data for custom slope calculations, visibility analyses, and more. Simultaneously, API proxy security layers ensure strict management of data access and usage.

2. Temporal Analysis Using Landsat Imagery
Temporal analysis of Landsat imagery involves handling large volumes of time-series data. Here, openEO acts as the analytical hub, allowing users to submit analysis requests (e.g., calculating the Normalized Difference Water Index (NDWI)) via the API. The framework then automatically invokes OGC services for data processing and result generation.

Conclusion

The proposed datacube framework successfully integrates openEO API and OGC services, offering a scalable, interoperable, and high-performance solution. As a unified data access and analytical interface, openEO provides flexible and robust tools that significantly simplify complex data processing workflows. By lowering technical barriers and enhancing analytical accessibility, the framework delivers unprecedented convenience for geospatial data analysis, making it a key tool in research and decision-making processes.

How to cite: Hao, C.-Y., Chang, J.-Y., Shih, I.-L., and Change, Y.-C.: Scalable and Interoperable Datacube Framework for Advanced Geospatial Data Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16708, https://doi.org/10.5194/egusphere-egu25-16708, 2025.

EGU25-18594 | Orals | ESSI3.1

FAIRification of sensor-based time-series data – a demonstration of the Helmholtz DataHub digital ecosystem  

Benjamin Louisot, Roland Koppe, Robin Heß, Ulrich Loup, Jürgen Sorg, Marc Adolf, Claas Faber, Andreas Lehmann, Nils Brickmann, Marc Hanisch, David Schäfer, Linda Baldewein, Ulrike Kleeberg, Marie Ryan, Sabine Barthlott, Christof Lorenz, Florian Obersteiner, and Hylke van der Schaaf

In environmental sciences, observational data remains indispensable for monitoring and understanding of natural processes, validating earth system models and remote sensing products, and training of data driven methods. However, unified standards and interfaces for ensuring that such data is consistently available, usable, and compliant with FAIR and Open Science principles are still lacking.

The so-called DataHub initiative of the Helmholtz Research Field Earth & Environment, involving seven large environmental research Centers across Germany, addresses this gap by collaboratively developing a cohesive and unified research data space, including consistent data formats, metadata standards, tools, interfaces and services.

Since the beginning of the DataHub, we have been particularly focusing on unifying time-series data from environmental sensor systems, which are operated across all participating Centers. In this context, we have developed a digital ecosystem, that enhances and links existing and established research data infrastructures with well-defined interfaces and metadata standards. This ecosystem now covers the full processing chain from the integration of new sensor systems and their metadata over automatic and manual quality checks and flagging schemes to the visualization via dashboards and data portals or the usage in data analysis frameworks. In particular, our framework consists of multiple independent tools and services like the Sensor Management System (Brinckmann et al., 2024) as dedicated system for managing sensor metadata, the System for Automated Quality Control (SaQC, Schäfer et al. 2024) as common framework for QA/QC, a tailored metadata profile which adapts the SensorThings API (STA) from the Open Geospatial Consortium to common requirements from environmental sciences (Lorenz et al. 2024), the Earth Data Portal (https://earth-data.de) as overarching data portal and visualization suite as well as tools and services that link all these different building blocks.

While the first concepts of this ecosystem were based on temporary tools and interfaces, we have now reached a level of maturity, that allows us to confidently scale our solutions to new communities and user groups. In this presentation, we will hence give a brief overview of our ecosystem as well as the integrated tools and services. The main focus will be on a hands-on demonstration of the full workflow from deploying a new sensor system, the integration into the contributing services, the (meta)data provision via STA as well as the integration in different downstream systems like the Earth Data Portal for data visualization.

By this, we want to promote the potential of a decentralized research data infrastructure, that has been developed and adopted across multiple research Centers and reach out for new communities and user groups for ultimately creating a FAIR and inter-institutional open data space for our environmental sciences.

How to cite: Louisot, B., Koppe, R., Heß, R., Loup, U., Sorg, J., Adolf, M., Faber, C., Lehmann, A., Brickmann, N., Hanisch, M., Schäfer, D., Baldewein, L., Kleeberg, U., Ryan, M., Barthlott, S., Lorenz, C., Obersteiner, F., and van der Schaaf, H.: FAIRification of sensor-based time-series data – a demonstration of the Helmholtz DataHub digital ecosystem , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18594, https://doi.org/10.5194/egusphere-egu25-18594, 2025.

EGU25-18601 | ECS | Orals | ESSI3.1

OceanUCA: Technological innovation for the management and communication of coastal data in Andalusia through numerical modelling and open source technology. 

Jerimar Vasquez Rojas, Juan Carbone, Alfredo Izquierdo González, Javier Benavente González, Jesús Gómez Enri, Tomás Fernández -Montblanc, Flavio Martins, William Cabos Narvaez, Carlos Yagüe, Carlos Román-Cascón, Oscar Álvarez, Caio Fonteles, Bruno Marques, and Francisco Campuzano

The main objective of the OceanUCA project is the modernization of the technological infrastructure of the University of Cadiz in relation to atmospheric and hydrodynamic numerical modeling specifically configured to simulate physical processes on the coast of Andalusia (Spain).
The initiative focuses on the improvement of modeling systems (oceanographic and atmospheric) and the modernization of servers, mainly THREDDS and ERDDAP. THREDDS facilitates connectivity between scientific data providers and end users, while ERDDAP simplifies the sharing and visualization of time series data through common formats, graphics and maps. The project aims to optimize access, organization and storage of data, create a complete data bank and standardize protocols.
For the storage of data from numerical models, a file server is acquired that allows the custody of large volumes of information related to simulated physical processes, especially focused on the Andalusian coasts. In the future, this server will also facilitate the storage of data from other sources for further calculation, processing and sampling. This acquisition contributes to centralizing the files, currently distributed across different storage sources, and to improving communication across the THREDDS/ERDDAP servers.
The project includes a web application that presents the models in a user-friendly and interpretable format, especially for the scientific community, through the visualization of images.
The technological infrastructure will allow significant advances by facilitating the download of numerical data and taking advantage of graphical processing and high-performance computing to process large data sets. This approach improves the scalability and resolution of forecasts, making them more accessible to the public. By adopting an open-source framework, the project promotes collaboration and knowledge sharing at national and international scales, empowering both the scientific community and the public to use coastal and atmospheric data for informed decision-making and sustainable resource management.

How to cite: Vasquez Rojas, J., Carbone, J., Izquierdo González, A., Benavente González, J., Gómez Enri, J., Fernández -Montblanc, T., Martins, F., Cabos Narvaez, W., Yagüe, C., Román-Cascón, C., Álvarez, O., Fonteles, C., Marques, B., and Campuzano, F.: OceanUCA: Technological innovation for the management and communication of coastal data in Andalusia through numerical modelling and open source technology., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18601, https://doi.org/10.5194/egusphere-egu25-18601, 2025.

EGU25-19278 | Orals | ESSI3.1

One platform will not solve everything: How FID GEO strengthens Germany’s Open Science Landscape for the geosciences. 

Melanie Lorenz, Kirsten Elger, Inke Achterberg, and Malte Semmler

The Specialized Information Service for Geosciences (FID GEO) is a German Research Foundation (DFG)-funded initiative that has been serving the geoscience community in Germany for almost a decade. FID GEO provides essential publication services through its partner domain repositories GFZ Data Services (for research data and software) and GEO-LEOe-docs (for text publications). Beyond these repositories, FID GEO actively supports the digital transformation and supports researchers in adopting Open Science practices mainly through workshops, publications, conference contributions and active participation in topic-specific meetings. 

Collaboration is a cornerstone of FID GEO’s work. It engages closely with geoscientific societies, national infrastructures and initiatives such as the German National Research Data Infrastructure (NFDI), while also contributing to policy-making processes such as the planned German Research Data Act. Recognizing the inherently global nature of geosciences, FID GEO further aligns its activities with international developments, striving to synchronize national progress with global standards and best practices for data management and distribution. FID GEO acts as an interface between scientists, libraries, repositories and the world of digital data management and thus support the transformation of the publication culture in the geosciences at national and international level.

For many years, FID GEO has received feedback from researchers expressing a strong desire for a ‘single source’ platform to manage and share their increasingly large datasets, publications, and projects. At the same time, researchers often feel overwhelmed by the complexity and number of existing infrastructures. However, not only does a one-size-fits-all solution appear technically out of reach, it also faces issues in scalability and sustainable maintenance. A viable way forward is the widespread implementation of machine-readable (meta)data standards that also enable the connection between distributed data systems. Additional metadata properties enable persistent digital links between different research outputs and the unique identification of authors and institutions through persistent identifiers. Another significant challenge within the research landscape is the often competing nature of infrastructures, driven by limited funding opportunities and overlapping goals. Through its extensive network and active collaborations, FID GEO addresses these challenges by guiding researchers through this complex landscape and demonstrates practical ways to make their scientific outputs visible, reusable, and aligned with the FAIR and Open Science principles.

This presentation will share best practices, lessons learned, and future directions for fostering a collaborative and open research environment. FID GEO envisions a geoscience community empowered by shared data and cooperative infrastructures, better equipped to address pressing global challenges.

How to cite: Lorenz, M., Elger, K., Achterberg, I., and Semmler, M.: One platform will not solve everything: How FID GEO strengthens Germany’s Open Science Landscape for the geosciences., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19278, https://doi.org/10.5194/egusphere-egu25-19278, 2025.

EGU25-19453 | Orals | ESSI3.1

ESGF Next Generation and preparations for CMIP7 

Rhys Evans, David Poulter, Philip Kershaw, Ian Foster, Rachana Ananthakrishnan, Forrest Hoffman, Aparna Radhakrishnan, Stephan Kinderman, Sasha Ames, and Daniel Westwood

The Earth System Grid Federation (ESGF) is the international partnership responsible for the distribution, cataloging and archiving of both the Coupled Model Intercomparison Project (CMIP) and the Coordinated Regional Climate Downscaling Experiment (CORDEX). In operation since 2009, it was the first decentralised climate data repository of its kind, storing and serving many petabytes of data across tens of global and region data centre partners.

Over the last five years, the system has been fully rearchitected, introducing a cloud-ready deployment architecture and a new system for distributed search, fundamental to ESGF’s federated model for data access. This has involved innovations, translating successful experience with the STAC (Spatio-Temporal Asset Catalogue) specification from the EO world and developing a profile for its use with global climate projections data. Providing a STAC interface to ESGF archives has allowed us to explore alternate access methods for cloud-accessible analysis-ready data ready formats through the use of tools such as Kerchunk, a lightweight non-conversion approach for referencing existing data, which works with open-source python packages like fsspec and Xarray. Use of STAC also provides the potential for greater integration between EO and climate modelling domains essential for the validation of model outputs.

ESGF has traditionally used a distributed model for search services which though powerful has led to challenges around consistency of search content. Over the last twelve months, in preparation for CMIP7, a further fundamental innovation has been made in the architecture to address these issues. The new system adopts a centralised model, with two search nodes, one in the US and one in Europe each hosted on public cloud. These two nodes are synchronised together using a new event-driven architecture. This approach, driven by a shared messaging framework between the nodes, ensures eventual-consistency across the nodes, to reduce or eliminate errors caused by individual node down time and simplify processes such as the replication and retraction of data from the archives distributed at sites across the federation.

The move to a message based, event driven architecture has been integrated with STAC records and services. In ESGF-NG data is shared between nodes as messages in the form of STAC Item records, ensuring a consistent, publicly documented archive distributed across many nodes. The ESGF team have contributed several changes to the STAC project to facilitate this change. Looking forward, we see potential in this new event driven architecture for search systems as a means to integrate across federations - in the European context this could include the ESA Climate Change Initiative open data portal, work with the Copernicus Climate Data Store and DestinE.

How to cite: Evans, R., Poulter, D., Kershaw, P., Foster, I., Ananthakrishnan, R., Hoffman, F., Radhakrishnan, A., Kinderman, S., Ames, S., and Westwood, D.: ESGF Next Generation and preparations for CMIP7, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19453, https://doi.org/10.5194/egusphere-egu25-19453, 2025.

EGU25-20533 | Orals | ESSI3.1

A Grass-roots Standard for Time Series Data in any Domain: HAPI 

Jon Vandegriff, Robert Weigel, Jeremy Faden, and Alexander Antunes

We describe a simple interface for accessing time series numeric data: the Heliophysics Application Programmer's Interface (HAPI). Although it started in NASA's Heliophysics domain, no Heliophysics idioms are present in the standard, and HAPI can be used to serve any tabular, numeric data that is indexed by time.  HAPI was the result of a community push to standardize similar access methods at multiple data centers, and it is now in use at 12 data centers around the world, with over 12,000 datasets available in a standard way. HAPI offers a more conceptual view of the data, independent of the storage arrangements at a server. It also is not intended to replace an existing server's API, but to sit alongside that API.  The project is mature, with a reference server available, as well as clients in multiple programming languages.  We will present an overview of the API and compliance with FAIR principles. We also will describe some of the visualization and analysis tools being developed now that standardized access is becoming a reality. We invite discussion with other time series data providers in other domains.

How to cite: Vandegriff, J., Weigel, R., Faden, J., and Antunes, A.: A Grass-roots Standard for Time Series Data in any Domain: HAPI, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20533, https://doi.org/10.5194/egusphere-egu25-20533, 2025.

EGU25-20550 | Orals | ESSI3.1

Are Publicly Funded Data-Infrastructures Reliable? 

Hans Pfeiffenberger and David Carlson

As founders and former chief editors of Earth System Science Data (ESSD), the authors are concerned about the reproducibility and availability of important scientific sources and findings, and about timely access to scientific data and data-related services. We are discussing (1) incidents with the availability of DOIed datasets and their metadata and (2) a recent outage of an important data infrastructure.

Both observations are considered sufficiently serious that the authors wonder why the underlying facts and realities are not discussed widely in this community.

1) The most cited dataset published through ESSD is the series of yearly reports on the Global Carbon Budget, e.g. the latest, https://doi.org/10.5194/essd-2024-519. These articles are cited in scientific publications by the hundreds of times and routinely inform the United Nations climate change conferences (COPs). The first datasets of the series were held and provided DOIs by the Carbon Dioxide Information Analysis Center (CDIAC), which was hosted by the Oak Ridge National Laboratory. When CDIAC was shut down in 2017, the datasets were transferred to a repository at another US National Lab, loosing most of the metadata in the process, most notably authorship. Thankfully, hosting of post-2017 additions to the dataset series has been taken over by the Integrated Carbon Observing System (ICOS) and DOIs to all elements of the series still resolve (albeit, in a sloppy manner for pre-2018 data). One could argue that the most reliable holder of metainformation about this – not just scientifically – important data are not the repositories but ESSD, operated by a commercial publisher, Copernicus. 

2) When tropical storm Helene hit North Carolina, in September 2024, power and internet connectivity went out from the Asheville headquarter site of NOAA’s NCEI, an aggregator, archive and service provider for environmental data. Although NCEI is hosted at four geographically dispersed sites, NCEI data ingest and services came to a halt for several weeks. It appears that most data from the period during and after Helene have been collected retroactively, and services are fully available again. While NOAA’s real-time weather services, important to deal with the emergency, seem to have been available during Helene, one is tempted to ask if they could become interrupted under similar circumstances.

Both these and some other observations – which will be discussed at EGU2025 - create the uncomfortable impression that the huge efforts of this community wrt. the FAIRness of data and in the creation of a multitude of publicly funded infrastructure elements do not achieve to meet today’s needs, and possibly may not meet them tomorrow. If government labs and agencies of a rich nation cannot achieve this – who can?

(Part of this work has been presented before, at a pre-conference workshop to RDA20, Gothenburg, 2023)

How to cite: Pfeiffenberger, H. and Carlson, D.: Are Publicly Funded Data-Infrastructures Reliable?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20550, https://doi.org/10.5194/egusphere-egu25-20550, 2025.

EGU25-21042 | Posters on site | ESSI3.1

Efforts of the Small Bodies Node in Providing “Analysis Ready Data” to Support Open Science 

Eric E. Palmer, Kristina Lopez, and Mike Drum

The Planetary Data System (PDS) provides key structural support for Open Science by meeting the tenants of "Free, unrestricted access1.” Here we will discuss the need to expand our offerings by improving support for the OS tenants of "Ease of use.”

Analysis-ready data (ARD) provides data in formats that, while different than what was provided by the mission team, are orders of magnitude more useful to scientific researchers. 

For NASA Planetary Science Missions, the data is provided to us in stable and long-term formats that are well documented.  However, the data formats for each mission are typically different.  Additionally, many processing steps are not done by the science team for the archived products, such as ortho-rectification, geospatial positioning, or co-alignment with digital terrain models.  Additionally, there is little consensus within Planetary Science for a standard format for almost any data type, for example images that can be in FITS, VICAR, custom IMG formats, or sometimes JPEG.

PDS nodes have begun to host such ARD as either part of the official archive or outside of the archive using the new PDS annexes2.  We have several initiatives to support ARD.  These include the Small Bodies Image Browser and digital terrain models in both ISIS and GeoTiff formats. While generated data in these formats initially requires additional effort, once created they continuously provide value to the data user community.

Analysis-ready data can significantly increase "ease of use" in many different ways.  They typically will be preprocessed, saving data users significant effort that they would have spent learning how to process the data themselves. This preprocessing also lowers the technical barriers and eases the use of complex data sets. In addition to the preprocessing, datasets can be provided in standardized, commonly used data formats that are more useable and accessible than many of the current formats. Streamlining the ARD would greatly ease both researchers' and the public’s ability to use data spanning many different missions in ways that is not currently possible. Focusing on providing the most interoperable and usable data to the community also enables more interdisciplinary collaboration and increases reproducibility — all key goals of Open Science.  

Analysis-ready data in the PDS will be essential to create more open and usable data. As the complexity of planetary mission data increases, ARD can allow the PDS to maximize the scientific return of these valuable datasets.

References:
[1] NASA Science Mission Directorate. (2023). Open-Source Science Guidance, Version 2.1.
[2] Mouginis-Mark, P., Williams, D., Bleacher, J., et al. (2023). Analysis Ready Data (ARD) within the Planetary Data Ecosystem: Benefits for the Science Community. 54th Lunar and Planetary Science Conference.

How to cite: Palmer, E. E., Lopez, K., and Drum, M.: Efforts of the Small Bodies Node in Providing “Analysis Ready Data” to Support Open Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21042, https://doi.org/10.5194/egusphere-egu25-21042, 2025.

EGU25-21367 | Orals | ESSI3.1

AuScope’s Research Data Systems: Operationalising FAIR place-based research through collaboration 

Rebecca Farrington, Lesley Wyborn, Jo Croucher, Anusuriya Devaraju, Alex Hunt, Hannes Hollmann, Jens Klump, Angus Nixon, Alexander Prent, Sara Polanco, Nigel Rees, and Tim Rawling

Addressing global environmental and societal challenges requires robust, interdisciplinary data ecosystems that support collaboration across geographic, cultural, and disciplinary boundaries. AuScope, Australia’s National Research Infrastructure (NRI) provider for the geoscience community, collaboratively tackles interdisciplinary grand challenges such as climate change, natural resource security, and natural hazards. AuScope is funded by the Australian Government’s National Collaborative Research Infrastructure Strategy (NCRIS) and integrates tools, data, analytics, and services across Australian research and government agencies, in particular, partnering with organisations at the forefront of research data systems and infrastructure.

Through collaborations with CSIRO, Geoscience Australia, state and territory geological surveys, universities, and other NCRIS facilities, including the National Computational Infrastructure (NCI), the Terrestrial Ecosystem Research Network (TERN), and the Australian Research Data Commons (ARDC), AuScope is addressing the complexities of modern FAIR data management at scales ranging from small scale local installations to co-located High Performance Compute and Data (HPCD) Platforms. Key AuScope initiatives such as Geophysics 2030 Collections (https://ardc.edu.au/project/2030-geophysics-collections/), AusGeochem (https://ausgeochem.auscope.org.au/), the Modelling Atlas of The Earth (M@TE; https://mate.science), and the AuScope Data Repository (https://repository.data.auscope.org.au/) exemplify how the FAIR principles can be operationalised to support impactful research both within and beyond the geosciences and at multiple scales.

Nationally, AuScope collaborates with other Earth and Environmental Research Infrastructure providers, working to transform Australia’s research capabilities through, for example, Coastal Research Infrastructure (CoastRI) and implementing the National Digital Research Infrastructure Strategy (NDRI). Globally, AuScope contributes to initiatives like OneGeochemistry, the CODATA-led WorldFAIR Plus project, EarthScope (US), EPOS, Geo-INQUIRE, and ChEESE (Europe), ensuring compatibility with international research infrastructures, data standards, and best practices while at the same time, aligning with Australia’s geoscience priorities. 

This presentation will highlight how AuScope is progressively operationalising the FAIR and TRUST principles across its investments by focusing on place-based research to foster interoperability, strategic collaboration, and Open Science practices. By aligning with the CARE principles as well as advancing collaborative data infrastructure, AuScope creates trusted, interoperable data ecosystems that empower researchers to effectively and efficiently address pressing interdisciplinary societal challenges at both a national and international scale.

How to cite: Farrington, R., Wyborn, L., Croucher, J., Devaraju, A., Hunt, A., Hollmann, H., Klump, J., Nixon, A., Prent, A., Polanco, S., Rees, N., and Rawling, T.: AuScope’s Research Data Systems: Operationalising FAIR place-based research through collaboration, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21367, https://doi.org/10.5194/egusphere-egu25-21367, 2025.

EGU25-2295 | Orals | ESSI3.2

Data Lifecycle Management for Field Campaigns: Welcome to the Earth Observing Laboratory Field Catalog and Archive 

Jacquelyn C. Witte and the Data Management and Services Team

The NSF NCAR Earth Observing Laboratory (EOL) has supported over 600 national and international field campaigns which represent half a century of field-based observational science. Our mission is to provide responsive, high quality data services to researchers in field campaigns including pre-field phase planning, real-time decision-making tools, and long-term data curation to support the complete project life cycle. Such support includes (1) serving as the online hub for field campaign operations with access to real-time mission coordination displays and communication tools, (2) ensuring a secure, easily accessible archive of campaign observations, and (3) providing long-term stewardship and curation of observational datasets. All datasets in the EOL’s Field Data Archive are publicly accessible and findable at https://data.eol.ucar.edu/.  

 

EOL data management services are continuously evolving as we pursue FAIR and TRUSTed principles based on industry standards, user feedback and the desire to increase data discovery and accessibility to the broader scientific community. The management of our field campaign data is an iterative, human-driven and agile process. Thus, to address challenges arising from data preparation, preservation, and provenance metadata as the volume and variety of our data grows, EOL has developed tools and workflows that track and maintain the collection of data. In this presentation we will introduce highlights and functionalities of the Field Catalog and the Field Data Archive that together provide end-to-end customized data management services for field campaigns.

How to cite: Witte, J. C. and the Data Management and Services Team: Data Lifecycle Management for Field Campaigns: Welcome to the Earth Observing Laboratory Field Catalog and Archive, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2295, https://doi.org/10.5194/egusphere-egu25-2295, 2025.

EGU25-2803 | Orals | ESSI3.2

GOYAS: A FAIR-by-Design System for Innovative remote-sensing data products 

Fernando Aguilar Gómez, Verónica González-Gambau, Cristina González-Haro, Aina García-Espriu, Eva Flo, Estrella Olmedo, Isabel Caballero, Evgeniia Makarova, Marcos Portabella, Daniel García-Díaz, and Isabel Afán

The Geospatial Open Science Yielding Applications (GOYAS) project, under the umbrella of the Horizon Europe project “OSCARS”, proposes a new approach for open science and open data in remote-sensing, integrating FAIR principles (Findable, Accessible, Interoperable, and Reusable) from the initial design phase. GOYAS provides innovative and/or experimental Earth Observation (EO) data and open science practices to address diverse environmental challenges, delivering advanced geospatial products that are tailored to meet the needs of multiple stakeholders, including researchers, decision-makers, and environmental managers.

GOYAS focuses on generating innovative and accessible remote sensing products for a variety of applications: monitoring water quality parameters, such as turbidity or chlorophyll-a; deriving high-resolution bathymetric maps over coastal regions based on optical instruments; assessing oceanographic variables like sea surface temperature and salinity; improving ocean and atmosphere forecasting capabilities with enhanced sea-surface wind & stress products; and supporting ecosystem monitoring and management in protected areas such as Doñana National Park. These products are generated through the integration of multi-source EO data, including Copernicus Sentinel satellites and complementary datasets, with advanced processing pipelines built on machine learning algorithms and geospatial standards.

A core strength of the GOYAS project lies in its FAIR-by-design system architecture, which prioritizes:

  • Findability: Metadata-rich datasets indexed through open repositories and geospatial catalogues to enhance discoverability.

  • Accessibility: FAIR-compliant platforms with user-friendly interfaces that provide seamless access to data products, ensuring usability across diverse technical expertise levels. GOYAS aims at facilitating the access providing data in common formats and contextualizing them with proper metadata.

  • Interoperability: Adoption of open geospatial standards (e.g., OGC, INSPIRE) to ensure compatibility with existing systems and facilitate data exchange, specially under the context of Research Infrastructure hubs like ENVRI.

  • Reusability: Comprehensive documentation and adherence to open licenses that allow users to adapt and build upon project outputs.

Key innovations include the automated processing of remote-sensing data to extract actionable insights and the application of machine learning to improve the accuracy and reliability of derived parameters. For example, GOYAS employs advanced spectral analysis techniques to calculate shallow bathymetry with sub-meter precision in coastal environments, as well as algorithms for near-real-time detection of water quality anomalies in inland waters.

The system also provides support for the monitoring and management of sensitive ecosystems. In Doñana National Park, GOYAS enables the identification of changes in hydrological regimes or vegetation health through the integration of long-term EO datasets with local ecological studies. Similar applications extend to marine protected areas, where GOYAS aids in monitoring oceanographic dynamics and ecosystem responses to climate change.

This presentation will detail the design, architecture, implementation, and outcomes of the GOYAS project, emphasizing its alignment with FAIR principles and its transformative potential for environmental monitoring. By fostering interoperability and collaboration across disciplines, GOYAS serves as a model for how open science and advanced remote sensing can drive innovation, sustainability, and informed decision-making in geospatial research.

How to cite: Aguilar Gómez, F., González-Gambau, V., González-Haro, C., García-Espriu, A., Flo, E., Olmedo, E., Caballero, I., Makarova, E., Portabella, M., García-Díaz, D., and Afán, I.: GOYAS: A FAIR-by-Design System for Innovative remote-sensing data products, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2803, https://doi.org/10.5194/egusphere-egu25-2803, 2025.

EGU25-4203 | ECS | Orals | ESSI3.2

Escaping from the 1600s: Advancing FAIR scientific knowledge with reborn articles 

Lauren Snyder, Hadi Ghaemi, Ricardo Perez-Alvarez, and Markus Stocker

Text-based literature remains the primary expression of scientific knowledge. Since the first scientific article published in 1665, we have managed the switch from physically printed articles to PDFs, but nothing more. While PDF publications can be easily shared electronically, they remain unstructured text-based documents that machines cannot easily interpret (i.e., they are not machine-reusable). This limits our ability to use digital support tools to efficiently extract and organize knowledge from scientific articles. Rather, to reuse most scientific results (e.g., for synthesis research), we must first extract them from articles and organize them into databases, which is time consuming and prone to error. 

Here, we present reborn articles, which offer a novel approach to producing scientific knowledge. By integrating with programming languages commonly used for data analysis, like R and Python, reborn articles allow researchers to produce scientific results in a machine-reusable format from the outset. This means subsequent data users can download the results of a reborn article as a CSV file with just a click of a button and bypass post-publication data extraction. To support the production, publication, and reuse of reborn article data, we developed ORKG reborn, a FAIR knowledge online infrastructure. 

Using an ecological dataset, we showcase the production of a reborn article, and its impact on knowledge integration and synthesis. Building on the author’s original data analyses conducted in R, we developed an accompanying R script to produce machine-reusable descriptions of the original statistical models that were automatically harvested by ORKG reborn, eliminating manual data entry. We envision that the use of programming languages, like R, to facilitate the production of machine-reusable scientific knowledge could feasibly be streamlined into existing FAIR data management requirements that are already implemented by many academic publishers. Broad adoption of the approach across research communities could transform the way we share and synthesize scientific knowledge. 

How to cite: Snyder, L., Ghaemi, H., Perez-Alvarez, R., and Stocker, M.: Escaping from the 1600s: Advancing FAIR scientific knowledge with reborn articles, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4203, https://doi.org/10.5194/egusphere-egu25-4203, 2025.

EGU25-9401 | Orals | ESSI3.2

Scalable Solutions for Urban Data Spaces: Insight from the USAGE blueprint 

Piotr Zaborowski, Francesca Noardo, Giacomo Martiano, and Danny Vandenbroucke

The USAGE objective is to identify, implement, and demonstrate an architecture and solutions for a data space supporting the European Green Deal priorities. It implements the methodology based on the data USAGE data space framework built around specific use cases in the context of local and European policies and guidelines as well as digitalization agendas. Use cases, considered the primary value proposition for the data uptake, are developed and maintained in the USAGE framework. They cover  urgent municipalities scenarios like heat islands, clean energy, qir quality and mobility. Target requirements are translated into data and service requirements expressed in the ISO catalog-based model tailored to the specific data quality measures for the Decision Ready Information. Implementation of the value chain goes across various data inputs including satelite and airborne images, local sensors and citizen science data, surface and urban models producing intermediary and end user products and services. Disciplined and tool-supported collection of the data and application assets consistent with the INSPIRE-compliant schemas and data requirements model which allows them to leverage the solutions' potential and implement the value proposition for their providers. Profiled models create the frames of the data value chain, documenting processing steps from the data requirements through BPMN data flow models linking to the used and produced assets. In addition, licensing schema, including the constraints model, allows for data sovereignty and trust among the data space actors.

The outcome blueprint for the urban data space goes beyond the USAGE pilots to test scalable solutions based on adopting the proposed set of standards coming mainly from ISO, OGC, W3C, OASC and their extensions. It is built in the European initiatives and legal references (i.e., the European strategy for data, the European interoperability framework, the European interoperability reference architecture), and reviewed several projects and initiatives results contributing to shaping data spaces: Open DEI design principles, the International Data Spaces Association (IDSA) reference architecture, Gaia-X architecture, Data Spaces Business Alliance (DSBA) documents, the Data Spaces Support Centre (DSSC) results, Data Space for Smart and Sustainable Cities and Communities (DS4SSCC) outcomes, and the GREAT project Technical Blueprint. Presentation goes across the best practices and guidances extracted from the implementation of the FAIR dataspace and considerations given defined frameworks.

How to cite: Zaborowski, P., Noardo, F., Martiano, G., and Vandenbroucke, D.: Scalable Solutions for Urban Data Spaces: Insight from the USAGE blueprint, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9401, https://doi.org/10.5194/egusphere-egu25-9401, 2025.

EGU25-9739 | Posters on site | ESSI3.2

Reducing the Pain of Data Discovery in Earth System Science 

Aenne Loehden, Claudia Martens, and Andrea Lammert

Ontologies offer significant potential for advancing Earth System Science (ESS) by improving the discoverability and usability of complex datasets and tools. This poster builds on last year’s comic, which illustrated the foundational benefits of ontologies, and presents the first steps in implementing generic tools from already existing terminology services designed to enhance data findability and data comprehension. These tools enable scientists to easily search for appropriate data and retrieve information about data from specific repositories, thus supporting the FAIR (Findable, Accessible, Interoperable, and Reusable) principles in ESS.

Key aspects of terminologies include the clear and consistent description of scientific terms, their relationships, and the unambiguous identification of terms to prevent inconsistencies. By using terminologies we can ensure that terms are defined in a way that is both standardized and interoperable across different datasets and research communities. Concrete examples will be drawn from the World Data Center for Climate (WDCC), where first steps have been taken to implement generic tools and extend the application of terminologies, and to thus enhance data discoverability and facilitate better searchability of climate-related information.

How to cite: Loehden, A., Martens, C., and Lammert, A.: Reducing the Pain of Data Discovery in Earth System Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9739, https://doi.org/10.5194/egusphere-egu25-9739, 2025.

EGU25-10627 | Orals | ESSI3.2

FAIR EU soil vocabularies: an overview of joint efforts from some EU Soil Mission projects 

Mickaël Beaufils, Paul van Genuchten, Fenny van Egmond, and Kathi Schleidt

Vocabularies or thesauri, lists of terms with their definitions and unique ID, like a dictionary of a language, play a critical role in the domain of soil science, providing a standardized framework for accurately documenting and communicating soil characteristics. In soil science, the use of precise and consistent terminology ensures the effective exchange of data, promoting interoperability among researchers, practitioners, and decision-makers. A well-structured vocabulary, part of soil classification or soil description standards, facilitates the classification of soil properties, such as texture, structure, fertility, and organic content, allowing combining data from different sources but described in a similar way. And thereby enabling reliable comparison and interpretation across different regions and time periods. Furthermore, these vocabularies enable and support the development of standardized databases, soil datasets and soil monitoring systems, which are essential for environmental management, land use planning, and agricultural practices. Inaccurate or ambiguous soil descriptions can lead to misinformed decisions, making the establishment of clear, universally accepted vocabularies crucial for advancing soil science, conservation efforts, and sustainable land management practices. Such practices would greatly enhance the FAIRness of the data being managed, ensuring data conservation over time.

Soil vocabularies come from many sources, some national or regional, some from international organizations such as the Food and Agriculture Organization of the United Nations (FAO) or the International Union of Soil Sciences (IUSS), e.g. World Reference Base for Soil Resources (WRB) or FAO Guidelines on Soil Description. Several initiatives worked on the identification and provision of agreed vocabularies in order to ensure the interoperability of their results at different scales (national, EU, international). This includes work by standard setting organizations (eg. ISO TC190), legislation (eg. EU INSPIRE Directive) and of course numerous collaborative projects, such as SIEUSOIL, EJP SOIL, ISLANDR, SoilWise, SPADES, Soil Mission Support and MARVIC. At present, many existing vocabularies have not been exposed in a referenceable and machine-readable manner, and instead remain “trapped” within PDF documents. Extracting the relevant concepts and exposing them in both human and machine readable forms on persistent URIs would be a valuable step towards soil data harmonization.

The European Mission: A Soil Deal for Europe, with currently about 50 research projects and a network of 100 living-labs and lighthouses, offers an interesting environment and opportunity for the co-creation of a harmonised framework for soil vocabulary description. Due to the diversity of Soil Mission Projects, gaps in existing vocabularies can be identified and experience can be gained in how to best present vocabularies for both data annotation as well as data discovery.

In this presentation we will share the current status on this topic, offering a non-exhaustive yet hopefully informative overview on existing materials (vocabularies and associated technologies to share them), on-going work and key challenges for achieving better soil data interoperability.

This study was made possible through funding from the EU's Horizon Europe program, specifically the ISLANDR and SoilWise projects.

How to cite: Beaufils, M., van Genuchten, P., van Egmond, F., and Schleidt, K.: FAIR EU soil vocabularies: an overview of joint efforts from some EU Soil Mission projects, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10627, https://doi.org/10.5194/egusphere-egu25-10627, 2025.

EGU25-11467 | ECS | Posters on site | ESSI3.2

Advancing FAIR geochemical data: 25 Years of GEOROC database service 

Leander Kallas, Marthe Klöcking, Kirsten Elger, Bärbel Sarbas, Adrian Sturm, Stefan Möller-McNett, Matthias Willbold, and Gerhard Wörner

The GEOROC synthesis database, a pioneering open-access resource for geochemical and isotopic data, marks 25 years of service to the geoscience community. Over its history, GEOROC has compiled data from more than 22,750 publications in the field of geochemistry, and provides free access to over 39 million individual data values, primarily on igneous and metamorphic rocks, minerals and their inclusions. As a cornerstone for interdisciplinary research, GEOROC is complementary to other geochemical synthesis databases like PetDB, AstroMat and GeoReM, in facilitating reuse of data for innovative studies that leverage data analytics and machine-learning approaches across geoscientific disciplines and beyond.

The Digital Geochemical Data Infrastructure (DIGIS) project for GEOROC 2.0 is providing an up-to-date IT infrastructure that aligns GEOROC with the FAIR principles. Data findability and accessibility are ensured through the newly developed API and the improved GEOROC web interface that allows users to retrieve a variety of distinct data products and services, including a fully customizable search functionality. Interoperability is achieved via implementation of a feature-based data model compatible with the OGC Observations and Measurements standard and controlled, machine-readable vocabularies that harmonize geospatial, analytical and sample-related metadata, and enabling seamless integration in multiple databases and portals (e.g., EarthChem). Reusability is further supported by archiving time-stamped GEOROC data products in the DIGIS Data Repository, hosted by GFZ Data Services, where datasets with digital object identifiers (DOIs) are archived for the long-term. Additionally, researchers are encouraged to directly submit new or already “published” datasets to this domain repository—through standardized (meta-)data templates, ensuring high-quality data submissions that facilitate data quality assessment and reuse.

In collaboration with national and global initiatives, such as OneGeochemistry and NFDI4Earth, the DIGIS project further promotes practical approaches to the FAIR principles for geochemistry by developing unified controlled vocabularies for geochemical data and their metadata (e.g., analytical methods, sample description, location). These vocabularies also integrate external standards, such as the International Mineralogical Association’s "List of Minerals" and MinDat’s "Subdivisions of Rocks," alongside newly developed (and published) frameworks for categories such as geological setting and analytical methods (collaboration with EarthChem). By harmonizing metadata across geospatial, analytical and sample-related categories, these efforts ensure consistency, improve data quality assessment and control and enhance interoperability across data systems, including but not limited to GEOROC, PetDB, and AusGeochem. Such advancements expand the potential applications of geochemical data, fostering innovation in fields such as environmental science, remote sensing, archaeology and geohealth.

With 25 years of experience and ongoing innovation through the DIGIS project, the GEOROC database exemplifies how operationalizing the FAIR principles enhances its value as a critical resource for the geoscience community. By providing both FAIR and open data, GEOROC empowers researchers to conduct reproducible, impactful studies and fosters interdisciplinary collaboration, driving innovation and advancing progress across the geosciences.

How to cite: Kallas, L., Klöcking, M., Elger, K., Sarbas, B., Sturm, A., Möller-McNett, S., Willbold, M., and Wörner, G.: Advancing FAIR geochemical data: 25 Years of GEOROC database service, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11467, https://doi.org/10.5194/egusphere-egu25-11467, 2025.

EGU25-14320 | ECS | Orals | ESSI3.2

EarthBank by AuScope: Building FAIR research data infrastructure for the global geochemical community 

Angus Nixon, Bryant Ware, Brent McInnes, Fabian Kohlmann, Moritz Theile, Wayne Noble, Yoann Gréau, Hayden Dalton, Halimulati Ananuer, Malcolm McMillan, and Ashley Savelkouls

The geochemical community increasingly generates and requires large volumes of analytical data from a wide array of acquisition methods, analytical scales, and sample types in order to address broad research applications. Resulting datasets are commonly collected and reported through non-standardised protocols and reporting formats, if indeed standards are applied at all, which inhibits easy sharing of data during collaborative research projects or repurposing of legacy data. Existing repository services do not presently satisfy requirements for Findable, Accessible, Interoperable and Reusable (FAIR) data, and especially contain significant flaws as to the reuse and interoperability of geochemical data. Generalist repositories such as Zenodo or Figshare do not provide consistent data structures or curation, hence data held within these services is highly variable with regard to format, parameters reported and potentially quality. While domain repositories commonly do implement internally consistent data formats and a level of curation, data within repositories is gathered from published sources which may be incomplete or unstructured, and hence often lack the complete information (metadata) required to appropriately describe the data and allow it to be confidently reused. 


To truly unlock the potential of the ever expanding wealth of geochemical data and meet FAIR requirements, improvements to the data infrastructure landscape are clearly required. The AuScope Geochemistry Network (AGN) is an Australian-based collaboration of geoscientists producing bespoke data resources and infrastructure for the international community to capture, normalise, and share geochemical data resources. These resources include best practice data reporting schema and vocabularies for a variety of data types, produced through collaborations with expert advisory groups and, where available, following or expanding on existing international community recommendations. These data resources have been implemented to the EarthBank platform (formerly AusGeochem), an open web service designed by the AGN to capture, share, store and evaluate geochemical data and metadata. Unlike many other services, researchers are able to upload data prior to publication which can assist both in allowing researchers to compare their data with other existing resources prior to submission, but importantly also improves the likelihood of capturing the full data and metadata associated with analyses required for reuse. Once data is uploaded to this service it may be associated with a dataset DOI to support data access requirements for publication, in order to streamline the publication process and provide a domain specific repository for supplemental data. Data models for U/Pb, fission track, (U-Th-Sm)/He, 40Ar/39Ar and inorganic major and trace geochemistry data types are presently implemented within EarthBank, allowing users to freely upload generated research data for these systems, or explore and integrate existing datasets. Best practice templates for upload are openly available through the EarthBank platform, and vocabularies are openly discoverable through the Research Vocabularies Australia (RVA) service. These resources may be used not only to upload data, but also to develop cross-walks for machine-to-machine interoperability with other repository services to build a global FAIR compliant infrastructure required to maximise data access and improve research outcomes.

How to cite: Nixon, A., Ware, B., McInnes, B., Kohlmann, F., Theile, M., Noble, W., Gréau, Y., Dalton, H., Ananuer, H., McMillan, M., and Savelkouls, A.: EarthBank by AuScope: Building FAIR research data infrastructure for the global geochemical community, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14320, https://doi.org/10.5194/egusphere-egu25-14320, 2025.

EGU25-14828 | Orals | ESSI3.2

Uplifting and streamlining FAIR data implementation for Australia’s climate modelling outputs 

Kelsey Druken, Clare Richards, Romain Beucher, Johanna Basevi, Chris Bull, Claire Carouge, Martin Dix, Aidan Heerdegen, Paul Leopardi, Davide Marchegiani, Heidi Nettelbeck, Anton Steketee, Charles Turner, Marc White, and Spencer Wong

Australia’s Climate Simulator (ACCESS-NRI) is a national research infrastructure established to support the Australian Community Climate and Earth System Simulator (ACCESS) modelling system. Since its launch in 2022, ACCESS-NRI has focused on modernising climate modelling software and data practices for ACCESS. Guided by the needs of our community, our goal is to make the modelling framework and data outputs more FAIR (Findable, Accessible, Interoperable, and Reusable) and easier to use.  

One of the key challenges in achieving FAIR for ACCESS data is the reliance on often optional post-processing steps to meet most of the FAIR guidelines. While ACCESS model outputs generally follow community standards (e.g., CF-Conventions), their implementation can be inconsistent across modelling components (e.g., atmosphere, ocean, and land models) as well as among individual data generators. As a result, using direct model output data frequently requires users to have previous knowledge and understanding of the specific climate models and leads to significant overheads for compatibility with data discovery and evaluation tools (e.g., Intake, ESMValTool). 

As a new infrastructure dedicated to Australian climate software and data, ACCESS-NRI has a unique opportunity to uplift and directly embed FAIR practices into the climate modelling software components we maintain and support. Building on successes and lessons learned from participation in global intercomparison activities such as CMIP6, ACCESS-NRI is working to apply similar data standardisation practices for the lower-level model outputs in a way that enhances consistency and usability. The effort involves close collaboration with the research community, identifying gaps and commonalities to establish a data specification that can be versioned and linked to future ACCESS model releases. This includes minimum and recommended requirements for file and dataset metadata such as: controlled vocabularies, file and variable naming conventions, provenance statements, and other critical elements to ensure data consistency and usability across all ACCESS components.    

By embedding FAIR principles directly into the ACCESS modelling system, ACCESS-NRI is not only addressing current challenges but is also future-proofing Australia’s climate modelling capabilities to meet the evolving needs of the research community. This approach will make data and tools more accessible, reduce research overheads, and enhance the adaptability of the infrastructure to future changes and new technologies. 

How to cite: Druken, K., Richards, C., Beucher, R., Basevi, J., Bull, C., Carouge, C., Dix, M., Heerdegen, A., Leopardi, P., Marchegiani, D., Nettelbeck, H., Steketee, A., Turner, C., White, M., and Wong, S.: Uplifting and streamlining FAIR data implementation for Australia’s climate modelling outputs, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14828, https://doi.org/10.5194/egusphere-egu25-14828, 2025.

EGU25-16485 | Orals | ESSI3.2

Improving the accessibility of ECMWF open weather forecast data and charts: maintenance challenges 

Milana Vuckovic, Emma Pidduck, Cihan Sahin, and Iain Russell

ECMWF's move towards an extensive free and open data policy is approaching its final phase, extending its user base far beyond operational forecasters in Member and Co-operating States and other licensed customers. Beginning in 2020, the first phase saw the opening of hundreds of web forecast charts (www.charts.ecmwf.int) and made archived data available under a Creative Commons (CC BY 4.0) open licence. This transition continued in January 2022 with the introduction of a free and open subset of real-time forecast data, with ongoing updates incorporating new parameters and datasets. Notably, the latest updates in 2024 included increasing the resolution from 0.4° to 0.25° and including the new Artificial Intelligence Forecasting System (AIFS) forecast data.
This phased move towards free and open data supports the UN EW4All initiative and also aims to support creativity, innovation and reproducibility in scientific research and weather applications. However, this can not be achieved by only opening the real time and archived data. The users need to be able to find and easily use the data and integrate it into their own research work or application workflows.
To address this, additional efforts are underway to improve the data's FAIR (Findable, Accessible, Interoperable and Reusable) attributes. Key developments include the creation of open source Python libraries for data downloading, processing and visualisation under the EarthKit umbrella, alongside the introduction of a set of Jupyter notebooks, each of which is reproducing one open weather forecast chart - from the downloading the data to processing and visualisation.
However, the tools and data constantly change, and keeping up with these changes in the example Jupyter notebooks presents a significant challenge if not designed with the maintenance in mind.
This talk will provide an overview of the open forecast web charts and the use of Jupyter notebooks for their reproduction, followed by an exploration of the maintenance challenges and future plans.

How to cite: Vuckovic, M., Pidduck, E., Sahin, C., and Russell, I.: Improving the accessibility of ECMWF open weather forecast data and charts: maintenance challenges, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16485, https://doi.org/10.5194/egusphere-egu25-16485, 2025.

EGU25-19266 | Posters on site | ESSI3.2

Establishing FAIRness through all-actor approaches to data pipelines: Frameworks for successful development of data standards and pipelines at the UK’s National Centre for Atmospheric Sciences 

Graham Parton, Barbara Brooks, Wendy Garland, Joshua Hampton, David Hooper, Nicholas Marsden, Hannah Price, Hugo Ricketts, Dave Spronson, Ag Stephens, and Chris Walden

The FAIR data principles are a common theme in many discussions and focus of work within research data management. Such work often focuses on particular parts of the data management lifecycle, for example: FAIR through data management planning, FAIR through data discovery and, more recently, areas of FAIR as applied to software and machine learning. 

However, whilst there are many successful attempts at enhancing metadata and data FAIRness for specific parts of the data lifecycle, there may be issues that only arise when considering the overall interconnections between the various stages and the associated actors. For example, a domain may follow common file and metadata conventions for data interoperability, such as CF conventions, enabling research to take place utilising multiple data sources, but pertinent metadata to long-term curation or wider end-usability may not be presented or indeed captured at source. This can have ongoing issues around the level that wider (true?) FAIRness that can be reached and present additional overheads for other actors wishing to handle such data resources, such as manual effort needed for full long-term curation or missed opportunities for data re-use in other spheres.

Recognising these issues and, crucially,  the interplay between all actors along the data lifecycle, the UK’s National Centre for Atmospheric Science (NCAS) have developed the frameworks to ensure all actors’ needs are considered. These are succinctly captured in the ‘NCAS Data Pyramid’, where each corner represents a given actor (data provider, long-term archive, those creating tools aiding data flows and utilisation, end-user community), whilst the sides explore the interconnections between these actors. All parts of the pyramid (corners and sides) provide a range of use-cases and requirements that need to be supported. This approach has enabled NCAS to then develop a range of data standards to enhance data FAIRness for surface and remote sensing data (including from ships and aircraft), imagery data and, in due course, laboratory data.

Furthermore, to aid establishing new data standards NCAS has developed data standards development framework, utilising the ‘Scope -> Define -> Develop -> Sustain’ data standard lifecycle:

  • Scope: Identify community groups. Assess their needs. Determine the scope for the standard.
  • Define the standard by: ensuring all stakeholder needs are covered; defining user-focused data products that it will deliver; and the underpinning standards to be drawn on for wider interoperability. 
  • Develop: provider tools (including checkers for compliance); data delivery pipelines (including those workflows to capture internal/external metadata required for data use/contextualisation of data (e.g. project info); develop end-user data exploitation(visualisation) tools
  • Sustain: having developed standards and workflows have a governance structure to maintain and manage future iterations of the standards development cycle. This must ensure that it refers back to the community groups (as in step 1). 

The approach also keeps wider inter-standards interoperability a key focus throughout. The success of this approach is demonstrated through the establishment of data pipelines aiding data to flow with associated metadata from provider to end-user and has seen wider adoption of NCAS data standards within the wider atmospheric community.

How to cite: Parton, G., Brooks, B., Garland, W., Hampton, J., Hooper, D., Marsden, N., Price, H., Ricketts, H., Spronson, D., Stephens, A., and Walden, C.: Establishing FAIRness through all-actor approaches to data pipelines: Frameworks for successful development of data standards and pipelines at the UK’s National Centre for Atmospheric Sciences, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19266, https://doi.org/10.5194/egusphere-egu25-19266, 2025.

EGU25-19487 | Orals | ESSI3.2 | Highlight

Challenges and opportunities in implementing open and FAIR data in Intergovernmental Panel on Climate Change (IPCC) Seventh Assessment Report (AR7)  

Xiaoshi Xing, Gian Carlo Delgado Ramos, Azra Alikadic, April Lamb, Martina Stockhause, Lina E. Sitz, and Adam Milward

Intergovernmental Panel on Climate Change (IPCC) authors of assessment reports (ARs) and special reports (SRs) use a huge volume of input data, generate a great deal of intermediate data in the process, and produce a large amount of final data for figures and annexes in the published reports. In previous assessment cycles before the Sixth Assessment Report (AR6), only a limited amount of IPCC data were archived and made publicly available. There was  great progress in the AR6, but many critical data sets were still not properly curated. This resulted in a data rescue effort during the transition from AR6 to AR7, supported by the IPCC and government fundings. The challenges encountered during the data rescue effort included missing or lost data after the report publication, missing data licensing agreements, version control issues, and missing data quality assurance/quality control (QA/QC) so that some data did not match the published figures. Addressing these issues demanded significantly more resources than the regular process to track, retrieve, archive, and resolve the legal and technical issues.

In the Seventh Assessment Report (AR7), IPCC progressively promotes the FAIR data principles (Findable, Accessible, Interoperable, and Reusable) through the IPCC Task Group on Data Support for Climate Change Assessments (TG-Data) and the Data Distribution Centre (DDC) (1, 2). The Working Group Technical Support Units (TSUs) have also designated data specialists in the TG-Data (3). This provides opportunities to support authors in implementing open and FAIR data in IPCC AR7. For example, in Chapter 2 of the Special Report on Climate Change and Cities (SRCities), there is an area of focus on “Data, information, tools accessibility/availability/usability/transparency" (4). By collaborating the TSUs and DDC can provide a coordinated approach that supports authors with training and tools on data workflow, metadata schema, data provenance, licensing and citation, persistent identifiers, etc., to improve the data curation process and to avoid the issues encountered in previous cycles.

References:

  • 1. Intergovernmental Panel on Climate Change. (2023). TG-Data Recommendations for AR7 (1.0). Zenodo. https://doi.org/10.5281/zenodo.10059282
  • 2. Stockhause M, Huard D, Al Khourdajie A, Gutiérrez JM, Kawamiya M, Klutse NAB, Krey V, Milward D, Okem AE, Pirani A, Sitz LA, Solman SA, Spinuso A, Xing X. (2024).  Implementing FAIR data principles in the IPCC seventh assessment cycle: Lessons learned and future prospects. PLOS Climate 3(12): e0000533. https://doi.org/10.1371/journal.pclm.0000533
  • 3. https://www.ipcc.ch/data/ (2025)
  • 4. IPCC Special Report on Climate Change and Cities (SRCities) report outline. (2024). https://www.ipcc.ch/site/assets/uploads/2024/08/IPCC-61_decisions-adopted-by-the-Panel.pdf

How to cite: Xing, X., Delgado Ramos, G. C., Alikadic, A., Lamb, A., Stockhause, M., Sitz, L. E., and Milward, A.: Challenges and opportunities in implementing open and FAIR data in Intergovernmental Panel on Climate Change (IPCC) Seventh Assessment Report (AR7) , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19487, https://doi.org/10.5194/egusphere-egu25-19487, 2025.

EGU25-19699 | Orals | ESSI3.2

Building an EOSC based virtual research environment to support the adoption of FAIR and Open Science practices in Climate Change Adaptation communities 

Raúl Palma, Malgorzata Wolniewicz, Adam Rynkiewicz, José Manuel Gómez, Andres Garcia Silva, Daniel Garijo, Esteban Gonzalez Guardia, and Anne Fouilloux

During the last years, Open Science has been gaining increasing attention from research communities and policy makers because of the benefits it can provide not only to scientists, but also to society in general, as it can accelerate the production of science and the quality of results. Open science is a policy priority for the European Commission (EC) and the standard method of working under its research and innovation funding programmes. Thus, the EC initiated the European Open Science Cloud (EOSC) initiative, which aims to create a virtual environment for sharing and accessing research data across borders and scientific disciplines, aligning with Open Science and FAIR principles. EOSC specified a layered approach with a set of core services at its center, a federated data layer, a rich set of exchange services to expand the capabilities offered to researchers across disciplines, plus a set of thematic/discipline-specific services. To fully realise EOSC’s vision, it is envisioned as a federation of distributed systems, combined into a system of systems, consisting of multiple Nodes’. At the end of last year, the first of such nodes (EOSC EU node) was launched featuring the core services enabling scientific research infrastructures to federate and a set of common exchange “horizontal services” for end-users to benefit from. 

Based on the integration of thematic, horizontal, and core resources, the goal is that EOSC enables the creation of thematic execution environments/VREs. A VRE is an online support system for researchers,  encompassing online tools, network resources and technologies interoperating with each other to ease/enhance the research process within and across institutional boundaries, facilitating collaboration, data management, analysis, and other research-related activities in one online space.

To build an EOSC-based VRE, we have leveraged and integrated different core and exchange services. At the center of the proposed VRE are RO-Crate based research objects (providing an implementation of the FAIR digital object), as well as the associated technological support (provided by ROHub platform), to manage the research lifecycle and the associated scientific resources used and produced. The VRE leverages data cubes services for efficient and scalable structured data access and discovery, AI-based text mining services  that extract machine-readable metadata from scientific resources supporting recommendations and comprehension analysis, and FAIR assessment tools supporting researchers in the FAIRification of their outcomes. Additionally, the VRE relies on EOSC services for authentication and authorization to enable seamless access to different services, the computing platforms to execute computational methods, and data repositories to store and/or share their data in their personal/community workspaces or general repositories. The VRE also connects DMP platforms to enable the creation of machine-actionable plans, and with the scientific knowledge graph to enable the discovery of resources by different communities. In the FAIR2Adapt project, such environment is being enhanced with a set of added-value services (e.g., search and discovery using NL questions, multilingual semantic enrichment, sentence detection, FAIRness-aware search and recommendations, and multilingual generative question answering) and adapted to boost FAIR adoption in Climate Change Adaptation communities and research.

How to cite: Palma, R., Wolniewicz, M., Rynkiewicz, A., Manuel Gómez, J., Garcia Silva, A., Garijo, D., Gonzalez Guardia, E., and Fouilloux, A.: Building an EOSC based virtual research environment to support the adoption of FAIR and Open Science practices in Climate Change Adaptation communities, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19699, https://doi.org/10.5194/egusphere-egu25-19699, 2025.

Assessing the status and trends of water quality in inland water bodies requires access to reliable water quality monitoring data and associated metadata such as the monitoring locations, sampling methods, monitoring equipment and analytical methods. Many environmental agencies and research organizations collect water quality monitoring data, but unlike in other environmental domains and due to a lack of common best practices and standards, most organizations use their own data models, formats and controlled vocabularies to store and share these data. As a result, large-scale water quality analyses with a transboundary, continental or global scope require significant efforts to collect the necessary monitoring data from different sources and to harmonize the different data structures. Several international initiatives such as the UNEP Global Environment Monitoring System for Freshwater (GEMS/Water)1 or research activities such as the Global River Water Quality Archive (GRQA)2 have compiled global water quality datasets to facilitate large-scale hydrological studies, all facing the same challenges and often duplicating data processing efforts.
Over the last 20 years, the observing community has developed data models and semantic ontologies such as the OGC Observations, Measurements, and Samples (OMS)3 standard or the OGC/W3C Semantic Sensor Network (SSN)4 ontology to describe observations and associated metadata. These form the basis of several standards for the exchange of hydrological observation data such as the WaterML 2.0 family of standards. However, water quality specific aspects such as the description of sampling activities and associated metadata have not yet been included in these water specific standards. 
To address this issue, several government agencies and research organizations have started a Water Quality Interoperability Experiment (WQIE) within the Open Geospatial Consortium (OGC) in 2022. Several use cases for the exchange of water quality monitoring data of physical and chemical parameters monitored in surface and groundwater bodies using in-situ (sensor) or ex-situ (laboratory) monitoring were developed and described as object diagrams in UML based on the OMS conceptual model. Based on this exercise, a physical data model was developed by extending the OGC SensorThingsAPI (STA)5 with a plugin for the open source FROST server6. Several WQIE participants deployed pilot instances of water quality enabled FROST servers, making their water quality data publicly available. A web client was developed to facilitate access to the various STA endpoints and to enable data visualisation7
This presentation will give an overview of the developments of the OGC Water Quality Interoperability Experiment, highlighting achievements, outstanding challenges and future development plans. 

References:

1 https://www.unep.org/explore-topics/water/monitoring-water-quality

2 Virro, H., Amatulli, G., Kmoch, A., Shen, L., and Uuemaa, E.: GRQA: Global River Water Quality Archive, Earth Syst. Sci. Data, 13, 5483–5507, https://doi.org/10.5194/essd-13-5483-2021, 2021.

3 https://docs.ogc.org/as/20-082r4/20-082r4.html

4 https://www.w3.org/TR/vocab-ssn/

5 https://www.ogc.org/publications/standard/sensorthings/

6 https://github.com/hylkevds/FROST-Server.Plugin.WaterQualityIE/tree/main

7 https://api4inspire.k8s.ilt-dmz.iosb.fraunhofer.de/servlet/is/226/ 

How to cite: Heinle, M. and Saile, P.: A step towards FAIR water quality data – lessons learned from the OGC Water Quality Interoperability Experiment, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19719, https://doi.org/10.5194/egusphere-egu25-19719, 2025.

Since their development, the FAIR principles have been met with broad acceptance in the scientific community. Tools based on various approaches are available to assess the FAIRness of individual data sets. These range from qualitative assessments based on questionnaires to automated quantitative measurements of fairness. As the FAIR principles are rather vaguely formulated, these approaches are based on individual, often differing, interpretations of the FAIR principles. In addition, the authors of the FAIR principles explicitly recognize the different implementations of FAIR within the various specialist communities. This makes it necessary to develop community-specific metrics and tests and to adapt FAIR assessment tools accordingly.

This diversity of methods for assessing FAIR is encouraging, as it sheds light on a variety of aspects of FAIR. However, this also sometimes leads to different, divergent results from these tools, which is difficult for users to work with. In addition, the measurement of FAIRness of individual datasets is heavily dependent on various technical implementations on the part of the data providers and their service providers. Numerous, possibly unintentional restrictions on the accessibility of datasets can influence or falsify FAIR measurements. 

In this presentation, we would like to report on our experiences with the applied FAIR assessment within this context. We will report on the further development of F-UJI, in particular our experiences with discipline-specific FAIR metrics and their implementation. Furthermore, we will discuss the limitations of FAIR measurements and try to delineate FAIR from aspects of data quality and accessibility and how to derive informative holistic assessments of datasets that include all these aspects in the future.

How to cite: Huber, R.: Opportunities and limitations of applied FAIR evaluation of data sets, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20064, https://doi.org/10.5194/egusphere-egu25-20064, 2025.

EGU25-20123 | Orals | ESSI3.2

Automatic annotation following the I-ADOPT framework 

José Manuel Gómez Pérez and Andrés García

The fulfillment of the FAIR principles is a central requirement in modern research. Data findability and reusability are highly dependent on the quality and interoperability of their metadata. Among other attributes in earth and environmental sciences, FAIR metadata should ensure consistent and uniquely referenceable naming of geoscientific variables that support machine-interpretable semantic annotations. But in practice, most terminologies used to describe datasets and observed variables vary wildly in their granularity, quality, governance and interconnectivity which, in turn, limits their interoperability. The RDA endorsed I-ADOPT Framework addresses this issue by breaking down descriptions of observed variables into five well-defined atomic components ObjectofInterest, Property, Matrix, Constraint and Context anticipating their annotation with generic terms from FAIR semantic artefacts. As of today, the I-ADOPT decomposition is still a highly manual process that requires semantic and domain skills. Here, we propose the application of Large Language Models (LLM) to transform scientific terms into I-ADOPT-aligned descriptions. This model will enable the transformation into machine-interpretable representations by simply using natural language descriptions of observational research provided by domain experts. We will leverage the existing set of high-quality, human-made formalizations of I-ADOPT variables to adjust the LLM for this task. We will consider LLM in zero-shot scenarios where the LLM is used in its pretrained version and in-context learning where the LLM sees some examples of the task. We will also consider training specialist LLM where the LLM is further fine-tuned for this task, although the success of this approach depends on the amount of training data available. For developing this model and a first demonstrator, we will build on our previous experience in developing the I-ADOPT Framework, in transfer learning and fine-tuning neural networks, FAIR data stewardship, research data infrastructures and research software engineering. Our project will be further linked to several other ongoing activities and initiatives both on a national and also European level, which allows us to directly evaluate the performance of our LLM by potential end-users and communities. Such a service will be integrated into platforms like RoHub to help scientists make research datasets FAIR.

How to cite: Gómez Pérez, J. M. and García, A.: Automatic annotation following the I-ADOPT framework, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20123, https://doi.org/10.5194/egusphere-egu25-20123, 2025.

EGU25-20132 | Orals | ESSI3.2

How domain repositories support reusable data: metadata tools from GFZ Data Services 

Marcel Meistring, Holger Ehrmann, Jana Franz, Simone Frenzel, Ali Mohammed, and Kirsten Elger

The availability of reusable data and their associated metadata is increasingly demanded to address global societal challenges. Research data repositories and databases are the primary access points for geosciences data, and especially domain repositories are known to publish well documented and reusable data. This is due to a thorough data and metadata curation provided by the repository staff that usually includes domain scientists. Overall, the documented publication of a complex data set via a domain repository often takes time and additional preparation by the scientists, but the results clearly show a significant increase of the metadata and data quality, including the provision of cross-references to other publications, datasets, code and originating physical samples.

The largest challenge for domain repositories is to provide incentives to the researchers that reduce their workload and in the same time ensure a high quality of metadata and data documentation already at an early stage of a planned data publication. This challenge is especially high in repositories with a focus on the highly variable and usually small data from so-called “long-tail communities”. GFZ Data Services is a domain repository for DOI-referenced geosciences data and scientific software, hosted at the GFZ Helmholtz Centre for Geosciences. The repository has both a focus on the curation of long-tail data, and offers data publication services for international projects and services in the geosciences. To support researchers with the provision of descriptive metadata and receive structured data documentation, GFZ Data Services has developed an online metadata editor and data description templates. This presentation will focus on these support tools and demonstrate how both help the researchers and in the same time reduce the data curation workload.

A major focus will lay on our new metadata editor that is currently jointly developed between the University of Applied Sciences Potsdam and GFZ Data Services. The new metadata editor will enhance the support of users in data entry, so that the manual curation effort by the GFZ Data Services is reduced, and the metadata quality is improved at the same time. Technically, it has a responsive design and offers a dark mode. New facets include the ability to retrieve specific information, e.g., affiliations from the ROR API via a dropdown menu. Keywords are made uniquely identifiable through the automatic storage of schema names and uniform resource identifiers of the specific terms. All integrated thesauri can be updated via API calls. Real time validation of the input fields prevents the submission of incomplete or incorrect entries, so that significantly less work is required in data curation. The integrated help guide supports users to fill in the input fields.

The data description templates collect additional technical description in a structured form and are essential for data reuse. They are available in “commented” and “usable” versions and ensure that the descriptions meet our requirements (for many researchers the data documentation is new), offer clear instructions and even reduce the workload of the curators, because the descriptions are already provided at a very high level of content.

How to cite: Meistring, M., Ehrmann, H., Franz, J., Frenzel, S., Mohammed, A., and Elger, K.: How domain repositories support reusable data: metadata tools from GFZ Data Services, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20132, https://doi.org/10.5194/egusphere-egu25-20132, 2025.

The Intergovernmental Panel on Climate Change (IPCC) Data Distribution Centre (DDC) serves as a critical registry for climate change data, providing a shared infrastructure to ensure data quality and accessibility for the scientific community. Managing data to support IPCC reports presents challenges due to its multidisciplinary nature and diverse sources.

Key to this effort is the curation of metadata, particularly developing a metadata schema that enables data to be FAIR (Findable, Accessible, Interoperable, and Reusable). This presentation examines the IPCC's experience over the past four years in curating and preserving digital objects, focusing on the implementation of FAIR and open data principles. We will explore the successes and setbacks of the AR6 experience, with particular attention to the development and application of a metadata schema. Finally, we will offer recommendations for consolidating and expanding this approach for AR7 to enhance transparency, reproducibility, and reusability of assessment outcomes.

This initiative aims to increase the transparency of IPCC's work, improve the reproducibility and reusability of assessment outcomes, optimize the utilization of the IPCC DDC's services, and promote compliance with open science best practices.

How to cite: Milward, D., Milward, A., and Xing, X.: Managing a FAIR Climate Change Data Catalogue: Lessons Learned from IPCC AR6 and Recommendations for AR7, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20454, https://doi.org/10.5194/egusphere-egu25-20454, 2025.

EGU25-20583 | Orals | ESSI3.2

International Earth, space, and environmental coordination of data and software management efforts 

Shelley Stall, Danie Kinkade, Natalie Raia, Lesley Wyborn, and Pedro Corrêa

The international Earth, space, environmental sciences informatics community has recently formed a new Research Data Alliance Community of Practice. Here we are focused on improving data and software management and sharing practices that result in our researchers having access to community informatics resources that support their research.  This community of practice will provide a place for teams and organizations in the Earth, space, and environmental research ecosystem to coordinate on common challenges, share information, review and consider RDA recommendations, seek leading practices, and work towards finding approaches to discipline-specific challenges and issues around data and software management and sharing. The international Earth, space, and environmental community is broad and includes researchers, data managers, data curators, institutions, instrument creators and manufacturers, software developers, tools, repositories, journal editors and more. 

An RDA community of practice is where those with common interests can collaborate on complex challenges that need multiple stakeholders to work through the layers of a solution. It is a place where projects can be highlighted and shared for the benefit of building collaboration and connection.     

Join us for this session and learn more about how we envision supporting the many global data and software management efforts.

How to cite: Stall, S., Kinkade, D., Raia, N., Wyborn, L., and Corrêa, P.: International Earth, space, and environmental coordination of data and software management efforts, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20583, https://doi.org/10.5194/egusphere-egu25-20583, 2025.

EGU25-21162 | Posters on site | ESSI3.2

Implementing FAIR Principles for Earth System Data: Insights from the European Eddy-Rich Earth-System Models (EERIE) project 

Heinrich Widmann, Chathurika Wickramage, and Fabian Wachsmann

We attempt to make EERIE data FAIR (Findable, Accessible, Interoperable, and Reusable) to enhance its scientific impact and utility. These principles of FAIRness ensure global access, integration, and reuse by researchers and decision-makers, thereby promoting collaboration and innovation.

Findability is enhanced through persistent identifiers such as DOIs and PIDs, ensuring data remains reliably locatable. Metadata standards, including CF conventions and CMIP standard names, ensure precise and efficient searchability. We enhance findability through data catalogs produced in the EERIE and nextGEMS projects, as well as platforms like World Data Center for Climate (WDCC) and DOKU. The WDCC ensures long-term storage with a focus on FAIRness, quality control, and DOI assignment following CF standards. Our EERIE data is also archived on DOKU with PIDs to ensure discoverability.

Accessibility is ensured by providing data through open protocols with clear terms of use. While accessibility does not always mean free access, it guarantees transparency and ease of use. Open-access repositories such as EERIE Cloud, Earth System Grid Federation (ESGF), and, WDCC combination with standardized formats such as NetCDF and Zarr, ensure broad accessibility. Additionally, tools like Zarr provide API access via HTTP, facilitating seamless and efficient data retrieval.

Interoperability is fundamental for integrating datasets across disciplines and platforms. The EERIE project supports this by linking datasets through initiatives such as EERIE Cloud, FREVA and by using standards such as CF conventions to ensure compatibility, facilitating multidisciplinary research.

Reusability is supported through detailed metadata, clear licensing models like CC-BY and CC0, and strong version control practices (e.g, v20240304). Documentation platforms such as easy.gems.dkrz.de assist users to understand and reproduce results. The maintenance of high data quality and the emphasis on archival and replication further enhance the long-term scientific use of these datasets.

Despite these efforts, the implementation of the FAIR data principles in a comprehensive manner poses significant challenges. In the EERIE project, for instance, we work with vast amounts of data, and standardizing it (e.g., CMORizing) can be complex. Obtaining CF-compliant names for all variables is particularly difficult, as there is often no one-to-one documentation from modeling groups. In some cases, this requires manually analyzing code to determine the correct definitions for certain variables.

For climate science, the application of FAIR principles is transformative. These efforts promote global collaboration, enhance the transparency of climate models, and equip policymakers with reliable data to address critical challenges such as climate adaptation and mitigation. Initiatives like EERIE cloud, ESGF and advancements in data processing, such as kerchunking massive datasets, further enhance the FAIRness of climate data, driving innovation and impact.

How to cite: Widmann, H., Wickramage, C., and Wachsmann, F.: Implementing FAIR Principles for Earth System Data: Insights from the European Eddy-Rich Earth-System Models (EERIE) project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21162, https://doi.org/10.5194/egusphere-egu25-21162, 2025.

EGU25-21472 | Orals | ESSI3.2

Community Support and Engagement for FAIR Science in Climate Change Adaptation 

Erik Schultes and Barbara Magagna

Global climate change requires urgent and actionable adaptation planning.

Current Climate Change Adaptation (CCA) strategies often lack the necessary data and other relevant information to be scientifically competent. These limitations can complicate effective action and evaluation locally, and in combination with other regions. The recently awarded FAIR2Adapt Project aims to establish a comprehensive FAIR and open data framework for CCA and to demonstrate the impact of FAIR data on CCA strategies. By making CCA data FAIR, FAIR2Adapt will accelerate adaptation actions that are tailored to local needs.

Next to the technical development of FAIR data and services, a key issue in the effective uptake  of FAIR is the transfer of knowledge regarding FAIR practices, and in many cases hands-on skills related to the design, creation and governance of domain-relevant FAIR Enabling Resources.  Beginning in February 2025, the FAIR2Adapt, stakeholders (including members of it’s six use cases) will participate in FAIR awareness and training based on the GO FAIR Foundation’s FAIR Capacity Building Programme [https://zenodo.org/records/14187859]. This will include general FAIR Awareness workshops, training on the creation of FAIR Implementation Profiles and community-specific metadata and vocabulary in Metadata for Machine workshops. In addition to this, special attention will be given to the identification and prioritization of user requirements (both the technical approach in FAIR2Adapt as well as the case studies). Having both the technical expertise and building up the salient knowledge and skills, the FAIR2Adapt community will be well positioned to co-design, implement and share CCA related data and services that can accelerate meaningful and customized CCA. In this presentation, we will report the first draft user requirements for FAIR2Adapt and the emerging list of CCA community-specific FAIR Enabling Resources.  

 

How to cite: Schultes, E. and Magagna, B.: Community Support and Engagement for FAIR Science in Climate Change Adaptation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21472, https://doi.org/10.5194/egusphere-egu25-21472, 2025.

EGU25-21516 | Orals | ESSI3.2

Advancing Data Infrastructure for Chemical Risk Assessment and Exposome Research: The GENASIS Platform in the Context of FAIR Principles 

Katarína Řiháčková, Jana Borůvková, Zdenka Bednářová, Richard Hůlek, and Jana Klánová

Excellence in exposome research and chemical risk assessment (CRA) relies on robust capacities, innovative technologies, and skilled human resources. Research infrastructures are vital in providing access to these resources and driving innovation. Over recent decades, Europe has developed numerous research infrastructures, including EIRENE RI (Research Infrastructure for Environmental Exposure Assessment in Europe), the first EU research infrastructure dedicated to the human exposome. EIRENE RI aims to integrate interdisciplinary data, offering harmonized workflows and services to users across various sectors. Other initiatives, such as the Partnership for the Assessment of Risks from Chemicals (PARC), work on advancing harmonization and innovation in CRA.

A robust data infrastructure aligned with FAIR data and Open Science principles is essential for these research infrastructures. Mapping and evaluating the current data landscape is a critical step toward enhancing FAIR implementation and machine actionability. This contribution highlights existing strategies for harmonizing and managing global data on chemical occurrences developed through two decades, using the use case of the GENASIS information system.

GENASIS information is a platform originally developed for storing, harmonizing, and visualizing global environmental monitoring data. Over time, it has expanded to include data on chemical occurrences in indoor environments, consumer products, and human matrices. Today, it hosts over 3 million harmonized records on more than 800 chemicals, described with rich metadata, and it is continuously expanding. This enables the identification of gaps, locality comparisons, and evaluation of global trends in chemical concentrations in the environemnt and humans. GENASIS also serves as a model and sister database for the Global Monitoring Plan Data Warehouse of the Stockholm Convention and supports the United Nations Environment Programme in managing environmental and human monitoring data to evaluate the effectiveness of global treaties on chemical pollutants. GENASIS’ ongoing development and associated services contribute to the European Open Science Cloud (EOSC) in the Czech Republic, EIRENE RI and PARC initiatives.

This contribution evaluates GENASIS in terms of FAIR principles, detailing its current status, roadmap for further FAIR implementation, efforts to enhance machine actionability, and challenges encountered. The discussion is framed within the broader context of initiatives such as PARC, EIRENE RI, and EOSC CZ, emphasizing their role in advancing exposome research and CRA in Europe.

Acknowledgement: This project was supported from the European Union’s Horizon 2020 research and innovation programme under grant agreement No 857560 (CETOCOEN Excellence), and from the Horizon Europe programme under grant agreements No 101057014 (PARC) and 101079789 (EIRENE PPP). Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union, European Health and Digital Executive Agency (HADEA) or European Research Executive Agency (REA). Neither the European Union nor the granting authorities can be held responsible for any use that may be made of the information it contains. Authors thank the RECETOX Research Infrastructure (No LM2023069) financed by the Ministry of Education, Youth and Sports.

How to cite: Řiháčková, K., Borůvková, J., Bednářová, Z., Hůlek, R., and Klánová, J.: Advancing Data Infrastructure for Chemical Risk Assessment and Exposome Research: The GENASIS Platform in the Context of FAIR Principles, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21516, https://doi.org/10.5194/egusphere-egu25-21516, 2025.

EGU25-21605 | Orals | ESSI3.2

Status, issues and challenges with FAIRness of seismological waveform data and beyond 

Florian Haslinger, Lesley Wyborn, Rob Casey, Helle Pederson, Elisabetta D’Anastasio, Javier Quinteros, Jonathan Hanson, and Jerry Carter

Driven by the scientific need for global exchange of data to study earthquakes and related phenomena, community standards and best practices have evolved in seismology for decades. These developments are largely driven by operational and scientific requirements coming directly from the community of academic research and seismological monitoring, and have resulted in standardised data formats, data models and services for data access and exchange.

Initial developments, promotion and further evolution of these standards are coordinated mainly within the International Federation of Digital Seismic Networks (FDSN, https://fdsn.org), a commission of IASPEI (International Association of Seismology and Physics of the Earth's Interior, httwww.iaspei.org) that is one of eight associations of the IUGG (International Union of Geodesy and Geophysics, https://iugg.org).   

With the introduction of the FAIR (Findable, Accessible, Interoperable, Reusable) principles in 2016 and the subsequent appearance of FAIR assessment methods and tools it became clear that these seismological community standards only cover parts of the FAIR principles. Interoperability remains challenging, for example, due to the lack of community standardised FAIR vocabularies, and the lack of a harmonised and consistently applied data license policy impacts Reproducibility.

The emergence of new data types and the drastic increase in data volumes due to new measurement techniques require updates and evolution of the existing community standards, highlighting another general challenge:  Who are the recognised and appropriate governance bodies for curation and further development of 'relevant community standards' (as required by the FAIR principles)?

In this presentation we describe the current status of FAIRness for seismological waveform data and beyond, also looking towards seismology in general, geodesy and some other fields of geophysics. Based on our assessment of current challenges we discuss open questions and possible ways forward. We look at FAIR-relevant development and governance of standards, the potential role of existing international organisations like FDSN, IASPEI and IUGG, and the possibility and need to coordinate across domains for harmonisation as well as demarcation.   

How to cite: Haslinger, F., Wyborn, L., Casey, R., Pederson, H., D’Anastasio, E., Quinteros, J., Hanson, J., and Carter, J.: Status, issues and challenges with FAIRness of seismological waveform data and beyond, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21605, https://doi.org/10.5194/egusphere-egu25-21605, 2025.

EGU25-1511 | ECS | Orals | ESSI3.3

A workflow for cloud-based and HPC simulations with the NEMO ocean model using containers 

Aina Gaya-Àvila, Bruno de Paula Kinoshita, Stella V. Paronuzzi Ticco, Oriol Tintó Prims, and Miguel Castrillo

In this work, we explored the deployment and execution of the NEMO ocean model using Singularity containers within the EDITO Model Lab, implementing the European Digital Twin of the Ocean. The Auto-NEMO workflow, a fork of Auto-EC-Earth used to run NEMO workflows using the NEMO Community reference code, was adapted to run simulations using containers. The use of a Singularity container ensures consistent execution by packaging all dependencies, making it easier to deploy the model across various HPC systems.

The containerized approach was tested on multiple HPC platforms, including MareNostrum5 and LUMI, to evaluate scaling performance. Our tests compared the use of mpich and openmp libraries, providing insights into how communication strategies impact the computational performance of the model in containerized setups. In addition, the runs are orchestrated by a content workflow manager, in this case Autosubmit, deployed in a cloud infrastructure in EDITO-Infra, making the entire solution (workflow manager and workflow itself) portable end-to-end. The benefits of portability and reproducibility make containers an attractive solution for streamlining workflows in diverse computational environments.

A comparison between containerized and non-containerized runs highlights the trade-offs involved. Direct execution may provide slightly better performance in some cases, but the containerized approach greatly reduces setup complexity. These findings demonstrate the potential of containerization to enhance efficiency and accessibility in large-scale ocean modeling efforts.

How to cite: Gaya-Àvila, A., de Paula Kinoshita, B., Paronuzzi Ticco, S. V., Tintó Prims, O., and Castrillo, M.: A workflow for cloud-based and HPC simulations with the NEMO ocean model using containers, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1511, https://doi.org/10.5194/egusphere-egu25-1511, 2025.

EGU25-2142 | ECS | Posters on site | ESSI3.3

Enhancing Data Provenance in Workflow Management: Integrating FAIR Principles into Autosubmit and SUNSET 

Albert Puiggros, Miguel Castrillo, Bruno de Paula Kinoshita, Pierre-Antoine Bretonniere, and Victòria Agudetse

Ensuring robust data provenance is paramount for advancing transparency, traceability, and reproducibility in climate research. This work presents the integration of FAIR (Findable, Accessible, Interoperable, and Reusable) principles into the workflow management ecosystem through provenance integration in Autosubmit, a workflow manager developed at the Barcelona Supercomputing Center (BSC), and SUNSET (SUbseasoNal to decadal climate forecast post-processing and asSEmenT suite), an R-based verification workflow also developed at the BSC.

Autosubmit supports the generation of data provenance information based on RO-Crate, facilitating the creation of machine-actionable digital objects that encapsulate detailed metadata about its executions. Autosubmit integrates persistent identifiers (PIDs) and schema.org annotations, making provenance records more accessible and actionable for both humans and machines.  However, the provenance metadata provided by Autosubmit through RO-Crate focuses on the workflow process and does not encapsulate the details of the data transformation processes. This is where SUNSET plays a complementary role. SUNSET’s approach for provenance information is based on the METACLIP (METAdata for CLImate Products) ontologies. METACLIP offers a semantic approach for describing climate products and their provenance. This framework enables SUNSET to provide specific, high-resolution  provenance metadata for its operations, improving transparency and compliance with FAIR principles. The generated files provide detailed information about each transformation the data has undergone, as well as additional details about the data's state, location, structure, and associated source code, all represented in a tree-like structure.

The main contribution of this work is the generation of a comprehensive provenance object by integrating these tools. SUNSET uses Autosubmit to parallelize its data processing tasks, with Autosubmit managing SUNSET jobs. As part of this process, an RO-Crate is automatically generated describing the overall execution. This object encapsulates detailed provenance metadata for each individual job within the workflow, using METACLIP's semantic framework to represent each SUNSET execution process. Certain schema.org entities are introduced to have the RO-Crate created by Autosubmit link with the provenance details generated by SUNSET. This integrated approach provides a unified hierarchical provenance record that spans to both the workflow management system and the individual job executions, ensuring that provenance objects are automatically generated for each experiment conducted.

This work demonstrates the practical application of FAIR principles in climate research by advancing provenance tracking within complex workflows. It represents an initial step to obtain and share metadata about the provenance of the data products that a workflow provides. The integration of RO-Crate and METACLIP not only enhances the reproducibility of climate data products but also fosters greater confidence in their reliability. To our knowledge, this is the first effort in the climate domain to combine different provenance formats into a single object, aiming to obtain a complete provenance graph with all the metadata. 

How to cite: Puiggros, A., Castrillo, M., de Paula Kinoshita, B., Bretonniere, P.-A., and Agudetse, V.: Enhancing Data Provenance in Workflow Management: Integrating FAIR Principles into Autosubmit and SUNSET, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2142, https://doi.org/10.5194/egusphere-egu25-2142, 2025.

EGU25-4355 | ECS | Posters on site | ESSI3.3

Generic State Vector: streaming and accessing high resolution climate data from models to end users 

Iker Gonzalez-Yeregi, Pierre-Antoine Bretonnière, Aina Gaya-Avila, and Francesc Roura-Adserias

The Climate Adaptation Digital Twin (ClimateDT) is a contract under the Destination Earth initiative (DestinE) that aims to develop a digital twin to account for climate change adaptation. This is achieved by running high-resolution simulations with different climate models by making use of the different EuroHPC platforms. In addition to the climate models, applications that consume data from models are also developed under the contract. A common workflow is used to execute the whole pipeline from the model launching to the data consumption by the applications in a user-friendly and automated way.

One of the challenges of this complex workflow is to handle the different outputs that each of the climate models initially offered. Each model works with its own grid, vertical levels, and variable set. These differences in format make it very complicated for applications to consume and compare data coming from different models in an automated and timely manner. This issue is resolved by introducing the concept of Generic State Vector (GSV), which defines a common output portfolio for all models to ensure a homogeneous output between models. The conversion from the model's native output to the GSV happens before the data is written in the HPC and it is automated in the workflow allowing transparent access to the data changing only the name of the model in the call.

Data in the GSV format can be read using a newly designed dedicated Python tool: the GSV Interface. This tool links the model part of the workflow with the applications part of the workflow, enabling running everything in a single complex workflow (end-to-end workflow). The GSV Interface allows to read data that has been previously converted to GSV, adding proper metadata. It also offers some extra features like interpolation to regular grids and area selection. All the workflow components that read data from the models rely on the GSV Interface. In addition to that, the GSV Interface can also be used to transparently retrieve and process data from the public Destination Earth Service Platform.

How to cite: Gonzalez-Yeregi, I., Bretonnière, P.-A., Gaya-Avila, A., and Roura-Adserias, F.: Generic State Vector: streaming and accessing high resolution climate data from models to end users, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4355, https://doi.org/10.5194/egusphere-egu25-4355, 2025.

EGU25-4466 | ECS | Posters on site | ESSI3.3

ClimateDT Workflow: A containerized climate workflow 

Francesc Roura-Adserias, Aina Gaya-Avila, Leo Arriola i Meikle, Iker Gonzalez-Yeregi, Bruno De Paula Kinoshita, Jaan Tollander de Balsch, and Miguel Castrillo

The Climate Adaptation Digital Twin (ClimateDT), a contract (DE_340) inside the Destination Earth (DestinE) flagship initiative from the European Commission, is a highly collaborative project where climate models are executed in an operational manner on different EuroHPC platforms. The workflow software supporting such executions, called ClimateDT Workflow, contains a model component and an applications component. The applications can be seen as elements that consume the data that is provided by the climate models. They aim to provide climate information to sectors that are critically dependent on climate change, such as renewable energy or wildfires, among others. This workflow relies on the Autosubmit workflow manager and is executed over different EuroHPC platforms that are part of the contract.

There are six lightweight applications that are run in this workflow, in parallel to the model and in a streaming fashion. Setting up and maintaining an environment for these applications for each EuroHPC platform (plus the development environments) is a time-consuming and cumbersome task. These machines are shared by multiple users, have different operating systems and libraries, some do not have internet access for all users on their login nodes, and there are different rules to install and maintain software on each machine.

In order to overcome these difficulties all the application-required dependencies of the workflow are encapsulated beforehand in a Singularity container and therefore the portability to the different platforms becomes merely an issue with path-binding inside the platform. Through the use of Singularity containers, their execution does not require administrator permissions, which allows anyone with access to the project to execute the desired application either on the EuroHPC machines, or on their local development environment.

This work shows the structure of the ClimateDT workflow and how it uses Singularity containers, how they contribute not only to portability but also to traceability and provenance, and finally the benefits and issues found during its implementation. We believe that the successful use of containers in this climate workflow, where applications run in parallel to the climate models in a streaming fashion and where the complete workflow runs on different HPC platforms, presents a good reference for other projects and workflows that must be platform-agnostic and that require agile portability of their components.

How to cite: Roura-Adserias, F., Gaya-Avila, A., Arriola i Meikle, L., Gonzalez-Yeregi, I., De Paula Kinoshita, B., Tollander de Balsch, J., and Castrillo, M.: ClimateDT Workflow: A containerized climate workflow, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4466, https://doi.org/10.5194/egusphere-egu25-4466, 2025.

In an era of unprecedented availability of Earth Observation (EO) data, the Copernicus Data Space Ecosystem (CDSE) emerges as a vital platform to bridge the gap between data accessibility and actionable insights. With petabytes of freely accessible satellite data at our fingertips and multiple operational data processing platforms in place, many of the foundational challenges of accessing and processing sensor data have been addressed. Yet, the widespread adoption of EO-based applications remains below expectations. The challenge lies in the effective extraction of relevant information from the data. While numerous R&D projects demonstrate the possibilities of EO, their results are often neither repeatable nor reusable, primarily due to prototype-level implementations and overly tailored, non-standardized workflows.  

CDSE tackles these barriers by adopting common standards and patterns, most notably through openEO, an interface designed to standardize EO workflow execution across platforms. openEO enables the development of reusable workflows that are scalable and transferable, paving the way for systematic and objective monitoring of the planet. CDSE has already integrated openEO as a core processing interface, and further advancements are underway, including the integration of Sentinel Hub to support openEO. This integration will enhance instantaneous visualization, synchronous API requests, and batch processing, as well as support openEO process graphs within the Copernicus Browser, bringing the simplicity and speed of Sentinel Hub’s synchronous engine to the openEO ecosystem.  

CDSE’s openEO capabilities are already validated through large-scale operational projects such as ESA WorldCereal and Copernicus Global Land Cover and Tropical Forestry Mapping and Monitoring Service (LCFM), which leverage its robust, scalable, and reliable infrastructure. Additionally, the openEO Algorithm Plaza fosters collaboration by enabling the easy sharing and reuse of processing workflows, while the Bring Your Own Data feature allows users to integrate their datasets into the ecosystem, promoting data interoperability and collaborative advancements.  

CDSE is embracing a federated approach, allowing additional data or service providers to become part of the ecosystem. This inclusivity ensures a growing network of interoperable services while maintaining technical and operational stability—a cornerstone for broad adoption and long-term sustainability.  

By addressing the need for operational and reusable workflows with openEO and related initiatives, CDSE is not only advancing the technical landscape of EO but also fostering a culture of repeatable, scalable, and impactful science. Through this session, we aim to spark a discussion on how to make EO applications more accessible, reusable, and impactful for the global community.

How to cite: Sharma, P.: How openEO standardizes workflows for scalable and reusable EO data analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5593, https://doi.org/10.5194/egusphere-egu25-5593, 2025.

EGU25-6201 | Orals | ESSI3.3

Advancing Computational Workflow Sharing in Earth Science: Insights from DT-GEO and Geo-INQUIRE 

Marco Salvi, Rossana Paciello, Valerio Vinciarelli, Kety Giuliacci, Daniele Bailo, Pablo Orviz, Keith Jeffery, Manuela Volpe, Roberto Tonini, and Alejandra Guerrero

The increasing complexity and volume of data in Solid Earth Science necessitate robust solutions for workflow representation, sharing, and reproducibility. Within the DT-GEO (https://dtgeo.eu/) project, we addressed the challenge of creating interoperable and discoverable representations of computational workflows to facilitate data reuse and collaboration. Leveraging the EPOS Platform (https://www.epos-eu.org/), a multidisciplinary research infrastructure focused on Solid Earth Science, we aimed to expose workflows, datasets, and software to the community while adhering to the FAIR (Findable, Accessible, Interoperable, and Reusable) principles. While the EPOS-DCAT-AP (https://github.com/epos-eu/EPOS-DCAT-AP) model, already used in EPOS, can effectively represent datasets and software, it lacks direct support for computational workflows, necessitating the adoption of alternative standards.

To overcome this limitation, we employed the Common Workflow Language (CWL, https://www.commonwl.org/) to describe workflows, capturing their structure, software, datasets, and dependencies. The developed CWL representations are "abstract" focusing on general workflow structures while omitting execution-specific details to prioritize interoperability. To package these workflows along with metadata, we utilized Workflow Run Crate, an extension of the RO-Crate (https://www.researchobject.org/ro-crate/) standard. Together, these technologies enable workflows to become self-contained entities, simplifying sharing and reuse. 

This approach not only aligns with community standards but also benefits from a mature ecosystem of tools and libraries, ensuring seamless integration and widespread applicability. Initial implementations within the DT-GEO project serve as a model for adoption in related initiatives such as Geo-INQUIRE (https://www.geo-inquire.eu/), where similar methodologies are being used to share workflows derived from the Simulation Data Lake (SDL) infrastructure. These implementations pave the way for broader integration within the EPOS Platform, enhancing access to advanced workflows across disciplines.

Our contribution highlights the value of adopting standardized tools and methodologies for workflow management in Solid Earth Science, showcasing how CWL and RO-Crate streamline interoperability and foster collaboration. These advances address challenges in data and computational management, contributing to the scalable FAIR workflows essential for tackling the complexities of Solid Earth Science. Moving forward, the integration of these standards across projects like DT-GEO and Geo-INQUIRE will further enhance the EPOS Platform's capabilities, offering a unified gateway to reproducible, secure, and trustworthy workflows that meet the evolving needs of the scientific community.

How to cite: Salvi, M., Paciello, R., Vinciarelli, V., Giuliacci, K., Bailo, D., Orviz, P., Jeffery, K., Volpe, M., Tonini, R., and Guerrero, A.: Advancing Computational Workflow Sharing in Earth Science: Insights from DT-GEO and Geo-INQUIRE, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6201, https://doi.org/10.5194/egusphere-egu25-6201, 2025.

EGU25-6216 | Posters on site | ESSI3.3

CAMELS-PLUS: Enhancing Hydrological Data Through FAIR Innovations. 

Carlos Zuleta Salmon, Mirko Mälicke, and Alexander Dölich

The CAMELS-PLUS initiative is revolutionizing the way hydrological, and Earth System Science (ESS) data are processed, shared, and utilized by enhancing the widely-used CAMELS-DE dataset. While Germany boasts one of the richest hydrological datasets globally, CAMELS-DE has faced challenges due to its reliance on fragmented, manual workflows, which are error-prone and hinder collaboration. CAMELS-PLUS introduces a groundbreaking solution: a standardized framework for containerized scientific tools that embed rich metadata, ensuring provenance, reusability, and seamless integration across diverse scientific domains.

A key innovation of CAMELS-PLUS lies in its ability to bridge the gap between disciplines by implementing a fully containerized pipeline for dataset pre-processing. This approach allows researchers in meteorology, forestry, and other ESS subdomains to easily contribute and extend CAMELS-DE without the complexity of navigating storage systems or inconsistent workflows. The initiative’s metadata schema, implemented as YAML files with JSON-based tool parameterization, enables tools to "speak the same language," ensuring they are interoperable and aligned with FAIR principles.

Key Deliverables:

  • Updated CAMELS-DE Dataset: Incorporates new precipitation sources and enhanced metadata for seamless integration with the NFDI4Earth Knowledge Hub.
  • Standardized Scientific Containers: A community-adopted specification for containerized tools, promoting accessibility and reusability across disciplines.
  • Interactive Community Engagement: Extensions to camels-de.org, transforming it into a hub for exploring workflows and fostering interdisciplinary collaboration.

What makes CAMELS-PLUS particularly compelling is its potential to democratize access to cutting-edge hydrological datasets. By enabling non-specialists to contribute and utilize CAMELS-DE through intuitive, containerized workflows, the initiative reduces barriers to entry and accelerates innovation in data-driven hydrology and beyond. This project not only sets a new standard for dataset management in ESS but also creates a replicable model for tackling similar challenges across other scientific domains. CAMELS-PLUS is poised to inspire transformative changes in how large-sample datasets are curated, shared, and advanced for global scientific impact.

How to cite: Zuleta Salmon, C., Mälicke, M., and Dölich, A.: CAMELS-PLUS: Enhancing Hydrological Data Through FAIR Innovations., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6216, https://doi.org/10.5194/egusphere-egu25-6216, 2025.

EGU25-6544 | Posters on site | ESSI3.3

PyActiveStorage:  Efficient distributed data analysis using Active Storage for HDF5/NetCDF4 

Bryan N. Lawrence, David Hassell, Grenville Lister, Predoi Valeriu, Scott Davidson, Mark Goddard, Matt Pryor, Stig Telfer, Konstantinos Chasapis, and Jean-Thomas Acquaviva

Active storage (also known as computational storage) has been a concept often proposed but not often delivered. The idea is that there is a lot of under-utilised compute power in modern storage systems, and this could be utilised to carry out some parts of data analysis workflows. Such a facillity would reduce the cost of moving data, and make distributed data analysis much more efficient.

For storage to be able to handle compute, either an entire compute stack has to be migrated to the storage (with all the problems around security and dependencies) or the storage has to offer suitable compute interfaces. Here we take the second approach, borrowing the concept of providing system reduction operations in the MPI interface of HPC systems, to define and implement a reduction interface for the complex layout of HDF5 (and NetCDF4) data.

We demonstrate a near-production quality deployment of the technology (PyActiveStorage) fronting JASMIN object storage, and describe how we have built a POSIX prototype. The first provides compute “near” the storage, the second is truly “in” the storage. The performance with the object store is such that for some tasks distributed workflows based on reduction operations on HDF5 data can be competitive with local workflow speeds, a result which has significant implications for avoiding expensive copies of data and unnecessary data movement. As a byproduct of this work, we have also upgraded a pre-existing pure python HDF5 reader to support lazy access, which opens up threadsafe read operations on suitable HDF5 and NetCDF4 data.

To our knowledge, there has previously been no previous practical demonstration of active storage for scientific data held in HDF5 files. While we have developed this technology with application in distributed weather and climate workflows, we believe it will find utility in a wide range of scientific workflows.

How to cite: Lawrence, B. N., Hassell, D., Lister, G., Valeriu, P., Davidson, S., Goddard, M., Pryor, M., Telfer, S., Chasapis, K., and Acquaviva, J.-T.: PyActiveStorage:  Efficient distributed data analysis using Active Storage for HDF5/NetCDF4, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6544, https://doi.org/10.5194/egusphere-egu25-6544, 2025.

EGU25-7056 | Orals | ESSI3.3

Reliable and reproducible Earth System Model data analysis with ESMValTool 

Valeriu Predoi and Bouwe Andela

ESMValTool is a software tool for analyzing data produced by Earth System Models (ESMs) in a reliable and reproducible way. It provides a large and diverse collection of “recipes” that reproduce standard, as well as state-of-the-art analyses. ESMValTool can be used for tasks ranging from monitoring continuously running ESM simulations to analysis for scientific publications such as the IPCC reports, including reproducing results from previously published scientific articles as well as allowing scientists to produce new analysis results. To make ESMValTool a user-friendly community tool suitable for doing open science, it adheres to the FAIR principles for research software. It is: - Findable - it is published in community registries, such as https://research-software-directory.org/software/esmvaltool; - Accessible - it can be installed from Python package community distribution channels such as conda-forge, and the open-source code is available on Zenodo with a DOI, and on GitHub; - Interoperable - it is based on standards: it works with data that follows CF Conventions and the Coupled Model Intercomparison Project (CMIP) Data Request, its reusable recipes are written in YAML, and provenance is recorded in the W3C PROV format. It supports diagnostics written in a number of programming language, with Python and R being best supported. Its source code follows the standards and best practices for the respective programming languages; - Reusable - it provides a well documented recipe format and Python API that allow reusing previous analyses and building new analysis with previously developed components. Also, the software can be installed from conda-forge and DockerHub and can be tailored by installing from source from GitHub. In terms of input data, ESMValTool integrates well with the Earth System Grid Federation (ESGF) infrastructure. It can find, download and access data from across the federation, and has access to large pools of observational datasets. ESMValTool is built around two key scientific software metrics: scalability and user friendliness. An important aspect of user friendliness is reliability. ESMValTool is built on top of the Dask library to allow scalable and distributed computing, ESMValTool also uses parallelism at a higher level in the stack, so that jobs can be distributed on any standard High Performance Computing (HPC) facility; and software reliability and reproducibility - our main strategy to ensure reliability is modular, integrated, and tested design. This comes back at various levels of the tool. We try to separate commonly used functionality from “one off” code, and make sure that commonly used functionality is covered by unit and integration tests, while we rely on regression testing for everything else. We also use comprehensive end-to-end testing for all our “recipes” before we release new versions. Our testing infrastructure ranges from basic unit tests to tools that smartly handle various file formats, and use image comparison algorithms to compare figures. This greatly reduces the need for ‘human testing’, allowing for built-in robustness through modularity, and a testing strategy that has been tailored to match the technical skills of its contributors.

How to cite: Predoi, V. and Andela, B.: Reliable and reproducible Earth System Model data analysis with ESMValTool, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7056, https://doi.org/10.5194/egusphere-egu25-7056, 2025.

EGU25-7070 | Posters on site | ESSI3.3

EarthCODE - a FAIR and Open Environment for collaborative research in Earth System Science  

Chandra Taposeea-Fisher, Garin Smith, Ewelina Dobrowolska, Daniele Giomo, Francesco Barchetta, Stephan Meißl, and Dean Summers

The Open Science and Innovation Vision included in ESA’s EO Science Strategy (2024) addresses 8 key elements: 1) openness of research data, 2) open-source scientific code, 3) open access papers with data and code; 4) standards-based publication and discovery of scientific experiments, 5) scientific workflows reproducible on various infrastructures, 6) access to education on open science, 7) community practice of open science; and 8) EO business models built on open-source. EarthCODE (https://earthcode.esa.int) is a strategic ESA EO initiative to support the implementation of this vision. 

EarthCODE (Earth Science Collaborative Open Development Environment) will form part of the next generation of cloud-based geospatial services, aiming towards an integrated, cloud-based, user-centric development environment for European Space Agency’s (ESA) Earth science activities. EarthCODE looks to maximise long-term visibility, reuse and reproducibility of the research outputs of such projects, by leveraging FAIR and open science principles and enabling, thus fostering a sustainable scientific process. EarthCODE proposes a flexible and scalable architecture developed with interoperable open-source blocks, with a long-term vision evolving by incrementally integrating industrially provided services from a portfolio of the Network of Resources.  Additionally, EarthCODE is a utilisation domain of EOEPCA+, contributing to the development and evolution of Open Standards and protocols, enabling internationally interoperable solutions.  

EarthCODE will provide an Integrated Development Platform, giving developers tools needed to develop high quality workflows, allowing experiments to be executed in the cloud and be end-to-end reproduced by other scientists. EarthCODE is built around existing open-source solutions, building blocks and platforms, such as the Open Science Catalogue, EOxHub and EOEPCA. It has additionally begun to integrate platform services from DeepESDL, Euro Data Cube, Polar TEP and the openEO federation on CDSE platforms, with more being added annually through ESA best practices. With it’s adopted federated approach, EarthCODE will facilitate processing on other platforms, i.e. DeepESDL, ESA EURO Data Cube, Open EO Cloud/Open EO Platform and AIOPEN/AI4DTE.   

The roadmap for the portal includes the initial portal release by end of 2024, followed by the capability to publish experiments in Q1 2025 (including development, publishing, finding and related community engagement), and by mid-2025 to have a further release with reproducibility capabilities around accessibility and execute functionalities.  

Collaboration and Federation are at the heart of EarthCODE. As EarthCODE evolves we expect providing solutions allowing federation of data and processing. EarthCODE has ambition to deliver a model for a Collaborative Open Development Environment for Earth system science, where researchers can leverage the power of the wide range of EO platform services available to conduct their science, while also making use of FAIR Open Science tools to manage data, code and documentation, create end-to-end reproducible workflows on platforms, and have the opportunity to discover, use, reuse, modify and build upon the research of others in a fair and safe way. Overall, EarthCODE aims to enable elements for EO Open Science and Innovation vision, including open data, open-source code, linked data/code, open-access documentation, end-to-end reproducible workflows, open-science resources, open-science tools, and a healthy community applying all the elements in their practice.

How to cite: Taposeea-Fisher, C., Smith, G., Dobrowolska, E., Giomo, D., Barchetta, F., Meißl, S., and Summers, D.: EarthCODE - a FAIR and Open Environment for collaborative research in Earth System Science , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7070, https://doi.org/10.5194/egusphere-egu25-7070, 2025.

EGU25-8114 | ECS | Orals | ESSI3.3

Flexible and scalable workflow framework HydroFlows for compound flood risk assessment and adaptation modelling 

Willem Tromp, Dirk Eilander, Hessel Winsemius, Tjalling De Jong, Brendan Dalmijn, Hans Gehrels, and Bjorn Backeberg

Flood risk assessments are increasingly guiding urban developments to safeguard against flooding. These assessments, consisting mainly of hazard and risk maps, make use of interconnected models consisting of a chain of climate, hydrological, hydraulic, and impact models, which are increasingly run interactively to support scenario modelling and decision-making in digital twins. To maintain interoperability, transparency, and reusability of this chain and the assessments themselves, using a workflow manager to manage the inter-model dependencies is a natural fit. However, composing and maintaining workflows is a non-trivial, time-consuming task, and they often have to be refactored for new workflow engines, or when changing compute environments, even if the workflow conceptually remains unchanged. These issues are particularly relevant in the development of digital twins for climate adaptation, where flood risk assessments serve as input to indicate high-risk areas. The complex model chain underpinning such digital twins can benefit greatly from transparent workflows that can be easily reused across different contexts.

To address these challenges, we developed the HydroFlows Python framework for composing and maintaining flood risk assessment workflows by leveraging common patterns identified across different workflows. The framework allows users to use one of the many steps available in the library or define workflow steps themselves and combine these into complete workflows which are validated on the fly. Available workflow steps include building, running, and postprocessing of models. Execution of the workflow is handled by one of the workflow managers to which our workflow description can be exported, such as Snakemake or tools with CWL support. This flexibility allows users to easily scale their workflows to different compute environments whenever the computational requirements demand so.

We demonstrate the flexibility of the HydroFlows framework by highlighting how it can be used to create complex workflows needed for digital twins supporting climate adaptation. HydroFlows not only enhances the flexibility and portability of the digital twin modelling workflows but also facilitates the integration of digital twin tooling and advanced computing and processing solutions to support interactive flood risk assessments in federated compute and data environments.

How to cite: Tromp, W., Eilander, D., Winsemius, H., De Jong, T., Dalmijn, B., Gehrels, H., and Backeberg, B.: Flexible and scalable workflow framework HydroFlows for compound flood risk assessment and adaptation modelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8114, https://doi.org/10.5194/egusphere-egu25-8114, 2025.

EGU25-8305 | ECS | Posters on site | ESSI3.3

Enabling reliable workflow development with an advanced Testing Suite 

Alejandro Garcia Lopez, Leo Arriola Meikle, Gilbert Montane Pinto, Miguel Castrillo, Bruno de Paula Kinoshita, Eric Ferrer Escuin, and Aina Gaya Avila

Climate simulations require complex workflows that often integrate multiple components and different configurations per experiment, typically involving high-performance computing resources. The exhaustive testing required for these workflows can be time and resource consuming, presenting significant challenges in terms of computational cost and human effort. However, robust Continuous Integration (CI) testing ensures the reliability and reproducibility of such complex workflows by validating the codebase and ensuring the integrity of all the components used when performing climate simulations. Additionally, CI testing facilitates both major and minor releases, enhancing the efficiency of the development lifecycle.

To address these challenges, we present our Testing Suite software, designed to automate the setup, configuration, and execution of integration tests using Autosubmit, a workflow manager developed at the BSC. Autosubmit is typically used for climate modelling experiments, but also atmospheric composition ones, and also constitutes the backbone of some operational systems and Digital Twin initiatives. The Testing Suite software allows Autosubmit commands to be executed in batches and the responses from the Workflow Manager to be bypassed in a structured manner. By streamlining this process, it minimizes the effort required for exhaustive testing while ensuring reliability.

Beyond integration testing, the Testing Suite offers advanced capabilities for scientific result verification. By automatically comparing output data bit by bit, it swiftly detects regressions during test execution. Additionally, it provides CPMIP performance metrics, offering insights into the efficiency of the workflows.

As a result, the Testing Suite plays an important role in quality assurance, particularly during releases, where extensive testing ensures the workflow meets required functionality and performance standards across different configurations. These integration tests act as a checkpoint, validating the stability and robustness of the software before release. They also identify stable points in the main codebase, enabling developers to create new branches with confidence. This approach minimizes compatibility issues and facilitates a smoother development process.

In conclusion, the Testing Suite is a crucial part of the development lifecycle for climate simulations. It mitigates risks, ensures stability, and fosters innovation, all while maintaining a robust and reliable foundation for scientific research and development.

How to cite: Garcia Lopez, A., Arriola Meikle, L., Montane Pinto, G., Castrillo, M., de Paula Kinoshita, B., Ferrer Escuin, E., and Gaya Avila, A.: Enabling reliable workflow development with an advanced Testing Suite, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8305, https://doi.org/10.5194/egusphere-egu25-8305, 2025.

EGU25-8621 | ECS | Posters on site | ESSI3.3

Auto-EC-Earth: An automatic workflow to manage climate modelling experiments using Autosubmit 

Eric Ferrer, Gilbert Montane, Miguel Castrillo, and Alejandro Garcia

The European community Earth system model EC-Earth is based on different and interoperable climate components simulating different processes of the Earth system. This makes it a complex model that requires multiple input data sources for its various model components, which can be run in parallel with multiple configurations and resolutions, demanding different computational resources in each case.

The EC-Earth software contains a minimum set of scripts to manage the compilation and execution of the simulations, but these are not enough to perform all the tasks that experiments demand nor to guarantee the traceability and reproducibility of the entire workflow in a high-productivity scientific environment. For that matter, the Auto-EC-Earth software has been developed at the Earth Sciences department of the Barcelona Supercomputing Center (BSC-ES) relying on Autosubmit, a workflow manager also developed at BSC-ES.

We take advantage of the automatization provided by the workflow manager that allows us to configure, manage, orchestrate and share experiments with different configurations and target platforms. The workflow manager allows the user to split the run into different tasks that are executed on different local and remote machines, like the HPC platform where the simulation needs to be performed. This is achieved in a seamless integration between Autosubmit, the EC-Earth tools, and the different machines where the scripts run, all without any user-input required after the initial setup and the launch of the experiment thanks to the workflow developments. Autosubmit also allows to ensure traceability of the actual runs, to have all the required data available for different kinds of experiments separated and well documented.

However, running the main part of the simulation is a cooperative task between the Autosubmit workflow manager and the different tools used for each model version. Auto-EC-Earth workflow has evolved to adapt the best possible to the EC-Earth model scripts that are present to help with the model runs. In EC-Earth 4, ScriptEngine is used to manage the run, and it has been fully integrated into the Auto-EC-Earth 4 workflow and used to set up the environment, while Autosubmit still manages the submission of jobs to the HPC and the dependencies between them.

Auto-EC-Earth is a great example of a workflow system that has been developed and used throughout the years, well established within the BSC-ES and used in multiple production cases, like multiple CMIP exercises as well as a reference for newer ESM workflows like the one developed in the Destination Earth project. It has also allowed the BSC-ES to collaborate with the EC-Earth community through the testing of the new releases of the model.

How to cite: Ferrer, E., Montane, G., Castrillo, M., and Garcia, A.: Auto-EC-Earth: An automatic workflow to manage climate modelling experiments using Autosubmit, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8621, https://doi.org/10.5194/egusphere-egu25-8621, 2025.

EGU25-9175 | Posters on site | ESSI3.3

Enhancing Earth system models efficiency: Leveraging the Automatic Performance Profiling tool 

Roc Salvador Andreazini, Xavier Yepes Arbós, Stella Valentina Paronuzzi Ticco, Oriol Tintó Prims, and Mario Acosta Cobos

Earth system models (ESMs) are essential to understand and predict climate variability and change. However, their complexity and computational demands of high-resolution simulations often lead to performance bottlenecks that can impede research progress. Identifying and resolving these inefficiencies typically require significant expertise and manual effort, posing challenges for both climate scientists and High Performance Computing (HPC) engineers.

We propose automating performance profiling as a solution to help researchers concentrate on improving and optimizing their models without the complexities of manual profiling. The Automatic Performance Profiling (APP) tool brings this solution to life by streamlining the generation of detailed performance reports for climate models.

The tool ranges from high-level performance metrics, such as Simulated Years Per Day (SYPD), to low-level metrics, such as PAPI counters and MPI communication statistics. This dual-level reporting makes the tool accessible to a wide range of users, from climate scientists seeking a general understanding of the model efficiency, to HPC experts requiring granular insights for advanced optimizations.

Seamlessly integrated with Autosubmit, the workflow manager developed at the Barcelona Supercomputing Center (BSC), APP ensures compatibility with complex climate modelling workflows. By automating the collection and reporting of key metrics, APP reduces the effort and expertise needed for performance profiling, empowering users to enhance the scalability and efficiency of their climate models.

APP currently supports multiple models, including the EC-Earth4 climate model and the NEMO ocean model, and is compatible with different HPC systems, such as Marenostrum 5 and ECMWF’s supercomputer. Furthermore, the modular design of the tool allows adding new models and HPC platforms easily.

How to cite: Salvador Andreazini, R., Yepes Arbós, X., Paronuzzi Ticco, S. V., Tintó Prims, O., and Acosta Cobos, M.: Enhancing Earth system models efficiency: Leveraging the Automatic Performance Profiling tool, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9175, https://doi.org/10.5194/egusphere-egu25-9175, 2025.

Geo-simulation experiments (GSEs) are experiments allowing the simulation and exploration of Earth’s surface (such as hydrological, geomorphological, atmospheric, biological, and social processes and their interactions) with the usage of geo-analysis models (hereafter called ‘models’). Computational processes represent the steps in GSEs where researchers employ these models to analyze data by computer, encompassing a suite of actions carried out by researchers. These processes form the crux of GSEs, as GSEs are ultimately implemented through the execution of computational processes. Recent advancements in computer technology have facilitated sharing models online to promote resource accessibility and environmental dependency rebuilding, the lack of which are two fundamental barriers to reproduction. In particular, the trend of encapsulating models as web services online is gaining traction. While such service-oriented strategies aid in the reproduction of computational processes, they often ignore the association and interaction among researchers’ actions regarding the usage of sequential resources (model-service resources and data resources); documenting these actions can help clarify the exact order and details of resource usage. Inspired by these strategies, this study explores the organization of computational processes, which can be extracted with a collection of action nodes and related logical links (node-link ensembles). The action nodes are the abstraction of the interactions between participant entities and resource elements (i.e., model-service resource elements and data resource elements), while logical links represent the logical relationships between action nodes. In addition, the representation of actions, the formation of documentation, and the reimplementation of documentation are interconnected stages in this approach. Specifically, the accurate representation of actions facilitates the correct performance of these actions; therefore, the operation of actions can be documented in a standard way, which is crucial for the successful reproduction of computational processes based on standardized documentation. Aprototype system is designed to demonstrate the feasibility and practicality of the proposed approach. By employing this pragmatic approach, researchers can share their computational processes in a structured and open format, allowing peer scientists to re-execute operations with initial resources and reimplement the initial computational processes of GSEs via the open web.

How to cite: Zhu, Z. and Chen, M.: Reproducing computational processes in service-based geo-simulation experiments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9791, https://doi.org/10.5194/egusphere-egu25-9791, 2025.

EGU25-10981 | Orals | ESSI3.3

yProv: a Software Ecosystem for Multi-level Provenance Management and Exploration in Climate Workflows 

Fabrizio Antonio, Gabriele Padovani, Ludovica Sacco, Carolina Sopranzetti, Marco Robol, Konstantinos Zefkilis, Nicola Marchioro, and Sandro Fiore

Scientific workflows and provenance are two faces of the same medal. While the former addresses the coordinated execution of multiple tasks over a set of computational resources, the latter relates to the historical record of data from its original sources. As experiments rapidly evolve towards complex end-to-end workflows, handling provenance at different levels of granularity and during the entire analytics workflow lifecycle is key for managing lineage information related to large-scale experiments in a flexible way as well as enabling reproducibility scenarios, thus playing a relevant role in Open Science.

The contribution highlights the importance of tracking multi-level provenance metadata in complex, AI-based scientific workflows as a way to foster documentation of data and experiments in a standardized format, strengthen interpretability, trustworthiness and authenticity of the results, facilitate performance diagnosis and troubleshooting activities, and advance provenance exploration. More specifically, the contribution introduces yProv, a joint research effort between CMCC and University of Trento targeting multi-level provenance management in complex, AI-based scientific workflows. The yProv project provides a rich software ecosystem consisting of a web service (yProv service) to store and manage provenance documents compliant with the W3C PROV family of standards, two libraries to track provenance in scientific workflows at different levels of granularity with a focus on AI models training (yProv4WFs and yProv4ML), and a data science tool for provenance inspection, navigation, visualization, and analysis (yProv Explorer). Activity on trustworthy provenance with yProv is also ongoing to fully address end-to-end provenance management requirements.

The contribution will cover the presentation of the yProv software ecosystem and use cases from the interTwin (https://www.intertwin.eu/) and ClimateEurope2 (https://climateurope2.eu/) European projects as well as from the ICSC National Center on HPC, Big Data and Quantum Computing targeting Digital Twins for extreme weather & climate events and data-driven/data-intensive workflows for climate change. 

How to cite: Antonio, F., Padovani, G., Sacco, L., Sopranzetti, C., Robol, M., Zefkilis, K., Marchioro, N., and Fiore, S.: yProv: a Software Ecosystem for Multi-level Provenance Management and Exploration in Climate Workflows, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10981, https://doi.org/10.5194/egusphere-egu25-10981, 2025.

EGU25-11937 | Posters on site | ESSI3.3

DAM2 — A Scalable and Compliant Solution for Managing enriched Infrared images as FAIR Research Data  

Jean Dumoulin, Thibaud Toullier, Nathanael Gey, and Mathias Malandain

Abstract

Efficient and secure dataset management is a critical component of collaborative research projects, where diverse data types, sharing requirements, and compliance regulations converge. This work presents a dataset management tool entitled DAM2 (Data and Model Monitoring) developed within the Chips Joint Undertaking (Chips JU) funded European BRIGHTER project [1], to address these challenges. It provides a robust and adaptable solution for handling private and public ground based measurements datasets throughout the project lifecycle. These datasets combine infrared images (e.g. multispectral ones), with visible images, local weather measurements, labeled data, etc.

The tool is designed to ensure rights management, enabling selective data sharing among authorized partners based on predefined permissions. It incorporates secure access controls to safeguard sensitive data and meets GDPR (General Data Protection Regulation) requirements to guarantee compliance with European privacy standards. For public datasets, the tool integrates with Zenodo, an open-access repository, to support long-term storage and accessibility, aligning with the principles of open science. Key technical features include the usage of an open source, S3 compatible object storage server (MinIO [2]) providing scalability to manage large volumes of data. Additionally, the use of Zarr [3] data format behind the scene offers significant advantages for this cloud-based data management tool, including efficient storage of large datasets through chunking and compression, fast parallel read and write operations, and compatibility with a wide range of data analysis tools. The tool adheres to FAIR (Findable, Accessible, Interoperable, Reusable) principles, storing metadata alongside datasets to enhance usability and interoperability.

Developed as an open-source platform, the tool promotes transparency and collaboration while providing a complete and well-documented API for seamless integration with other systems. A user-friendly interface ensures accessibility for stakeholders with varying technical expertise, while the tool remains flexible to accommodate additional file formats as required. The development process incorporates insights from relevant COFREND (French Confederation for Non-Destructive Testing) working groups, to ensure alignment with broader initiatives in data management, interoperability and durability.

This paper addresses the design, study and developed platform. First operational functionalities are demonstrated through the manipulation of first BRIGHTER and other research project datasets.

In conclusion, DAM2 is a comprehensive solution for managing diverse datasets in collaborative projects, balancing security, compliance, and accessibility. It provides a foundation for efficient, compliant, and interoperable data handling while supporting the principles of open science and FAIR data management.

Perspectives include expanding interoperability with additional repositories, incorporating advanced analytic and visualization features, and integrating AI-driven automation.

Acknowledgments

Authors would like to acknowledge the BRIGHTER HORIZON project. BRIGHTER has received funding from the Chips Joint Undertaking (JU) under grant agreement No 101096985. The JU receives support from the European Union’s Horizon Europe research and innovation program and France, Belgium, Portugal, Spain, Turkey.

References

[1] Brighter --- Project-Brighter. https://project-brighter.eu/, accessed on January 2025.

[2] MinIO, Inc. MinIO S3 Compatible Storage for AI --- Min.Io. https://min.io/, accessed on January, 2025.

[3] Zarr --- Zarr.dev. https://zarr.dev/, accessed on January, 2025.

How to cite: Dumoulin, J., Toullier, T., Gey, N., and Malandain, M.: DAM2 — A Scalable and Compliant Solution for Managing enriched Infrared images as FAIR Research Data , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11937, https://doi.org/10.5194/egusphere-egu25-11937, 2025.

EGU25-13604 | ECS | Orals | ESSI3.3

Streamlining configurations of process-based models through extensible and free workflows 

Kasra Keshavarz, Alain Pietroniro, Darri Eythorsson, Mohamed Ismaiel Ahmed, Paul Coderre, Wouter Knoben, Martyn Clark, and Shervan Gharari

High-resolution and high-complexity process-based hydrological models play a pivotal role in advancing our understanding and prediction of water cycle dynamics, particularly in ungauged basins and under nonstationary climate conditions. However, the configuration, application, and evaluation of these models are often hindered by the intricate and inconsistent nature of a priori information available in various datasets, necessitating extensive preprocessing steps. These challenges can limit the reproducibility, applicability, and accessibility of such models for the broader scientific user community. To address these challenges, we introduce our generalized Model-Agnostic Framework (MAF), aimed at simplifying the configuration and application of data-intensive process-based hydrological models. Through a systematic investigation of commonly used models and their configuration procedures, we provide workflows designed to streamline the setup process for this category of hydrological models. Building on earlier efforts, this framework adheres to the principle of separating model-agnostic and model-specific tasks in the setup procedure of such models. The model-agnostic workflows focus on both dynamic datasets (e.g., meteorological data) and static datasets (e.g., land-use maps), while the model-specific components feed preprocessed, relevant data to the hydrological models of interest. Our initial prototypes of MAF includes recipes for various static and dynamic datasets and also tailored model-specific workflows for MESH, SUMMA, and HYPE process-based modelling frameworks. We demonstrate the effectiveness of these novel workflows in reducing configuration complexity and enhancing the reproducibility of process-based hydrological models through test applications in high-performance computing environments. The framework automates numerous manual tasks, significantly saving time, and enabling continuity in research efforts. Moreover, by minimizing human error and enhancing reproducibility, this research has fostered collaboration with several Canadian government entities, leveraging sophisticated process-based models to address complex environmental challenges.

How to cite: Keshavarz, K., Pietroniro, A., Eythorsson, D., Ahmed, M. I., Coderre, P., Knoben, W., Clark, M., and Gharari, S.: Streamlining configurations of process-based models through extensible and free workflows, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13604, https://doi.org/10.5194/egusphere-egu25-13604, 2025.

EGU25-18040 | ECS | Posters on site | ESSI3.3

Workflows for numerical reproducibility in the OceanVar data assimilation model 

Francesco Carere, Francesca Mele, Italo Epicoco, Mario Adani, Paolo Oddo, Eric Jansen, Andrea Cipollone, and Ali Aydogdu

Numerical reproducibility is a crucial yet often overlooked challenge in ensuring the credibility of computational results and the validity of Earth system models. In large-scale, massively parallel simulations, achieving numerical reproducibility is complicated by factors such as heterogeneous HPC architectures, floating point intricacies, complex hardware/software dependencies, and the non-deterministic nature of parallel execution.

This work addresses the challenges of debugging and ensuring bitwise reproducibility (BR) in parallel simulations, specifically for the MPI-parallelised OceanVar data assimilation model. We explore methods for detecting and resolving BR-related bugs, focusing on an automated debugging process. Currently mature tools to automate this process are lacking for bugs due to MPI-parallelisation, making automatic BR verification in scientific workflows involving such codebases a time-consuming challenge.

However, BR is sometimes considered unrealistic in workflows involving heterogeneous computing architectures. As an alternative, statistical reproducibility (SR) is proposed and explored by various research groups in the Earth system modelling community, for which automated tools have been developed. For example, the scientific workflow of CESM supports automatic verification of SR using the CESM-ECT framework/PyCECT software. In case of failure of SR a root-cause analysis tool exists, CESM-RUANDA, albeit currently not fully functional. We explore SR as an alternative and complementary approach to of BR focusing on its potential to support numerical reproducibility in workflows involving heterogeneous computing architectures.

How to cite: Carere, F., Mele, F., Epicoco, I., Adani, M., Oddo, P., Jansen, E., Cipollone, A., and Aydogdu, A.: Workflows for numerical reproducibility in the OceanVar data assimilation model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18040, https://doi.org/10.5194/egusphere-egu25-18040, 2025.

EGU25-18890 | Posters on site | ESSI3.3

Research data management for numerical simulations in Earth-System Science 

Klaus Getzlaff and Markus Scheinert

One of today's challenges is the effective access to scientific data either within research groups or across different institutions to increase the reusability of the data and therefore their value. While large operational modeling and service centers have enabled query and access to data via common web services, this is often not the case for smaller institutions or individual research groups. Especially the maintenance of the infrastructure and the simplicity of the workflows, in order to make the data and their provenance available and accessible, are common challenges for scientists and data management.

At GEOMAR there are several data steward positions to support RDM for special disciplines and formats. They are also connected across centres to work on common standards, e.g. the netcdf standard working group in the Helmholtz Earth and Environment DataHUB.

Here we will present the institutional approach on research data management for numerical simulations in earth system science. The data handling, especially the possibilities for data sharing, publication and access, which is in today’s focus, is realized by using persistent identifier handles in combinations with a modern http web server index solution and a THREDDS server allowing remote access using standardized protocols such as OPeNDAP, WMS. By cross-linking this into the central institutional metadata and publication repositories it allows the re-usability of the data by scientists from different research groups and backgrounds. In addition to the pure data handling the documentation of the numerical simulation experiments is of similar importance to allow re-usability or reproducibility and to provide the data which will be addressed too.

How to cite: Getzlaff, K. and Scheinert, M.: Research data management for numerical simulations in Earth-System Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18890, https://doi.org/10.5194/egusphere-egu25-18890, 2025.

EGU25-19655 | Posters on site | ESSI3.3

Multi-faceted habitat connectivity: how to orchestrate remote sensing with citizen science data? 

Ivette Serral, Vitalii Kriukov, Lucy Bastin, Riyad Rahman, and Joan Masó

In the era of declining biodiversity, global climate change and transformations in land use, terrestrial habitat connectivity is one of the key parameters of ecosystem management. In this regard, the land-use/land-cover (LULC) dynamics is crucial to detect the spatiotemporal trends in connectivity of focal endangered species and to predict the effects for biodiversity for planned or proposed LULC changes.

Apart from the LULC derivatives of remote sensing, connectivity analysis and scenarios modelling can also benefit from citizen science datasets, such as Open Street Map and GBIF species occurrence data cubes in which aggregated data can be perceived as a cube with three dimensions - taxonomic, temporal and geographic. The synthetic LULC datasets which cover Catalonia every 5 years (1987-2022) were enriched via developed Data4Land harmonisation tool harnessing Open Street Map (through Overpass Turbo API) and World Database on Protected Areas. Two outstanding well-known tools, Graphab and MiraMon GIS&RS (using the Terrestrial Connectivity Index Module - ICT), were used to create the overarching dataset on terrestrial habitat connectivity in Catalonia (2012-2022) for target species and broad land cover categories, forests. Significant decline trends in forest habitat connectivity are observed for Barcelona metropolitan area, and vice versa in the Pyrenees mountain corridor and protected areas. According to the local case study on the connectivity of Mediterranean turtle in the Albera Natural Park, general positive trend was affected by massive fires in 2012.

To ensure the replicable results, the pipeline to create reliable metadata in accordance with FAIR principles, especially data lineage, is being developed, as well as the high performance computing pipeline for Graphab.

How to cite: Serral, I., Kriukov, V., Bastin, L., Rahman, R., and Masó, J.: Multi-faceted habitat connectivity: how to orchestrate remote sensing with citizen science data?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19655, https://doi.org/10.5194/egusphere-egu25-19655, 2025.

EGU25-21553 | Posters on site | ESSI3.3

European Digital Twin of the Ocean: the integration with EuroHPC platforms 

Stella Valentina Paronuzzi Ticco, Simon Lyobard, Mathis Bertin, Quentin Gaudel, Jérôme Gasperi, and Alain Arnaud

The EDITO platform serves as the foundational framework for building the European Digital Twin of the Ocean, seamlessly integrating oceanographic data, processes and services on a single and comprehensive platform. The platform provides scalable computing resources interconnected with EuroHPC supercomputing centers. We have developed a mechanism that allows users to remotely execute functions (processes) on HPCs and store the resulting output at the location of their choice (e.g. EDITO personal storage, third parties S3 buckets, etc.). This output can then be leveraged as input for subsequent processes, fostering a streamlined and interconnected workflow. Our presentation will delve into the technical details to achieve such an integration between cloud and HPC systems. 

How to cite: Paronuzzi Ticco, S. V., Lyobard, S., Bertin, M., Gaudel, Q., Gasperi, J., and Arnaud, A.: European Digital Twin of the Ocean: the integration with EuroHPC platforms, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21553, https://doi.org/10.5194/egusphere-egu25-21553, 2025.

EGU25-74 | ECS | PICO | ESSI3.4

An open-source, QGIS-based solution for digital geological mapping: GEOL-QMAPS 

Julien Perret, Mark W. Jessell, and Eliott Bétend

Digital geological mapping has progressed significantly with the advent of commercial GIS, GPS technologies, and portable devices over the past three decades. However, many software and app tools that enhance field data collection remain proprietary and specific to mapping projects or organisations, which limits their integration and sharing within the geoscientific community. This presentation will introduce GEOL-QMAPS, an open-source, QGIS-based solution designed for flexible and harmonised digital geological mapping, developed as part of the West African eXploration Initiative 4. GEOL-QMAPS includes a QGIS field data entry template and a custom QGIS plugin, both available on open-access online repositories. It supports fieldwork on tablets or mobile devices via QField app, integrates with desktop QGIS, and facilitates the creation of new and legacy field databases according to user-defined guidelines. The presentation will cover the general workflow for implementing GEOL-QMAPS and provide examples demonstrating its effectiveness both in the field and in office settings.

How to cite: Perret, J., Jessell, M. W., and Bétend, E.: An open-source, QGIS-based solution for digital geological mapping: GEOL-QMAPS, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-74, https://doi.org/10.5194/egusphere-egu25-74, 2025.

Spatial heterogeneity (SH), known as the second law of geography, has been a topic of extensive research. One common approach to analyzing SH involves comparing variances between and within strata to assess the impact of independent variables on the dependent variable. This method, known as spatial stratified heterogeneity (SSH) analysis, is often performed using the geographical detector model. Over time, several optimized versions of geographical detectors have emerged, focusing on discretizing single or dual variables. However, methods for discretizing three or more variables are still limited to the interaction detector, with research on spatial scale effects mainly focused on single factors. To overcome these limitations, an optimal multivariate-stratification geographical detector (OMGD) model has been developed. This model includes two additional modules: factor discretization optimization and scale detector. Fine-tuning factor discretization involves using five univariate and five cluster-based stratification methods to automatically explore the optimal discretization scheme for single factors or multi-factor combinations based on the Geodetector 𝑞 statistics. The scale detector can then iterate through various spatial scales to identify the optimal spatial scale for SSH analysis. Furthermore, the developed OMGD model has been tested with multiple case datasets to validate its applicability and robustness. The findings demonstrate that the OMGD model can effectively extract the main attributes of single factors and multi-factor combinations, providing a better explanation for geographical phenomena. It can also automatically determine the best spatial scale for SSH analysis, thereby enhancing the overall capability of conducting SSH analysis with the geographical detector.

How to cite: Guo, Y.: An optimal multivariate-stratification geographical detector model for revealing the impact of multi-factor combinations on the dependent variable, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-507, https://doi.org/10.5194/egusphere-egu25-507, 2025.

EGU25-1311 | PICO | ESSI3.4

Open-source 3D tools developed for the CO3D mission and beyond 

David Youssefi, Valentine Bellet, Alexandre Constantin, Dimitri Lallement, Emmanuel Dubois, and Yannick Tanguy
After several years of development, mid-2025 should see the launch of the CO3D1 (Optical Constellation in 3D) mission. The CO3D project is a public-private partnership between CNES (French space agency) and AIRBUS (aerospace company).

The mission aims to launch pairs of optical satellites to photogrammetrically reconstruct the Digital Surface Model (DSM) of the Earth's land surface. A notable innovation is the ability to manage stereo acquisition in synchronous mode for capturing moving elements.

As part of the CO3D program, CNES is developing the 2D and 3D product processing chains integrated into the mission's ground segment, alongside an Image Calibration Center (ICC) for radiometric and geometric calibration. These systems leverage a suite of open-source tools created by CNES teams. CNES chooses to release these tools prior to the mission's launch to gather user feedback and improve the quality of final CO3D products.

  • CARS2 is the Multiview Stereo Framework (MVS). It produces DSM from satellite images acquired from different view angles.
  • PANDORA3 is the stereo matching framework. It combines both switchable similarity measures at the pixel level and regularizers. The CARS pipeline integrates PANDORA.
  • BULLDOZER4 removes above-ground elements (e.g., trees, buildings) from DSM generated by CARS to produce Digital Terrain Models (DTM).
  • XDEM5 evaluates and validates the 3D quality of the generated models (i.e., the CARS DSM and the Bulldozer DTM).
  • SLURP6 produces land cover masks from very high-resolution (VHR) images. The CARS pipeline integrates these masks to generate sharper 3D reconstruction.

CO3D mission represents a significant step forward in Earth observation, offering innovative tools for producing accurate digital surface models. By engaging with the user community early, CNES aims to ensure the mission delivers results that fit user needs.

[1] Lebegue, L. et al. (2024). CO3D Products Qualification Forecast. IGARSS 2024 - 2024 IEEE International Geoscience and Remote Sensing Symposium, 8555-8559.
[2] Youssefi, D. et al. (2020). CARS: A Photogrammetry Pipeline Using Dask Graphs to Construct A Global 3D Model. IGARSS 2020 - 2020 IEEE International Geoscience and Remote Sensing Symposium, 453-456.
[3] Cournet, M. et al. (2020). Ground-truth generation and disparity estimation for optical satellite imagery. ISPRS - International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences.
[4] Lallement, D. et al. (2023). Bulldozer, a free open source scalable software for DTM extraction. The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences.
[5] Hugonnet, R. et al. (2022). Uncertainty Analysis of Digital Elevation Models by Spatial Inference From Stable Terrain. IEEE Journal Selected Topics in Applied Earth Observations and Remote Sensing, 15, 6456–6472.
[6] Tanguy, Y. et al. (2024). Smart Land Use Masks: A Simple and Robust Approach to Produce Low/High Vegetation Masks from a Single High Resolution Satellite Image. IGARSS 2024 - 2024 IEEE International Geoscience and Remote Sensing Symposium, 4164-4168.

How to cite: Youssefi, D., Bellet, V., Constantin, A., Lallement, D., Dubois, E., and Tanguy, Y.: Open-source 3D tools developed for the CO3D mission and beyond, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1311, https://doi.org/10.5194/egusphere-egu25-1311, 2025.

EGU25-8703 | ECS | PICO | ESSI3.4

OceanUCA: Enhancing Coastal Observation and Forecasting in Andalucía (Spain) through Data Visualisation and Communication 

Juan Carbone, Jerimar Vasquez-Rojas, Alfredo Izquierdo, Javier Benavente, Jesús Gómez-Enri, Tomás Fernández-Montblanc, Flávio Martins, William D. Cabos Narvaez, Carlos J. González, Carlos Yagüe, Carlos Román-Cascón, and Oscar Alvarez

The OceanUCA project aims to develop an operational platform that enhances the existing tools created by the University of Cádiz, integrating new observational systems and high-resolution numerical models, both atmospheric and hydrodynamic, to achieve maximum resolution along the Andalusian coast (Spain). This platform utilizes open-source software and WebGIS services, enabling flexible and accessible geospatial data visualization and analysis for a wide range of stakeholders.

By enhancing computational resources with state-of-the-art open-source GIS tools, the platform provides solutions for addressing coastal environmental challenges such as oil spills, marine heatwaves, and extreme event tracking. The system integrates output from high-resolution oceanographic models, the atmospheric Weather Research and Forecasting (WRF) model, and observational data from various sources, including meteorological weather stations, surface fluxes instruments, oceanographic buoys, satellite imagery, high-frequency coastal radars, and data from oceanographic and atmospheric field campaigns. This data is presented through an interactive GIS interface, facilitating real-time model evaluation and decision-making.

The project also leverages high-performance computing and graphic processing units to enable the processing of large datasets, which improves the scalability and resolution of the forecasts. The platform aims to enhance environmental protection, support conservation efforts, and provide early-warning products that contribute to effective coastal management. Through its open-source approach, the system fosters collaboration and knowledge transfer with stakeholders at both national and international levels, enabling the wider community to access, engage with, and utilize coastal and atmospheric data for informed decision-making.

 

How to cite: Carbone, J., Vasquez-Rojas, J., Izquierdo, A., Benavente, J., Gómez-Enri, J., Fernández-Montblanc, T., Martins, F., Cabos Narvaez, W. D., González, C. J., Yagüe, C., Román-Cascón, C., and Alvarez, O.: OceanUCA: Enhancing Coastal Observation and Forecasting in Andalucía (Spain) through Data Visualisation and Communication, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8703, https://doi.org/10.5194/egusphere-egu25-8703, 2025.

EGU25-9616 | PICO | ESSI3.4

I/O Profiling and Optimisation to Improve Energy Efficiency: Insights from the Climate Digital Twin project 

Carlos Peña, Razvan Aguridan, Xavier Yepes, and Mario Acosta

The continuous increase in resolution and complexity of Earth System Models (ESMs), aimed at improving the accuracy of simulations, significantly increases data handling demands. For this reason, scalable and efficient I/O solutions are critical to ensure that data storage, processing, and transfer do not become bottlenecks that hinder overall simulation throughput. Optimising this workflow is essential not only for improving performance but also for reducing the energy footprint of large-scale Earth system simulations.

Many ESMs, including the IFS-NEMO coupled model used in our case study within the Destination Earth Climate Digital Twin project, adopt a client-server I/O architecture to address these challenges. In this scheme, the model sends generated data to a server that handles complex post-processing tasks, such as interpolation, encoding, and data writing, while the model continues simulating the next time steps. However, this approach also requires continuous optimisation to ensure the overall efficiency of the output pipeline.

To achieve this, detailed I/O profiling was conducted, with a focus on the MultIO library, which manages the output pipeline, including ocean server creation and ocean data transport. On the client side, frequent and costly access to metadata was identified as a major source of I/O overhead, while on the server side, high interpolation times were observed, prompting further analysis and targeted optimisations that achieved a sixfold speedup in ocean interpolation. Additional pipeline actions were reviewed and optimised, contributing to a more efficient and scalable output workflow. Combined with tests to optimise server resource allocation, these efforts resulted in an overall efficiency improvement of up to 6.7% in simulation performance.

How to cite: Peña, C., Aguridan, R., Yepes, X., and Acosta, M.: I/O Profiling and Optimisation to Improve Energy Efficiency: Insights from the Climate Digital Twin project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9616, https://doi.org/10.5194/egusphere-egu25-9616, 2025.

EGU25-12051 | PICO | ESSI3.4

Use of Open Source Software in the ESA Planetary Science Archive 

Fran Raga, Mark Bentley, Daniela Coia, Ruben Docasal, Emmanuel Grotheer, David Heather, Tania Lim, Joana Oliveira, Jose Osinde, Thomas Cornet, Jaime Saiz, and Gemma Ramos

The European Space Agency (ESA) integrates open-source software to manage, visualize, and distribute planetary data, focusing on Mars and supporting global scientific collaboration through the Planetary Science Archive (PSA). This advanced infrastructure combines cutting-edge tools and technologies to enhance data accessibility and promote international research.

Tools and Technologies

  • OpenLayers: Powers 2D interactive maps, offering scientists an intuitive interface for exploring planetary geospatial data.
  • GeoServer: Shares spatial data via standard protocols like WMS, serving base maps and integrating advanced caching techniques.
  • Three.js: Facilitates 3D visualization of celestial objects, such as comets and asteroids, providing dynamic exploration capabilities.
  • PostgreSQL and PostGIS: Stores and manages complex geospatial datasets, enabling advanced spatial queries and integration with other GIS tools.

Collaborative Efforts

ESA contributes to open-source projects to solve unique planetary data challenges:

  • Astroquery: A Python library for accessing astronomical databases, ensuring efficient integration of planetary mission data into research workflows.
  • Antimeridian: Addresses geospatial data crossing the 180° longitude line, ensuring accurate planetary mapping.

PSA Interface and GIS Architecture

The PSA’s new interface integrates 2D and 3D visualizations, data filtering, and real-time access to information. Scientists can overlay geological, topographical, and spectral data layers, analyze specific regions, and download curated datasets. The GIS architecture combines:

  • GeoServer: Distributes optimized base maps of planetary surfaces.
  • OpenLayers and Three.js: Provides a seamless 2D and 3D visualization experience.
  • PostgreSQL/PostGIS: Manages and analyzes large geospatial datasets.
  • Astroquery and Antimeridian: Enhances data accessibility and accuracy.

Benefits for the Scientific Community

This open-source approach promotes transparent, collaborative research. Tools like Antimeridian address specific planetary data issues, enabling continuous and precise geospatial representation. The PSA allows researchers to cross-reference data from various missions and instruments efficiently, accelerating planetary science advancements.

In conclusion, ESA’s adoption of open-source tools and robust GIS architecture provides an accessible, powerful platform for planetary research, fostering innovation and collaboration across the global scientific community.

How to cite: Raga, F., Bentley, M., Coia, D., Docasal, R., Grotheer, E., Heather, D., Lim, T., Oliveira, J., Osinde, J., Cornet, T., Saiz, J., and Ramos, G.: Use of Open Source Software in the ESA Planetary Science Archive, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12051, https://doi.org/10.5194/egusphere-egu25-12051, 2025.

EGU25-15828 | ECS | PICO | ESSI3.4

Modelling Forest Fire Spread in the SEQ Region Using Meteorological and Environmental Datasets: A Cellular Automaton Approach 

Harikesh Singh, Li-minn Ang, and Sanjeev Kumar Srivastava

This study develops a Cellular Automaton (CA) model to predict forest fire spread in the Sunshine Coast region, utilizing diverse meteorological and environmental datasets. Key variables such as temperature, wind speed, rainfall, soil moisture, solar radiation, vegetation type, slope, elevation, and proximity to roads and streams are integrated to simulate fire dynamics with high spatial resolution. Historical fire occurrence data are used for model calibration and validation, ensuring accuracy in replicating fire spread patterns. The CA model operates through iterative cell-based transitions, governed by rules reflecting the complex interplay of environmental and meteorological factors. Results highlight the significant influence of wind, vegetation type, and topography on fire behaviour, with simulations effectively capturing spatial variability and spread dynamics. The findings underscore the model's potential as a robust, scalable tool for wildfire management, enabling data-driven planning for prescribed burns and risk mitigation. This research offers valuable insights into forest fire behaviour, contributing to sustainable ecosystem management and resilience planning in subtropical regions such as the southeast part of Queensland Australia.

How to cite: Singh, H., Ang, L., and Srivastava, S. K.: Modelling Forest Fire Spread in the SEQ Region Using Meteorological and Environmental Datasets: A Cellular Automaton Approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15828, https://doi.org/10.5194/egusphere-egu25-15828, 2025.

EGU25-17728 | PICO | ESSI3.4

The libvips image processing library 

Kirk Martinez, John Cupitt, Lovell Fuller, and Kleis Auke Wolthuizen

Libvips is an open source image processing library originally created for museum imaging projects through a range of EU-funded projects from 1989 onwards. The challenges of processing images which were much larger than available RAM as well as coping with multi-band multi-format pixels led to an extremely efficient software design. It also makes automatic use of multi-core CPUs. Today libvips is used by many websites due to its speed and low memory use (from Wikipedia to shopify and booking.com). It is OSS Fuzz tested by Google as it is classed as essential Internet software. It has been used in many museum imaging projects to stitch X-ray images and process massive scans (e.g. The battle of Murten 1.6 TerraPixel scan). Tiled pyramidal tiff images made for multi-resolution web-browsing are easily made and handled by libvips and its viewer vipsdisp. A spreadsheet-like GUI called nip is also useful for experimenting with image processing. These features make it a useful tool for processing images in the earth sciences, especially when sizes are larger than 32 GiB when most desktop or laptop computers struggle with typical software. Python can use the library (pyvips) which makes it easy to use with other tools but can also be used from C, C++, Ruby and Javascript.

How to cite: Martinez, K., Cupitt, J., Fuller, L., and Wolthuizen, K. A.: The libvips image processing library, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17728, https://doi.org/10.5194/egusphere-egu25-17728, 2025.

Accurately monitoring terrestrial ecosystems is essential for addressing global environmental challenges, including deforestation, biodiversity loss, and climate change. NASA's Global Ecosystem Dynamics Investigation (GEDI) mission has revolutionized ecosystem monitoring by providing near-global, high-resolution data on vegetation structure and terrain elevation using spaceborne LiDAR. However, spaceborne LiDAR data often require correction due to various sources of error, such as instrument inaccuracies, atmospheric conditions (e.g., dense cloud cover), and spacecraft platform instability. A primary challenge in utilizing GEDI data is its horizontal geolocation error, which has an accuracy of approximately 10 meters for calibrated final products (Version 2). These errors, particularly in heterogeneous landscapes, can significantly compromise the accuracy of canopy height and terrain elevation estimates.

To address these challenges, the scientific community has developed methods to enhance GEDI’s geolocation accuracy. Notably, the GEDI Simulator tool, created by the GEDI Science Team, applies orbit-level systematic corrections using small-footprint ALS data. This approach assumes a uniform systematic error across the orbit and determines a single coordinate offset to correct horizontal deviations, which can often fail in complex and heterogeneous landscapes. Consequently, alternative methods, such as beam-level corrections (calculating an independent offset for each beam track) and footprint-level corrections (computing individual offsets for each footprint), have emerged. Despite their potential, these methods, including the GEDI Simulator, face practical limitations such as complexity, computational inefficiency, and a lack of user-friendly interfaces, restricting their broader adoption for remote sensing applications.

To overcome these limitations, we introduce GEDICorrect, an open-source Python framework for precise beam and/or footprint-level geolocation correction, designed with simplicity and accessibility in mind. GEDICorrect integrates multiple methods, criteria, and metrics, including waveform matching, terrain matching, and relative height (RH) profile matching, to achieve refined geolocation accuracy at the orbit, beam, or footprint levels. By leveraging advanced similarity metrics - such as Pearson and Spearman waveform correlations, Curve Root Sum Squared Differential Area (CRSSDA), and Kullback-Leibler divergence - GEDICorrect ensures precise alignment between GEDI measurements and simulated data.

Additionally, GEDICorrect incorporates parallel processing strategies using Python’s multiprocessing capabilities, enabling efficient handling of large-scale GEDI and ALS datasets. This scalability makes the framework practical for global-scale applications while maintaining accuracy and computational efficiency. By addressing critical barriers in geolocation correction with an open-source, user-friendly design, this framework enables a better assessment of canopy structure that can be applied to a wide range of fields, from advancing our understanding of carbon sequestration to supporting more informed planning and conservation efforts.

How to cite: Corado, L. and Godinho, S.: GEDICorrect: A Python Framework for GEDI Geolocation Correction Using Multiple Criteria and Parallel Processing Methods, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18367, https://doi.org/10.5194/egusphere-egu25-18367, 2025.

EGU25-18460 | ECS | PICO | ESSI3.4

A systematic methodology for directive-based GPU porting: NEMO ocean model as a case study 

Rommel Quintanilla, Alexey Medvedev, Xavier Yepes-Arbós, Razvan Aguridan, and Mario C. Acosta

The constant growth in computational demands of scientific applications, combined with energy efficiency requirements, makes GPU acceleration an important factor to consider in high-performance computing environments. Therefore, GPU porting has become essential for efficiently utilizing modern heterogeneous systems that currently provide both multicore CPUs and GPU-accelerated partitions.

While small, relatively new projects might be candidates for complete rewrites in low-level GPU languages like CUDA or HIP, this approach becomes impractical for larger and more complex codebases. Thus, a more convenient way is provided by the directive-based approach, which allows developers to maintain their original C++ or Fortran code while adding extra OpenACC/OpenMP directives to generate energy-efficient GPU code. 

However, this seemingly straightforward method often presents significant challenges. For instance, dealing with code that employs layouts that are not well-suited for GPU architectures, such as deeply nested loop structures or complex memory access patterns that result in suboptimal performance, might lead to the need to reorganize the initial code after all.

In this work, we present a systematic approach to performing the GPU code transition through compiler directives. This several-step incremental process seeks to reach a significant performance and energy consumption improvement while preserving code maintainability, portability, and output accuracy. We demonstrate the effectiveness of our approach through a detailed case study of our ongoing project porting the NEMO ocean model, which represents an interesting example of a complex scientific Fortran code with numerous common computational patterns. Finally, we discuss the experiences, limitations, and trade-offs encountered during this process, providing useful insights for other porting efforts that could face similar GPU migration challenges.

How to cite: Quintanilla, R., Medvedev, A., Yepes-Arbós, X., Aguridan, R., and Acosta, M. C.: A systematic methodology for directive-based GPU porting: NEMO ocean model as a case study, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18460, https://doi.org/10.5194/egusphere-egu25-18460, 2025.

The Doppler Weather Radar (DWR) plays an important role in providing valuable 3D observations of precipitation systems. The advent of radar polarimetry enhances radar capabilities by providing detailed precipitation target characteristics like shape, size, etc. However, the complexities associated with radar data processing pose several challenges to its effective and widespread use. To address these challenges in radar data handling and analyzing, several open-source Python modules have been developed to facilitate radar data processing and analysis, such as WRADLIB, PYART, PYCWR, etc. One such value addition to these open-source tools is an in-house developed Python Indian weather radar toolkit (PYIWR). The toolkit incorporates standard procedures for processing and visualizing polarimetric weather radar data, making it easier for radar users to work with raw radar data, mitigating various challenges due to different data structures and formats. The present work focuses on integrating a novel ground clutter mitigation algorithm developed into the PYIWR framework. The algorithm leverages the statistical properties of long-term radar observations to identify persistent ground clutter using a probabilistic clutter map. It has been extensively tested and evaluated using long-term data from the C-band Doppler Weather Radar at the Thumba Equatorial Rocket Launching Station (TERLS) in Thiruvananthapuram, Kerala, India, spanning 2017 to 2024. Quantitative evaluation of the clutter removal ratio demonstrates that the proposed technique outperforms existing methods, like standalone Gabella filter and fuzzy logic approaches, in mitigating persistent ground clutter, especially in complex terrain. The integration of this newly developed algorithm into the PYIWR framework significantly enhances its capabilities for radar data quality control, making it a more robust and effective tool for the radar user community.

How to cite: Tyagi, V. and Das, S.: Integration of Clutter Mitigation Algorithm into PYIWR Framework: A Python Toolkit for Analyzing Weather Radar Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20297, https://doi.org/10.5194/egusphere-egu25-20297, 2025.

EGU25-20575 | PICO | ESSI3.4

GeoSight: An Open-Source library for Geospatial Analysis and Visualization 

Hamed Vaseghnia, Nestor Fernando Cardozo Diaz, and Enrico Riccardi

GeoSight is an open-source Python package developed for geospatial analysis, processing, and visualization.

GeoSight’s development has been shaped by ongoing collaboration between lecturers, students, and industry professionals. Students have contributed, using and improving the package while learning Python programming, object-oriented design, and addressing real-world geoscience challenges.  

With a focus on accessible, structured and modular programming, the package is user-friendly, flexible, and easily extendable. It is distributed under a public license to encourage collaboration and widespread use.

Key features of GeoSight include 2D and 3D geospatial data visualization, contour and slope analysis, noise filtering, and the integration of satellite imagery into terrain models. Advanced capabilities, such as machine learning for classification and geological computations such as strike and dip measurements, might serve geoscientists, engineers, and urban planners. For an efficient processing, GeoSight supports GPU/CPU architectures.

We here present GeoSight features, benefits, and limitations, discussing its modular design, its support for student learning, and adaptability to changing geospatial standards. We plan in keep expanding its analytical tools, its cross-platform compatibility, and adding further features such as real-time visualization and multiple data-source integration

How to cite: Vaseghnia, H., Fernando Cardozo Diaz, N., and Riccardi, E.: GeoSight: An Open-Source library for Geospatial Analysis and Visualization, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20575, https://doi.org/10.5194/egusphere-egu25-20575, 2025.

Recent advances in remote sensing and computer vision have reshaped how geospatial data is captured, visualized, and analyzed. Drones equipped with high-resolution cameras and sensors enable rapid, detailed surveys from multiple angles, providing near-real-time insights into complex environments. While photogrammetry-based workflows yield accurate 3D meshes, they often demand substantial computational power and lengthy processing times.

Emerging techniques such as Gaussian splatting and Neural Radiance Fields (NeRFs) offer compelling alternatives to traditional mesh-based methods. Gaussian splatting represents 3D scenes as point-based “splats” with mathematical distributions, enabling faster, photorealistic rendering. NeRFs employ neural networks to generate volumetric reconstructions from sparse image inputs, capturing intricate lighting and geometry with minimal manual intervention. Together, these methods reduce post-processing complexity and enhance visual fidelity.

In this session, we demonstrate novel applications of Gaussian splats and NeRFs within ArcGIS and discuss how these approaches can integrate with familiar mesh-based workflows. We also explore ways to extend existing 3D streaming standards, such as OGC’s Indexed 3D Scene Layers (I3S), to incorporate these emerging data types. Finally, we showcase real-world examples demonstrating how blending Gaussian splats and conventional meshes enables richly detailed, interactive visualizations at multiple scales. This convergence promises more efficient collaboration, cost-effective workflows, and deeper insights into rapidly evolving built and natural environments.

How to cite: Belayneh, T.: Next-Generation Geospatial Visualization: From Traditional Meshes to Gaussian Splats and NeRFs, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21310, https://doi.org/10.5194/egusphere-egu25-21310, 2025.

ESSI4 – Advanced Technologies and Informatics Enabling Transdisciplinary Science

EGU25-1768 | ECS | Orals | ITS1.17/ESSI4.1

Electrical Structure of the Crust and Mantle in Southwestern Australia 

Li Lili, Cai Hongzhu, Wang Xinyu, Liu Lichao, and Hu Xiangyun

Funding: This research is funded by the National Natural Science Foundation of China (42274085).

Abstract: The Yilgarn Craton in Western Australia, one of the world's oldest cratons, is rich in mineral resources and provides significant opportunities for research into geothermal energy, crustal dynamics, and mineral exploration. To investigate the electrical structure of southwestern Western Australia, we interpret magnetotelluric data using a finite element-based inversion algorithm we developed, complemented by Bouguer gravity anomaly data, and perform a detailed analysis of the crust-mantle electrical structure. We rigorously validate model sensitivity and cross-verify the inversion results with those obtained using ModEM and Bouguer gravity anomaly interpretations. Our findings identify the Darling Fault and the southern Manjimup Fault as critical structural boundaries that delineate distinct geological features in the study area. All three methods consistently reveal low-resistivity anomalies in the asthenosphere at depths shallower than 100 kilometers. By integrating these results with insights from seismology, gravity, geodynamics, and geochemistry, we suggest that significant geological activity occurs beneath the ancient crust of the Yilgarn Craton. The observed low-resistivity anomalies likely result from the influence of the Darling Fault, the southern Manjimup Fault, and early magmatic processes associated with the craton’s evolution.

 

How to cite: Lili, L., Hongzhu, C., Xinyu, W., Lichao, L., and Xiangyun, H.: Electrical Structure of the Crust and Mantle in Southwestern Australia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1768, https://doi.org/10.5194/egusphere-egu25-1768, 2025.

EGU25-3456 | Posters on site | ITS1.17/ESSI4.1

Workshops, Personalised Training and summer school of Geo-INQUIRE EU-project - Enhancing cross-disciplinary research 

Mariusz Majdanski, Iris Christadler, Giuseppe Puglisi, Jan Michalek, Stefanie Weege, Artur Marciniak, Sylwia Dytłow, Fabrice Cotton, Angelo Strollo, Mateus Litwin Prestes, Helle Pedersen, Laurenciu Danciu, Marc Urvois, Stefano Lorito, Daniele Bailo, Otto Lange, and Gaetano Festa

The Geo-INQUIRE (Geosphere INfrastructure for QUestions into Integrated REsearch) project, supported by Horizon Europe, aims to improve geoscience research infrastructures and services to make high-level data and products available to the broad geosciences research community. The goal of the Geo-INQUIRE project is to encourage curiosity-driven research to understand Geosystem processes at the interface of the solid Earth, oceans and atmosphere using big data sets, high-performance computing methods and state-of-the-art facilities.

The project places great emphasis on supporting the dynamic development of data and services through the effective use of Research Infrastructures such as EPOS, EMSO, ECCSEL and ChEESE. Training, networking and community building are the key to supporting it. The methodology ensures the strengthening of the participation of both young and experienced researchers and the inclusion ofoften underrepresented communities. It incorporates also new and cross-cutting perspectives, while addressing current major environmental and economic challenges as well as stimulating curiosity-based and interdisciplinary research.

Project dissemination activities include a series of open online training and more specialized on-site workshops focusing on data, data products and software solutions. Scientists, early-career scientists and students are communities that are able to explore various fields of science related to the geosphere, even those not directly related to their field, with possible connection through research infrastructures. Through lectures and use cases, we show and teach how to use data and information from interdisciplinary research infrastructures. We raise awareness on the potential and possibilities of Research Infrastructures contributing to Geo-INQUIRE, as well as data integration and the importance of FAIR principles. The training offer is constantly updated on the project website www.geo-inquire.eu.

In autumn 2025 the second summer schools will be organised in Catania, Sicily, and will be dedicated to cross-disciplinary interactions of solid Earth with marine science and with atmospheric physics. The second call of the personalised training program, supporting short research stays, will be announced in 2025. Moreover, after the first two successful calls, the 3rd call for Transnational Access to Research Facilities is open until the end of February 2025. The final 4th call will open in late spring/early summer. Data and products generated through Transnational Access will be made available to the scientific community at large in strict adherence to the FAIR principles.

Geo-INQUIRE is funded by the European Commission under project number 101058518 within the HORIZON-INFRA-2021-SERV-01 call.

How to cite: Majdanski, M., Christadler, I., Puglisi, G., Michalek, J., Weege, S., Marciniak, A., Dytłow, S., Cotton, F., Strollo, A., Prestes, M. L., Pedersen, H., Danciu, L., Urvois, M., Lorito, S., Bailo, D., Lange, O., and Festa, G.: Workshops, Personalised Training and summer school of Geo-INQUIRE EU-project - Enhancing cross-disciplinary research, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3456, https://doi.org/10.5194/egusphere-egu25-3456, 2025.

EGU25-5533 | Orals | ITS1.17/ESSI4.1

Tectonic Influence on Ecosystem Dynamics in the Kenya Rift and Tanzanian Craton 

Simon Kübler, Beth Kahle, Mjahid Zebari, Chintan Purohit, Donjá Aßbichler, and Stephen Rucina

Whilst earthquakes cause destruction, the faults along which they occur are responsible for building varied landscapes and influencing ecosystems by controlling topography, hydrology, soil properties, and vegetation patterns. Faulting acts as both a water conduit and a hydrological barrier, channeling groundwater and creating localized zones of moisture retention. Surface faulting and the resulting topographic complexity contribute to heterogeneous vegetation patterns, with denser vegetation often developing along steep fault escarpments where grazing and agricultural activities are limited. Erosion along fault scarps enriches soils with nutrients and clays, supporting vegetation growth, while also posing risks such as the release of harmful substances like fluoride and arsenic, especially in geothermal regions.

We carry out a broad interdisciplinary study within the East African Rift System to explore the connections between tectonic processes and ecosystem dynamics. By combining geomorphological analysis, soil and geochemical studies, and remote sensing techniques, we investigate how faulting shapes soil fertility, hydrology, and vegetation patterns in these regions. Here, we focus on three illustrative case studies: the southern and central Kenyan Rifts and the Serengeti-Mara ecosystem.

In the southern Kenya Rift, fault-driven erosion and volcanic ash deposition around Lake Magadi enhance soil fertility, sustaining vegetation in this climatically vulnerable area. In contrast, uplifted footwalls and eroded substrates exhibit nutrient deficiencies, limiting ecological productivity. In the central Kenya Rift, near Lake Nakuru, elevated fluoride levels in ground- and surface water are among the highest globally and pose significant health risks to humans and animals. Fluoride concentrations are driven by the naturally high fluoride content in trachytic pyroclastics, which leach into the hydrological system through geothermal activity along active normal faults.

The Serengeti-Mara ecosystem is largely situated on the ancient continental crust of the Tanzanian Craton, where fault activity in the northern and southeastern sectors locally enhances soil moisture and vegetation stability. These tectonically influenced areas provide fertile hotspots within a landscape otherwise characterized by highly dynamic seasonal vegetation patterns. This patchy nutrient distribution is crucial for grazing animals, whose migrations are shaped by the shifting availability of fertile areas, driving ecological connectivity and long-term resource distribution.

Our studies highlight the dual role of fault activity in sustaining biodiversity while presenting challenges through earthquake activity and the release of potentially harmful elements. These findings contribute to a broader understanding of the interplay between geological processes and ecological resilience in tectonically active landscapes.

 

How to cite: Kübler, S., Kahle, B., Zebari, M., Purohit, C., Aßbichler, D., and Rucina, S.: Tectonic Influence on Ecosystem Dynamics in the Kenya Rift and Tanzanian Craton, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5533, https://doi.org/10.5194/egusphere-egu25-5533, 2025.

During space weather events, electric currents in the magnetosphere and ionosphere induce telluric currents near the Earth’s surface, which in turn generate disturbances of the local magnetic field that perturb the detection systems of broad-band sensors. Seismologist consider this effect as noise masking low frequency seismic waves. However, records of this interference can become an opportunity to study in greater detail the evolution of magnetic events and their effects on Earth.

The May 2024 solar storm, the largest in recent decades, has provided an excellent opportunity to analyze these signals. Thanks to their wide global distribution and their availability through platforms such as EPOS, broad-band seismometers provide extensive coverage of the magnetic signals associated with the solar storm. As an example, more than 310 seismometers have clearly recorded the solar storm in Europe, compared to the few tens of magnetometers available in the Intermagnet network in the same region. This geomagnetic storm has been recorded by broad-band seismometers distributed around the world for a time interval of more than 55 hours. Signals related to magnetic field variations can be identified in seismic data for frequencies below 10 mHz, but are clearer between 1.5 and 5 mHz, the frequency band corresponding to Pc5 magnetic pulsations. In the case of magnetic and seismic signals acquired at close locations, there is an excellent correlation between the seismic records and the time derivative of the magnetic field. The number of seismological stations that detect the signals varies significantly between the various seismic networks analyzed, depending on factors such as the presence of magnetic insulation systems or the bandwidth of the sensor.

Our study shows that the recording of magnetic events in broad-band seismometers can be affected by local effects that modify their amplitude and/or polarity, making a detailed calibration of each seismometer necessary before using seismic data to model the waveforms and amplitudes of the magnetic pulsations. However, broad-band data facilitate the monitoring of the temporal variations of the magnetic field disturbances in a large number of sites around the world, hence providing valuable information to complement data acquired by magnetometers.

How to cite: Diaz, J.: On the use of broad-band seismometers to monitor the temporal evolution of magnetic storms; the case of the May 2024 solar storm, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6076, https://doi.org/10.5194/egusphere-egu25-6076, 2025.

EGU25-6796 | Posters on site | ITS1.17/ESSI4.1

Investing in Strategic Infrastructures for Geohazard Prevention in the Eastern Mediterranean Region: The CyCLOPS Integrated GNSS/InSAR Permanent Network and the Cyprus Ground Motion Service 

Chris Danezis, Dimitris Kakoullis, Kyriaki Fotiou, Christopher Kotsakis, Miltiadis Chatzinikos, Michael Eineder, Ramon Brcic, Nerea Ibarrola Subiza, George Ioannou, Marios Tzouvaras, and Diofantos Hadjimitsis

CyCLOPS is a strategic research infrastructure unit led by the Cyprus University of Technology Laboratory of Geodesy, designed in collaboration with the German Aerospace Center (DLR), and supported by government agencies and European initiatives. CyCLOPS is Cyprus’ first and only Tier-1/ Class-A permanent GNSS station network designed to monitor geohazards and densify global and regional frames in the country. Its main objectives are to precisely estimate ground displacements at the national level, bolster resilience to seismic and geological threats, and establish Cyprus as a dedicated calibration site for SAR satellite missions.

Co-located with highly precise tiltmeters, weather stations, and calibration-grade corner reflectors, this infrastructure provides millimeter-level positioning and velocity estimates, revealing critical insights into the island’s geodynamic regime. The CyCLOPS strategic research unit integrates GNSS data with InSAR products — DInSAR, PSI, and SBAS — to expand deformation spatial resolution beyond GNSS’s single-point observations. Using corner reflectors designed with the DLR, CyCLOPS enables multi-track calibration for ascending and descending satellite orbits, which are particularly important for studying tectonic shifts along the Eurasian-African plate boundary. To date, the infrastructure has identified the tectonic motion of Cyprus and monitored active landslides with considerable weekly velocity.

In addition to its permanent segment, CyCLOPS features a mobile segment equipped with GNSS stations of the same grade, tiltmeters, and electronic corner reflectors that can be swiftly deployed to areas prone to geohazard risks, such as landslides or rockfall zones. The Operations Center (OC) manages storage, analysis, and dissemination of both GNSS and InSAR data products. Leveraging a cloud-supported, mixed microservices architecture, the OC delivers daily positions of ground stations and their quality assessment, real-time alerts for abrupt events, and monitors the infrastructure operating status.

Beyond national priorities, another key objective of CyCLOPS is to support regional and global infrastructure initiatives. To that end, CyCLOPS already contributes three GNSS CORS to the European Plate Observing System (EPOS), and one station to EUREF’s EPN.

CyCLOPS+ marks the next expansion phase aiming to establish a continuously updated Cyprus Ground Motion Service (CyGMS) by densifying the existing network with at least five new Tier-1/Class A GNSS CORS sites, and more electronic corner reflectors (ECR-C) to enhance both real-time and post-processing analysis. An important outcome of CyCLOPS+ will be a robust national velocity model that will complement and calibrate the European Ground Motion Service (EGMS) by filling spatial coverage gaps and addressing reference frame challenges. Finally, CyCLOPS+ aims to improve national disaster preparedness, inform infrastructure planning, and provide critical data to authorities responsible for safeguarding communities against seismic hazards, landslides, and other geological threats. Through close collaboration with government agencies and stakeholders, CyCLOPS+ aims to position Cyprus at the forefront of integrated ground motion monitoring in the Eastern Mediterranean region.

Acknowledgements:

  • The authors would like to acknowledge the 'CyCLOPS+' (RIF/SMALL SCALE INFRASTRUCTURES/1222/0082) project, which is co-financed by the European Regional and Development Fund and the Republic of Cyprus through the Research and Innovation Foundation in the framework of the Cohesion Policy Programme "THALIA 2021-2027" and by national resources.

How to cite: Danezis, C., Kakoullis, D., Fotiou, K., Kotsakis, C., Chatzinikos, M., Eineder, M., Brcic, R., Ibarrola Subiza, N., Ioannou, G., Tzouvaras, M., and Hadjimitsis, D.: Investing in Strategic Infrastructures for Geohazard Prevention in the Eastern Mediterranean Region: The CyCLOPS Integrated GNSS/InSAR Permanent Network and the Cyprus Ground Motion Service, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6796, https://doi.org/10.5194/egusphere-egu25-6796, 2025.

EGU25-6911 | ECS | Posters on site | ITS1.17/ESSI4.1

Regional building damage assessment in 2023 Turkey-Syria earthquake doublet based on strong-motion records 

Guan Chen, Siau Chen Chian, and Shengji Wei

An earthquake doublet with magnitudes Mw7.8 and Mw7.6 struck southeastern Turkey on February 6, 2023, causing widespread loss of life and property. To evaluate the seismic damage across 11 affected provinces, we conducted a comprehensive analysis of strong motions and building damage. Specifically, we analyzed the statistical and spatial ground motion intensity measures, along with special characteristics of near-fault pulse-like ground motion. Based on nonlinear seismic analysis, fragility functions were developed to assess the damage states of buildings, where five types of structures are adopted to represent the most common buildings and infrastructures in Turkish cities. Furthermore, the spatial distributions of ground motion intensities and building damage states were validated using official damage reports and field surveys. Results indicate that our model aligns well with these reports and surveys, provided that sufficient seismic records are available. Extensive building damage in the earthquake is primarily attributed to the high intensities of strong motion, construction quality and building resonance, with additional contributions from earthquake-induced geological and geotechnical hazards. Moreover, near-fault regions experienced greater damage due to stronger pulse-like ground motions, fault displacements, and geohazards, all closely associated with fault ruptures. By providing insights into special seismic impacts in near-fault regions and the real characteristics of ground motions, this work contributes to the advancement of ground motion modeling, seismic risk analysis, and disaster management, ultimately supporting the development of more resilient communities.

How to cite: Chen, G., Chian, S. C., and Wei, S.: Regional building damage assessment in 2023 Turkey-Syria earthquake doublet based on strong-motion records, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6911, https://doi.org/10.5194/egusphere-egu25-6911, 2025.

Case Study of Vertical Interoperability Between Research Tools Enabling an End-to-End Sample Workflow from Collection, to Management, to Archiving

Vertical Interoperability

In recent years, interoperability has taken the forefront of discussions on research data management, whether related to research tools, data, or metadata. When it comes to research tool interoperability, the focus so far has been horizontal, improving the flows between tools that serve the same category: GREI’s standardisation of generalist repository metadata [1], a DMP Common Standard [2].

However, the data and metadata is also going to flow vertically, across tools used in very different stages of the research process. These tools will naturally have different requirements, focuses, and functionality from each other, especially differing between domains. How can we enable information to flow between these tools while ensuring FAIR principles are upheld? How can we facilitate researcher processes while ensuring traceability and no metadata loss? What considerations need to be taken into account by institutions and tool developers to design a flexible solution that satisfies user needs? 

Case Study - Fieldmark, RSpace, repositories

In this presentation, we will provide an update on the development of our end-to-end, integrated research data management workflow for samples. We integrate three tools, covering sample collection, processing, storage, and archiving:

  • Fieldmark, an offline sample metadata collection tool
  • RSpace, an ELN and sample management system and RDM platform
  • Generalist and domain-specific data repositories

The presentation will also explain how consistent use of IGSN IDs (the material sample persistent identifier) in every tool and at every stage of the process acts as an integrating force and enhances data discovery.

We wish to present both practical recommendations, as well as higher-level reflections on how to approach thinking and developing vertical interoperability at an institution, and its benefits for researchers and RDM as a whole. We will also cover planned support for PIDINST. We hope that attendees will gain a strengthened mental model of how their tools ecosystem could interact, and how to approach building greater interoperability in their workflows.

References

[1] Curtin, L., Feri, L., Gautier, J., Gonzales, S., Gueguen, G., Scherer, D., Scherle, R., Stathis, K., Van Gulick, A., & Wood, J. (2023). GREI Metadata and Search Subcommittee Recommendations_V01_2023-06-29. Zenodo. https://doi.org/10.5281/zenodo.8101957

[2] https://github.com/RDA-DMP-Common/RDA-DMP-Common-Standard

How to cite: Plankytė, V., Edmunds, R., and Macneil, R.: Case Study of Vertical Interoperability Between Research Tools Enabling an End-to-End Sample Workflow from Collection, to Management, to Archiving, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9046, https://doi.org/10.5194/egusphere-egu25-9046, 2025.

EGU25-9063 | Posters on site | ITS1.17/ESSI4.1

Improving the findability of legacy laboratory data: enrich metadata using controlled vocabularies 

Laurens Samshuijzen, Otto Lange, Ronald Pijnenburg, Richard Wessels, and Maik Nothbaum

The EPOS TCS Multi-Scale Laboratories (MSL) collects and harmonizes both available and newly emerging laboratory (meta)data, thereby aiming to generate data products that are easily Findable, Accessible, Interoperable and Reusable (FAIR) for future research, notably into Geo-resources, Geo-storage, Geo-hazards and Earth System Evolution. Key for discovery of MSL data is the use of well-established and openly published controlled community vocabularies. These vocabularies provide all terms for a full contextual description of a conducted laboratory experiment (e.g., materials used, apparatus, etc.). To improve the findability of future data publications we provide (metadata) editor components which connect to the community vocabularies. These vocabularies themselves are openly accessible and ready for incorporation in existing data publication chains at data repositories.

Challenges arise especially with respect to legacy content stemming from the long tail of science, i.e. data that were published before the MSL community standards for metadata and vocabularies became available. In many of such cases the presence of standardized metadata for discovery and provenance is often limited. To improve the findability of these valuable but non-harmonized data publications we developed a strategy which makes use of the MSL vocabularies. With this strategy we demonstrate how controlled vocabularies can be used for filling metadata gaps in older data publications and as such can be useful not merely for new data publications, but for the improvement of FAIRness for older sets as well.

The first challenge we faced concerned the identification of relevant legacy content that had to be discovered within the large offering at repositories. Using controlled term recognition we were able to identify a large set of data publications that appeared to be relevant to the MSL community. The second issue to solve was the enrichment of metadata to improve the findability of the identified publications. The use of the MSL vocabularies in combination with a textual analysis of the collected abstracts and titles allowed for an hierarchical description of the data, the experiment itself, and the equipment used. The result was an improvement of the findability through an extension of the initial metadata.

The extended metadata is shared via the EPOS Platform (https://www.ics-c.epos-eu.org/) and the MSL community data catalogue (https://epos-msl.uu.nl) which guides users in finding data publications through the provision of hierarchical filtering options with increasing granularity. The methodology we describe could be applied in broader contexts within the solid Earth sciences.

How to cite: Samshuijzen, L., Lange, O., Pijnenburg, R., Wessels, R., and Nothbaum, M.: Improving the findability of legacy laboratory data: enrich metadata using controlled vocabularies, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9063, https://doi.org/10.5194/egusphere-egu25-9063, 2025.

EGU25-9291 | Posters on site | ITS1.17/ESSI4.1

Advancing cross-disciplinary FAIR data practices: Harmonization, assessment, and continuous improvement in Geo-INQUIRE 

Otto Lange, Laurens Samshuijzen, Enoc Martínez, Stefano Rapisarda, Javier Quinteros, Helle Pedersen, Angelo Strollo, Carine Bruyninx, Florian Haslinger, Marc Urvois, Laurentiu Danciu, and Anna Miglio

It is widely acknowledged that ‘putting FAIRness into practice’ with respect to cross-disciplinary data sharing demands overcoming domain-specific practices regarding data dissemination.  I.e., communities may rely on specialized standards for describing and sharing data (metadata, vocabularies, services) that do not always easily allow for successful reuse in other domains and may as such not be directly fitted for cross-disciplinary research. In the Geo-INQUIRE project (https://www.geo-inquire.eu/) the European ESFRI landmark research infrastructures EPOS, EMSO, and ECCSEL, the Center of Excellence ChEESE, and the ARISE infrasound community collaborate in overcoming cross-domain barriers, especially the land-sea-atmosphere environments, thereby exploiting innovative data management techniques. As such, one of the strategic priorities of the project is to ‘enhance FAIRness of all data and data products’ for the research infrastructures involved. This concerns not merely a one-time application of the FAIR principles as far as possible, but also measuring the impact for research communities through the establishment of a feedback loop and the measurement of appropriate performance indicators which must be taken from a feasible metrics framework for FAIRness. This approach allows for a constant improvement of data and data products from the FAIR perspective.

The challenges that follow from this ambition are three-fold: 1) In the light of the variety of specialized sub-communities there is the demand to decide about the distinction of intermediate levels for harmonization of metadata, vocabularies, and services design; 2) An instrument is required to perform the actual assessment on the basis of the adopted FAIR metrics framework (thereby following the harmonized standards at the appropriate level), and which must be ready for use by data and/or installation managers; 3) A feedback loop must be configured to support the monitoring of impact and improvement with respect to FAIRness.

To meet these challenges within Geo-INQUIRE we used valuable outcomes from external initiatives (e.g., FAIRsFAIR, GoFAIR). For the FAIR assessment we developed the Geo-INQUIRE FAIRness Assessment Pipeline, a system that evaluates the FAIRness of multiple datasets over time by means of the F-UJI tool in the background, while providing a GUI to analyze the results through multiple dimensions and levels of classification (e.g. discipline). Evaluation over time tracks improvement in a quantitative manner and provides a powerful instrument for creating increased awareness.

For the process of community harmonization at the appropriate intermediate levels we turned to the use of FAIR Implementation Profiles (FIPs). The results we share offer an interesting example of an approach that could easily be transferred to many different cross-disciplinary contexts.

How to cite: Lange, O., Samshuijzen, L., Martínez, E., Rapisarda, S., Quinteros, J., Pedersen, H., Strollo, A., Bruyninx, C., Haslinger, F., Urvois, M., Danciu, L., and Miglio, A.: Advancing cross-disciplinary FAIR data practices: Harmonization, assessment, and continuous improvement in Geo-INQUIRE, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9291, https://doi.org/10.5194/egusphere-egu25-9291, 2025.

The Earth System is a complex and dynamic system that encompasses the interactions between the atmosphere, oceans, land, and biosphere. Galaxy is an open, comprehensive, and sustainable web platform for understanding and analyzing data from the Earth System sciences, which is essential, for example, to study the impacts of climate change.

Therefore, Galaxy can be used as an IT toolkit for multidisciplinary and interdisciplinary studies with a set of tools for data visualization, analysis, and processing across various scientific fields such as oceanographic, atmospheric, land sciences, and more. By design, Galaxy manages data by sharing and publishing results, workflows, and visualizations, ensuring reproducibility by capturing the necessary information to repeat and understand data analyses. Thus, Galaxy for the Earth System sciences aim at directing users toward standardized tools that can be plugged into cross-domains workflows.

Fully integrated into the work area, the Galaxy Training network (available at training.galaxyproject.org) is an initiative that aims at making the Galaxy platform accessible to a wide audience by providing free and open educational resources. It offers an extensive collection of detailed and reviewed tutorials authored by administrators, developers, and scientists. These tutorials serve as valuable resources for individuals seeking to learn how to navigate Galaxy, employ specific functionalities like tools or execute workflows for specific analyses. By mixing trainings and tools in the same friendly user webapp, Galaxy is a tool perfectly suited for open science.

As part of the FAIR-EASE project, we have deployed a Galaxy adaptation for Earth System studies (earth-system.usegalaxy.eu) with dedicated models, data, tools and data visualisation.  We want to use this opportunity to present during your session a set of workflows and trainings mixing in-situ and biogeochemical ocean data, atmospheric volcanoes data, and marine biodiversity data. Our goal is to showcase the possibility to have multiple scientific domains studied and visualise several data types of the same geographical area in one virtual research environment.

How to cite: Jossé, M. and Detoc, J.: Galaxy for Earth System Science: Integrating Data, Tools, and Training for Open Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9941, https://doi.org/10.5194/egusphere-egu25-9941, 2025.

EGU25-10649 | Orals | ITS1.17/ESSI4.1

Optimizing Transnational and Virtual Access: A Data-Driven Framework for Managing Geoscience Research Infrastructure 

Juliano Ramanantsoa, Daniele Bailo, Jan Michalek, Sven Peter Näsholm, Rossana Paciello, and Angelo Strollo

The rapid evolution of cross-disciplinary research in geoscience has led to an exponential increase in complex data production, significantly challenging the data research experts as well as the data repositories management. This complexity is evident in large-scale data infrastructure projects like the EU-funded Geo-INQUIRE project, which includes five major Research Infrastructures (RIs) in geoscience, namely EPOS-ERIC, EMSO-ERIC, ECCSEL-ERIC, ARISE and ChEESE, offering both Transnational Access (TA) and Virtual Access (VA).

Integrating data from TA into a unified VA systems often presents challenges, particularly in multi-institutional projects. This process requires significant expert intervention and frequently results in excessive meetings and potential integration failures.

To address this, the current contribution proposes a novel data science-driven method targeting research infrastructure governance challenges. The approach introduces an automated analytical framework to guide the integration of TA assets into VA systems. Leveraging Large Language Models (LLMs) for semantic embedding, the method transforms unstructured metadata from VA and TA sources into structured data vectorizations. This cohesive data frame then undergoes a series of similarity analysis techniques based on cross-semantic embedding evaluations. Using data from the multidisciplinary Geo-INQUIRE project, the method's is tested for its ability to manage complex asset integration across five major geoscience RIs.

The primary finding offers a preemptive framework streamlining connections for integrating TA assets into appropriate VA systems, facilitating decision-making on asset integration flow.

The resulting mapping not only optimizes TA-VA asset matching but also uncovers cross-connections between installations (services), inter-RIs, and potential multi-institutional collaborations. Furthermore, the research presents complex scenarios, through idealized simulations based on TA-VA metadata variable changes, proposing alternative integration pathways when minor asset adjustments or asset enhancements are implemented at the VA installation level.

This contribution is a proof-of-concept research based on a data-driven solution aimed at streamlining data integration in large-scale geoscience projects. It could potentially reduce expert intervention, enhance cross-disciplinary research opportunities, and improve overall efficiency in managing complex, multi-institutional data infrastructures.

How to cite: Ramanantsoa, J., Bailo, D., Michalek, J., Näsholm, S. P., Paciello, R., and Strollo, A.: Optimizing Transnational and Virtual Access: A Data-Driven Framework for Managing Geoscience Research Infrastructure, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10649, https://doi.org/10.5194/egusphere-egu25-10649, 2025.

EGU25-11539 | Posters on site | ITS1.17/ESSI4.1

 EPOS-Norway – Research Infrastructure for Geohazards (EPOS-NG) 

Mathilde Sørensen and Juliano Ramanantsoa and the EPOS-NG team

The EPOS-Norway – Research Infrastructure for Geohazards (EPOS-NG) will be established, starting from spring 2025, with funding from the Research Council of Norway’s Infrastructure program. EPOS-NG aims to be the go-to infrastructure for research on geohazards in Norway (i.e., landslides, tsunamis, earthquakes, and cryospheric hazards). Complementary to EPOS ERIC and building on research infrastructure developed during EPOS-Norway (EPOS-N) phase 1, EPOS-NG will establish new pools of instruments that are easily accessible to all geoscientists in Norway. We will develop an enhanced and extended state-of-the-art data portal to provide nationwide access to a range of geoscience data as well as computational and visualisation services. The EPOS-NG instrument pools include rapid-deployable seismometers, ocean bottom seismographs, Distributed Acoustic Sensing and Distributed Temperature and Strain Sensing instrumentation, Transient Electromagnetic measurement capacity, piezometers, self-potential sensors and ground-based interferometric radar systems. The new instruments will facilitate research on a wide range of processes including seismicity, slope stability and landslides, groundwater and soil conditions, permafrost and cryospheric processes. Combined with new services for tsunami hazard assessment, as well as novel datasets on InSAR displacement trends and historical and palaeoseismological events, new links can be established through comprehensive, multidisciplinary studies. Effective data integration and visualisation will be achieved via the EPOS-N portal, which was developed in EPOS-N phase 1 and will be substantially enhanced in close dialogue with the users in EPOS-NG. The portal combines data from distributed monitoring networks, innovative services for advanced data analysis and national databases within geosciences into a single national e-infrastructure, following FAIR principles. EPOS-NG thus represents a unifying nationwide research infrastructure, including all the relevant physical infrastructures and providing a national hub for solid Earth science data and services.

How to cite: Sørensen, M. and Ramanantsoa, J. and the EPOS-NG team:  EPOS-Norway – Research Infrastructure for Geohazards (EPOS-NG), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11539, https://doi.org/10.5194/egusphere-egu25-11539, 2025.

EGU25-12709 | Posters on site | ITS1.17/ESSI4.1

Advancing FAIRness and National Collaboration in Geosciences: The Role of EPOS-Spain in Open Science and Data Integration 

Adelina Geyer, Olaya Dorado, Noah Schamuells, José Luis Fernández-Turiel, and Claudia Prieto-Torrell

The European Plate Observing System (EPOS) (https://www.epos-ip.org/) is Europe’s foremost infrastructure for multidisciplinary and global research in Earth Sciences. Serving as a unique gateway, EPOS offers access not only to raw data but also to data products, services, software, and research facilities, facilitating inter- and transdisciplinary collaboration across the geosciences. EPOS-Spain (https://epos-es.org) plays a critical role in implementing EPOS at the national level, aligning its efforts with the broader goals of the EPOS framework while addressing specific national needs. It focuses on strengthening and integrating the Spanish nodes within EPOS’s thematic core services, while advancing Open Science principles by enhancing the accessibility, interoperability, and reusability of geoscientific data and services. EPOS-Spain has developed innovative digital infrastructures and implemented resources designed to improve the FAIRness (Findability, Accessibility, Interoperability, and Reusability) of geoscientific data. This commitment ensures that researchers within Spain can seamlessly discover, access, and utilize data, fostering greater collaboration and innovation at the national level. EPOS-Spain actively promotes the use of EPOS resources in national research projects, educational programs, and capacity-building initiatives, particularly benefiting Early Career Scientists. By engaging stakeholders from diverse backgrounds, including geologists, engineers, and policymakers, EPOS-Spain facilitates interdisciplinary workflows and collaborative approaches to address societal challenges such as risk mitigation and urban planning. These efforts are complemented by initiatives to strengthen ties among Spanish institutions, creating a robust and cohesive network of geoscientific research within the country.

Through its initiatives under the EPOS-SpN RED2022-134516-E project, EPOS-Spain strengthens researchers’ ability to integrate geoscientific data across disciplines and domains. Notable efforts include the development of educational and outreach materials, such as postcards and videos that explain FAIR principles and their application across various branches of geosciences. The EPOS-ES website is regularly updated and now features a dedicated blog that delves deeper into key concepts, enhancing accessibility and engagement. Additionally, EPOS-Spain organizes events like Summer Schools, which foster training opportunities and collaboration between Early Career Scientists and established researchers. Regular meetings with the national Thematic Core Services (TCS) facilitate continuous dialogue and integration within the geoscientific community. These efforts collectively contribute to fostering groundbreaking inter- and transdisciplinary studies, enabling innovative solutions to both scientific and societal challenges. By facilitating the discovery, sharing, and analysis of geoscientific data, EPOS-Spain exemplifies the transformative potential of integrated research infrastructures, advancing Earth Sciences while supporting the broader goals of Open Science.

These activities are supported by the EPOS-SpN RED2022-134516-E grant funded by MICIU/AEI/10.13039/501100011033.



How to cite: Geyer, A., Dorado, O., Schamuells, N., Fernández-Turiel, J. L., and Prieto-Torrell, C.: Advancing FAIRness and National Collaboration in Geosciences: The Role of EPOS-Spain in Open Science and Data Integration, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12709, https://doi.org/10.5194/egusphere-egu25-12709, 2025.

EGU25-13128 | Orals | ITS1.17/ESSI4.1

The Irpinia Near Fault Observatory: A Cutting-Edge Infrastructure Exploring the Interplay Between Earthquakes, Deep Fluids and Climate Forcing 

Gaetano Festa, Aldo Zollo, Luca Elia, Francesco Scotto di Uccio, Claudio Strumia, Simona Colombelli, Grazia De Landro, Titouan Muzellec, Matteo Picozzi, Antonio Scala, Nicola D'Agostino, Gilberto Saccorotti, Stefania Tarantino, Alister Trabattoni, Francesco Carotenuto, Antonio Giovanni Iaccarino, Mauro Palo, Raffaello Pegna, and Guido Russo

The Irpinia Near Fault Observatory (INFO) is a state-of-the-art infrastructure for monitoring seismic activity in the Southern Apennines, a region of high seismic hazard that experienced the 1980 M 6.9 Irpinia earthquake. Managed by the University of Naples, the observatory operates the dense ISNet seismic network,  including 30 strong-motion and short-period sensors, 9 broadband seismometers, as well as geodetic and geochemical stations from INGV. Data and products are openly shared through the EPOS platform and the FRIDGE community portal. INFO also serves as a testbed for the Geo-Inquire project, providing unique transnational access for geophysical surveys and real-time analysis.

Fifteen years of continuous seismic monitoring have uncovered a strong correlation between the hydrological loading of shallow karst aquifers, GNSS-measured surface deformation, changes in elastic properties of subsurface, and seismicity rates at depths where large historical earthquakes have nucleated. Velocity and attenuation tomography have further revealed the pervasive presence of deep fluids, with evidence of reservoirs likely containing CO₂ and brine.

Despite these findings, the background seismicity in the area appears sparse, with hypocenters distributed irregularly within the graben system bounded by the faults responsible of the 1980 earthquake. To better understand the microseismicity pattern and its relationship with the major fault structures, we deployed a temporary dense network of 20 arrays (10 stations each) for one year (DETECT experiment), alongside with a Distributed Acoustic Sensing (DAS) system monitoring a 20 km fiber-optic cable.

Advanced machine learning detection techniques, applied to data from the dense monitoring network, expanded the standard seismic catalog by a factor of eight, producing a dataset comparable to a decade of traditional observations. The  enhanced catalog revealed that seismic events follow the seasonal hydrological loading, predominantly cluster at depth, forming small sequences of aftershocks (magnitude <1) that trace a 20–30 km long structure with a stepover. The DAS system has provided coherent recordings of deep phases, likely reflecting the interface between the carbonate plate and the crystalline basement. These insights have paved the way for the installation of permanent arrays and DAS systems in the area, expected for 2025, enhancing the observatory's capability to unravel the complex interplay between seismicity, deep fluids, and external forcing mechanisms.

How to cite: Festa, G., Zollo, A., Elia, L., Scotto di Uccio, F., Strumia, C., Colombelli, S., De Landro, G., Muzellec, T., Picozzi, M., Scala, A., D'Agostino, N., Saccorotti, G., Tarantino, S., Trabattoni, A., Carotenuto, F., Iaccarino, A. G., Palo, M., Pegna, R., and Russo, G.: The Irpinia Near Fault Observatory: A Cutting-Edge Infrastructure Exploring the Interplay Between Earthquakes, Deep Fluids and Climate Forcing, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13128, https://doi.org/10.5194/egusphere-egu25-13128, 2025.

EGU25-13208 | Orals | ITS1.17/ESSI4.1

Cross-fertilization across research infrastructures within Geo-INQUIRE: plans, ongoing activities and future perspectives 

Angelo Strollo, Fabrice Cotton, Mateus Litwin Prestes, Elif Türker, Stefanie Weege, Arnau Folch, Carmela Freda, Kety Giuliacci, Enoc Martinez, Aljaz Maslo, Klaus Tobias Mosbacher, Sven Peter Näsholm, and Ingrid Puillat and the Geo-INQUIRE project management board

The Geo-INQUIRE (Geosphere INfrastructures for QUestions into Integrated REsearch) project, launched in October 2022, fosters collaboration between several European research infrastructures, including three key ESFRI European Research Infrastructure Consortia (ERICs), to enhance geoscientific research and innovation. We highlight here those activities within the project that promote synergies between EPOS ERIC (European Plate Observing System), EMSO ERIC (European Multidisciplinary Seafloor and Water Column Observatory), ECCSEL ERIC (European Carbon Dioxide Capture and Storage Laboratory Infrastructure), ChEESE (Centre of Excellence for Exascale in Solid Earth) and the ARISE (Atmospheric Dynamics Research Infrastructure in Europe) infrasound community. 

The cross-fertilization approach makes use of EPOS's extensive geophysical and geological data, EMSO's ocean and seafloor observation capabilities, ECCSEL's expertise in carbon capture and storage technologies, ChEESE's advanced pre-exascale computing capabilities for hazard and risk assessment, and ARISE's atmospheric monitoring technologies. Through shared data platforms, interoperable tools and collaborative research workflows, Geo-INQUIRE advances the understanding of Earth processes in both terrestrial and marine domains. Key developments include improved assessments of selected geohazards, insights into marine ecosystems, responses to carbon sequestration, and the integration of innovative deep-sea and subsurface monitoring technologies.

These advances will be made possible by providing users with enhanced services integrating new multidisciplinary FAIR data, integrated workflows, training modules, transnational access at key testbed sites, and management policies and KPIs essential for infrastructure governance. This collaborative framework demonstrates how coordinated efforts between research infrastructures (ERICs) can strengthen the European geoscience research landscape and foster multidisciplinary approaches to address critical global challenges. The project highlights the importance of open data sharing and interoperability standards to maximise the societal and scientific impact of research infrastructures.

The presentation will describe the envisaged approach during the proposal preparation phase, the current state of play, and finally, highlight the challenges with the evolving landscape of the project, with use cases shifting from an early emphasis on FAIR data only to a growing focus on AI-driven applications. In addition, the project addresses the rapid updates of data management policies in different communities, while providing a common framework of Key Performance Indicators (KPIs) for data providers, infrastructure operators and other stakeholders. The scientific focus is also evolving during implementation, from an initial focus only on the land/sea interface to also preparing for future climate and biological applications through AI-ready geoscientific data and services, which are becoming a critical asset for understanding the drivers of climate change.

How to cite: Strollo, A., Cotton, F., Litwin Prestes, M., Türker, E., Weege, S., Folch, A., Freda, C., Giuliacci, K., Martinez, E., Maslo, A., Mosbacher, K. T., Näsholm, S. P., and Puillat, I. and the Geo-INQUIRE project management board: Cross-fertilization across research infrastructures within Geo-INQUIRE: plans, ongoing activities and future perspectives, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13208, https://doi.org/10.5194/egusphere-egu25-13208, 2025.

EGU25-13529 | Orals | ITS1.17/ESSI4.1 | Highlight

Multidisciplinary exploitation of spaceborne DInSAR data for investigating volcanoes and seismic areas 

Francesco Casu, Manuela Bonano, Teresa Bortolotti, Sabatino Buonanno, Federica Casamento, Federica Cotugno, Claudio De Luca, Marianna Franzese, Adele Fusco, Riccardo Lanari, Michele Manunta, Fernando Monterroso, Pasquale Noli, Giovanni Onorato, Francesco Poggi, Yenni Roa, Pasquale Striano, Muhammad Yasir, Giovanni Zeni, and Ivana Zinno

Spaceborne Differential SAR Interferometry (DInSAR) is a widely exploited technique that allows measuring ground displacements with centimeter/millimeter accuracy at a large spatial scale. The recent availability of worldwide DInSAR measurements, as well as their standardization in terms of format and access procedures, has further pushed this technique toward its application and integration with other data sources for carrying out multidisciplinary analysis of natural and anthropogenic surface deformation phenomena. In the following, we show some examples carried out in volcanic and seismic areas, testifying the capability of the DInSAR technique to be exploited in multidisciplinary contexts.

For what concerns volcanic scenarios, we focus on the Campi Flegrei caldera (Italy) which is experiencing a continuous ground uplift since 2005, with a main radial pattern centered in the Rione Terra district of Pozzuoli. The analysis of detailed DInSAR measurements, retrieved by processing image time series acquired by the Copernicus Sentinel-1 and the Italian COSMO-SkyMed SAR constellations, allowed the identification of a geodetic anomaly in the Campi Flegrei long term uplift pattern, i.e. an area that shows a deficit in the uplift. The amount of this deficit has been analyzed by also considering other data sources, such as the seismicity of the area, showing a high correlation factor. In addition, the location and spatial extension of the anomaly have been further demonstrated to be related to the geology of the area. These findings provide intriguing insights into the volcanic evolution process and the related hazard.

With reference to the development of seismic analysis, we concentrate on EPOSAR, which is an operative service based on Copernicus Sentinel-1 data deployed by CNR-IREA, that allows generating, at the global scale and in a systematic way, co-seismic DInSAR ground displacement measurements once the satellite data are available after a major earthquake (Mw>5.5, ipocenter depth < 20km) occurrence. These products are automatically provided to the scientific community through the EPOS data portal according to a defined standard. The availability of these kinds of measurements also allowed the development, in collaboration with INGV, of a new service that operates in a cascade to the previous one and retrieves the seismic source that generated the earthquakes. To this aim, the DInSAR measurements are jointly exploited with the available seismic moment tensors provided by the main global seismic services (e.g., USGS and INGV). This automatic service is another example of multidisciplinary data integration and it is worth noting that it strongly benefits from the open access and interoperability policies adopted by the respective data providers.

 

This work has been carried out with the support of: IREA-DPC agreement; HE EPOS-ON (GA 101131592); PNRR MEET (IR00000025); PNRR CN-HPC (CN00000013); PNRR GeoSciences (IR00000037); PNRR MOST (CN00000023).

How to cite: Casu, F., Bonano, M., Bortolotti, T., Buonanno, S., Casamento, F., Cotugno, F., De Luca, C., Franzese, M., Fusco, A., Lanari, R., Manunta, M., Monterroso, F., Noli, P., Onorato, G., Poggi, F., Roa, Y., Striano, P., Yasir, M., Zeni, G., and Zinno, I.: Multidisciplinary exploitation of spaceborne DInSAR data for investigating volcanoes and seismic areas, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13529, https://doi.org/10.5194/egusphere-egu25-13529, 2025.

EGU25-13550 | Orals | ITS1.17/ESSI4.1

Enhancing cross-domain data access in georesources and bridging EPOS and ECCSEL Research Infrastructures: contribution from Geo-INQUIRE project 

Marc Urvois, Salsabyl Benlalam, Franck Chan Thaw, Caroline Correia, Joanna Kocot, Marco Pantaloni, J. Román Hernández Manchado, Agnieszka Mtupa-Ndiaye, Volker Röhling, Jean Schmittbuhl, Andrea Travan, and Lucas Valarcher

The Geo-INQUIRE (Geosphere INfrastructure for QUestions into Integrated Research project - www.geo-inquire.eu) aims to foster the curiosity-driven research about solid Earth. Monitoring dynamic processes within the geosphere requires facilitated access to data, data products and services in a wide range of geoscientific disciplines.

A particular focus on georesources is addressed by using two operational research infrastructures, EPOS (European Plate Observing System) and ECCSEL (European Carbon Dioxide Capture and Storage Laboratory Infrastructure) with innovative activities to extend and enrich the existing underlying thematic data services. While EPOS provides virtual access to data and information over large territories in Europe and worldwide, ECCSEL primarily produces local experimental datasets at lab facilities level in Europe. Four thematic communities teamed up to concretise the cross-domain scientific activities, both from the data provider and end user sides: EPOS -geology, induced seismicity, geodesy- and ECCSEL -permanent CO2 storage, temporary subsurface feedstock storage (H2 and derivates, heat, air, CO2), geothermal energy-.

Halfway through the project implementation, the collaborative work of the stakeholders results in strengthening the respective data contents and management structures enabling their connections. In France, the induced seismicity fact sheets recorded in the CDGP (Data Centre for Deep Geothermal Energy) are now better documented with geological maps and boreholes as well as geodesy and petrophysical properties. The anthropogenic hazards events capitalised and disseminated through the EPISODES platform offer access to episodes and information about boreholes located in their vicinity, being both the source of seismicity and monitoring locations. This enhanced virtual access to these induced events will soon be available on the EPOS data portal. The bridge between EPOS and ECCSEL research infrastructures is now enabled through the integration of a first set of boreholes and experimental data of two platforms in Norway and Italy to be accessible on the EPOS data portal through the national borehole database e-nodes.

The presentation will also expose how this cross-domain data access is enabled through semantic and technical interoperability in line with the FAIR principles to guarantee an efficient and reliable access to research contents.

How to cite: Urvois, M., Benlalam, S., Chan Thaw, F., Correia, C., Kocot, J., Pantaloni, M., Hernández Manchado, J. R., Mtupa-Ndiaye, A., Röhling, V., Schmittbuhl, J., Travan, A., and Valarcher, L.: Enhancing cross-domain data access in georesources and bridging EPOS and ECCSEL Research Infrastructures: contribution from Geo-INQUIRE project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13550, https://doi.org/10.5194/egusphere-egu25-13550, 2025.

EGU25-15362 | ECS | Orals | ITS1.17/ESSI4.1

Bringing in-situ data to light: A formal approach to bridging user needs and provider capacities for enhanced data availability 

Alba Brobia, Joan Masó, Javiera Crisóstomo, Carsten Iversen, and Jean-Philippe Aurambout

When referring to Earth Observation data, we consider two sides of the same coin: space-based data and in-situ data collected on or near to the ground. While satellite-derived data benefits from a consolidated data management and sharing practices, in-situ data is more complex, highly heterogeneous by nature, involving a wide range of actors and data sources, which creates significant challenges in making this data standardised, integrated and interoperable, and ultimately, accessible and usable.

Willing to address these challenges, the InCASE project —supported by the European Environment Agency and funded by the European Commission as a contribution to the Group on Earth Observations (GEO)— developed the Geospatial in-situ requirements (G-reqs) tool. Designed primarily to support GEO Work Programme activities but open to contributions beyond GEO, G-reqs acts as a database and a standard methodology to collect and manage user requirements for in-situ datasets.

The development of G-reqs was done with the hope that the content generated will help in identify shared requirements across domains, detect barriers and gaps, and act as a bridge between user demands and data providers by facilitating the matchmaking between the required and the produced data, or even to prioritize new in-situ data collection strategies. During the last year, the focus was on engaging with the user community to collect as many requirements as possible trying to avoid bias in particular theme, backgrounds, or geographic regions.

In this communication we analyse the content of the G-reqs and discuss to what extent it can fulfil the hopes described before via a series of showcases and statistical overalls. The presented approach demonstrates user-driven solutions and the significance of initiatives like GEO in advancing Open Science and extract new knowledge enabling cross-domain interaction for environmental research and decision-making.

How to cite: Brobia, A., Masó, J., Crisóstomo, J., Iversen, C., and Aurambout, J.-P.: Bringing in-situ data to light: A formal approach to bridging user needs and provider capacities for enhanced data availability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15362, https://doi.org/10.5194/egusphere-egu25-15362, 2025.

EGU25-15535 | ECS | Posters on site | ITS1.17/ESSI4.1

Greek earthquake impact database (GEID): AD 1800-2024 

Ioanna Triantafyllou, Ioannis Koukouvelas, and Efthimios Lekkas

Earthquakes can affect societies causing dramatic effects in both the built and the natural environments. Greece is characterized by the highest seismicity in the Mediterranean region, with a record of earthquakes and associated phenomena from antiquity up to the present. We organized for the first time a unified earthquake impact database covering the Greek territory from AD 1800 up to 2024, which include building damage and rates of fatalities and injuries. Data about earthquake secondary effects have also been inserted in the database concerning several types of ground failures, such as co-seismic landslides, soil liquefaction, surface fault traces, ground fissures, other environmental changes and tsunamis. The new Greek earthquake impact database (GEID), apart from the descriptive information of an earthquake, also provides parametric attributes such as earthquake epicentre, focal depth magnitude and intensity.   The GEID is of great importance since it may help in studies such as a better understanding of the seismic hazard and risk in Greece and its surroundings.

How to cite: Triantafyllou, I., Koukouvelas, I., and Lekkas, E.: Greek earthquake impact database (GEID): AD 1800-2024, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15535, https://doi.org/10.5194/egusphere-egu25-15535, 2025.

EGU25-16779 | Posters on site | ITS1.17/ESSI4.1

Three-dimensional high resolution joint inversion of gravity and magnetotelluric data  

Hongzhu Cai, Wang Xinyu, Li Lili, Huang Sining, Liu Lichao, and Hu Xiangyun

Funding: This research is funded by the National Natural Science Foundation of China (42274085).

Abstract:Gravity and magnetotelluric methods are pivotal geophysical techniques used to study the distribution of density and electrical conductivity within the Earth's interior. These methods have been widely used in multi-scale explorations for various engineering and academic applications. Considering the varying resolution capabilities of different geophysical methods in delineating near-surface geological structures, we propose a three-dimensional parallel joint inversion framework for gravity and MT data, based on Gramian structural constraints. The framework discretizes the inversion model with an unstructured tetrahedral mesh, enhancing the efficiency of forward modeling and sensitivity calculations for both gravity and MT data via a parallelized approach. To achieve sharper and more focused subsurface imaging, we incorporate a zero-order minimum entropy constraint into the objective function of the joint inversion. The objective function is minimized using the Gauss-Newton method, with model updates facilitated by the MINRES solver and line search techniques. Results from synthetic models show that joint inversion significantly improves the results for gravity and MT data, revealing a stronger correlation between residual density and resistivity. The zero-order minimum entropy constraint delivers more distinct model boundaries compared to traditional regularization method.

 

How to cite: Cai, H., Xinyu, W., Lili, L., Sining, H., Lichao, L., and Xiangyun, H.: Three-dimensional high resolution joint inversion of gravity and magnetotelluric data , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16779, https://doi.org/10.5194/egusphere-egu25-16779, 2025.

EGU25-18507 | Posters on site | ITS1.17/ESSI4.1

EPOS-GNSS Data Gateway: News and Novelties 

Mathilde Vergnolle, Jean-Luc Menut, Eric Marin-Lamellet, Guillaume Verbiese, and Imène Thiellement

The EPOS-GNSS Data Gateway (DGW) is the European thematic gateway to GNSS data distributed within the European Plate Observing System - EPOS framework. Thanks to this portal, all interested parties have free access to metadata and data from over 2,000 European GNSS stations.

The information system is based on a network of servers, the nodes, connected to a main server, the DGW. The main showcase is the DGW's graphical interface (https://gnssdata-epos.oca.eu/), which enables all the data and metadata in the EPOS-GNSS data infrastructure to be browsed and downloaded. It conceals a complex system of multiple software enabling the integration and synchronization of metadata between the DGW and the nodes. The development and population of this system is the result of a team effort involving the development team, node managers and the node infrastructure and DGW operation coordination team (https://gnss-epos.eu).

New features for 2024 include the integration of two new nodes (CEGNxEPOS, Italy and SONEL, France), filling a gap in Central Europe (Northern Italy, Austria, Slovenia) and opening up to other scientific communities, such as those working on long-term sea level trends as part of GLOSS (Global Sea Level Observing System). Their deployment and population, at record speed, demonstrate the commitment of the new partners, the robustness of the system and the efficiency of the procedures. Next, the level of data completeness at the DGW in relation to the stations proposed to EPOS is becoming very good. Finally, the number of files not validated at the nodes, according to the EPOS-GNSS procedure, and therefore not transmitted to the DGW, is now very low.

On the other hand, there are some important novelties worth highlighting. All the monitoring tools needed to check that the entire system is working properly are now operational. These tools focus on monitoring all system elements and their interaction at the DGW, comparing metadata between the DGW and the nodes that highlights metadata and synchronization issues, monitoring availability statistics for each DGW-hosted service and user statistics. The system also now gives the opportunity to publish hourly High-Rate GNSS data that are accessible at both the DGW and the EPOS multidisciplinary platform. In early 2025, a new version of the graphical interface, developed using a different technology, will be deployed, enabling easier customization of the interface by node managers, in particular to better acknowledge all contributors to the EPOS-GNSS system.

How to cite: Vergnolle, M., Menut, J.-L., Marin-Lamellet, E., Verbiese, G., and Thiellement, I.: EPOS-GNSS Data Gateway: News and Novelties, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18507, https://doi.org/10.5194/egusphere-egu25-18507, 2025.

EGU25-20216 | Posters on site | ITS1.17/ESSI4.1

MEET hydrogeochemical monitoring platform for data analytics 

Carlo Cipolloni, Valerio Comerci, Antonio Scaramella, and Fabrizio Terzoni

ISPRA has developed a platform for the collection and analysis of data from a continuous hydrogeochemical monitoring network in the framework of MEET (Monitoring Earth’s Evolution and Tectonics) project, funded by the Ministry of Research (MUR) through the National Recovery and Resilience Plan (Mission 4, Component 2, Investment Line 3.1). The platform allows for near real-time transmission of physico-chemical parameters such as water level, temperature, and electrical conductivity from wells and springs monitored using automated instrumentation.

Hydrogeochemical data, when systematized and integrated with geophysical and geological parameters, are useful for understanding seismic and volcanic activity at different temporal scales, as well as for monitoring water quality and quantity, i.e. environmental protection purposes. Hydrological variations (piezometric levels, spring flow rates, chemical and temperature changes) can reflect changes in the stress field within the Earth's crust. For instance, significant hydrological variations were observed during major earthquakes such as L'Aquila (2009), Emilia (2012), and Amatrice-Norcia (2016), as well as during historical events. However, in non-volcanic areas of Italy, systematic and prolonged monitoring of these parameters is still lacking.

Recent advances in geophysical prospecting and the analysis of hydrogeochemical variations related to volcanic and seismic phenomena have provided valuable information for identifying possible precursors. The existing monitoring network will be expanded with new stations provided by INGV, located at sites identified by ARPAs.

The collected data will be stored in a hybrid cloud system, based in ISPRA, to ensure access, interoperability, and continuous sharing of data at a transnational level, complying with INSPIRE technical standards and the FAIR principle. A new architecture has been designed to collect historical and real-time data, ensuring high quality and compliance. This includes an innovative engine for data storage, validation, and querying, which serves as the core of the system.

The system uses a No-SQL database with native APIs, enabling the publication of data through interoperable OGC INSPIRE services and interactive access via a responsive platform. The choice of a flexible search engine was driven by the need to handle an increasing volume of real-time data while maintaining high performance. An ETL (Extract, Transform, Load) procedure was implemented to transform the relational model into a document-based model, optimizing indexing and enhancing system performance. This approach allows response times up to ten times faster than the previous system.

Data governance is a critical aspect: a well-documented process has been defined to ensure quality and efficiency throughout the entire production cycle. The integration of these technologies significantly enhances monitoring and analysis capabilities, contributing to the development of a national and transnational network for hydrogeochemical and environmental monitoring.

How to cite: Cipolloni, C., Comerci, V., Scaramella, A., and Terzoni, F.: MEET hydrogeochemical monitoring platform for data analytics, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20216, https://doi.org/10.5194/egusphere-egu25-20216, 2025.

Climate change significantly impacts land degradation, posing substantial threats to ecosystems and human livelihoods. This paper proposes an outline to create a digital twin aiming to integrate various datasets for monitoring land degradation and supporting stakeholders. The Digital Twin of Land Degradation (DTLD) will integrates satellite and drone imagery, real-time data, IoT, and field collection data to establish a dynamic and real-time simulation of land degradation at the affected regions. This framework implementation can be used by different end-users like scientific community, private sector, NGO, public community, decision-makers and internation organizations for analysis, report, and support the effective of land management practices. By implementing a nested method, the interconnected areas and their associated physical processes contribute to a larger twinning of the entire region. This is achieved by collecting geospatial information from different sources at different scales and modeling this data in real-virtual time. Ultimately, the DTLD approach should proves invaluable in pinpointing vulnerable areas and informing targeted mitigation strategies. This study underscores the critical need for advanced monitoring tools in combating land degradation and highlights the potential of digital twins in environmental research.

How to cite: Labbaci, A., Barton, K., and Lee, J.: Evaluating the Land Degradation trends: A Digital Twin Approach for Enhanced Environmental Monitoring , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1300, https://doi.org/10.5194/egusphere-egu25-1300, 2025.

EGU25-1584 | Orals | ESSI4.4

World Data Center for Climate – Repository for the Earth System Sciences 

Eileen Hertwig, Andrea Lammert, and Andrej Fast

The World Data Center for Climate (WDCC) is a repository hosted by the German Climate Computing Center (DKRZ) in Hamburg, Germany. It provides access to and offers long-term archiving for data relevant for climate and Earth System research in a highly standardized manner following the FAIR principles. The repository is an accredited regular member of the World Data System (WDS) since 2003 and WDCC is certified as a Trustworthy Data Repository by CoreTrustSeal (https://www.coretrustseal.org).

WDCC services are aimed at both scientists who produce data (e.g. long-term archiving to fulfill the guidelines of good scientific practice) and scientists who re-use published data for new research. In Earth System Sciences often a big quantity of data is produced and needed for climate research. This is especially true for climate model output. To enable scientists to re-use data also across domains, it is essential that data is archived including rich metadata. Before data is published in WDCC, it undergoes multiple checks and curation. Recently WDCC has established its own standard for NetCDF file headers, so that only data fulfilling this standard are accepted for publication.

The WDCC minimal standard is based on CF metadata conventions, but goes beyond the requirements of most conventional CF checkers to ensure rich metadata in file headers. Requirements for the minimal standard have been published in the WDCC user guide to provide clear instructions for our users. A tool to check for WDCC minimal standard in NetCDF files has recently been released for internal use and is currently under development for public use as well. This will make it easier for data producers to check their data before submission to WDCC.

The overall goal of establishing the WDCC minimal standard is to keep and improve high metadata standards at WDCC, to ease the submission process as well as to improve re-usabilty of published data.

How to cite: Hertwig, E., Lammert, A., and Fast, A.: World Data Center for Climate – Repository for the Earth System Sciences, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1584, https://doi.org/10.5194/egusphere-egu25-1584, 2025.

EGU25-2821 | Orals | ESSI4.4

Establishing a Terminology Service for the Earth System Sciences 

Anette Ganske, Markus Stocker, and Angelina Kraft

Earth System Science (ESS) encompasses scientists from different scientific disciplines who use a multitude of heterogeneous terms to describe processes and data. This volume of often ambiguous, duplicate, inconsistent terms presents numerous challenges regarding interoperability and standardisation, e.g. for automated data selection or searching for data in a repository. Terminologies such as ontologies, thesauri and controlled vocabularies can enable scientists and infrastructure providers to realise a machine-processable expression of the information contained in their research data and other scholarly outputs. However, selecting the most appropriate ontology is often difficult and requires support for data producers and curators.

The  BITS1 project is building a Terminology Service (TS) for Earth System Sciences (ESS TS) as part of the existing Terminology Service2 (TIB TS) of the TIB – Leibniz Information Centre for Science and Technology3. It has implemented the ESS collection4 within the TIB TS, which already contains 40 terminologies that are relevant for the ESS and to which further relevant terminologies will be added. All terminologies in the TIB TS are quality controlled. New terminologies for the ESS collection can be suggested at any time via the ESS homepage. New terms for terminologies hosted on Github can also be suggested and forwarded to the developers of that terminology.

One possible use case for the ESS TS is to support the annotation of data with terms from terminologies. A major challenge in this case is the breaking of annotations: this can happen if the term of a terminology used for the annotation is deleted - e.g. in a subsequent version of a terminology. Therefore, BITS is conducting a feasibility study: can we assign persistent handles to all classes and individuals of each future version of a terminology in the  ESS TS collection, so that the handles redirect to a landing page of the respective terms? These handles could then be used in annotations and the TIB will ensure that they are persistent.

The integration of the ESS TS into the two different data repositories of the German Climate Computing Centre (DKRZ) and the Senckenberg - Leibniz Institute for Biodiversity and Earth System Research (SGN) is another task of BITS. The experience gained will be used to develop blueprints for connecting other ESS repositories to the TS. We also work closely with NFDI4Earth and the wider ESS community, and with the BASE4NFDI basic service TS4NFDI. Feedback from the wider ESS community on their expectations and needs for such a service is welcome and necessary for the project. Our goal is a terminology service that serves as a valuable resource for researchers, students, professionals and developers in ESS, providing them with accurate and consistent terminologies to enhance their work, improve communication and data sharing, and advance knowledge in their respective fields.

1: https://projects.tib.eu/bits/home 

2: https://terminology.tib.eu/ts 

3: https://www.tib.eu/en/

4:  https://terminology.nfdi4earth.de 

How to cite: Ganske, A., Stocker, M., and Kraft, A.: Establishing a Terminology Service for the Earth System Sciences, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2821, https://doi.org/10.5194/egusphere-egu25-2821, 2025.

EGU25-4319 | Orals | ESSI4.4

Bridging metadata gaps for FAIR multidisciplinary data access in Virtual Research Environments - Insights from Blue-Cloud2026 and FAIR-EASE 

Paul Weerheim, Peter Thijsse, Dick Schaap, Tjerk Krijger, Alexandra Kokkinaki, and Enrico Boldrini

With Virtual Research Environments (VRE) and digital twins getting more and more common to support multidisciplinary Open Science, there is an ever growing need for the clear discovery and accessibility of data from different domains, FAIR for humans and machines. However, FAIR data and services are not yet the standard. Each domain, and within the domain the different research and data infrastructures, have different API’s, different metadata models, and semantics in place. In order to support multidisciplinary case studies, we need to succeed to bridge the gaps between these different domain-specific (meta)data standards and provide the scientists with a harmonised way of finding, accessing and processing this varying (meta)data.  This can be done in several ways:

  • By creating a common metadata profile, suitable for machine-to-machine communication, to publish (bottom up!) the information about the data access service and the metadata of the datasets, taking also into account domain specific semantics (e.g. to describe parameters, units, etc). Such practices and standards should be tested and afterwards promoted widely, e.g. supported by EOSC. 
  • Since the above profile is not yet implemented, except parts in certain subdomains, syntactic and semantic brokers need to be deployed to harmonise the various models into a central system with its own universal target model.
  • By bringing the various metadata, via brokering, into one place this offers the opportunity to validate and check the metadata for completeness and suitability. Reports can be created to feed back FAIRness levels and improvements to the data publishers.

In this session, several promising activities will be presented related to work done in the Blue-Cloud2026 (marine) and FAIR-EASE (multi-disciplinary) projects, with all activities focusing on use cases requiring data access in a VRE and being linked to EOSC. The activities include:

  • Set-up of a FAIR-EASE RDF Metadata model (DCAT-FE) to describe multi- disciplinary (meta)data uniformly, aiming to bridge the gaps between domain-specific (meta)data standards. 
  • Development of an (Interdisciplinary) Data Discovery and Access Service ((I)DDAS) based on CNR’s Data Access Broker (DAB) that maps the harvested metadata into an ISO19139 target model, including domain specific vocabularies in order to provide harmonized discovery and access. 
  • Reports and semantic analyser: Using the DAB as a broker the various metadata models and their content can be analysed for completeness and existence of semantics via a built in reporting service. Within FAIR-EASE NOC-BODC has developed a semantic analyser that checks for the existence of vocabulary terms, even if not expressed as such in the metadata.

The above work has much overlap with thematics at EOSC level where there are similar challenges in providing data access to the large variety of datasets in the data infrastructures. Both the Blue-Cloud2026 and FAIR-EASE teams are involved in EOSC Task Forces, Opportunity Area working groups and RDA to further promote results and support EOSC in solving this challenge.

How to cite: Weerheim, P., Thijsse, P., Schaap, D., Krijger, T., Kokkinaki, A., and Boldrini, E.: Bridging metadata gaps for FAIR multidisciplinary data access in Virtual Research Environments - Insights from Blue-Cloud2026 and FAIR-EASE, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4319, https://doi.org/10.5194/egusphere-egu25-4319, 2025.

EGU25-4504 | Orals | ESSI4.4

Linking multi-disciplinary data with webODV 

Reiner Schlitzer and Sebastian Mieruch-Schnülle

webODV is the online version of the popular Ocean Data View (ODV) desktop software for analysis and visualization of marine and other environmental data used by more than 10,000 researchers worldwide. Like ODV, webODV provides an interactive graphical user interface and offers rich feature sets via context specific menus. While desktop ODV requires all datasets to reside on the end user machine, webODV works differently. All datasets as well as a special version of the ODV software reside and run on dedicated webODV servers. Users do not have to install any software or download the sometimes-bulky datasets. Instead, users simply connect to datasets using their web-browser. New browser tabs open for every opened dataset, each tab providing an “ODV-like” interactive user interface. Previous ODV users will find it very easy to work with webODV. Concise getting-started documents help guide new users.

Large volumes of important environmental datasets for all parts of the Earth System are accessible from webODV servers at https://explore.webodv.awi.de/ and https://webodv-egi-ace.cloud.ba.infn.it. This includes global TS- and BGC-Argo profile data, GLODAP carbon system data, SOCAT surface fCO2 data, MEOP marine mammals data, and the World Ocean Atlas 2023. In addition, we also provide global collections of historical meteorological as well as river-runoff data.

In this presentation we focus on new developments that greatly facilitate connecting these diverse, multidisciplinary datasets, linking, for instance, ocean interior data with satellite observations, numerical model output as well as with meteorological and river-runoff data. A number of use cases will be shown, including (1) using trusted reference data for quality control of new BGC-Argo float data, (2) calculating and displaying the difference between ocean section repeats, and (3) the analysis and display of model/data differences.

How to cite: Schlitzer, R. and Mieruch-Schnülle, S.: Linking multi-disciplinary data with webODV, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4504, https://doi.org/10.5194/egusphere-egu25-4504, 2025.

EGU25-5136 | ECS | Posters on site | ESSI4.4

Developing Central Support Services for the German National Research Data Infrastructure in Earth System Sciences through a Community-Driven Effort 

Christin Henzen, Anna Brauer, Jonas Grieb, Ralf Klammer, Markus Konkol, Roland Koppe, Kemeng Liu, Johannes Munke, Tom Niers, Daniel Nüst, Tim Schäfer, and Alexander Wellmann

In Earth System Sciences, the landscape of resources for research data management, ranging from specific datasets to RDM services, is very diverse to tackle the heterogeneous and domain-specific (sub) community needs. Although marketplaces for services and central access points for specific data types exist, researchers need assistance discovering resources for research data management tasks without having to navigate multiple, differently designed and not integrated platforms.

By engaging with the communities, NFDI4Earth aims to address the needs and requirements of the end users and includes the following objectives:

  • map the existing landscape of sustainable services and reuse them as data sources or user interface components;
  • align strategies across infrastructure providers/across NFDI4Earth community members to curate provided information for data sources with the providers and share compute resources;
  • develop metadata harvesters or required adaptions together with the service provider's developers, and share the source code with an open license; and
  • reuse open standards and specifications for all (service) interfaces and share own specifications openly.

We apply this strategy for the implementation of our two central support services.

  • The Knowledge Hub (https://knowledgehub.nfdi4earth.de) is a knowledge graph database and acts as an aggregator. It harvests curated information for relevant resources from various sources and makes it available as semantically enabled open data. The solution builds on an open-source data management system and a triple store. Fostering the community effort, we encourage and investigate whenever possible the implementation across Knowledge Hub and data source developers, such as with the Helmholtz Data Hub Earth and Environment. By actively collaborating with the original sources on the curation of the harvested information, e.g., with re3data for repository information, we ensure the distribution of high-quality information via the data sources and the NFDI4Earth Knowledge Hub.
  • The OneStop4All (https://onestop4all.nfdi4earth.de) and the integrated Living Handbook facilitate the discovery of ESS-relevant RDM resources, such as datasets or software, and act as a resource catalog based on the Knowledge Hub data. Through the publication, curation, and integration of related information, such as best practices or showcase articles for published resources, the OneStop4All also acts as a community portal. This approach is implemented by managing Living Handbook articles in an open repository, editing them in an open review process, harvesting them through the Knowledge Hub, and making them visible in a user-friendly presentation. In addition, the OneStop4All offers several search functionalities providing benefits for differently skilled users including RDM novices and experts. In the following steps, we will integrate existing frontend services, e.g., Earth Data Portal map viewers, to facilitate researchers using a single entry point for their RDM tasks.

Our FAIR-by-design services will be iteratively developed with an open and distributed developer team focussing on data access to data publications, with an eye on using the content for advanced data analytics use cases.

How to cite: Henzen, C., Brauer, A., Grieb, J., Klammer, R., Konkol, M., Koppe, R., Liu, K., Munke, J., Niers, T., Nüst, D., Schäfer, T., and Wellmann, A.: Developing Central Support Services for the German National Research Data Infrastructure in Earth System Sciences through a Community-Driven Effort, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5136, https://doi.org/10.5194/egusphere-egu25-5136, 2025.

Climate science relies heavily on the effective creation, management, sharing, and analysis of massive and diverse datasets. As the digital landscape evolves, there is a growing need to establish a framework that ensures FAIRness in handling climate science digital objects. In particular, the machine-to-machine actionability of digital objects will be a crucial step towards future AI-assisted workflows. Illustrated by a use case, this contribution proposes adopting the Fair Digital Object (FDO) standard in synergy with the emerging InterPlanetary File System (IPFS) protocol to address the challenges associated with the interdisciplinary reuse of climate model simulation outputs and observational data.

FDOs are encapsulations of data and their metadata, made accessible via persistent identifiers, ensuring that data and its context remain a complete unit as the FDO travels through cyberspace and time. They represent a paradigm shift in data management, emphasizing machine-actionability principles of FAIRness and the requirements for enabling cross-disciplinary research. The FDO concept can be applied to various digital objects, including data, documents, and software across different research disciplines and industry areas.

IPFS is a peer-to-peer network protocol that enables decentralized file storage and sharing by assigning each file a unique content identifier. This system facilitates efficient, tamper-resistant storage across a distributed network, inherently supporting immutability and version control. Employing IPFS as the access layer for FDOs adds scalability, security, and redundancy to data management frameworks, while FDOs themselves contribute a semantically structured approach to defining, accessing, and linking digital objects.

This work presents a prototypical implementation and highlights the immediate benefits of the described concept when applied to manage data derived from the ORCESTRA (Organized Convection and EarthCARE Studies over the Tropical Atlantic) campaign. The campaign, conducted in August and September 2024, involved gathering data from multiple measurement platforms, including satellite observations, airborne instruments, ground-based systems, and climate model data. From a data management perspective, this multi-sensor campaign offers a valuable opportunity to test and refine concepts for handling large, heterogeneous datasets. As part of this work, selected datasets from the campaign were ingested and transferred via IPFS and included in a public catalog adhering to the FDO standard.

In conclusion, IPFS and FDOs establish a decentralized, verifiable, and interoperable ecosystem for digital objects, effectively addressing the requirements for interdisciplinary scientific data sharing and management. Together, these innovative concepts can significantly enhance the reproducibility of research workflows and strengthen the consolidation of scientific results in the societally and economically critical domain of weather and climate research.

How to cite: Kulüke, M., Peters-von Gehlen, K., and Anders, I.: Advancing Climate Data Use by Leveraging the Synergy of the Fair Digital Object Standard and the InterPlanetary File System Protocol, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5788, https://doi.org/10.5194/egusphere-egu25-5788, 2025.

EGU25-6087 | Posters on site | ESSI4.4

The DATA-TERRA Research Infrastructure : from data to services for integrated use of environmental data under the scope of success interdisciplinary achievements  

Sebastien Payan, Frédéric Huynh, Anne Puissant, Jean-François Faure, Patrice Henry, Emmanuel chaljub, Erwann Quimpert, and Yvan Lebras

The consequences of global change on the Earth system are multiple such as increase in air temperature and sea level, stronger weather events, impacts on ecosystem biodiversity and natural hazards. But the detection of changes and impacts is still difficult because of the diversity and variability of the Earth environments (oceans, land surfaces, atmosphere, solid Earth). While there has been a clear increase in the number of environmental observations, whether by in situ, laboratory or remote sensing measurements, each data is both costly to acquire and unique. The number and variety of data acquisition techniques require efficient methods of improving data availability via interoperable portals, which facilitate data sharing according to FAIR principles for producers and users.

In this context, DATA-TERRA Research Infrastructure (RI) for Earth data, is the entry point to access all the French environmental observation data. As a digital infrastructure, DATA-TERRA works closely with Earth Observation research infrastructures and space agencies. It is backed by a continuum of distributed and interconnected platforms, proposing services that span the full data cycle from access to value-added processing, thus enabling the exploitation of large volumes of data - notably satellite data- and the generation of information through advanced on-demand and systematic processing services. At national, European and international levels, it is advancing the development of open science, implementation of FAIR* approaches, contributing to space missions and applications and to the initiative to generate digital twins of the Earth.

The objective of this talk is to present the DATA-TERRA strategy atmosphere data and products from space to Erath, in the domain of marine sciences (ODATIS Data hub), continental surfaces sciences (THEIA), atmospheric sciences (AERIS), solid Earth science (FormaTerre), and ecological science (PNDB). DATA-TERRA relies on several scientific consortia in order to promote and develop innovative processing methods and products with a focus on success interdisciplinary achievements.

How to cite: Payan, S., Huynh, F., Puissant, A., Faure, J.-F., Henry, P., chaljub, E., Quimpert, E., and Lebras, Y.: The DATA-TERRA Research Infrastructure : from data to services for integrated use of environmental data under the scope of success interdisciplinary achievements , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6087, https://doi.org/10.5194/egusphere-egu25-6087, 2025.

EGU25-8950 | Posters on site | ESSI4.4 | Highlight

VOLCPLUME, an interactive open access web platform for the multiscale monitoring of volcanic emissions and their impacts on the atmosphere 

Marie Boichu, Raphael Grandin, Théo Mathurin, Nicolas Pascal, Christine Deroo, Colette Brogniez, Maximilien Patou, Sylvain Neut, Cédric Tétard, Jérôme Riedi, Luc Blarel, and Philippe Goloub

The open access VOLCPLUME web platform (https://volcplume.aeris-data.fr) is part of the Volcano Space Observatory portal under development within the framework of the Horizon Europe EOSC FAIR EASE project. This web interface aims at supporting the near-real-time monitoring of volcanic emissions and the multi-scale analysis of volcanic plumes in the atmosphere from local to global scales (Boichu and Mathurin, 2022).

To reach this goal, VOLCPLUME allows users to jointly analyse a broad set of satellite and ground-based active/passive remote sensing observations of both volcanic gas and particles, including Low Earth and Geostationary Orbit imagery, spaceborne and ground-based lidar, as well as photometric measurements. The platform also gives access to in-situ ground-level data from air quality monitoring networks. This synergy aims at facilitating the assessment of the multiscale impacts of volcanic plumes on atmospheric chemistry, air quality, aviation safety and climate.

The « SO2 Flux Calculator » (https://dataviz.icare.univ-lille.fr/so2-flux-calculator), a companion web application, also allows for automating the computation of daily SO2 gas flux emissions from Sentinel-5P/TROPOMI observations with a robust noise estimation (Grandin et al. 2024). Regarding volcano monitoring and initialisation of atmospheric models, such interactive tools allow for remotely tracking changes in the degassing or eruptive activities of any isolated or non-instrumented volcano.

For illustration, we present different case-studies including the eruptions of La Palma/Cumbre Vieja, Piton de La Fournaise, Soufrière Saint-Vincent and Hunga Tonga. 

 

Boichu, M. and Mathurin, T. (2022). VOLCPLUME, an interactive web portal for the multiscale analysis of volcanic plume physico-chemical properties [Interactive Web based Ressource], AERIS, DOI : 10.25326/362, Platform access: https://volcplume.aeris-data.fr, Homepage: https://www.icare.univ-lille.fr/volcplume/

Grandin, R., Boichu, M., Mathurin, T. and Pascal, N. (2024). Automatic estimation of daily volcanic sulfur dioxide gas flux from TROPOMI satellite observations: application to Etna and Piton de la Fournaise. J. Geophys. Res. https://doi.org/10.1029/2024JB029309

How to cite: Boichu, M., Grandin, R., Mathurin, T., Pascal, N., Deroo, C., Brogniez, C., Patou, M., Neut, S., Tétard, C., Riedi, J., Blarel, L., and Goloub, P.: VOLCPLUME, an interactive open access web platform for the multiscale monitoring of volcanic emissions and their impacts on the atmosphere, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8950, https://doi.org/10.5194/egusphere-egu25-8950, 2025.

EGU25-10434 | Posters on site | ESSI4.4

Fostering Open Science through the Arctic Permafrost Geospatial Center (APGC) 

Sebastian Laboor, Tillmann Lübker, Joshua Hashemi, Antonie Haas, Andreas Walter, Patrick Willner, and Guido Grosse

Open-access thematic data portals play a critical role in advancing an open data culture by addressing knowledge gaps and reducing uncertainties in research. The Arctic Permafrost Geospatial Center (APGC) serves as a pivotal platform for providing high-quality geospatial data to the permafrost research community. By facilitating easy access to diverse data products, APGC supports multi-scale, interdisciplinary analyses that integrate field observations, remote sensing, and modeling efforts.
APGC’s primary objectives are to ensure (i) the delivery of data that is impactful, user-friendly, and scientifically valuable, and (ii) to streamline data discovery, visualization, and metadata exchange through its comprehensive data catalog, accessible at https://apgc.awi.de. This portal is well-suited to support research initiatives and fulfills requirements for publishing and visualizing data and metadata, a growing necessity in securing project funding.
The APGC catalog accommodates datasets of varying formats, spatial scales, temporal extents, and themes. Users can search datasets by geographic location—either through spatial keywords or interactive map selection—as well as by category, product type, project, tags, keywords, licensing, or data format. Download links provide direct access to repositories such as PANGAEA, ensuring seamless data acquisition.
Built on the open-source CKAN framework and utilizing the DCAT metadata standard, the catalog adheres to FAIR (Findable, Accessible, Interoperable, Reusable) data principles. Each dataset is accompanied by detailed metadata, a concise abstract, and a preview option, with metadata available in multiple formats such as RDF/XML, JSON, or Turtle.
Initially established through the ERC PETA-CARB and ESA GlobPermafrost projects, APGC now hosts over 360 curated datasets from various sources. These datasets encompass a wide range of permafrost-related themes, including surface and subsurface characteristics such as soil temperature, carbon content, ground ice, land cover, vegetation, periglacial landforms, and subsidence. Data formats range from vector and raster to time series and are derived from in-situ measurements, Earth observation, and modeling studies. WebGIS tools further enhances user engagement by enabling interactive exploration of most datasets (https://maps.awi.de & https://apgc-map.awi.de/).
APGC encourages data contributions from individual researchers and project consortia. Submitted datasets undergo evaluation based on criteria such as relevance to permafrost research, scientific significance, accessibility, quality, and metadata completeness. To ensure long-term preservation and accessibility, datasets must be archived in repositories like PANGAEA.
By integrating tools for data publication, visualization, and long-term preservation, APGC provides an essential service for research projects, enabling them to meet funding requirements while advancing the understanding of permafrost systems across the Arctic and beyond.

How to cite: Laboor, S., Lübker, T., Hashemi, J., Haas, A., Walter, A., Willner, P., and Grosse, G.: Fostering Open Science through the Arctic Permafrost Geospatial Center (APGC), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10434, https://doi.org/10.5194/egusphere-egu25-10434, 2025.

EGU25-11410 | Posters on site | ESSI4.4

GALAXY as a Virtual Research Environment Advancing FAIR Principles 

Erwan Bodéré

Galaxy is an open-source Virtual Research Environment (VRE) that empowers researchers by enabling reproducible, collaborative, and efficient workflows across disciplines.Designed to align closely with the FAIR principles, Galaxy integrates diverse tools, scalable data storage solutions, and authentication mechanisms, creating an ecosystem that enhances the accessibility and utility of scientific research Complementing Galaxy, platforms like Zenodo play a role in ensuring the findability and reusability of research outputs, such as datasets and workflows, by providing persistent identifiers (DOIs) and rich metadata.

Galaxy enhances the findability of research artifacts by supporting Persistent Identifiers (PIDs) for tools and workflows, ensuring long-term discoverability. Its integration with RO-Crate facilitates the packaging and description of data and workflows, enabling cataloging and retrieval. Thus, Galaxy fosters reusability by prioritizing reproducibility and modularity. Researchers can share, clone, and adapt workflows enriched with metadata and version histories, enabling replication of results. Searchable repositories such as the Galaxy Toolshed (the appstore where all the tools are listed) further ensure discoverability, allowing researchers to access shared tools and workflows across domains. These capabilities help scientists locate resources efficiently. Furthermore, its open-source nature further encourages community-driven contributions, ensuring Galaxy’s continuous adaptation to evolving research needs.

Accessibility is an important aspect of Galaxy’s architecture. It provides free access to many publicly available instances and supports integration with scalable cloud and object storage systems like S3, ensuring reliable access to datasets regardless of size. Galaxy’s use of Pulsar, a lightweight distributed computing system, extends computational capabilities by enabling workflows to run across diverse remote resources. Security is enhanced through federated identity management, such as EGI Check-in, which facilitates secure access for users worldwide.

Interoperability is at the heart of Galaxy’s design, achieved through compliance with community-driven standards like CWL (Common Workflow Language) and the Galaxy API, ensuring integration with other platforms. Built-in tools like JupyterLab and remote desktop environments enhance the platform’s capacity to interact with external analytical environments and visualization frameworks. Tool containers leveraging Docker and Singularity ensure portability and reproducibility, allowing workflows to operate across heterogeneous infrastructures.

With a vibrant global community driving its development, Galaxy represents a cutting-edge platform that embodies the FAIR principles. By integrating interoperable tools, scalable storage, robust authentication systems, and collaborative frameworks, Galaxy supports researchers in creating, sharing, and reusing scientific knowledge. 

How to cite: Bodéré, E.: GALAXY as a Virtual Research Environment Advancing FAIR Principles, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11410, https://doi.org/10.5194/egusphere-egu25-11410, 2025.

EGU25-12566 | ECS | Posters on site | ESSI4.4

Assessing Warming Trends in the Mediterranean Sea: A Workflow-Based Approach 

Enrico Baglione, Simona Simoncelli, Paolo Oliveri, Marjahn Finlayson, and Sissy Iona

The Mediterranean Sea is warming at a rate faster than the global ocean average, as recent research highlights. This region is particularly vulnerable to climate change due to its distinctive topography and thermohaline circulation patterns. Observational evidence and model-based analyses have revealed considerable shifts in the properties of Mediterranean water masses.

A crucial metric for tracking this phenomenon is the Ocean Heat Content (OHC). This study addresses the challenge of devising a cloud-based workflow to estimate OHC, enabling analysis of its trends across user-defined sub-regions and depth layers within the Mediterranean basin. Developed within the framework of the EU Blue Cloud 2026 project, this application integrates machine-to-machine access to various blue data infrastructures. 

The OHC indicator will be part of a generalized outcome made up of several Marine Environmental Indicators managed by a Virtual Research Environment (VRE) infrastructure that enables users to assess and monitor marine environmental conditions, offering crucial support for informed decision-making in ocean management. By integrating multiple data sources, the future platform will deliver a centralized data analysis service, enabling online computation of indicators through digital tools.

This case study focuses on the use of World Ocean Database temperature data but it can be easily adapted to other data sources such as SeaDataNet, EuroArgo and the Copernicus Marine Service.

The workflow employs the DIVAnd tool to interpolate historical in situ temperature data onto a regular grid, deriving sliding annual and seasonal decadal temperature fields. These fields will be validated with the World Ocean Atlas 2023 and the results will be compared against ocean reanalysis datasets provided by INGV and the Copernicus Marine Service. The primary goal is to uncover OHC trends: this will help to better understand their impact on the regional climate system. 

The anticipated findings will shed light on the spatial heterogeneity of warming trends across different sub-regions and depth layers, emphasizing the intricate relationship between climate change and hydrodynamic processes in shaping the thermal structure of the Mediterranean Sea.

Additionally, the workflow ensures that critical ocean variables are regularly updated and validated using the latest community standards. This advancement will enable rapid and reliable updates of OHC as a key indicator, fostering informed decision-making and efficient responses.

How to cite: Baglione, E., Simoncelli, S., Oliveri, P., Finlayson, M., and Iona, S.: Assessing Warming Trends in the Mediterranean Sea: A Workflow-Based Approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12566, https://doi.org/10.5194/egusphere-egu25-12566, 2025.

EGU25-14957 | ECS | Orals | ESSI4.4

Adapting a Key Semantic Interoperability Innovation from e-Health to Earth Informatics: Are Two-Level Information Models Relevant? 

Moazzam Shah Bukhari Syed, Paula Kelly, Paul Stacey, and Damon Berry

As the current state-of-the-art, ontologies have supported the need for semantic interoperability and data reusability by enabling consistent descriptions of information in the Earth Sciences domain. However, ontologies alone cannot enable semantic interoperability (or data reusability) in the Earth Sciences domain, particularly in the complex modelling of dynamic environments. A challenge lies in integrating heterogeneous Earth Observation data, where high semantic standards, scalability, and flexibility are essential to handle the continuously expanding (and ever-changing) data volume, complex descriptions, and relations of the multifaceted real-world entities, and improve their usability across platforms. Ontologies typically represent comparatively static knowledge which affects their utility in situations like these where modelling dynamic, ever-evolving processes is required on an ongoing basis. They also, at times, lack the necessary descriptive constructs for documenting complex, real-life applications, affecting data interoperability and reusability. In information systems, particularly those in the Earth Sciences, the documentation aspect of defining the concepts, relationships, and terminologies for real-world entities in a dataset is integral to enable semantic interoperability and data reusability. In such continuously evolving environments, Archetypes metadata constructs that can act as high-quality and rigorous documentation templates, exhibit their potential to work together with ontologies to address the challenge of semantic interoperability and maximise the reuse potential of Earth science data. In the archetype-based two-level information models that have been developed over the past twenty years in e-health, the reference model is used to define the core structure and relationships of data while the archetype model is used to provide the domain-specific details separately. A consistent reference model ensures a standardised data format which reduces inconsistencies in data. The archetype model allows definitions that are tailored for a specific domain where archetypes are designed to represent all the necessary information within that domain. The archetypes ensure that all required data fields are collected consistently as defined in the archetype model, ensuring uniform data collection across the domain. The archetypes further help with metadata management by defining data elements and attributes needed for each domain. Within their structure, archetypes also allow the binding of ontologies to provide classification (and associated meaning) of the entities. Over two decades of experience with the use of two-level models for e-health documentation has shown that combining ontologies in an archetype-based two-level information model helps create well-structured and meaningful data and associated information systems. The archetype-based two-level information model approach ensures consistency, enhanced data integration, reusability, and semantic interoperability. This approach, if explored and applied on a domain-wide level, has the potential to make Earth information systems flexible, scalable, more reliable, and efficient in supporting decision-making surrounding environmental sustainability.

How to cite: Syed, M. S. B., Kelly, P., Stacey, P., and Berry, D.: Adapting a Key Semantic Interoperability Innovation from e-Health to Earth Informatics: Are Two-Level Information Models Relevant?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14957, https://doi.org/10.5194/egusphere-egu25-14957, 2025.

EGU25-16118 | Orals | ESSI4.4

Seizing the FAIR-EASE project interdisciplinary opportunity to investigate the Ocean Biogeochemical data in the vicinity of the 2022 record breaking Hunga Tonga Eruption 

Catherine Schmechtig, Marie Boichu, Thierry Carval, Delphine Dobler, Raphael Grandin, Theo Mathurin, Nicolas Pascal, Virginie Racapé, Catalina Reyes, Raphaelle Sauzede, and Reiner Schlitzer

Under the umbrella of the EOSC ecosystem, the FAIR-EASE project funded under HORIZON-INFRA-2021-EOSC-01-04 aims to facilitate access to interoperable data and services for earth and environmental multi-disciplinary use cases, demonstrating the capabilities to support open science (https://fairease.eu/). Based on three of its pilots more specifically: the Volcano Space Observatory pilot, the Ocean Biogeochemical Observations pilot and the Coastal Dynamic pilot, the FAIR-EASE partners would like to highlight both the synergy and the new emerging interdisciplinary collaborations and progresses that can be achieved in the framework of such a European project promoting FAIR principles. 

Indeed,

* The Volcano Space Observatory Pilot supports the implementation of innovative web services (notably here the open access VOLCPLUME web platform) displaying a broad range of satellite and ground-based data relevant to the characterization of volcanic gas and particle properties for the near real-time monitoring of volcanic activity and atmospheric hazards.

* The Ocean Biogeochemical (BGC) Observations aims to provide a common QA/QC (Quality Assessment /Quality Control) platform to the whole BGC community to enhance the BGC data quality and address fundamental scientific questions. 

* The webODV software, part of the Coastal Water Dynamic pilot tools, allows to display and superimpose very heterogeneous datasets (i.e. satellite surface data vs. in situ profiles data, climatology vs. in situ profiles data, observations vs. model simulations in general).

Taking as a starting point, the eruption of the Hunga Tonga-Hunga Ha’apai volcano on January 15, 2022, and the availability of various satellite observations of volcanic plumes and ocean surface properties together with in situ Argo (Argo is an international program that collects information from inside the ocean using a fleet of robotic instruments that drift with the ocean currents) floats measuring BGC variables such as the chlorophyll-a and suspended particles in the eruption area, FAIR-EASE partners aim to investigate the potential impacts of such a major stratospheric eruptions a record breaking eruption in the satellite era, on the marine ecosystem. Volcano and BGC community expertise as well as tools developed and pooled on Galaxy Europe platform (Galaxy is an open-source Virtual Research Environment) during the FAIR-EASE project support scientists in their investigation.

How to cite: Schmechtig, C., Boichu, M., Carval, T., Dobler, D., Grandin, R., Mathurin, T., Pascal, N., Racapé, V., Reyes, C., Sauzede, R., and Schlitzer, R.: Seizing the FAIR-EASE project interdisciplinary opportunity to investigate the Ocean Biogeochemical data in the vicinity of the 2022 record breaking Hunga Tonga Eruption, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16118, https://doi.org/10.5194/egusphere-egu25-16118, 2025.

EGU25-16885 | Posters on site | ESSI4.4

A flexible open brokering framework supporting distributed semantic discovery  

Enrico Boldrini, Roberto Roncella, Fabrizio Papeschi, Paolo Mazzetti, Alexandra Kokkinaki, Gwenaëlle Moncoiffé, Tjerk Krijger, Paul Weerheim, and Dick Schaap

The Discovery and Access Broker (DAB) technology is implemented and deployed in the context of several European Union research projects and international initiatives, such as the Global Earth Observation System of Systems (GEOSS) and WMO Hydrological Observing System (WHOS), to enable discovery and access amongst distributed geospatial data provider services, along the line of fair and open approaches. 

Lately the DAB has been employed in the context of two European Open Science Cloud (EOSC) EU-funded projects FAIR-EASE, and Blue-Cloud 2026, characterized by a strong synergy between them. In FAIR-EASE, a customized DAB instance enables harmonized semantic searches amongst 15 data sources: EuroArgo, the Copernicus Marine Environment Monitoring Service (CMEMS), EasyData, ELIXIR-ENA, EurOBIS, European Environment Agency SDI Catalog, EMODnet, ICOS (Data Portal and SOCAT), Joint Research Centre Data Catalog, SeaDataNet (open datasets and products), US NODC collections, VITO/Copernicus Global Land Services, and WEkEO, totaling 156,000 datasets available for search. 

These services publish data using different service interfaces and data models. The DAB can seamlessly connect to each of them and map results to the interface and data model required by the users, leveraging a flexible and extensible harmonized internal data model based on ISO 19115. 

The broker discovery services are accessible through multiple interfaces, such as OGC CSW, OpenSearch API, and SPARQL endpoint. In particular, the SPARQL endpoint interface follows a linked data approach that is further leveraged by the FAIR-EASE architecture. To realize the SPARQL endpoint, a mapping from ISO 19115 to the FAIR-EASE DCAT profile was implemented. Concept URIs found in the original metadata (e.g., linking parameters, stations, instruments, keywords from online vocabularies) can be easily represented as properties and relations of the mapped FAIR-EASE dataset, supporting further reasoning and harmonized semantic query. 

The connection with the Semantic Analyzer component provided by BODC has been explored and will be the subject of further research. This tool can analyze FAIR-EASE datasets by connecting to the DAB CSW interface, extracting a list of terms and scanning known online vocabularies and ontologies to identify possible concept URIs matches. These URIs can finally replace natural text occurring in the original metadata, enhancing overall semantics and quality. 

How to cite: Boldrini, E., Roncella, R., Papeschi, F., Mazzetti, P., Kokkinaki, A., Moncoiffé, G., Krijger, T., Weerheim, P., and Schaap, D.: A flexible open brokering framework supporting distributed semantic discovery , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16885, https://doi.org/10.5194/egusphere-egu25-16885, 2025.

The French National Institutional Facility for Shared Access to Satellite Imagery (DINAMIS) is a cross-functional component of the National Research Infrastructure DATA-TERRA. It was created in 2018 under the impetus of several funding French organizations (CNES, CNRS, IGN, INRAE, IRD, CIRAD). Its mission is to organize and operate the procurement and the access to both Very-High resolution satellite data (SPOT6-7, Pléiades, among other) and to advanced products derived from Earth observation resources.

To meet these objectives DINAMIS ensures the supply and dissemination of very high resolution datasets meeting the needs (a) of the French scientific communities of Land surfaces (THEIA), Ocean (ODATIS), Solid Earth (FormaTerre), Atmosphere (AERIS) and Biodiversity (PNDB), (b) of French Governmental Organizations, NGOs invested of a public mission, and some private companies involved in R&D projects. In complement, DINAMIS is opened to international collaborations with possible limited access for any scientists belonging to public research laboratories after membership certification. DINAMIS further applies an Open data policy to all Spot 6-7 orthorectified datasets over France (with more than 10 annual coverages of the territory at 1.5 m resolution).

DINAMIS acquires and disseminates to its certified users Very High-Resolution Imagery from past satellites (Spot 1-5, through the SWH/Spot World Heritage repository) and to active satellites (Spot 6-7, Pléiades, Pleiades Neo); it will further be plugged in 2025/26 to up-coming programs (PWH/Pléiades World Heritage; CO3D). All products are pooled in a unified Catalog. The Catalog is regularly updated with a gross 1 Million square kilometers per year with annual Spot 6 national coverages, results from users tasking or Archive requests of Pléiades and Spot 6-7 in response to specific needs. All products are associated to a tailor-made License enabling all certified users to access the datasets.

In complement to the satellite data access, DINAMIS is interfaced to on-demand processing services operated by the Thematic Datahubs of DATA-TERRA. For instance, DINAMIS provides seamless access to the DSM-OPT webservice of FormaTerre allowing the calculation of Digital Surface Models, true ortho-images and ortho-mosaics from Pléiades and Spot 6-7 stereo/tri-stereo images. This services hosts currently ca. 200 users with 15 to 20 on-line processing per week. All generated surface models are distributed with a CC-BY-NC license through the FormaTerre and THEIA catalogs, allowing to promote open science principles.

How to cite: Faure, J.-F., Mallet, J.-P., Pointal, E., Debard, S., and David, M.: DATA-TERRA -DINAMIS/ForMaTerre: a Very High Resolution Satellite Data Facility for sharing access to optical Earth Observation resources and to on-line processing services for Digital Surface Models creation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16948, https://doi.org/10.5194/egusphere-egu25-16948, 2025.

EGU25-17005 | Posters on site | ESSI4.4

Building semantic bridges between multi-domain scientific data resources 

Alexandra Kokkinaki, Gwenaelle Moncoiffe, Christelle Pierkot, and Guillaume Alviset

Data silos exist as a result of domain-specific semantic and syntactic legacy but they continue being created as an inevitable consequence of the exploratory and experimental nature of scientific investigation. New technologies are developed and adopted, new variables are measured, new terms and concepts are defined, new formats are required, and all this within typically fairly narrow and highly specialised fields of research. Data sharing technologies must adapt to this evolving environment. Flexibility and connectivity between neighbouring or overlapping fields of research is key. Bridging the semantic gaps and discrepancies to enable seamless discovery, sharing and exploitation of data is the challenge. 

Achieving cross-domain interoperability requires the establishment and harmonisation of both the syntax and semantics of datasets. The Semantic Analyser was developed to address the semantic challenge by scanning metadata records and data files to identify and analyse the semantics used for specific metadata elements, focusing on instruments, parameters, platforms, and keywords.

To determine whether the values for these metadata elements originated from semantic artefacts, we initially explored leveraging existing large semantic repositories, such as Earth Portal and BioPortal. These repositories offered extensive semantic artifacts, potentially reducing the effort required to match terms. However, this approach presented two significant challenges: (1) implementing a federated service for term matching against these repositories proved to be slow and inefficient, and (2) the large number of matched terms generated confusion among users, largely due to the difficulty of selecting appropriate vocabularies and ontologies for specific domains and targeted context. 

To overcome these obstacles, we decided to construct a dedicated knowledge base (KB) containing well-known vocabularies relevant to the datasets in focus. The KB was iteratively refined as new insights were gained, providing a streamlined and domain-specific solution for semantic harmonization and improving the usability and performance of the Semantic Analyser.

How to cite: Kokkinaki, A., Moncoiffe, G., Pierkot, C., and Alviset, G.: Building semantic bridges between multi-domain scientific data resources, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17005, https://doi.org/10.5194/egusphere-egu25-17005, 2025.

EGU25-17496 | Orals | ESSI4.4

Sea Observations Utility for Reprocessing, Calibration and Evaluation (SOURCE): software update for virtual environment workflows 

Simona Simoncelli, Claudia Fratianni, Paolo Oliveri, Paolo Frizzera, Charles Troupin, and Reiner Schlitzer

The Sea Observations Utility for Reprocessing, Calibration and Evaluation (SOURCE) is a software tool (DOI: 10.5281/zenodo.5008245) designed for web applications that permits to visualize and analyze in-situ observations, and at the same time, calibrate and validate ocean models in a selected sea region. Within the framework of the EU FAIR-EASE project (https://fairease.eu/), the SOURCE python code has been updated in GitHub repository and adapted in a Jupyter notebook for use in a cloud-based Coastal Water Dynamics Pilot focused in the Northern Adriatic Sea.

The workflow aims to provide the users with an interactive experience for the analysis of the coastal water dynamics through multiple data sources (in situ, remote sensed and model data). The deployment of SOURCE, together with DIVAnd and webODV tools on the Galaxy platform represents a significant advancement in enabling interactive workflows for the analysis of the coastal marine environment, where important processes, such as the evolution of plankton blooms or the transport and fate of nutrients, carbon, and contaminants take place and depend critically on many factors. 

Nowadays, in-situ observations and ocean models can be freely accessed from several marine data services together with the relative metadata information. FAIR data access services and input data allow the user to access and filter the data according to the level of quality required for the intended use. However, the in situ data might still contain some anomalous data (bad data flagged as good) or need a higher level of quality control to be fit for use. SOURCE uses moored temperature and salinity observations, ocean reanalysis distributed by the Copernicus Marine Service and climatologies computed with the DIVAnd tool (DOI: 10.5281/zenodo.1303229) within the same cloud-based environment.

SOURCE tool performs a re-processing of observations and model data, computes accurate model skill scores and provides output data for visualization and further analysis. Data reprocessing allows one to characterize temperature and salinity variability at each mooring site and continuously monitor the ocean state. The SOURCE output includes climatologies, trends, averages at different time scales and model skill scores at the mooring location, which could also be useful to visually inspect both model and mooring sensor performance.

The output files in NetCDF format can be visualized through web applications such as webODV and downloaded for further use. 

A demonstration of SOURCE capabilities will be presented.

How to cite: Simoncelli, S., Fratianni, C., Oliveri, P., Frizzera, P., Troupin, C., and Schlitzer, R.: Sea Observations Utility for Reprocessing, Calibration and Evaluation (SOURCE): software update for virtual environment workflows, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17496, https://doi.org/10.5194/egusphere-egu25-17496, 2025.

EGU25-19337 | Posters on site | ESSI4.4

Essential Variables as a semantic interoperability solution for the Green Deal Data Space 

Ivette Serral, Victoria Lush, Joan Masó, Lucy Bastin, Raul Palma, and Alejandro Villar

The Green Deal Data Space is born in the big data paradigm where many sensors produce constant streams of Earth observation data (remote sensing and in-situ). The traditional and manual organization of data in layers is no longer efficient as data is constantly evolving and combined in new ways. There is a need for a new organization and representation of the data. Data Spaces are intended to become the EC’s comprehensive solution to integrate data from different sources with the aim to generate and provide a more ready to use knowledge in support of the Green Deal priority actions on climate change, circular economy, pollution, biodiversity, and deforestation.

Semantics need to be moved from the layer level to the property level. The solution is not a comprehensive new data model, but a framework composed by a suite of ontologies implemented in line with best practices, reusing existing standard vocabularies, such as Essential Variables (EVs). These are used in Earth observation to define variables that correspond to high impact on the Earth systems and should be a priority for monitoring. EVs assume that there is a limited number of variables that are essential to characterize the state in a system without losing significant information on its past and future trends.

In AD4GD (Horizon Europe nº 101061001), we are describing EVs following the I-ADOPT ontology, starting with the Essential Biodiversity Variables, and extending the model to describe the concept of EV products, where products define spatial and temporal resolution constraints. I-ADOPT (Barbara Magagna et al, RDA) is an ontology framework designed to facilitate interoperability between existing variable description models across research domains. It provides a common set of core components and relations to define machine-interpretable variable descriptions that re-use FAIR (Findable, Accessible, Interoperable, Reusable) vocabulary terms.

I-ADOPT defines 5 classes (Variable, VariableSet, Property, Entity, and Constraint), and specifies several object properties. Thus, i-ADOPT enables the decomposition of complex observable properties into essential atomic parts represented through concepts in FAIR terminologies, serving as a common layer of abstraction to systematically align and extend concepts from different terminologies as needed.

EVs expressed under this new ontology schema are being translated as linked data to become available via the OGC RAINBOW Definition Server (https://defs.opengis.net/vocprez/). RAINBOW will ensure that Essential Variables and Products are assigned unique and persistent identifiers facilitating wider adoption and reuse.

How to cite: Serral, I., Lush, V., Masó, J., Bastin, L., Palma, R., and Villar, A.: Essential Variables as a semantic interoperability solution for the Green Deal Data Space, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19337, https://doi.org/10.5194/egusphere-egu25-19337, 2025.

EGU25-21496 | Orals | ESSI4.4

Enhancing Earth Science Research through the FAIR-EASE Data Lake Infrastructure: Integrating Diverse Data Sources for Advanced Computational Services 

Samuel Keuchkerian, Vincent Breton, David Sarramia, Marc Portier, Antoine Mahul, and Erwan Bodere

The FAIR-EASE Data Lake infrastructure is a pivotal development in Earth sciences, providing a deep depth in cloud approaches that significantly enhances the accessibility and utility of complex data for the earth science research community. At its core, the infrastructure integrates diverse data sources, including Copernicus, enabling comprehensive environmental and geophysical analyses, and several existing infrastructures. A key strength of the FAIR-EASE datalake lies in its sophisticated collaborative framework. It incorporates features from established environments like the European Grid Infrastructure (EG.euI), Galaxy.eu, D4Science, and the UCA test bed, along with several analytics tools that collectively enhance the infrastructure's operational and security capabilities. This setup will ensure high levels of interoperability and facilitate the usage of data and data sources across various scientific domains in the earth science domain. The integration among the five strategic pilot projects—Coastal Water Dynamics, Earth Critical Zones, Ocean Biogeochemical Observation, Marine Omics Observations, and Volcano Space Observatory—demonstrates the infrastructure's unique capability. These projects benefit from shared data access, whatever their format and access protocol and synergistic interactions among their data sources, allowing for innovative correlations, such as combining satellite data with biological data and in-situ data. This synergy provides new insights into biodiversity patterns or ecosystem health, showcasing the power of cross-disciplinary data integration. By providing discovery and access to diverse data sources and offering advanced analytical tools in a secure, collaborative environment, the FAIR-EASE Data Lake is pioneering new methodologies that transcend traditional disciplinary boundaries. It exemplifies the transformative potential of integrated data systems using distributed infrastructures in advancing our understanding of Earth’s dynamic systems. This has been done by identifying totechnical solutions tackling this distributed way of working used in other communities (such as Galaxy), identifying, integrating and deploying cloud data and data management tools (such as S3, Apache Iceberg). By developing tools enabling data discovery namely IDDAS for Interdisciplinary Data Discovery Access Service) and libraries namely UDAL for Uniform Data Access Layer, in combination with the proposal to use and include in users practices Amazon S3 API for data access, FAIR-EASE datalake has given the opportunity to include cloud technologies in pilots practices and to take advantages of distributed data resources in a very transparent way. In conclusion, the FAIR-EASE Data Lake infrastructure sets new standards for data-driven research and data-analytics in Earth sciences. By merging extensive data access with sophisticated computational resources and a robust collaborative framework, it empowers researchers to expand the frontiers of knowledge about Earth systems and their complex interactions.

 

 

 

How to cite: Keuchkerian, S., Breton, V., Sarramia, D., Portier, M., Mahul, A., and Bodere, E.: Enhancing Earth Science Research through the FAIR-EASE Data Lake Infrastructure: Integrating Diverse Data Sources for Advanced Computational Services, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21496, https://doi.org/10.5194/egusphere-egu25-21496, 2025.

EGU25-329 | ECS | Orals | ESSI4.5

High-resolution Spectral Characterization of Heterogenous Carbonate Lithofacies and their Controlling Factors 

Ahmed Hammam, Ardiansyah Koeshidayatullah, and Khalid Al–Ramadan

Discriminating carbonate lithofacies demands extensive fieldwork and costly geochemical analysis which is often constrained by outcrop accessibility and the number of studied sections. To overcome these challenges, the present work employs integration between spaceborne multispectral and lab hyperspectral datasets for the late Jurassic carbonate lithofacies discrimination and spectral characteristics. Carbonate minerals spectra show diagnostic absorption features in the shortwave infrared (SWIR) while impurities features exist in the visible near-infrared (VNIR)- SWIR wavelength region. Several chemical and physical factors affect the position and depth of carbonate minerals absorption features. In this study, Hanifa Formation in Central Saudi Arabia has gained economic significance as a key conventional and unconventional reservoir which is divided into Hawtah and Ulayyah members. The Minimum Noise Fraction (MNF) and Principal Component Analysis (PCA) image processing techniques were utilized on ASTER and Sentinel 2A multispectral data successfully differentiated the Hawtah member into three units and the Ulayyah member into twelve units for the first time which was verified by detailed microfacies analysis. Hyperspectral laboratory measurements of the newly identified units have characterized their absorption features, leading to the classification of four distinct spectrofacies within the Hanifa Formation. Each absorption feature corresponds to a specific mineralogy which shows high consistency with X-ray fluorescence, X-ray diffraction, and Scanning electron microscopy analyses. Also, Hyperspectral measurements showed that chemical factors, such as mineralogy, Mg, Fe, and clays, influence the depth and position of absorption features, while physical factors like grain size, porosity, and weathering primarily affect the reflectance values in the VNIR-SWIR wavelength ranges. This study demonstrates the effectiveness and high accuracy of using integrated multi and hyperspectral data to distinguish and characterize carbonate lithofacies. These methods can be applied worldwide for reservoir/ aquifer characterization, as well as for in-situ limestone quality control in the cement industry.

Keywords

Carbonates lithofacies, Absorption features, Hyperspectral, Multispectral, Spectrofacies.

How to cite: Hammam, A., Koeshidayatullah, A., and Al–Ramadan, K.: High-resolution Spectral Characterization of Heterogenous Carbonate Lithofacies and their Controlling Factors, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-329, https://doi.org/10.5194/egusphere-egu25-329, 2025.

EGU25-4007 | Posters on site | ESSI4.5

Using active and passive remote sensing techniques to quantify the surface deformation and lithology of salt diapirs, Zagros Mountains, southern Iran 

Stefanie M. Rieger, Mugabo Dusingizimana, Prokop Závada, Christina Plattner, Ramon Brcic, Beth Kahle, and Anke M. Friedrich

Salt diapirs and their caprocks are strategically significant for natural resource exploration and as potential sites for nuclear waste and CO2 storage. However, direct study of these systems is challenging because most diapirs are not exposed at the Earth’s surface. The Zagros Mountains in Iran, with their numerous exposed salt diapirs and caprocks, provide a rare and valuable opportunity to investigate the dynamics of active diapir-caprock systems.

In this study, we combine traditional fieldwork, space-based geodetic mapping, remote spectral analysis, and petrology to analyze the active processes and driving forces that shape salt diapir surfaces within the interconnected climate-diapir-caprock system.

The quantification of surface deformation of salt diapirs and their composition is challenging to map in field campaigns due to their rough terrain and remote location in the Zagros Mountains, southern Iran. To better understand patterns of the salt diapir’s surface deformation and composition active and passive remote sensing techniques are essential. However, the contemporary vertical surface deformation pattern is difficult to detect and interpret along disciplinary boundaries. With the aid of high-resolution PSI measurements and multispectral imagery analysis we detected high-precision spatiotemporal deformation patterns of the surfaces of several salt diapirs. In addition, time-series analysis helped to distinguish between salt-supply-driven domal uplift and vertical surface modification induced by precipitation, dissolution, and erosion.

We analysed Sentinel-1 PSI time-series, processed by the German Aerospace Center (DLR), to obtain the highest available spatiotemporal resolution of the vertical surface-deformation pattern across three diapirs – Karmostaj, Siah Taq, and Champeh – in the Zagros Mountains. We then correlated the Persistent Scatterers to the respective diapir’s composition based on multispectral analysis of Advanced Spaceborne Thermal Emission and Reflection Radiometer (ASTER) satellite images. Preliminary results indicate that the deformation pattern of the salt diapirs does not correlate with seasonal effects, such as precipitation and heat. The vertical surface deformation pattern on these three diapirs implies that these diapirs are active. We conclude that the strategic integration of space-based geodesy and remote spectral analysis provides an effective method for interpreting the complex surface deformation patterns of salt diapirs. The activity of salt diapirs should be considered a key factor in resource exploration, as well as in the evaluation of sites for nuclear waste and CO2 storage.

How to cite: Rieger, S. M., Dusingizimana, M., Závada, P., Plattner, C., Brcic, R., Kahle, B., and Friedrich, A. M.: Using active and passive remote sensing techniques to quantify the surface deformation and lithology of salt diapirs, Zagros Mountains, southern Iran, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4007, https://doi.org/10.5194/egusphere-egu25-4007, 2025.

Technological development allows new information to be derived from existing data. This saves financial resources while developing and detailing geological products such as maps. Using the data obtained for the geological map of the seabed and applying a new approach to data classification and visualisation, an information layer was created that represents a new approach to geomorphological mapping of the seabed of Poland (Southern Baltic Sea).

Data from extensive studies carried out by the Polish Geological Institute - NRI were used to produce this layer. The bathymetric background was obtained from the European Marine Observation and Data Network (EMODnet) project. Vocabularies and manuals developed by the EMODnet project team were also used to classify and visualise the data.

In the first phase of the work, bathymetric information was used. This was used to create a morphometric model using the Benthic Terrain Modeler (BTM) 3.0 for ArcGIS tool to determine the shapes present on the seafloor. The next step was to combine the resulting morphometric model with geological and genetic information from geological studies. The third step was to classify the extracted bathymetric-geological forms according to the EMODnet vocabularies and manuals. The fourth step was to generalise and unify the data for dissemination through a dedicated EMODnet Map Viewer.

How to cite: Pączek, U., Kaulbarsz, D., and Szarafin, T.: A New Perspective on Marine Geomorphological Mapping through Advanced Data Visualization in the southern Baltic Sea area (preliminary results), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5534, https://doi.org/10.5194/egusphere-egu25-5534, 2025.

EGU25-6065 | ECS | Posters on site | ESSI4.5 | Highlight

A New Geological Map of the Apollo 15 Landing Site and Its Implications 

Wajiha Iqbal, James W. Head, David R. Scott, Carolyn H. van der Bogert, Lukas Wueller, and Harald Hiesinger

The Apollo mission data and samples have led to substantial advancements in our understanding of the Moon's geological history and processes. By incorporating new data from recent orbital missions, we systematically developed high-resolution geological maps for each Apollo landing site [e.g., 1-3]. The present study offers a detailed geological map for the Apollo 15 site. The Apollo 15 mission is noteworthy for its significant contributions to lunar geology, leading to substantial advancements in our understanding of volcanic activity, impact cratering, and the Moon's thermal evolution. Notwithstanding this progress, there are as yet unanswered scientific questions, which have been articulated as objectives for future missions such as the 500-day Hadley Max design reference mission (DRM) [4,5].

The Apollo 15 landing site is located east of Hadley Rille on mare basalts that border the Imbrium basin. A thorough geological mapping of the area has revealed the presence of multiple units associated with the Imbrium basin, including its rim and ejecta deposits. These units have been classified based on their distinguishing topographic features. The surrounding area also contains plains deposits, such as Imbrian light plains, along with several mare basalt units of Eratosthenian and Imbrian age [6]. Materials from nearby craters, Autolycus and Aristillus [7,8], also contribute to the region's geological diversity. The linear rilles in proximity to the site have been mapped and categorized by age, employing a combination of stratigraphic relationships and morphological analysis.

The newly developed maps have enhanced the measurement of crater-size frequency distributions (CSFDs), leading to improved N(1) values and a refined lunar cratering chronology [1-3]. Furthermore, the maps facilitate the identification of potential sample sources, thereby enhancing our comprehension of lunar stratigraphy [4,5]. Finally, these maps provide a fundamental framework for the evaluation of in-situ resources and the testing of novel technologies for forthcoming lunar missions [9].

[1] Iqbal et al. (2019) Icarus 333, 528-547.

[2] Iqbal et al. (2020) Icarus 352, 113991.

[3] Iqbal et al. (2023) Icarus 407, 115732.

[4] Daniti et al. (2024) LPSC 55, #1667.

[5] Iqbal et al. (2024) LPSC 55, #1010.

[6] Hiesinger et al. (2000) JGR 105, 29239-29275.

[7] Hiesinger et al. (2000) JGR 105, 29239-29275.

[8] Carr et al. (1971) USGS, I-723.

[9] van der Bogert, et al. (2020) LPSC 51, #1876.

How to cite: Iqbal, W., Head, J. W., Scott, D. R., van der Bogert, C. H., Wueller, L., and Hiesinger, H.: A New Geological Map of the Apollo 15 Landing Site and Its Implications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6065, https://doi.org/10.5194/egusphere-egu25-6065, 2025.

EGU25-7710 | Orals | ESSI4.5

LithoSpace: A Democratised Spatial Data Platform for Extraterrestrial Geoscience and Geochemical Mapping 

Wayne Noble, Fabian Kohlmann, Jack Gale, Ben Dib, Gale Iles, Brandon Mahan, and Moritz Theile

With renewed interest in space exploration and the search for extraterrestrial resources, visualising spatial data from celestial bodies like the Moon, Mars, asteroids and other celestial bodies is becoming increasingly important. LithoSpace provides a digital infrastructure to address this need. It allows users to visualise and analyse extra-terrestrial spatial data, including points of interest, collected samples, corresponding geochemistry, and other analytical data, as well as to collate existing data and generate new data and therefore value and insight.

LithoSpace builds upon the proven technology of LithoSurfer, designed for terrestrial data types. This presentation demonstrates how LithoSpace's advanced data analytics and exploration tools can benefit the expanding frontiers of extra-terrestrial resource exploration. The platform's highly detailed relational data models enable the collation and analysis of diverse data types, uncovering relationships and patterns in data collected from rovers or probes, satellite imagery, and topographic features. Standardised data formats empower researchers and explorers to leverage advanced algorithms for in-depth, automated exploration of these datasets.

This study showcases how LithoSpace's unique cloud-based geochemistry tools can visualise slight variations in geochemical composition using existing, standardised, and cleaned lunar and martian geochemical data. The analysis confirms previously known findings, such as the basaltic geochemical composition of Apollo 11 samples and the wide range of geochemical composition of rocks on Mars as analysed by the Curiosity rover. However, it more importantly highlights how LithoSpace facilitates improved, user-friendly analytics, enabling “on-the-fly” calculation, interpretations and rock classifications. As more data is collected, LithoSpace will enhance our ability to develop new theories about planetary formation and assist with improved geological mapping of extra-terrestrial bodies.

LithoSpace empowers users with the latest technology and data science to navigate the initial stages of lunar exploration for mineral resources. The robust toolkit developed for terrestrial samples can be readily applied to analyse the influx of data from upcoming missions, potentially leading to groundbreaking discoveries and unlocking the hidden resources of our celestial neighbors. Furthermore, standardised and cleaned datasets within LithoSpace (https://app.lithospace.com/) pave the way for the application of advanced machine learning and artificial intelligence, ultimately refining interpretations and creating models for future space exploration endeavors.

How to cite: Noble, W., Kohlmann, F., Gale, J., Dib, B., Iles, G., Mahan, B., and Theile, M.: LithoSpace: A Democratised Spatial Data Platform for Extraterrestrial Geoscience and Geochemical Mapping, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7710, https://doi.org/10.5194/egusphere-egu25-7710, 2025.

EGU25-9042 | ECS | Orals | ESSI4.5

New Geological Maps of the Amundsen and Rubin Crater Regions Near the Lunar South Pole. 

Lukas Wueller, Wajiha Iqbal, Thomas Frueh, Carolyn H. van der Bogert, and Harald Hiesinger

The lunar south pole region is a high-priority exploration target due to its unique geological history, potential resources in permanently shadowed regions (PSRs), and regions of nearly continuous sunlight [e.g., 1-5]. In this context, the Amundsen crater region offers significant opportunities to investigate impact processes, lunar evolution, and volatile distribution, which we have presented through geologic mapping at a regional scale of 1:100,000 for the Amundsen crater region [1] and at a local scale of 1:30,000 for Rubin crater [2].

We produced a 1:100,000 scale geomorphologic map delineating basin materials, crater materials, and modified surface units, allowing a comprehensive reconstruction of the geologic history of the region [1]. Amundsen’s proximity to the proposed outer rim of the South Pole-Aitken (SPA) basin suggests that its ejecta may have reworked ancient lunar rocks, making it an invaluable site for sampling and understanding south polar impact history. Crater size-frequency distribution (CSFD) measurements yield an Amundsen crater formation age of ~4.04 Ga [1].

Within the Amundsen crater region, we have identified five Areas of Interest (AoIs), which are scientifically valuable regions such as the plains near Idel’son L and Rubin crater on the Amundsen rim [1]. These AoIs meet critical technical criteria, including gentle slopes, sufficient solar illumination, and Earth visibility to ensure safe landings and operational feasibility. Because Rubin crater (~4 km diameter), located on the northwestern rim of Amundsen, offers perfect conditions for a safe landing and may have direct access to the SPA material, we present a higher-resolution (1:30,000) map of this region [2].

The high-resolution mapping shows that Rubin crater’s terrain hosts boulders, PSRs, and fresh craters, serving as prime sampling targets for robotic and human exploration. To minimize risk and optimize science return, we evaluated candidate landing sites and traverse options near Rubin crater, considering engineering constraints such as slope limits and energy requirements. Detailed geologic mapping of the Rubin ejecta and surrounding terrain [2] provides insight into its potential as a science- and resource-rich site and its role as a testbed for operations in more challenging polar terrain.

Our mapping and analysis of the Amundsen region highlight its ability to address key lunar science objectives [3,4]. Sampling of Amundsen and Rubin ejecta can refine the lunar chronology and improve our understanding of lunar differentiation and early Solar System dynamics. Additionally, the study of PSRs can reveal the composition, distribution, and stability of lunar volatiles, which is critical for resource utilization [4,6]. By integrating regional and site-specific geologic data, we provide a framework for mission planning that maximizes scientific return while ensuring safety. These efforts confirm the Amundsen region’s status as a key location for advancing lunar science and exploration.

 

 

[1] Wueller et al. (2024) PSJ 5(6).

[2] Wueller et al. (2025) submitted to Adv. In Space Res.

[3] National Research Council (2007) The Scientific Context for Exploration of the Moon, National Academic Press.

[4] Artemis Science Definition Team (2020) Artemis Science Definition team Report.

[5] Krasilnikov et al. (2023) Icarus, 394.

[6] Crawford et al. (2023) Reviews in Mineralogy and Geochemistry, 89(1), 829-868.

How to cite: Wueller, L., Iqbal, W., Frueh, T., van der Bogert, C. H., and Hiesinger, H.: New Geological Maps of the Amundsen and Rubin Crater Regions Near the Lunar South Pole., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9042, https://doi.org/10.5194/egusphere-egu25-9042, 2025.

This work follows on from previous work on 3D photogrammetric modelling of rock masses, using the Structure from motion (SfM) technique and dense correlation as a basis. 
Photogrammetric 3D modelling is a technique that uses photographs to create three-dimensional models of objects or environments. In the context of rock masses, this approach is especially relevant for understanding the geometry and structure of these formations.
Based on this concept, this work was developed with the integration of techniques, photogrammetric data acquisition and 3D photogrammetric modelling, from images acquired with an Unmanned Aerial Vehicle (UAV), and virtual reality (VR) technology, allowing users to explore the 3D model interactively.  In this way, a realistic virtual environment is created, with the feeling of being present and experiencing a full-scale simulated environment, which makes it possible to visualise the rock mass from different perspectives and assess its structural geological characteristics and monitoring.

The application of this technique to other areas of study (environmental changes and natural hazards), using other sensors (multispectral/hyperspectral optical, thermal sensors, LIDAR) that allow techniques for data fusion, is being studied considering the development of multi-platform and inter-disciplinary surveillance.

How to cite: Duarte, J. and Cardoso, J.: Integrated use of geotechnologies and virtual reality to visualize and evaluate rock masses. Case study: Fátima-Portugal, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9072, https://doi.org/10.5194/egusphere-egu25-9072, 2025.

EGU25-9490 | ECS | Posters on site | ESSI4.5

Mapping of Fractures at the ExoMars Rosalind Franklin Landing Site 

Andrea Apuzzo, Alessandro Frigeri, Monica Rasmussen, Maria Cristina De Sanctis, and Francesca Altieri

Fractures are ubiquitous in rocks, representing the mechanical stresses exerted on geological materials. They are also of considerable biological interest because of their pivotal role in facilitating fluid circulation within the subsurface. The search for signs of life beyond Earth drives the European Space Agency (ESA) ExoMars Rosalind Franklin (RF) rover mission, which selected the phyllosilicate-rich region of Oxia Planum (latitude 16 °-19 ° N, longitude 23 ° 28 ° W), Mars, as its landing site. In this context, the identification and characterization of fractures are critical in guiding the search for potential biosignatures. Fracture patterns, with spacings ranging from meters to tens of meters, are observable in the region through the Mars Reconnaissance Orbiter (MRO) HiRISE camera, which provides high-resolution optical remote sensing imagery at a resolution of 30 cm per pixel. While the ExoMars team conducted a geological survey focused on the "one-sigma" landing ellipse (approximately 66.75 × 5 km, corresponding to a 67% probability of landing), we initiated a systematic mapping of fractures using HiRISE data through a grid-based mapping approach (1 km by 1 km). Our 1:50,000 scale map represents the current understanding of the spatial distribution of fractures across the "three-sigma" landing ellipse (approximately 115 × 15 km, with a 99% probability of touchdown). Fractures are classified into three categories based on their visibility at 1:5,000 map scale: clearly observable, barely observable, and not observable. By using open geospatial formats, we ensure that datasets produced at different times and in different contexts remain comparable. In this study, we compare our map of fractures with the existing geological map of the Rosalind Franklin landing site, highlighting similarities and differences. By implementing a grid-based mapping approach, we aim to extrapolate additional information and extend the current understanding of the region, providing critical information to support the surface operations of the RF rover. This extended dataset will contribute to the planning of rover exploration activities, provide a framework for testing geological hypotheses about the formation and evolution of Oxia Planum, and facilitate the identification of astrobiologically significant terrains with the potential to preserve biosignatures.

Acknowledgments: This work is supported by the ASI-INAF Mars Exploration agreement code 2023-3-HH 0. 

How to cite: Apuzzo, A., Frigeri, A., Rasmussen, M., De Sanctis, M. C., and Altieri, F.: Mapping of Fractures at the ExoMars Rosalind Franklin Landing Site, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9490, https://doi.org/10.5194/egusphere-egu25-9490, 2025.

EGU25-9518 | Posters on site | ESSI4.5

Geologic Map of the Poincaré Region, Moon. 

Lukas Wueller, Tessa Theiner, Wajiha Iqbal, Carolyn H. van der Bogert, and Harald Hiesinger

The lunar South Pole-Aitken (SPA) basin, the oldest, largest, and deepest impact basin on the Moon, is a high-interest target for future lunar exploration missions due to its unique geology and insights into the formation and evolution of the Moon [1]. Existing large-scale geological maps of the SPA basin [2-6] have improved our understanding of regional and global geological units [7]. However, these large-scale maps do not provide sufficient information about the detailed geology within SPA, including superposed impact basins such as the Poincaré basin, which is critical for mission planning, landing site safety assessment, resource utilization potential, and traverse planning. This study presents a detailed 1:200,000 scale geological map of the Poincaré region (159°E to 179°W, 48° to 64°S). With a diameter of 349 km, the Poincaré basin is one of the largest multiring basins superposed on the southwestern floor of SPA [3, 8-10]. Our map aims to provide the necessary geologic context for detailed planning of future exploration missions, such as the Endurance mission [2, 7].

We utilized high-resolution images from the Lunar Reconnaissance Orbiter (LRO) Wide-Angle Camera (100 m/pixel) [11], topographic data from the LRO Lunar Orbiter Laser Altimeter (118 m/pixel) [12], and Clementine data (100 m/pixel) [13]. Our map follows the standards of the Federal Geographic Data Committee [14], following the stratigraphic scheme originally proposed by [6].

We identified 10 geologic units, categorized as terra, plains, and crater materials. Our stratigraphy is based on superposition relationships and degradation stages, with absolute model ages available from the literature [e.g., 4,6,9]. The Poincaré region has a complex history dominated by impact and volcanic processes. The southern central parts of Poincaré crater are crossed by the traverses designed for the NASA Endurance rover [2]. Hence, our geologic map can contribute to the development of this mission, the definition of its scientific objectives, and evaluating the different lithologies that could be sampled by this mission and eventually returned to Earth.

[1] Duke, (2003), Adv. Space Res. 31

[2] Keane et al. (2021), Endurance mission concept

[3] Poehler et al. (2020), LPSC 51 #1951

[4] Fortezzo et al. (2020), Unified Geologic Map of the Moon, 1:5M, USGS

[5] Yingst et al. (2017), LPSC 48, #1964

[6] Wilhelms et al. (1987), USGS Prof. Pap. 1348

[7] Mouginis-Mark et al. (2021), Bull. AAS 53

[8] Pasckert et al. (2018), Icarus, 299

[9] Poehler et al. (2021), EPSC2021-646

[10] Spudis (2008), Cambridge University Press

[11] Robinson et al. (2010), Space Sci. Rev., 150

[12] Barker et al. (2016), Icarus, 273

[13] Pieters et al. (1994), Science, 266

[14] FGDC (2006), FGDC-STD013-2006

How to cite: Wueller, L., Theiner, T., Iqbal, W., van der Bogert, C. H., and Hiesinger, H.: Geologic Map of the Poincaré Region, Moon., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9518, https://doi.org/10.5194/egusphere-egu25-9518, 2025.

EGU25-9649 | Posters on site | ESSI4.5

A New Geological Map of the Bose and Bhabha region in the South Pole-Aitken Basin, Moon 

Carolyn van der Bogert, Wajiha Iqbal, Julian Theiner, Lukas Wueller, Astrid Oetting, and Harald Hiesinger

We present a detailed geologic mapping of the Bose and Bhabha crater region within the South Pole-Aitken (SPA) Basin on the farside of the Moon. This area is of significant scientific interest due to the presence of diverse geological features. The SPA Basin is thought excavate material from the lunar crust or mantle [1,2], and has been a priority for exploration, prompting numerous mission studies [e.g., 3]. The mapping was conducted using standard geologic mapping protocols [4,5] with LROC Wide Angle Camera data [6], LOLA/SELENE digital elevation models [7], and Clementine spectral data [8]. The analysis identified various geological units, including Pre-Nectarian and Nectarian materials, which represent ancient highland remnants and heavily degraded craters. Additionally, Imbrian-aged light and dark plains were identified, with the dark plains being interpreted as volcanic in origin. Eratosthenian craters, distinguished by their relatively recent morphology, were were delineated. Secondary crater chains formed across several periods; however, they are primarily observed during the Copernican period. Nevertheless, no primary Copernican craters were observed in the mapping region at the map scale. This comprehensive mapping effort provides critical information about the geological evolution of the SPA Basin and supports future mission planning, such as NASA's Endurance mission, by identifying key terrains and features of interest for exploration.

[1] Pieters et al. [2001] JGR, 106(E11), 28, 001-28,22. [2] Petro et al. (2011) GSA, 477, 129-140. [3] Jawin et al. (2019) ESS, 6. [4] FGDC (2006) FGDC-STD-013-2016. [5] Skinner et al., (2022) USGS, TM11-B13. [6] Robinson et al. (2010) Space Sci. Rev. 150, 81-124.  [7] Barker et al. (2016) Icarus 273, 346-355.  [8] Pieters et al. (1994) Science 266, 1844-1848. [9] Keane et al. (2021) MCSR, NASA

How to cite: van der Bogert, C., Iqbal, W., Theiner, J., Wueller, L., Oetting, A., and Hiesinger, H.: A New Geological Map of the Bose and Bhabha region in the South Pole-Aitken Basin, Moon, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9649, https://doi.org/10.5194/egusphere-egu25-9649, 2025.

EGU25-11629 | Posters on site | ESSI4.5

One step beyond: The rocky path towards the new GSEU lithology vocabularies 

Kristine Asch, Hugues Bauer, Stefan Bergman, Anne-Cécile Flindt, Paul Heckmann, Cécile Le Guern, Hans-Georg Krenmayr, Zoltán Németh, Matevž Novak,, Marco Pantaloni, Kris Piessens, Robert Schäfer, and Urszula Stepien

Harmonisation of geological data, both semantically and geometrically, is key to foster the understanding of geological information across national borders. Hereby, the multitude of national borders in Europe, coupled with the intensity of geological mapping efforts, present a considerable challenge.

The building of geological databases for Europe started in 1995 with the project of the International Geological Map of Europe and Adjacent Areas (IGME 5000). For the first time a spatial geological database for the entire Europe was built which covered Europe’s on-shore and off-shore regions. The project was finished in 2005, and the map database is available online since 2006. In 2007, the European INSPIRE Directive came into force requiring standardized data availability within a pan-European geodata infrastructure of 34 themes, including geology, according to common standards and data specifications/vocabularies. The INSPIRE geology data specifications/vocabulary were based on those developed by the OneGeology-Europe project (OneG-E ) which had the aim to make European geological data interoperable, harmonize them as far as possible and make them available for free according to the FAIR data principles. One of the vocabularies described the lithology of rock units.

While these past projects were comprehensive, they showed a lack of
a) vocabularies to describe detailed spatial databases (e.g. geology), and
b) thematic properties such as anthropogenic units, lithotectonic features, metamorphic and textural features, etc.

In 2022, within the EU Horizon Europe programme, the project GSEU (Geological Service for Europe) started to build a geological framework. This encompasses to build a pan-European data model, a metadata system, methods to visualize 3-D models and the creation of hierarchical machine-readable vocabularies based on the earlier IGME 5000, OneG-E and INSPIRE Geology terminology.

Within GSEU, hierarchical scientific vocabularies for lithology, anthropogenic deposits and lithotectonic units are being set up for defining the concepts to which geometrical descriptions (lines, polygons, and volumes) can be linked. In future, these vocabularies will be made available in several languages to scientists in the field and in the office settings so that they can add the proper name to their mapped rock types in a harmonized way. This poster is focussing on the development of the lithology vocabularies.

The main challenges the endeavour is facing are:

  • to set up vocabularies that take into account differing nomenclatures which classify the same concept (term),
  • to cope with obsolete and strictly regional terms,
  • to take into account multiple hierarchies and
  • to include genetically related terms, qualifiers and compound names.

Custom programming scripts, written in Python and JavaScript help to automatise the data handling and visualisation of the hierarchical relations of the lithology concepts.

The poster presents the historical background of building pan-European geological vocabularies, demonstrates graphically the actual status of the created GSEU lithology vocabulary and provides an outlook to the future development.

How to cite: Asch, K., Bauer, H., Bergman, S., Flindt, A.-C., Heckmann, P., Le Guern, C., Krenmayr, H.-G., Németh, Z., Novak,, M., Pantaloni, M., Piessens, K., Schäfer, R., and Stepien, U.: One step beyond: The rocky path towards the new GSEU lithology vocabularies, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11629, https://doi.org/10.5194/egusphere-egu25-11629, 2025.

EGU25-11685 | ECS | Posters on site | ESSI4.5

The GEOS experiment onboard Amadee24 crewed simulated mission to Mars 

Selina Schindler, Alessandro Frigeri, Seda Özdemir-Fritz, and Gernot Grömer

The GEOS experiment onboard Amadee24 crewed simulated mission to Mars

Selina Schindler, Alessandro Frigeri, Seda Özdemir-Fritz, Gernot Grömer

During the AMADEE24 mission in Armenia, the GEOS experiment focuses on geologic survey activities at the simulated Martian landing site. GEOS applies classic geological field survey methods to a simulated mission to Mars, drawing on the experience of the lunar field survey built by Apollo missions.

The elements of GEOS are the mapping, the sampling, and the compositional measurements. The mapping phase involves developing mission-specific cartography from orbital remote sensing to large-scale mapping produced during and after the mission (Ozdemir et al., 2020). The real-time refinement of geological maps during the mission, using data from drones, rovers, and on-site observation, highlights the methodology's adaptability and receptivity. The core element of GEOS is the sampling, providing the ground truth of the remote sensing observation. AMADEE24 Rovers and Analog astronauts have done rock and terrain sampling along transects on base maps supplied by RSS (Remote Science Support) and Flight Planning (FP) for the Extra Vehicular Activities (EVAs). Part of the samples will return from the simulated Martian habitat, and made available for more advanced laboratory analyses. In-habitat compositional measurements offer a first estimate of the mineralogy and geochemistry of the samples. Specifically, AMADEE24 carried a RAMAN spectrometer in the field.

Here, we will present the results of AMADEE24/GEOS and the importance of collaborative efforts and innovative methodologies in remote science operations.

How to cite: Schindler, S., Frigeri, A., Özdemir-Fritz, S., and Grömer, G.: The GEOS experiment onboard Amadee24 crewed simulated mission to Mars, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11685, https://doi.org/10.5194/egusphere-egu25-11685, 2025.

A machine learning-based method for mineralization prediction is proposed, leveraging a 3D geological-geophysical model, aiming to achieve precise delineation of three-dimensional prospective mineral exploration targets. This approach integrates multi-source geophysical property parameters, such as density, magnetic susceptibility, and resistivity, with regional geological settings, ore deposit characteristics, and drilling data. A mineralization prediction model is established based on machine learning algorithms to address parameter overlap and inherent geological ambiguity. Algorithms such as Random Forest and Support Vector Machines are employed to achieve nonlinear mapping and efficient classification of the data, while grid search is used to optimize model parameters, leading to notable improvements in prediction accuracy and reliability. Model performance is evaluated through cross-validation, demonstrating its applicability. Applied to the Duobaoshan ore district in Heilongjiang Province, China, a well-known mineralized region, this method successfully delineated multiple 3D prospective exploration targets, showcasing its potential in the integrated analysis and 3D modeling of geological and geophysical data. This study provides new insights and technical support for mineralization prediction under complex geological conditions.

Keywords: Multi-source geological-geophysical data; 3D modeling; Machine learning; 3D targeting

How to cite: Wang, G. and Lv, X.: Research on Integrated Analysis of Geological and Geophysical Data and 3D Mineralization Prediction Based on Machine Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11762, https://doi.org/10.5194/egusphere-egu25-11762, 2025.

EGU25-13465 | Orals | ESSI4.5

Strategies for International Collaboration in Planetary Geologic Mapping  

Jeannette Luna, James Skinner, Wajiha Iqbal, Alessandro Frigeri, and Alexandra Huff

As international partnerships and commercial entities propose upcoming missions to the Moon and Mars, global scientific collaboration has never been more critical. Despite significant advances in planetary exploration, the lack of standardized international mapping practices and limited access to training has hindered truly global participation in planetary cartography. Planetary geologic mapping provides a common ground where diverse nations, institutions, and private organizations can unite to achieve shared scientific and exploratory goals. Maps play a pivotal role in advancing our understanding of planetary surfaces, supporting mission planning and operations, and ensuring the safety and success of exploration efforts. Based on feedback from recent international planetary mapping workshops and published community recommendations, we advocate here for the following strategies toward fostering international collaboration. First, we support education to train the next generation of planetary scientists and mappers, particularly in underrepresented regions, to ensure a diverse and capable workforce. In addition to university programs and certificates, virtual workshops, international exchange programs, and accessible educational resources are proven methods to democratize access to this field. Second, we encourage planetary mappers to share data and products through space agency archives and repositories following the FAIR (findable, accessible, interoperable, and reusable) principles, so that scientists worldwide can contribute their unique perspectives to solve geologic problems and investigate planetary phenomena. Third, we support the development of standardized methods for geologic mapping—particularly focusing on consistent crater age dating techniques and structural feature documentation—especially as they can be applied to terrestrial planets, moons, and small bodies. Aligning these standards with established terrestrial cartographic practices, while innovating and adapting them for extraterrestrial environments, will ensure consistency and comparability over the coming decades. We propose establishing an International Planetary Cartography Working Group to develop collaboration on cartographic conventions and symbology that will likewise enable seamless integration of map efforts across nations and industries. Fourth, we urge the scientific community to prioritize inclusive naming conventions that incorporate indigenous astronomical knowledge and multilingual perspectives to reflect the cultural and linguistic richness of Earth, promoting global representation in the naming process. Finally, we celebrate maps as visually compelling ways to share the importance of space exploration with humanity. We hope that over the next five years, the international community can collectively advance planetary geologic mapping, yielding benefits including enhanced mission return, scientific collaboration, and increased public engagement with planetary science through accessible, standardized mapping products.

How to cite: Luna, J., Skinner, J., Iqbal, W., Frigeri, A., and Huff, A.: Strategies for International Collaboration in Planetary Geologic Mapping , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13465, https://doi.org/10.5194/egusphere-egu25-13465, 2025.

EGU25-16351 | Orals | ESSI4.5

Unveiling Seabed Substrate Characteristics: Insights from EMODnet Geology 

Susanna Kihlman, Anu Kaskela, Aarno Kotilainen, Ulla Alanen, and Henry Vallius and the EMODnet Geology project consortium

Increasing activities and pressure on marine and coastal environments have made easily accessible, reliable, and suitable marine information essential and seabed substrate is one of the key elements in describing the marine environments. The EMODnet (European Marine Observation and Data network) Geology has been collecting and harmonizing geological data at different scales from all the European sea areas since 2009, at present with a collaboration of about 40 partners and subcontractors.

One part of the project is concentrating on seabed substrates and substrate characteristics, such as sedimentation rates, seabed erosion and other complementary information. During these years, a lot of work has been done to find a way to compile this scattered, heterogenous data harmonized, cross boundary datasets that could be used for different purposes. At the same time, the geographical scope has expanded beyond Europe, currently including the Caspian Sea and Caribbean Sea.

Multiscale Seabed substrate, harmonized from the national data by the sediment grain size, is one of the key data products of EMODnet Geology and has been collected since the beginning of the project, along with sedimentation rates information. The latest addition to the data product catalogue is the seabed erosion index database, a literature catalogue of erosion studies that includes known erosional studies and various erosional areas. These data products have evolved during the years based on the feedback from partners, collaborators, and stakeholders. For instance, the seabed substrate database includes information on the seabed surface characteristics that have significant to the marine environment but cannot be solely defined by grain size (e.g., seagrass meadows, moving sediments, ferromanganese concretion bottoms and bioclastic features). Overall, the usefulness and usability have been enhanced for example by adding new data attributes and by developing confidence assessment.

The latest development has focused on seabed dynamics and the potential to acquire the most practical and valuable data on the subject. In addition to the already published data products on sedimentation rates and erosion index layers, several case studies have been conducted since the project’s inception to develop and test tools for substrate modeling and sedimentation rates. The recent phase of this work aims to identify various sedimentary environments within national datasets and explore the potential for creating a broader, harmonized, and useful database.

Over fifteen years since the beginning of the project, EMODnet Geology has become a key producer of publicly available, harmonized seabed substrate datasets covering broad areas and the methodology is widely recognized. Besides collecting the seabed substrate data and update the existing databases, the development of the data products, improving old and creating new, will sustain the relevancy of the data in the future as well. At best, this kind of data is a valuable addition to understand and define marine environment in dealing with various challenges the future will hold us.

The EMODnet Geology project is funded by The European Climate, Environment, and Infrastructure Executive Agency (CINEA) through contract EASME/EMFF/2020/3.1.11 - Lot 2/SI2.853812_EMODnet – Geology.

How to cite: Kihlman, S., Kaskela, A., Kotilainen, A., Alanen, U., and Vallius, H. and the EMODnet Geology project consortium: Unveiling Seabed Substrate Characteristics: Insights from EMODnet Geology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16351, https://doi.org/10.5194/egusphere-egu25-16351, 2025.

EGU25-17425 | Posters on site | ESSI4.5

EMODnet Geology - Supporting sustainable use of the European maritime areas and their resources  

Anu Kaskela, Henry Vallius, Susanna Kihlman, and Aarno T. Kotilainen and the EMODnet Geology project consortium

The European Marine Observation and Data Network (EMODnet) is a long-term initiative funded by the European Commission to assemble and make accessible high-quality marine data from diverse sources across Europe. Since its beginning (2009), EMODnet has aimed to support sustainable marine and coastal management by providing open-access FAIR data and data products that are needed by scientific research, policymaking, and industry applications. Today, this network of more than 120 organizations covers several broad disciplinary themes: bathymetry, biology, chemistry, geology, human activities, physics, and seabed habitats. Each of these themes contributes to a comprehensive understanding of Europe’s marine environment and provides a wide range EMODnet datasets available through the EMODnet Central Portal (https://emodnet.ec.europa.eu/en).

EMODnet Geology, one of the thematics, focuses on the collection and harmonisation of marine geological data. This thematic provides extensive datasets on seabed substrate, sedimentation rates and seabed erosion index database, sea floor geology including lithology and stratigraphy, Quaternary geology and geomorphology, coastal behaviour, geological events such as submarine landslides and earthquakes, marine mineral resources, as well as submerged landscapes of the European continental shelf at various time frames. It is providing the full areal coverage of European seas as well as expanding to new areas, as also the Caspian and the Caribbean Seas are included in the geographical scope of the current project phase. EMODnet Geology focuses on delivering harmonised interpreted data layers (i.e., maps) rather than the underlying data. However, the metadata provides information on the data holder in case user needs to access the raw data. By integrating data layers from national geological surveys, research institutions, and marine organizations, EMODnet Geology ensures the availability of accurate and standardized geological information to support maritime spatial planning, environmental impact assessments, and resource management.

The current EMODnet Geology project phase (2023-2025) aims to further enhance data coverage and quality. It is coordinated by the Geological Survey of Finland (GTK), and it is executed by a consortium of 40 partners and subcontractors. The core of the partnership is formed by members of the EuroGeoSurveys network, supported by other partner organizations with valuable expertise and data.

EMODnet Geology also supports third-party data submission. Third party data can be submitted either straight to EMODnet Geology or through EMODnet Data Ingestion (www.emodnet-ingestion.eu), which is reaching out to potential data providers from private bodies and public. By facilitating data sharing and collaboration, EMODnet Geology continues to support informed decision-making and sustainable management of marine environments. It is a dynamic initiative where existing datasets are continuously updated with new data.

The EMODnet Geology project is funded by The European Climate, Environment and Infrastructure Executive Agency (CINEA) through contract EASME/EMFF/2020/3.1.11 - Lot 2/SI2.853812_EMODnet – Geology.

How to cite: Kaskela, A., Vallius, H., Kihlman, S., and Kotilainen, A. T. and the EMODnet Geology project consortium: EMODnet Geology - Supporting sustainable use of the European maritime areas and their resources , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17425, https://doi.org/10.5194/egusphere-egu25-17425, 2025.

EGU25-17849 | ECS | Posters on site | ESSI4.5

Lithological Characterization of Extruded Salt-Diapir-Caprock Systems in the Zagros Mountains in Iran Using Satellite-Based Multispectral and Hyperspectral Remote Sensing  

Mugabo Wilson Dusingizimana, Anke Friedrich, Beth Kahle, Stefanie Rieger, Soraya Heuss-Aßbichler, Prokop Závada, and Mjahid Zebari

 

Salt-diapir-caprock systems form both subsurface and surface halokinetic features. The world’s best exposed salt-diapir-caprock systems are hosted in the Zagros Mountains, in the arid and mountainous part of Iran. Due to their economic significance, subsurface salt-diapir-caprock systems in various tectonic settings have long been the focus of geosciences research. Biogeochemical subsurface processes, which are thought to be responsible for caprock formation, are also genetically linked to the formation of Pb-Zn deposits and some of the largest native sulfur deposits. In addition, the subsurface systems form hydrocarbon traps that are important for energy exploration. For this reason, extensive studies have been conducted on subsurface caprocks to establish a conceptual lithological model that describes the formation processes and the spatiotemporal relationships of salt-diapir-caprock facies. On the contrary, studies on fully extruded caprock systems remain limited. This scarcity hampers the comparative assessment of the lithological makeup of both extruded and subsurface salt-diapir caprock systems. It also restricts our understanding of the compositional evolution of salt-diapir and caprock materials as they diapirically extrude and become exposed to further modification by subaerial surface processes.

In this study, we explored the potential of satellite-based multispectral ASTER (Advanced Spaceborne Thermal Emission and Reflection Radiometer) and hyperspectral EnMAP (Environmental Mapping and Analysis Program) remote sensing for producing lithological maps of exposed salt-diapir-caprock features in the Zagros Region. We tested our method on three geomorphologically different salt diapirs ― Karmostaj, Siah Taq, and Champeh. We further examined the similarities and differences between our results and the established lithological model of subsurface salt-diapir-caprock systems.

Our results indicate that satellite-based remote sensing offers an efficient approach to producing lithological maps of exposed salt deposits and related caprocks, hence allowing the identification of caprock lithological facies. However, the accuracy of these maps depends on the spectral and spatial resolutions of satellite data. Furthermore, the results allow us to define the fundamental compositional differences between caprock formed under subsurface biogeochemical environments and caprock formed under the influence of surface processes. Specifically, subsurface salt dissolution results in the accumulation of a substantial anhydrite cap. Microbially-driven subsurface caprock-forming processes alter Ca-sulfates into an extensive calcite cap and simultaneously cause iron sulfide mineralization. As diapiring microbial iron sulfides reach shallow-depths and subaerial conditions, they alter into ferric oxides and ferric oxy-hydroxides. Therefore, together with microbial carbonate, the ferric oxides and oxy-hydroxides serve as diagnostic proxies for subsurface caprocks. In contrast, under surface conditions, microbial processes are likely to be unfavorable, leading to the limited amount or lack of biogenic calcite caprocks and iron sulfide mineralization. Caprocks formed under surface conditions thus predominantly comprise quartz- and clay-rich lithologies, which are the main residuals of the dissolution of salt-rich extruded materials, and a limited amount of the Ca-sulfates as surface processes hamper their accumulation.

How to cite: Dusingizimana, M. W., Friedrich, A., Kahle, B., Rieger, S., Heuss-Aßbichler, S., Závada, P., and Zebari, M.: Lithological Characterization of Extruded Salt-Diapir-Caprock Systems in the Zagros Mountains in Iran Using Satellite-Based Multispectral and Hyperspectral Remote Sensing , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17849, https://doi.org/10.5194/egusphere-egu25-17849, 2025.

Uncrewed Aircraft Systems (UAS) are becoming increasingly accessible to the masses.  Most small commercial drones are equipped with a camera system, and their operation is affordable by anyone.  Geological maps and models of Earth are commonly developed by a systematic investigation of the expressions of the geology outcropping on the topographic surface.

UAVs can move in three dimensions over the ground, offering the geologic surveyor a privileged point of view.

In November 2023, we organized a field campaign at the Rio Tinto area in southwestern Spain, which is considered a terrestrial analog of  Mars.  At Rio Tinto, we combined drone surveys with field investigations to understand the spatial relationship between rocks and biosignatures.  We used the drone in its basic functionalities: by acquiring overlapping images from the flying platform, we produced image mosaics and digital terrain models (DTMs) by applying Structure from Motion (SfM) algorithms.  Those images and digital terrain models become the basemaps of our large-scale geologic mapping of key portions of our study area.  Here, we will discuss the methods, the type and quality of data acquired, and the evolution of our knowledge of the problem during and after the campaign.  

Four years after the first flight on Mars by the Ingenuity Mars Helicopter, we are now experimenting on Earth with new methods for geologic survey and what will be the future of space explorations, where robotic systems will support human surveys. 

How to cite: Frigeri, A. and Skinner, J.: Geologic Mapping with UAS: New Perspectives in Geologic Surveying with the Support of Drones in Earth and Planetary Exploration., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18984, https://doi.org/10.5194/egusphere-egu25-18984, 2025.

EGU25-19408 | ECS | Posters on site | ESSI4.5

Mercury: explorative geological maps through unsupervised learning  

Natalia Amanda Vergara Sassarini, Cristina Re, Riccardo La Grassa, Adriano Tullo, and Gabriele Cremonese

The mapping of planetary surfaces represents a fundamental activity in planetary science, offering invaluable insights into the formation history, surface processes, and compositional variations of celestial bodies. In addition, accurate and detailed mapping are crucial for tasks ranging from identifying potential landing sites to planning future exploration missions. These maps are primarily constructed from visible image data sets, providing topographic and albedo information which is mostly used to delineate and define the stratigraphy of geomorphological units (i.e., morpho-stratigraphic maps). However, the creation of such maps requires the specialized knowledge of expert planetary scientists and constitutes a time-intensive and highly complex task. In addition, often these maps rely solely on a geomorphology‐led approach overlooking meaningful details about composition (i.e., multispectral data) and physical properties of the defined units, with spectral information usually supplementing rather than informing geomorphological data.

This work aims to create the first set of global, explorative classification maps of Mercury’ surface which incorporate both spectral and morpho-stratigraphic information using an unsupervised learning approach based on Gaussian Mixture Models. This work represents an ambitious and promising approach for facilitating the generation of comprehensive geological maps.

In addition, this classification will facilitate geological interpretation and enhance the mapping of the planet's unexplored regions, while enriching the understanding of already surveyed regions. Such advancements are pivotal for unraveling the complexities of Mercury's surface, contributing significantly to our understanding of the planet in anticipation of the new wave of data expected from SIMBIO-SYS (Cremonese et al., 2020) data on the BepiColombo's mission (Benkhoff et al., 2021)

How to cite: Vergara Sassarini, N. A., Re, C., La Grassa, R., Tullo, A., and Cremonese, G.: Mercury: explorative geological maps through unsupervised learning , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19408, https://doi.org/10.5194/egusphere-egu25-19408, 2025.

EGU25-19461 | ECS | Posters on site | ESSI4.5

Ingenii Basin: characterization and feasibility as a lunar landing site 

Gloria Tognon, Riccardo Pozzobon, Giacomo Melchiori, and Matteo Massironi

The renewed interest in the exploration of the Moon, both human and robotic, and the technological advancement made it more feasible to look at sites previously underrated for the paucity of high-resolution data available (e.g. polar regions) and the challenges in communication (e.g. far side), accessibility and trafficability (e.g. underground cave systems).

Being characterized by a smooth basaltic infilling, optimal for landing and roving, the far side Ingenii basin (20.4°S, 129.1°E) represents a high-profile objective for the unique presence of both extensive and complex swirls, namely features related to crustal magnetic anomalies [1,2,3], and a pit with overhanging roof possibly giving access to a lava tube [e.g. 4]. In this study, we characterized the area surrounding the Mare Ingenii Pit (MIP) and performed a feasibility study for a robotic mission with a rover-hopper [5] by considering traverses of varying lengths, all providing for a hopping phase inside the pit, and simulating the environmental conditions along their paths.

More in detail, we used the Lunar Reconnaissance Orbiter (LRO) Wide Angle Camera (WAC) mosaic (up to 100 m/px) [6] for a contextual interpretation of the area together with LRO Narrow Angle Camera (NAC) images (up to 0.5 m/px) [6] for a detailed characterization of the area surrounding the MIP. The Lunar Orbiter Laser Altimeter (LOLA) and Kaguya Terrain Camera merged Digital Elevation Model (DEM) provide the most up-to-date surface height and slope information (vertical accuracy of 3-4 m) [7] and a NAC-derived Digital Terrain Model (DTM) provides the best available elevation data for the area surrounding the pit.

We quantitatively assessed the topographical characteristics of the surface within a reasonable distance from the MIP for the positioning of landing ellipses (1500 and 500 m in diameter), and automatically detected boulders >1 m using a machine learning algorithm and NAC imagery [8]. We then planned short (up to 5 km), intermediate-length (up to 10 km) and long (up to 15 km) traverses and evaluated slope and elevation variations along the paths taking into account a typical slope tolerance of maximum 15°. Finally, we used an interactive tool provided by LROC Quickmap [9] to perform simulations of the environmental conditions along each traverse path and identify a mission operating window based on illumination and temperature conditions over a lunar day.

A lunar landing candidate site located on the far side sure entails a major effort in communicating with Earth, however, the scientific relevance and peculiarity of Ingenii basin and its optimal topographical characteristics make it a site to be considered for future exploration.

 

References

[1] Pinet et al. (2000) JGR, 105, 9457-9475. 

[2] Hood et al. (2001) JGR, 106, 27825-27839.

[3] Garrick-Bethell et al. (2011) Icarus, 212, 408-492.

[4] Miaja et al. (2022) Acta Astronautica, 192, 30-46.

[5] Rimani et al. (2023) Aerospace, 10(8), 669.

[6] Robinson et al. (2010) Space Sci. Rev., 150, 81–124.

[7] Barker et al. (2016) Icarus, 273, 346-355.

[8] Prieur et al. (2023) JGR, 128, e2023JE008013.

[9] https://quickmap.lroc.asu.edu/

How to cite: Tognon, G., Pozzobon, R., Melchiori, G., and Massironi, M.: Ingenii Basin: characterization and feasibility as a lunar landing site, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19461, https://doi.org/10.5194/egusphere-egu25-19461, 2025.

EGU25-20011 | ECS | Orals | ESSI4.5

Geological mapping of North-West Mount Sharp region (Gale crater, Mars) and connections with data from Curiosity rover 

Susanna Tonoian, Matteo Massironi, Riccardo Pozzobon, Nicolas Mangold, and Adriano Tullo

The exploration of Gale Crater on Mars has returned a great amount of data, offering insights into the planet's geological history. However, to fully understand its evolution, comprehensive analysis at various scales is essential. This study focuses on geological mapping, stratigraphy and structural analysis transitioning from a regional scale to a local scale to establish the evolution of Gale Crater. The primary goal of this research is to clarify the structural and stratigraphic relationships among the sedimentary layers within the North-West part of Gale Crater, providing insights into the sedimentary environment at the moment of their deposition as well as the deformational history while the sequence of Mount Sharp was built up and shaped. The dataset used for this study was acquired both by orbital [1] and rover missions [2]. Merged DEM and Orthophoto were made by MSL team from HiRISE stereopairs for the needs of the Curiosity mission with resolution 1 and 0.25 (m/pixel) respectively [2]. For obtaining colored information we used the CASSIS image. The alignment and pansharpening process was conducted using the open-source PANCO suite, which automates co-registration of CaSSIS multispectral data with reference panchromatic images through computer vision algorithms, refines results manually, and employs an adapted Gram-Schmidt Adaptive method [4]. This process improves the color image resolution up to 18 times using a panchromatic image mosaic orthophoto as the base. As the result of mapping, we updated the stratigraphic column with unit’s description combined with observation from ground data and the previous literature and research. The novelty of the stratigraphic column lies in its organization based not on the rover’s traverse sequence but on a rigorous stratigraphic order which allowed us to infer the sequence of geological events in the region. As already highlighted by previous authors [5] the area exhibits a transition from a lacustrine environment to an aeolian one. We propose three new members on the upper part of the Mirador formation. Additionally in the middle of Mirador formation we have recognized two erosional events preceding a significant climatic shift which led to the final part of the studied series made up of a sulphate rich sedimentary sequence interpreted as aeolian facies with local broad cross stratification. Structurally, the region shows a low average dip of approximately 5 degrees towards the NW, with local variations possibly caused by gentle folding both before and after the detected unconformities.

Acknowledgements: This study was carried out within the Space It Up project funded by the Italian Space Agency, ASI, and the Ministry of University and Research, MUR, under contract n. 2024-5-E.0 - CUP n. I53D24000060005

[1] University of Bern (2024) MY36_018708_356_0 https://observations.cassis.unibe.ch/. [2] MSL, NASA (2011). Curiosity Analyst’s Notebook https://an.rsl.wustl.edu/msl/AN/an3.aspx. [3] Calef III F. J. and P. T. (2016) PDS Annex, U.S.G.S. [4] Tullo A. (2024) PSS 105997. [5] Meyer, M. J. (2024). GSA b37355.1

 

How to cite: Tonoian, S., Massironi, M., Pozzobon, R., Mangold, N., and Tullo, A.: Geological mapping of North-West Mount Sharp region (Gale crater, Mars) and connections with data from Curiosity rover, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20011, https://doi.org/10.5194/egusphere-egu25-20011, 2025.

EGU25-20030 | Posters on site | ESSI4.5

EMODnet geology and data harmonisation: Seafloor mapping of the Western Baltic Sea  

Alexander Müller, Kristine Asch, Urszula Pączek, and Dorota Kaulbarsz

Harmonisation of marine geological data across the EEZ , both semantically and geometrically, is key to understand geological information across national borders. There is a multitude of EEZ boundaries in the European Seas which are partly intense and partly hardly mapped. This presents a considerable challenge.

The European EMODnet geology project is running since the year 2009. One of the aims is to provide geological data of the European Seas, harmonized as far as possible and available according to FAIR data principles.

Within the workpackage seafloor geology several international teams are working on semantic and geometric data harmonisation in seven marine areas, the so-called prototype areas. One of them focusses on the Western Baltic Sea with participants from Poland, Sweden, Denmark and Germany. The data  being compiled and harmonized centrally at BGR, Germany. The heterogenity of the geological information from each of the partners derive from the following reasons: data exist in some regions in patches or ribbons, e.g. along research vessels‘ tracks, the mapping results and classifications of terms differ due to different scientific approaches, the mapping took place in different scales and at differing ages. Thus, it was crucial to set up common standards, especially controlled vocabularies based on international standards (INSPIRE Directive, IUGS) as a base and to also practically discuss and agree on the continuation of geological structures across EEZ boundaries.

The poster demonstrates differing national classifications approaches, outlines the method of agreeing on the continuation and naming of geological structures and presents the first results of a geological map of the Quaternary of the Western Baltic Sea.

How to cite: Müller, A., Asch, K., Pączek, U., and Kaulbarsz, D.: EMODnet geology and data harmonisation: Seafloor mapping of the Western Baltic Sea , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20030, https://doi.org/10.5194/egusphere-egu25-20030, 2025.

The discipline of planetary geologic mapping has experienced a renaissance over the past two decades, driven by increasing spacecraft exploration and rapid development of geographic information system (GIS) technologies. Unlike terrestrial mapping based on direct field observations, planetary maps depend on remote data, requiring careful inference of lithology and formation processes. As NASA prepares for Artemis missions, traditional planetary geologic maps must evolve beyond documenting phenomena to
become operational tools that support mission planning and execution.

The Lunar Mapping Program (LMAP), a NASA pilot initiative supported by the USGS, is transforming planetary geologic map creation through an accelerated, team-based campaign approach. Using the Shackleton-de Gerlache ridge region as a test area - with its mix of smooth and rugged terrain (-4380 to 1959 m elevation) and extensive permanently shadowed regions (25.1%) - the team is developing mapping methods that can be applied anywhere in the lunar south pole region. LMAP employs a novel dual-scale mapping strategy, recognizing that no single map can tell the whole story. 

The project combines regional context mapping at 1:150,000 with detailed 1:30,000 scale products, drawing from high-resolution LROC NAC mosaics (1 m/pixel), DEMs (5 m/pixel), and ShadowCam data to characterize  surface properties that matter to both scientists and engineers. LMAP brings together five essential components: traditional geologic maps showing stratigraphy and history, surface feature maps for traverse planning, hazard assessments identifying slopes and boulder fields, resource locations highlighting water ice and construction materials, and logistics planning that brings it all together. Think of it as creating not just a single map, but rather a comprehensive atlas. The team is working closely with key partners including NASA Flight Operations, the Artemis Geospatial Science Team, and the Lunar Surface Innovation Consortium (LSIC) to ensure these maps will serve real operational needs for Artemis V and beyond. Through parallel mapping teams under a tight, structured timeline, we are streamlining traditional mapping processes without sacrificing scientific quality, backed by USGS expertise and thorough review.

This pilot project charts new territory in planetary geologic mapping, producing both immediate tools - published as USGS Open File Reports - and proven approaches for future rapid mapping campaigns. These innovations have the potential to shape how we map the Moon for years to come. By adapting existing cartographic standards to remotely predicted features and being clear about our confidence in geologic interpretations, LMAP establishes better ways to support the next generation of lunar exploration.

How to cite: Skinner, J.: Beyond Traditional Geologic Mapping: The NASA-USGS Lunar Mapping Program (LMAP) for Operational Success, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21718, https://doi.org/10.5194/egusphere-egu25-21718, 2025.

EGU25-1941 | Posters on site | ESSI4.7

Progress in Cloud MICAPS Engine 

Feng Xue

Cloud MICAPS Engine is a next-generation cloud and component-based application development framework for Meteorological Information Comprehensive Analysis and Processing System(MICAPS)  in China, which has the following features:

I.Microservice Architecture
It utilizes a hybrid microservices management framework composed of K8s (Kubernetes) and Spring Cloud. The primary computational services are based on the underlying K8s + container cloud to implement microservices, while the upper-layer business applications integrate business-oriented microservices functions through Spring Cloud.

II.Real-time Visualization and Analysis
Developed using B/S technology, all weather-related business data is encapsulated as meteorological data layers through methods such as OGC, resumable transmission, and streaming services for loading by front-end business applications. It includes professional analysis components for common meteorological business operations such as points, lines, and surfaces, as well as interactive analysis of time curves, vertical soundings, and arbitrary sections.

III.Real-time Visualization Rendering
Based on real-time drawing technologies such as WebGL and WebGPU, it supports real-time product map services for products with high access volumes through pre-processing and pre-service methods, forming a highly consistent map service data environment with an integrated approach.

IV.Collaborative Editing and Forecast Service
Based on a 2D/3D map engine, the high-resolution real-time data visualization rendering technology is supported by CogTiff data storage and service technology. Real-time synchronization of interactive editing operations is achieved through WebSocket to realize real-time collaboration among multi-terminals. The back-end data editing algorithm realizes the consistency of updated data in FIFO.


V.LLM-driven Interaction and Processing
Large language models technology is used, with aggregating algorithms in the whole process of intelligent digital forecast service business, including observation perception, analysis diagnosis, interactive judgment, processing and generation, inspection and evaluation. "AI Agent" is the core to drive human-computer intelligent interaction and information recommendation.

Cloud MICAPS Engine  will be released later in 2025.

How to cite: Xue, F.: Progress in Cloud MICAPS Engine, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1941, https://doi.org/10.5194/egusphere-egu25-1941, 2025.

The Korea Meteorological Administration has developed a tool that can visualize various numerical model data in three dimensions by combining them with spatial information to support forecasters' weather analysis work.
The 3D visualization tool is a precise analysis tool that can compare and analyze weather data according to time, space, and altitude.
An open-source 3D visualization engine (CesiumJS) was applied to display numerical model data on 3D spatial information.
Spatial information displays major spatial information (national and administrative boundaries, major roads, rivers, etc.) and a three-dimensional numerical elevation model.
The numerical model data was developed to express the precursor model (ECMWF) and the local model (LDAPS), but it was flexibly implemented to express other numerical model data.
It is expected that more three-dimensional and precise analysis will be possible by combining numerical model data with spatial information and analyzing it in three dimensions.

How to cite: Kim, S.: Development of 3D visualization tool for KMA (Korea Meteorological Administration) numerical model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2209, https://doi.org/10.5194/egusphere-egu25-2209, 2025.

EGU25-3120 | Posters on site | ESSI4.7

Innovative Web Frontends for a secure access to High-Performance Computing resources via DASF 

Philipp S. Sommer, Björn Lukas Saß, and Markus Benninghoff

The open-source Data Analytics Software Framework (DASF) available at https://dasf.readthedocs.io is an advanced remote procedure call (RPC) framework designed to abstract Python code and make it securely callable over the internet. This framework ensures that computing resources remain protected from direct exposure to the internet. The latest innovation within DASF is the development of a flexible framework to automatically generate a web frontend for computing resources. 

The basic concept of DASF lies in the transformation of original Python code into a JSON schema. This schema serves a dual purpose: it is used to generate the web frontend and to validate the input received via the web interface. By converting Python code into a JSON schema, DASF leverages the underlying code to create a dynamic and interactive web frontend. This approach simplifies the deployment of web interfaces and additionally ensures that the input data conforms to the expected format, thereby reducing the likelihood of errors and enhancing the overall user experience. The web interface includes forms, input fields, and other interactive elements necessary for user interaction. This automated generation process eliminates the need for manual coding of the web frontend, saving time and reducing the potential for human error, and makes it especially useful for scientists without background in web development. 

One of the key benefits of this approach is the seamless integration of the web frontend with the underlying computing resources. Users can interact with the web interface to submit data and trigger computations, all while the JSON schema ensures that the input data is correctly formatted and validated. This validation step is crucial, as it prevents invalid data from being processed, which could otherwise lead to errors or security vulnerabilities. This security feature is particularly important in a web-based environment, where the potential for unauthorized access and data breaches is higher. 

Our framework is especially useful for digital twins in the earth system sciences, where considerable amounts of data and computing resources are often required. Digital twins are virtual representations of physical systems, and they rely on such resources to function effectively. DASF is particularly well-suited for these applications because it allows the code to run on high-performance computing (HPC) resources without exposing them to the internet. This ensures that sensitive infrastructure remains protected while still providing the necessary computational power. 

In our presentation we show how DASF works and provide live examples of the frontend implementation. 

How to cite: Sommer, P. S., Saß, B. L., and Benninghoff, M.: Innovative Web Frontends for a secure access to High-Performance Computing resources via DASF, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3120, https://doi.org/10.5194/egusphere-egu25-3120, 2025.

In spite of the availability of satellite data, its complexity has prevented widespread adoption into the practices of small and independent farmers. We present in this paper a system (AgriVision AR) designed to aid user interaction with satellite data sources such as ESA’s Copernicus Data Space Ecosystem. AgriVision AR uses mobile augmented reality to simplify data visualisation through intuitive interaction. Augmented Reality (AR) is a technology that links together the physical and the virtual worlds. In AR, digital information seamlessly blend with the real world, creating a perception of immersion of the user. Geospatial information is overlaid over the real environment making more easily to understand the data and increasing user immersion.  The application allows users to visualise agricultural indices and other information gathered through satellite imagery in the form of an animated color overlay on small scale landscapes. This approach will enable farmers to make informed decisions about crop management. It will also allow to optimize resource allocation, and mitigate localized issues such as soil fertility and pest infestation. AgriVision AR represents a step towards empowering small-scale farmers with advanced technology, fostering sustainable agricultural practices in the era of precision farming.

How to cite: Bacu, V. and Radu, M.: Augmented reality solution for visualization of agricultural data gathered from satellite imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3976, https://doi.org/10.5194/egusphere-egu25-3976, 2025.

This study introduces a novel methodology to enhance the accuracy and efficiency of 87Sr/86Sr isoscape mapping by integrating deep learning (DL) techniques with geostatistical methods. Utilizing kriging-based data augmentation, the proposed framework addresses data scarcity by generating high-quality synthetic training data while incorporating spatial geological factors and geochemical elements as input variables to improve model performance. The study developed an isotopic basemap for South Korea using 409 soil samples and a feedforward deep neural network (FDNN) model. The FDNN model demonstrated superior accuracy (91.67%) compared to traditional kriging (76.8%) and convolutional neural network (CNN)-based models (86.14%). The robustness of the FDNN model was significantly enhanced by kriging-based data augmentation, which not only captured geological anisotropies but also incorporated uncertainty analysis to improve reliability. The resulting 87Sr/86Sr isoscape map revealed distinct isotopic distributions across South Korea, with higher ratios associated with metamorphic and granitic rocks, reflecting geological history and topographical influences. Notably, the predicted isotopic distributions closely aligned with the boundaries of tectonic provinces, underscoring the geospatial accuracy of the developed model. Validation using bone samples additionally confirmed the efficacy of the proposed method in accurately estimating isotopic levels. These findings highlight the potential of combining geostatistical and DL approaches to overcome traditional challenges in isotopic mapping, offering scalable solutions for applications in environmental monitoring, archaeology, and provenance studies.

How to cite: Lee, H. and Jeong, J.: Enhancing Deep Learning-based Strontium Isotopic Landscape Estimation Using Geostatistical Method: A Case Study in South Korea, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5423, https://doi.org/10.5194/egusphere-egu25-5423, 2025.

The application of complementary geochemical analysis alongside deep learning techniques serves as a powerful tool in identifying geographical origin of environmental samples on a national scale. This research presents methodologies aimed at enhancing the accuracy of origin determination for soil samples across South Korea, leveraging geochemical and geological data. It addresses challenges associated with integrating Sr isotopes and multivariate geochemical variables and preserving geological interpretability by incorporating Autoencoder deep learning algorithms,which facilitate efficient feature engineering for comprehensive data analysis. Through the analysis of 412 soil samples collected nationwide, a geographic origin distribution and classification model was developed, establishing a novel framework for environmental sample analysis. The analysis identified six origins within South Korea, each distinguished by its geological tectonic units, bedrock age, and bedrock type. Extensively wide areas with granite bedrock nationwide were mostly classified into the same origin, irrespective of their geological tectonic configurations. The findings highlight the efficacy of integrating isotopic with geochemical data through advanced analytical techniques, significantly improving origin tracing accuracy and efficiency. Such advancements have significant implications for disciplines including agriculture, forensics, and archaeology, showcasing the potential of these methodological innovations.

 

How to cite: Lee, S. and Jeong, J.: Utilizing deep learning-based feature engineering for effective geographical origin subdivision and classification of environmental soil samples in South Korea, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5488, https://doi.org/10.5194/egusphere-egu25-5488, 2025.

EGU25-5968 | Posters on site | ESSI4.7

Immersive Visualizations for Marine Geohazards in the Aegean Sea: Bridging Science and Stakeholder Engagement. 

Jan Oliver Eisermann, Felix Gross, Jannes Vollert, Josephin Wolf, Lennart Petersen, Heidrun Kopp, Christian Berndt, Klaus Reicherter, Sebastian Krastel, Christian Hübscher, Christian Wagner-Ahlfs, Tom Kwasnitschka, and Armin Bernstetter

Communicating marine geohazards to stakeholders can be challenging, and traditional media may not be sufficient to convey the full range of processes involved. In addition, news images of devastating events such as tsunamis and volcanic eruptions can be associated with prejudice and fear, hindering fact-based awareness-raising. With recent advances in computer graphics and virtual reality headset hardware, immersive visualization methods are becoming accessible to a wider scientific community. Virtual presentations of different local scenarios provide an opportunity to discuss with experts, policy makers and the general public, overcoming abstraction and prejudice and transforming scenarios into realistic and spatially explicit experiences.

The MULTI-MAREX collaborative project is establishing a living lab in the Aegean Sea to study extreme marine geological events and associated hazards, with the aim of developing the knowledge needed to manage geohazards at different scales. Digital reconstruction of real, physical study sites leads to and enhances situational awareness, resulting in a personalized, in-depth understanding of local scenarios. Accessibility of the communication format is important to reach the target user.

We are exploring a wide range of hardware options, from the portability and convenience of smartphones, to the highly immersive experiences offered by head-mounted displays, through immersive simulators such as video walls and dome theatres. These diverse platforms ensure that immersive visualizations are adaptable to different user needs and environments, facilitating greater accessibility and engagement across stakeholder groups. We focus on developing workflows for geoscientists to enable semi-automated, asset-enhanced, immersive visualization that synthesize collected remote sensing data, such as terrestrial and marine digital outcrop models, hydroacoustic and numerical simulations within popular game engines. The use of popular game engines to seamlessly integrate different data types enables dynamic and interactive environments where users can explore scenarios in real time, enhancing both scientific analysis and stakeholder engagement to bridge the gap between complex geohazard science and effective stakeholder understanding, enabling informed decision making and risk management.

How to cite: Eisermann, J. O., Gross, F., Vollert, J., Wolf, J., Petersen, L., Kopp, H., Berndt, C., Reicherter, K., Krastel, S., Hübscher, C., Wagner-Ahlfs, C., Kwasnitschka, T., and Bernstetter, A.: Immersive Visualizations for Marine Geohazards in the Aegean Sea: Bridging Science and Stakeholder Engagement., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5968, https://doi.org/10.5194/egusphere-egu25-5968, 2025.

EGU25-6265 | Orals | ESSI4.7

3D Characterization of Aquifers in the Cenozoic Sedimentary Formations of the Paris Basin 

Pascal Audigane, Justine Briais, Delphine Allier, Sandrine Grataloup, Thomas Klinka, and Alexandre Brugeron and the RGF Bassin parisien Team

The objective of this study is to synthesize the geological and hydrogeological characteristics of the main aquifers in the Paris Basin within the Cenozoic sedimentary formations. This work is part of the multi-annual scientific program RGF (Référentiel Géologique de la France, https://rgf.brgm.fr/page/bassin-parisien), which aims to update the representation and mapping of French geology on a national scale. The program also funds several PhD projects in collaboration with the academic community, focusing on the geometry, distribution of petrophysical properties, modeling, and mapping of key Cenozoic geological formations in the basin. Particular emphasis is placed on the major aquifers as part of this modeling effort.

The geology of the Paris Basin has been extensively documented by various authors (Pomerol, 1967; Mégnien, 1980; Gély and Lorenz, 1991; Briais, 2015). Data from well logs and core descriptions collected from over 2,000 petroleum wells have been used to reconstruct the stratigraphic surfaces of the main formations, while also identifying the large-scale geometries of associated aquifers and aquitards. Recent studies have enhanced the dating of specific stratigraphic markers (Marlot, 2023; Moreau, 2023), described the geometries of alluvial formations in the Seine River (Chourio-Camacho, 2024), advanced knowledge in structural geology (Brown, 2024), and provided petrophysical characterizations of reservoir rocks (Moreau, 2023; Marie, 2024).

The hydrogeology of the Paris Basin has been studied and modeled for decades (Mégnien, 1980; Goncalves, 2003; Lamé, 2013). The lateral extent of aquifers varies significantly across regions. For instance, in Île-de-France, the hydrosystem comprises six primary aquifers: the Alluvial, Brie, Champigny, Lutetian, Ypresian, and Chalk aquifers. However, lateral facies variations can significantly alter hydrogeological properties, influencing groundwater resource potential. In the Oligocene formations, the main aquifers are primarily located in Île-de-France and the northern part of the Centre-Val de Loire region: i) on the Beauce plateau, commonly referred to as the "Beauce aquifer," ii) in the Yvelines area, primarily associated with the Fontainebleau Sands, and iii) on the Brie and Bière plateaus, where they are predominantly contained within the Brie Limestone.

These lateral facies variations, coupled with the presence of fractures or karstification, result in substantial differences in the petrophysical properties of the identified aquifers and aquitards. Pumping test data have been compiled and converted into permeability and transmissivity coefficients, which were subsequently mapped along lateral transects in the Brie region of the basin (Marie, 2024).

This study will also contribute to the harmonization of the lithostratigraphic framework across the 187 geological maps covering the territory. Furthermore, the 3D model will facilitate vertical and lateral interpolation of hydrological reference data from the BDLISA database (https://bdlisa.eaufrance.fr/), which currently provides detailed mapping of water bodies at the scale of metropolitan France.

How to cite: Audigane, P., Briais, J., Allier, D., Grataloup, S., Klinka, T., and Brugeron, A. and the RGF Bassin parisien Team: 3D Characterization of Aquifers in the Cenozoic Sedimentary Formations of the Paris Basin, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6265, https://doi.org/10.5194/egusphere-egu25-6265, 2025.

EGU25-9972 | Orals | ESSI4.7

The geological Gordian knot - lithological challenges in the world of geological mapping 

Urszula Stępień, Kristine Asch, Stefan Bergman, Matevz Novak, Marco Pantaloni, Hugues Bauer, Paul Hackmann, and Hans-Georg Krenmayr

Lithology is one geological key attribute on geological maps, serving as a fundamental framework for understanding the composition and structure of the Earth's crust. The development of a common lithological vocabulary has been essential for the harmonisation of geological maps at national and international level, enabling the creation of a semantically consistent map of Europe. Several international projects such as OneGeology-Europe, IQUAME, GSEU and numerous national initiatives like Polish GeoTezaurus, have provided valuable experience in this area.

These projects have highlighted the need for a unified approach to lithological mapping, bringing together geologists and petrologists to address a number of challenges. The production of standardised lithological maps requires cooperation and expertise, particularly in dealing with the different interpretations and terminologies used in different regions and languages.

One of the main challenges is the existence of multiple lithological classifications, each reflecting a different geological perspective. Classifications can vary based on factors such as grain size, composition and structure, complicating the process of creating a consistent mapping standard. In addition, many lithological terms are closely linked to various geological interpretations such as genetic process and depositional environment, further complicating the mapping process.

Another major challenge is the problem that the same lithological term can have different meanings depending on the context in which it is used. Therefore, a clear definition of a concept is much more important than which exact label (term) is chosen to represent that concept. A particular problem arises when translating terms between languages, as the meaning of a term can shift or become ambiguous in different linguistic and geological contexts. In addition, the translation process itself often reveals further problems, as terms that accurately describe geological features in one language may not have an exact equivalent in another. Comparing geological maps, we can find differences even in national datasets, such as using the singular or plural form of terms. Some languages have additional complications due to a rich vocabulary and a multitude of synonyms. This can lead to the creation of new descriptive terms, or adjustments to existing terminology.

Furthermore, national languages present additional challenges in ensuring consistency and clarity in lithological definitions. The process of translating scientific terms into national languages often reveals subtle differences in meaning and interpretation that may not be immediately apparent. As a result, new descriptive terms may be required to accurately convey the intended geological concepts, adding another layer of complexity to the standardisation effort.

Special attention must also be given to lithogenetic terms, particularly those associated with Quaternary surficial deposits. These deposits often present unique classification challenges due to their complex nature. Replacing lithogenetic terms with their very rich meanings, which include lithology, environment, process and form, by litological ones only, reduces map data content.

In conclusion, the development of a standardised lithological vocabulary, both nationally and internationally, is a complex but essential task for the advancement of geological research and communication. By addressing these challenges through collaboration and expertise, the global geological community can work towards a more integrated and comprehensive understanding of the Earth's lithology.

How to cite: Stępień, U., Asch, K., Bergman, S., Novak, M., Pantaloni, M., Bauer, H., Hackmann, P., and Krenmayr, H.-G.: The geological Gordian knot - lithological challenges in the world of geological mapping, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9972, https://doi.org/10.5194/egusphere-egu25-9972, 2025.

EGU25-10206 | Posters on site | ESSI4.7

Zoom in - zoom out challenge: Semantically and visually coherent overview geological maps of Poland 

Katarzyna Jóźwik, Stępień Urszula, and Przasnyska Joanna

Standardisation of geological maps visualisation is crucial for improving data legibility and comparison across different scales and regions. In Poland, overview geological maps ranging from scales of 1:2,500,000 to 1:500,000 have been traditionally prepared using distinct graphical styles, each tailored to the particular characteristics of the mapped geological units. These maps used to employ individual patterns and colour palettes to enhance the visibility and readability of geological features, prioritising the requirements of printed editions. However, differences in the number and types of geological units across various maps led to inconsistent visual representations, limiting the ease of comparison between them.

To address the above, the Polish Geological Institute’s team embraced the idea of creating a unified, semantically harmonised graphical style for overview geological maps. The main objective was to develop a common style for all stratigraphic units, which could be applied across various maps of Poland, particularly those prepared for online publication in the frame of the Polish Geological Cartograhy Platform. This experiment aimed to standardise the colour and pattern schemes, building upon the stratigraphic classification system provided by the International Commission on Stratigraphy (ICS), but with necessary extensions to accommodate mixed stratigraphy.

While this approach slightly reduced the visibility of details in certain areas, it significantly enhanced the comparability of geological data across maps. By adopting a consistent visual language, the maps delivered a clearer cartographic message, particularly when zooming in and out in map viewers. The harmonisation of the graphical style also enhanced data visualisation across various scales, making it easier for geologists to interpret and compare geological units.

This initiative was inspired by earlier efforts, such as the OneGeology initiative and the INSPIRE Directive, both of which sought to standardise the visualisation of lithological and stratigraphic data. However, these frameworks primarily focused on older geological units and their principal lithologies, which could lead to potential misinterpretations of the data. For example, geological units spanning from the Cambrian to the Devonian period were often represented using a single colour, which could obscure their true geological diversity. To address this, PGI team proposed the use of distinct colours for each geological period, drawing inspiration from the Commission for the Geological Map of the World (CGMW) colour codes.

The results of this experiment demonstrate that a semantically harmonised approach to geological map visualisation not only enhances the clarity of individual maps but also makes data more comparable across different scales and regions. By providing a consistent and intuitive visual representation of geological units, this method helps to improve the overall understanding of geological data and facilitates its use in various scientific, educational, and practical contexts.

How to cite: Jóźwik, K., Urszula, S., and Joanna, P.: Zoom in - zoom out challenge: Semantically and visually coherent overview geological maps of Poland, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10206, https://doi.org/10.5194/egusphere-egu25-10206, 2025.

EGU25-10643 | Posters on site | ESSI4.7

Unravelling the complex volcanological evolution of Pantelleria Island (Channel of Sicily, Italy): new insights from the Italian Geological Mapping Project (CARG Project) 

Alessandra Pensa, Roberto Bonomo, Alessandra Cinquegrani, Valeria Ricci, Silvio Giuseppe Rotolo, Stefano Urbani, and Letizia Vita

The island of Pantelleria is an active volcanic complex, rising 800 m from sea level, situated on a continental rift within the NW sector of the Sicily Channel (Italy).

The island is characterized by a bimodal magmatism, mafic and felsic, with this latter -by far more abundant- including metaluminous trachyte and pantellerite (peralkaline rhyolites) magmas, erupted either as lava flows, pumice falls or pyroclastic currents.

This scenario, in addition to discontinuous field exposures and closely spaced (in space and time) explosive events produced by a multitude of eruptive centers producing compositionally similar deposits, makes challenging the detailed stratigraphic reconstruction of the volcanological evolution of the island.

In the last 40 years many studies have focused on specific volcanological, geochronological geochemical and petrological aspects of the island, unravelling peculiar eruptive dynamics and petrographic aspects of peralkaline magmatism. These efforts produced great steps forward in the knowledge of the volcanological evolution of the island, strictly tied to the peculiarities of peralkaline magmas. Nevertheless, a detailed geological map comprehensive the entire volcanological evolution (pre- and post -Green Tuff ignimbrite eruption) is still missing.

In the mid 70’s and 80’s unofficial schematic geological maps have been realized, mostly focused on lithological aspects of the erupted products and their areal distribution, summarized in the seminal paper of Mahood & Hildreth (1986) the first to define a comprehensive of pre- and post-Green Tuff stratigraphy. These studies generated discordant stratigraphic subdivisions and a not univocal stratigraphic nomenclature, until the work of Jordan et al. (2018) who defined an accurate stratigraphy and nomenclature of the pre-Green Tuff ignimbrite eruptions, adjuvated by progresses made by slightly earlier geochronological and paleomagnetic studies.  The stratigraphy of post-Green Tuff volcanism, mildly explosive to effusive, though much improved by recent Ar/Ar, field and petrographic studies, still has some doubtful points.

To summarize, harmonize and integrate all the existing data into a single detailed product, the Geological Survey of Italy (in collaboration with Palermo University) performed, between 2022 and 2024, a field survey at scale 1:10.000 in the framework of the Italian Geological Mapping Project 1:50.000 scale (CARG project).

Here, we present the preliminary geological field map of Pantelleria Island (according to CARG Project guidelines) corroborated by existent and new geochronological, geochemical and petrographic analyses on hundreds of samples collected from the different volcanic structures of the island.

Such detailed field mapping of the entire island, together with the geological survey of the offshore area (still in progress), updating the findings of previous studies, allows us to obtain the first official geological map of such an active volcanic island. This represents the first fundamental step for the development of future studies of volcanic hazard assessment.

References:

Mahood, G.A., Hildreth, W. Geology of the peralkaline volcano at Pantelleria, Strait of Sicily. Bull Volcanol 48, 143–172 (1986).

Jordan, N. J., Rotolo, S. G., Williams R., Speranza, F., McIntosh, W. C., Branney, M. J., Scaillet S. Explosive eruptive history of Pantelleria, Italy: Repeated caldera collapse and ignimbrite emplacement at a peralkaline volcano. JVGR, 349, 47-73, 2018

How to cite: Pensa, A., Bonomo, R., Cinquegrani, A., Ricci, V., Rotolo, S. G., Urbani, S., and Vita, L.: Unravelling the complex volcanological evolution of Pantelleria Island (Channel of Sicily, Italy): new insights from the Italian Geological Mapping Project (CARG Project), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10643, https://doi.org/10.5194/egusphere-egu25-10643, 2025.

EGU25-11034 | Posters on site | ESSI4.7 | Highlight

The geological reference system: from the concept of an integrated geological knowledge management to cartographic representation. 

Maxime Padel, Benjamin Le Bayon, Isabelle Bernachot, Florence Cagnard, Alexis Plunder, Benoit Issautier, Thierry Baudin, Hélène Tissoux, Frédéric Lacquement, Sandrine Grataloup, and Caroline Ricodel

Although digitized, geological maps have poorly evolved since their inception and still retain the constraints imposed by printing: heterogeneity between map sheets, limited point data, and challenging update procedure that remains essentially nonexistent.

The heterogeneity of geological maps also complicates their integration to Information System. Maps, based on rules and concepts that evolved over time, are no longer suited to the precision required by database structures. As concepts and models evolve in geology together with the scientific knowledge, the choice of the representation on the map also depends on the authors’ interpretations and the period of production. As a result, the mapped geological units reflect arbitrarily chosen characteristics of the rocks concerning either their protolith nature, their metamorphic characteristics or alteration transformations, considered as the most representative at the date of publication.

In France, 1:50,000 scale geological maps provide a full territory coverage but do not escape these heterogeneity issues. To address this problem, the BRGM (French Geological Survey) has implemented the RGF program (Référentiel Géologique de la France), which aims to develop a methodology to overcome these limitations and propose an innovative approach to represent the geological knowledge. In this research program new data are acquired by PhD students allowing to improve geological knowledge and to produce updated and harmonized information on geological maps and boreholes.

The tools developed and implemented over the past decade, based on the establishment of an Information System structured as a knowledge database called the Geological Reference System, now make it possible to offer a standardized representation of geological knowledge derived from traditional geological maps. This knowledge can be represented through different reference systems: lithostratigraphic unit, event (and thus the geological history), or domain and zone (e.g. lithotectonic units).

Here, we present some of the results obtained from works conducted in the Alps, the Pyrenees, and the Montagne Noire, illustrating how this structuring of geological knowledge into reference system enables the creation of new maps tailored to the geological information one wishes to represent and, consequently, to the scientific and societal challenges at hand.

How to cite: Padel, M., Le Bayon, B., Bernachot, I., Cagnard, F., Plunder, A., Issautier, B., Baudin, T., Tissoux, H., Lacquement, F., Grataloup, S., and Ricodel, C.: The geological reference system: from the concept of an integrated geological knowledge management to cartographic representation., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11034, https://doi.org/10.5194/egusphere-egu25-11034, 2025.

EGU25-11052 | ECS | Posters on site | ESSI4.7

Geometry of the Lampang Basin, Northern Thailand, Based on Gravity Data Analysis and 2D Modeling: Implications for CO₂ Storage Reservoirs from Mae Moh Coal Mine Emissions 

Jetnipit Muenjaem, Niti Mankhemthong, Rattanaporn Fongngern, Watchirachai Sukpa, Sarawute Chantranprasert, and Takonporn Kunpitaktakum

The Lampang Basin, a Tertiary intermontane basin in Northern Thailand, holds significant geological importance with potential implications for CO₂ storage reservoirs located approximately 15 km from the Mae Moh coal mine emissions. Despite its potential, publicly available subsurface geological data for the basin is limited due to its designation as a petroleum reserve by the Defense Energy Department. This study aims to develop a 2D geological model of the Lampang Basin using gravity data collected by the Department of Mineral Resources in 2015. Geosoft Oasis Montaj software was utilized to analyze and visualize the basin's subsurface structure. Interpretation and edge detection analysis of gravity anomalies reveal that the Lampang Basin comprises six sub-basins striking predominantly NE-SW, with the greatest depth of approximately 2.2 kilometers located south of Lampang City. These sub-basins are separated by an uplifted horst structure. The basin is bounded by Paleozoic sedimentary rocks to the west and Permo-Trisassic to late Triassic volcanic and sedimentary rocks of the Sukhothai Terrane to the east. The 2D modeling highlights the half-graben structures, with the western boundary fault playing a critical role in basin formation. By integrating gravity data with available 2D seismic sections, the model delineates sub-basin structures, identifying the sediment layers with potential for CO₂ storage. This research emphasizes the suitability of the Lampang Basin for secure and sustainable CO₂ storage, demonstrating its potential to mitigate emissions from industrial sources such as the Mae Moh coal mine. The findings contribute to environmental sustainability and resource management.

How to cite: Muenjaem, J., Mankhemthong, N., Fongngern, R., Sukpa, W., Chantranprasert, S., and Kunpitaktakum, T.: Geometry of the Lampang Basin, Northern Thailand, Based on Gravity Data Analysis and 2D Modeling: Implications for CO₂ Storage Reservoirs from Mae Moh Coal Mine Emissions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11052, https://doi.org/10.5194/egusphere-egu25-11052, 2025.

EGU25-11415 | Posters on site | ESSI4.7

Preserving and visualising detailed geological map-data through hierarchical spatial grids and controlled taxonomies 

Mikkel Lykke, Marie Katrine Traun, Henrik Kartin, Casper Bramm, and Søren Lund Jensen

Creating a geological map of the world is a challenge that many geologists and cartographers can attest to. Fundamentally, to map the whole world, geologists must make a globally applicable interoperative naming convention. Geology as a field mapping science has a strong history of local and regional multilingual naming traditions, that contain rich amounts of details, which are inevitable though necessarily lost in a global naming convention. However, accessing these local details from a global viewpoint can provide important insights for different use cases, for example locating mapped alteration zones in critical raw materials exploration.

At our company, Scandinavian Highlands, we have developed a framework to work with local geological maps, ie. shapefiles with lithology, age and description properties in multiple languages, and make their details searchable in a global way.

First, this is done by mapping local geological property descriptions from maps to a hierarchical taxonomy using multilingual controlled vocabularies of synonyms. Our taxonomy includes a Lithology-tree and modifiers to these (textures, minerals, etc.), a Geochronology-tree and identified stratigraphy and placename units. Thereby, we are not overwriting and simplifying original geological properties, but rather classifying them to a global taxonomy. Second, we use a hierarchical hexagonal discrete global grid system (H3), to represent the individual map polygons on a common coordinate system and to smoothly transition between global and local visual representations of the map-data. We grid the geographical extent of polygons to the finest resolution hexagons and then aggregate by compilation to the coarser resolutions.

Thus, using the hierarchical taxonomy and hierarchical spatial grid we can search in both the original property description and the global taxonomy groups, and view even the smallest local match on a global scale. This creates a highly flexible way to explore geological maps from industry and research perspectives and for global generalists to local expert users.

How to cite: Lykke, M., Traun, M. K., Kartin, H., Bramm, C., and Jensen, S. L.: Preserving and visualising detailed geological map-data through hierarchical spatial grids and controlled taxonomies, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11415, https://doi.org/10.5194/egusphere-egu25-11415, 2025.

EGU25-11428 | Posters on site | ESSI4.7

Development of geological reference systems: integrating databases and tools to improve geological knowledge and data harmonization 

Isabelle Bernachot, Benjamin Le Bayon, Maxime Padel, Florence Cagnard, Alexis Plunder, and Guillaume Dechambenoit

The RGF research program (French Geological Reference platform) aims to establish a continuous and coherent geological knowledge base covering the entire French territory. To achieve this, the program relies on the development and implementation of comprehensive geological reference systems designed to structure, harmonize, and manage geological data, including:

  • The lithostratigraphic reference system, which provides a hierarchical classification of geological units (super-group, group, sub-group, formation, member),
  • The geological events reference system, which organizes and ranks geological events to document and reconstruct the geological history of geological units,
  • The lithotectonic reference system, encompassing structural zones, hierarchically organized tectono-stratigraphic units, and paleogeographic domains,
  • The geological boundaries reference system, which catalogs structural and geologic contact information.

These reference systems are being conceived and structured in alignment with international standards such as GeoSciML and INSPIRE. Built as PostgreSQL databases, they are under development through close collaboration between geologists, computer scientists, and GIS experts to address scientific requirements and serve as a repository of geological knowledge.

At the same time, tools and applications are being developed to utilize these reference systems to constrain the attribution of geological elements across various datasets (e.g. boreholes, geological maps, 3D models), thus facilitating the harmonization, enrichment, and updating of legacy data. This includes, for example, the modernization of historical French geological maps at a 1:50,000 scale. The process involves linking map geometries to the reference systems through GIS tools and custom QGIS plugins developed by BRGM. This approach supports the transition from static maps to dynamic, multi-scale digital representations, and enables the creation of maps tailored to various scientific objectives and practical applications.

This presentation gives an overview of the lithotectonic and geological event reference systems, the underlying database model and the QGIS plugins developed for their application. Initially developed within the RGF and now within the Digital Earth project of the PEPR research program, these tools are also applicable to other projects, such as the ongoing development of the European Lithotectonic Map by the GSEU. The presentation will also give an overview of the work carried out in the Pyrenees region, demonstrating the dissemination of updated maps of lithostratigraphic units and the ability to query geological events associated with specific formations, all accessible through a dedicated ArcGIS mapping viewer.

How to cite: Bernachot, I., Le Bayon, B., Padel, M., Cagnard, F., Plunder, A., and Dechambenoit, G.: Development of geological reference systems: integrating databases and tools to improve geological knowledge and data harmonization, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11428, https://doi.org/10.5194/egusphere-egu25-11428, 2025.

EGU25-12511 | Orals | ESSI4.7

A framework for making available Europe’s treasure of geological basic information: A collaborative effort 

Hans Georg Krenmayr, Kristine Asch, Philippe Calcagno, Dana Čápová, Cecile Le Guern, Simon Lopez, Zoltán Németh, Kris Piessens, and Urzsula Stępień

Transboundary geological baseline information, such as geological maps, datasets and 3D models following the FAIR data principles, is still scarce at European level. This is due to (1) lack of transboundary harmonisation, (2) lack of FAIRness of existing harmonised data, (3) non-existence of such data and (4) lack of a suitable framework to provide such data.

The Geological Mapping & Modelling Expert Group of EuroGeoSurveys gathers the European Geological Surveys Organisations (GSO) and establishes the data collection and model production framework as mentioned in (4). This is being done as part of the ongoing Coordination & Support Action "Geological Service for Europe"(GSEU) of the EU Horizon Europe Framework Programme.

Ongoing actions include (1) collecting metadata of geological maps, datasets and 3D models of Europe, (2) establishing a conceptual and physical data model capable of accommodating multiscale basic geological data as well as applied geoscientific data in 2.5D, (3) creating improved or new scientific vocabularies for lithology, anthropogenic deposits and lithotectonic units based on linked data and SKOS technology, (4) building a lithotectonic spatial map database using the vocabularies developed in (3), and (5) sharing experience and best practices in 3D geomodelling to federate European GSO around common approaches, as much as possible employing open source tools.

In addition to the technical framework, we are also working on appropriate organisational structures and workflows for the maintenance and future updating of all elements of the framework and its datasets.

Research and innovation needs for the next EU Framework Programme in the field of geological mapping and modelling have recently been addressed in the Strategic Research and Innovation Agenda (SRIA) of EuroGeoSurveys. These include additional scientific vocabularies (e.g. for lithogenetic units, structural and geomorphological features), tools for (semi)automatic generalisation of datasets, and a collaborative GIS platform for the user community for efficient data harmonisation. In addition, the design of innovative incentive schemes to stimulate the collection of still missing data and new technologies for subsurface exploration and geomodelling to produce geological maps and 3D models, are important parts of the SRIA.

Working well together as a large group of people with a wide range of skills and experience is essential to provide applications, such as those dealing with resources and risk, with the quantified and reliable geological information they need for their processes.

How to cite: Krenmayr, H. G., Asch, K., Calcagno, P., Čápová, D., Le Guern, C., Lopez, S., Németh, Z., Piessens, K., and Stępień, U.: A framework for making available Europe’s treasure of geological basic information: A collaborative effort, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12511, https://doi.org/10.5194/egusphere-egu25-12511, 2025.

EGU25-12904 | ECS | Orals | ESSI4.7

Advancing Land Degradation Assessment and Restoration Planning in the Mediterranean through the LanDS Toolbox 

Elena Matta, Marco Micotti, Simone Corti, and Enrico Weber

The Mediterranean region faces critical challenges related to land degradation, requiring innovative and harmonized approaches for assessment and restoration. To address these challenges, the Land Degradation Decision-Support Toolbox (LanDS) has been developed as part of the REACT4MED project (https://react4med.eu/), funded by PRIMA (https://prima-med.org/). LanDS serves as a comprehensive and adaptable platform designed to evaluate land degradation and assess the impacts of restoration measures across diverse Mediterranean contexts.

LanDS exemplifies a novel approach to combating land degradation by bridging scientific research and practical application, promoting sustainable development, and supporting climate adaptation efforts in the Mediterranean region.

The toolbox integrates five core components:

  • Geo-referenced Data Repository: a centralized knowledge base that aggregates site-specific data and resources from the project’s ecosystem restoration living labs, alongside broader datasets from global and regional repositories and satellite-based indices.
  • Data Viewer: a suite of interactive visual analytics tools enabling effective data visualisation, sharing among project partners and stakeholders, and monitoring of restoration actions.
  • Indicators Library: a modular and adaptable code library offering a wide array of indicators supporting analysis and comparisons across different spatial and temporal scales, drawn from an extensive dataset built from global repositories and project’s data.
  • Machine-Learning-Based Procedure: a cutting-edge tool designed to identify and map potentially suitable areas for upscaling and outscaling restoration measures across the Mediterranean region.
  • Interactive Web Dashboard: a user-friendly interface that delivers harmonized assessments of land degradation and evaluates the effectiveness and impact of the project’s restoration measures, while supporting dissemination of project findings.

By synthesizing knowledge from global and regional datasets with insights from living labs in pilot areas, LanDS facilitates informed decision-making for land restoration and sustainable resource management. The platform fosters the development of policy recommendations and investment opportunities aimed at addressing land degradation in the Mediterranean. Further, it enables policymakers, stakeholders, and private actors to identify investments opportunities based on maximum cost-effectiveness and impact criteria.

Built on an open-source technology stack, the LanDS toolbox ensures accessibility and transparency and is freely available at http://lands.soft-water.it.

How to cite: Matta, E., Micotti, M., Corti, S., and Weber, E.: Advancing Land Degradation Assessment and Restoration Planning in the Mediterranean through the LanDS Toolbox, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12904, https://doi.org/10.5194/egusphere-egu25-12904, 2025.

EGU25-13319 | Orals | ESSI4.7

Towards a common geomodelling toolbox: sharing practices among European Geological Surveys 

Simon Lopez, Nicolas Clausolles, Christian Brogaard Pedersen, Philippe Calcagno, Montse Colomer, Chiara d'Ambrogi, Timothy Kearsey, Chris Heerema, Ignasi Herms, Monika Hölzel, Carlos Marin Lechado, Marton Palotaï, Laure Pizzella, Timo Spoerlein, Jan Stafleu, Urszula Stepien, Ondrej Svagera, Ewa Szynkaruk, Ricky Terrington, and Marianne Wiese

The Geological Service for Europe (GSEU) project is a five-year EU-funded Coordination and Support Action (CSA) that unites 35 geological surveys and 13 additional partners from 36 countries. Its primary objective is to establish a Geological Service for Europe (GSE), a geoscience-driven initiative delivering comprehensive data, information, and insights about Europe’s subsurface. The project addresses key challenges, including the sustainable management of critical raw materials, geothermal energy resources and storage, and groundwater systems. Work Package 6 (WP6), is dedicated to providing reference geological knowledge at European level: harmonized maps, concepts and 3D geological frameworks. The project is also anchored on the European Geological Data Infrastructure (EGDI) platform, which acts as a central repository for subsurface data collected before and during the project. This platform also serves as a dynamic resource, offering open access to a wide array of 3D geological models for diverse stakeholders.

Within WP6, Task 6.3 focuses on consolidating tools and best practices for designing and visualizing 3D geological models. This involves identifying and evaluating relevant tools, documenting their applications, and providing expert guidance to support the creation of shareable and reusable models. The task also tackles common challenges, emphasizes the strengths of individual tools, and explores opportunities for future advancements in modeling workflows.

To achieve these goals, we initiated efforts to inventory practices across European geological surveys and share in-house tools. Twice a year, we are also organizing open webinars on modeling workflows and related tools. These initiatives foster transparent and constructive discussions around 3D geological modeling methodologies. A core focus is ensuring accessibility and interoperability through the adoption of modular and open software components, enabling models and workflows to be easily adapted to diverse needs. The objective it to empower the geological surveys community to innovate and enhance existing frameworks.

The final deliverable of Task 6.3 is a comprehensive report that will consolidate best practices and documents available open-source tools. Beyond that, our action seeks to establish a sustainable network of 3D geological modeling experts across European geological surveys, fostering long-term collaboration and knowledge sharing.

How to cite: Lopez, S., Clausolles, N., Brogaard Pedersen, C., Calcagno, P., Colomer, M., d'Ambrogi, C., Kearsey, T., Heerema, C., Herms, I., Hölzel, M., Marin Lechado, C., Palotaï, M., Pizzella, L., Spoerlein, T., Stafleu, J., Stepien, U., Svagera, O., Szynkaruk, E., Terrington, R., and Wiese, M.: Towards a common geomodelling toolbox: sharing practices among European Geological Surveys, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13319, https://doi.org/10.5194/egusphere-egu25-13319, 2025.

EGU25-13479 | Orals | ESSI4.7

Lithotectonic map of Europe – methodology, contribution to geosciences and further inspiration for territories outside Europe 

Zoltán Németh, Kristine Asch, Hans-Georg Krenmayr, Manuel Pubellier, Kris Piessens, Francisco Javier Rubio Pascual, Maxime Padel, Stefan Luth, Ondrej Pelech, and Paul Heckmann

Since the defining of plate tectonics in the 1960s the knowledge on geodynamics of European territory registered a remarkable progress in understanding of the multiple – polyorogenic overprints within individual lithotectonic zones and units of Europe, becoming by this way an etalon for orogenic interpretations to the rest of the world.

According to the classification by Neuendorf et al. (2011) and in a slightly modified form by the INSPIRE Geology data specifications (http://inspire.ec.europa.eu/codelist/GeologicUnitTypeValue/lithotectonicUnit), a Lithotectonic unit is a geologic unit defined on basis of structural or deformation features, mutual relations, origin or historical evolution. Contained material may be igneous, sedimentary, or metamorphic.

The Lithotectonic map of Europe will represent a combination of a lithological and a tectonic map, showing a collage of lithotectonic units and their boundaries, highlighting the geodynamic aspects of more than 2.5 billion years of crustal evolution.

This novel type of map provides a wealth of information through annotated data, contributing to the development of applied research, including raw materials exploration, environmental geology, geo-energy, etc. The evolution of lithotectonic units can be placed in that of orogenic cycles, which include the Svecokarelian, Sveconorwegian, Cadomian, Caledonian, Variscan, Alpine and Hellenic cycles. Different orogenic phases will be discriminated in these orogenic cycles. For the lithotectonic framework it is sensible to emphasize those orogenic phases which differ from the standard orogenic (Wilson) cycle of the 1960s: the post-orogenic phases of unroofing, intraplate stress elimination and regional extension.

The Lithotectonic map of Europe that is currently being compiled will be based on International Geological map of Europe and Adjacent Areas (IGME 5000; Asch, 2005, BGR, Hannover). Co-funding is provided by the EC – CINEA HORIZON-CL5-2021-D3-D2 project 101075609 Geological Service for Europe (GSEU), led by EuroGeoSurveys and its Geological Mapping and Modelling Expert Group, Work Package WP6 – Geological framework for the European geological data & information system.

Reference

Neuendorf, K.K.E., Mehl Jr., J.P. & Jackson, J.A., 2011: Glossary of Geology. Fifth Edition. American Geosciences Institute, Alexandria, Virginia, 1–779.

How to cite: Németh, Z., Asch, K., Krenmayr, H.-G., Pubellier, M., Piessens, K., Pascual, F. J. R., Padel, M., Luth, S., Pelech, O., and Heckmann, P.: Lithotectonic map of Europe – methodology, contribution to geosciences and further inspiration for territories outside Europe, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13479, https://doi.org/10.5194/egusphere-egu25-13479, 2025.

EGU25-16190 | Posters on site | ESSI4.7

Manual data review and quality control – An add-on to SaQC 

Nicole Büttner, Benjamin Louisot, Christof Lorenz, David Schäfer, and Romy Fösig

The growing volume of high-resolution time-series data in Earth system science requires the implementation of standardised and reproducible quality control workflows to ensure compliance with the FAIR data standards. Automated tools such as SaQC1 address this need, but lack the capacity for manual data review and flagging. We are therefore planning to develop a Python-based tool with an intuitive graphical user interface (GUI) for local machines, thereby enhancing the functionality of SaQC. It is anticipated that the tool will be user-friendly, even for those with limited experience of Python. The GUI will be capable of interactively visualising the time-series data, highlighting the data that has already been automatically flagged. The selection of data points may be accomplished by clicking on them or via box-selection, and a flag may be assigned via a dropdown menu. An optional comment field can be used to record supplementary information, such as details of pollution events. Moreover, the option to unflag data that has failed the automated quality control process, but which is considered valid by the scientist, will be available.

The manual flagging tool will be based on SaQC, thereby facilitating a future integration into this software package. Consequently, integration into an existing SaQC workflow will be straightforward. It should be noted, however, that this is not exclusive to SaQC users; it can be easily applied to data created by another tool for automatic quality control. A simple conversion of the data via the pandas library will be sufficient for utilisation of the manual flagging tool. The flagging schemes can either be adopted from SaQC or user-specific schemes can be integrated. Once the flagging process is completed, the user is able to decide how to export the data set.

The manual flagging tool represents a valuable addition to existing toolkits for all scientists handling time-series datasets, effectively completing the data quality control process. From a scientific perspective, the benefits of this tool include increased efficiency and traceability in the data flow, as well as improved data quality through the fine-tuning of automatic controls based on experience and contextual knowledge.

 

1 Schäfer, David, Palm, Bert, Lünenschloß, Peter, Schmidt, Lennart, & Bumberger, Jan. (2023). System for automated Quality Control - SaQC (2.3.0). Zenodo. https://doi.org/10.5281/zenodo.5888547

How to cite: Büttner, N., Louisot, B., Lorenz, C., Schäfer, D., and Fösig, R.: Manual data review and quality control – An add-on to SaQC, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16190, https://doi.org/10.5194/egusphere-egu25-16190, 2025.

EGU25-17232 | Orals | ESSI4.7

Blue Insight – Using digital platforms for Arctic Ocean Science and sustainable fisheries 

Torill Hamre, Arne Johan Hestnes, Hanne Sagen, Leif Edvard Bildøy, Kjetil Haugvik Hansen, Espen Storheim, and Frode Monsen

Climate change in the Arctic is both a potential threat and a potential opener for new opportunities in sustainable development of the region. Better access to data, methods and tools, as well as better documentation of these are needed to advance science and support decision-making on public and private sector. While local, national and regional communities in different parts of the Arctic face individual challenges due to climate change, they experience similar problems when gathering and analysing data to address these issues.

The Arctic Ocean is one of the least explored oceans on the planet. There is lack of in situ observations in large parts of this region, especially on the seafloor and in the ocean under the sea ice. This means that the research communities working with climate, weather, ice-ocean processes, and geophysical hazards have limited knowledge about the processes below the sea ice. The lack of data limits the possibility to advance research in this region. It is therefore necessary to establish observing systems for data collection in the Arctic Ocean.

Fisheries plays a key role in the economy in many Arctic countries. With climate change affecting marine ecosystems and enabling accessibility of larger ocean areas in the region, there is a strong need for improved access to data and information tailored to the fisheries industry and the public sector monitoring and regulating Arctic Fisheries. Relevant data is available from many different sources, but efficient delivery chains for compiling, integrating, and analysing these in a common system is lacking.

These diverse stakeholder groups all need access to data from different sources, such as ice and underwater ocean observing systems, satellites, operational forecasting services and climate models, which provide data with different spatial and temporal resolution, in different formats and with varying levels of documentation. Furthermore, methods and tools are needed for data processing, analysis and visualisation to integrate heterogeneous multi-source data with reference and socio-economic data to support science driven as well as sector specific applications.

Blue Insight developed by Kongsberg Discovery AS offers a robust, modular platform designed for the processing, visualization, and sharing of ocean data. The core module integrates a cloud framework, data visualization tools, and comprehensive data storage and management capabilities. In addition, the system includes a data processing framework that facilitates the reuse of data processing methodologies. To enhance Blue Insight's functionality and cater to projects of varying scales, additional modules can be seamlessly integrated into this framework. This is done by means of container and workflow engine technologies, ensuring interoperability with existing digital twin for the ocean (DTO) and other research infrastructure initiatives through OGC standards.

The presentation will give an overview of the Blue Insight system and showcase how this digital platform is used and extended within the Horizon Europe HiAOOS project and the SBEP ARCFISH project.

How to cite: Hamre, T., Hestnes, A. J., Sagen, H., Bildøy, L. E., Hansen, K. H., Storheim, E., and Monsen, F.: Blue Insight – Using digital platforms for Arctic Ocean Science and sustainable fisheries, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17232, https://doi.org/10.5194/egusphere-egu25-17232, 2025.

EGU25-17590 | Posters on site | ESSI4.7

Streamlining geological field data collection, management, and integration with QField 

Juliette Stephan-Perrey, Isabelle Bernachot, Alexis Plunder, Maxime Padel, Benjamin Le Bayon, and Morgan Bezard

Field observations and measurements are essential for reconstructing the geological history of a region and for producing accurate maps and models. For national geological surveys, these raw data are critical resources that need to be efficiently managed, stored and disseminated for effective reuse in research. At BRGM, GeoField1, an application developed within the framework of the Référentiel Géologique de la France (RGF)2 and subsequently extended to support other geological projects, allows field data to be managed and capitalised according to the geological reference system, ensuring compatibility with internal and international standards. This system facilitates the compilation, sharing and reuse of data in accordance with the FAIR principles (Findable, Accessible, Interoperable, Reusable).

Recently, a data acquisition process was implemented to streamline field data collection and enable direct integration into GeoField. This workflow relies on QField3, an open-source mobile application built on the QGIS engine, leveraging key features such as seamless integration with QGIS projects, customizable data forms, the ability to use proper vocabularies, and offline mapping capabilities. A dedicated master project was designed to meet the specific needs of field geologists, enabling them to capture essential information through tailored forms, complete with dropdown lists to ensure consistency in terminology. Adapted symbology allows for real-time visualization of structural measurements on the map. The master project is made available to users through a QGIS plugin, which loads the project template, including a predefined database structure and up-to-date BRGM lexicons. Users can then customize their project by adding relevant layers, such as map backgrounds and other vector or raster data (e.g., DEM, geochemical analysis, geophysical data), and load the project onto their mobile devices for field acquisition.

Upon returning to the office, the plugin facilitates the automatic transfer of field data into GeoField, ensuring seamless integration into the BRGM central database. This workflow provides a robust, efficient, and standardized approach to geological data collection, capitalizing on the synergies between QGIS/QField and GeoField to enhance data management, sharing, and reuse within the geoscientific community.

 

1GeoField page: https://rgf.brgm.fr/page/geofield

2RGF page: https://rgf.brgm.fr/

3QField - Efficient field work built for QGIS, url: https://qfield.org/

How to cite: Stephan-Perrey, J., Bernachot, I., Plunder, A., Padel, M., Le Bayon, B., and Bezard, M.: Streamlining geological field data collection, management, and integration with QField, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17590, https://doi.org/10.5194/egusphere-egu25-17590, 2025.

EGU25-20452 | ECS | Orals | ESSI4.7

CRAFTY-UI: Bridging Complexity and Usability in Large-scale Agent-Based Land-Use Modelling for Earth System Science 

Mohamed Byari, Yongchao Zeng, Calum Brown, and Mark Rounsevell

Earth system science increasingly relies on complex, high-dimensional datasets from advanced modelling and remote sensing. Transforming these data into actionable insights for stakeholders and decision-makers beyond academia remains a major challenge. In response, we developed a novel, open-source graphical user interface (GUI) for the CRAFTY (Competition for Resources between Agent Functional Types) agent-based land-use model. Built using JavaFX and available on GitHub, this GUI bridges the gap between model complexity and user accessibility. Key features of CRAFTY-UI include: (1) an intuitive workflow for configuring Agent Functional Types (AFTs) that represent diverse land managers, (2) real-time visualization of simulation progress, and (3) input/output data analysis. Users can customize model parameters such as AFT behavior, productivity settings, societal demand for ecosystem services (including pricing), and multiple protected-area and policy restrictions. Additionally, they can select from different mechanisms such as competition algorithms, mutation intervals, and neighbourhood effects according to their specific research or policy scenarios.

CRAFTY relies on a broad range of input data including land resource maps (capital maps) for each year under multiple SSP-RCP scenarios, along with AFT parameterizations, ecosystem service demands, and protected-area restrictions. The CRAFTY-UI enables users to analyse these datasets across spatial, temporal, and scenario dimensions. In addition to monitoring simulations in real time, users can also import previous runs and employ difference analysis tools to compare outcomes under varying parameters and mechanisms. These features foster a deeper understanding of how social, economic, and environmental capitals interact to shape land-use trajectories. Beyond its advanced analytical capabilities, the CRAFTY-UI simplifies large-scale simulation exploration, offering adjustments that instantly update interactive charts and spatial outputs. By clarifying feedback loops between natural capital, socio-economic drivers, and institutional influences, the interface facilitates evidence-based decision making. Ultimately, the CRAFTY-UI bridges complexity and usability, enabling diverse stakeholders to engage with modelling outputs, enrich policy discussions, and collaboratively shape sustainable land-use strategies.

How to cite: Byari, M., Zeng, Y., Brown, C., and Rounsevell, M.: CRAFTY-UI: Bridging Complexity and Usability in Large-scale Agent-Based Land-Use Modelling for Earth System Science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20452, https://doi.org/10.5194/egusphere-egu25-20452, 2025.

EGU25-20681 | ECS | Posters on site | ESSI4.7

Improving Ice Segmentation in Permafrost Cores using Computed Tomography 

Mahya Roustaei, Jan Nitzbon, Jordan Harvey, Evan Francis, Steffen Schlueter, Julia Boike, and Duane Froese

The segmentation of ice in X-ray Computed Tomography (CT) scans of permafrost samples has traditionally relied on the Hounsfield Unit (HU) thresholding approach while their accuracy is often limited by overlapping density ranges in complex and heterogeneous samples. Recent advances, including automated thresholding algorithms and machine learning techniques, offer improved precision by leveraging texture, contrast, and morphological features in CT images. This study investigates the evolution of ice segmentation methodologies by applying multiple approaches to a 164 cm long permafrost core drilled from a Yedoma upland in north-eastern Siberia. The core was analyzed using traditional HU thresholding, automated thresholding methods (e.g., Otsu and adaptive histogram-based segmentation), and machine learning models (e.g., random forests and convolutional neural networks). The results from CT scans and segmentation methods were validated and compared against laboratory measurements of ice content and density, ensuring a robust evaluation of each technique's accuracy and reliability.

The results provide critical insights into the strengths, weaknesses, and suitability of different segmentation methods for permafrost cores. These findings contribute to the development of standardized, high-precision methodologies for non-destructive characterization of ice-rich soils, supporting geotechnical and climate change studies in permafrost regions.

How to cite: Roustaei, M., Nitzbon, J., Harvey, J., Francis, E., Schlueter, S., Boike, J., and Froese, D.: Improving Ice Segmentation in Permafrost Cores using Computed Tomography, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20681, https://doi.org/10.5194/egusphere-egu25-20681, 2025.

EGU25-20862 | ECS | Orals | ESSI4.7

B-QualMT - A software quality management tool to select borehole records for 3D models 

Elisabeth Schönfeldt, Thomas Hiller, Marcus Fahle, Jörg Giese, Mathias Hübschmann, and Friedemann Grafe

Exploration data (e.g. borehole records, geophysical sections) represent the essential input data for any geological model. Borehole records often play a decisive role in geological modeling. Usually, they contain descriptions and interpretations on petrography, lithology and stratigraphy. This information is crucial for modeling the spatial distribution of lithostratigraphic layers in three dimensions. However, these interpretations can be inconsistent or error-prone. The reasons include for instance, the date of recording (reflecting the prevailing state of knowledge at the time), the specific exploration target and techniques employed, the quality of digitalization, and the potential for human interpretative bias. Separating adequate from unsuitable borehole records is of great importance, yet rather difficult, especially evaluating large datasets. While visual inspection of the inferred geological model is a viable approach, it results in numerous iterations to identify inadequate drilling profiles, which is time-consuming and expensive.
               In order to streamline the process of testing the quality of the data from borehole records, we developed the Python-based software package B-QualMT (borehole quality management tool) that can filter borehole records based on user-adjustable standards. The tool has a given set of deterministic tests depending on the user’s auxiliary information (e.g. previous 3D-models) and knowledge of the regional geological settings (e.g. sequence of geological layers), which can be used to select divergent drilling profiles for geologically comparable regions.
               For our pilot study, we selected a former lignite mining area of Lusatia in the southeast of Germany bordering the Federal State of Brandenburg and the Free State of Saxony (Freistaat Sachsen). Here, the goal is to improve the previous geological model with 3000 additional borehole profiles from various exploration surveys spanning several decades. We show the evaluation process, how the deterministic tests work and will additionally give an outlook on the planned integration of machine learning algorithms identifying geological patterns in previously quality-tested borehole records.

How to cite: Schönfeldt, E., Hiller, T., Fahle, M., Giese, J., Hübschmann, M., and Grafe, F.: B-QualMT - A software quality management tool to select borehole records for 3D models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20862, https://doi.org/10.5194/egusphere-egu25-20862, 2025.

Satellite imagery is essential for continuously monitoring Earth phenomena, detecting disasters and hazards, and effectively identifying large and small-scale changes across wide areas. Over the past few decades, advancements in satellite technology have significantly increased the use of satellite imagery. In particular, in change detection studies or disaster monitoring research utilizing multi-temporal and multi-satellite imagery, the fusion of images from two or more time periods for the same region is indispensable. However, due to the inherent characteristics of satellite imagery being captured from a distance, geometric distortions are likely to occur, potentially resulting in misalignment between the images and the actual ground surface. The accuracy of high-resolution satellite imagery is determined by the precision of geometric corrections, which becomes an even more critical factor when using multi-satellite and multi-temporal imagery. Consequently, image registration is an essential process in studies that utilize the fusion of high-resolution satellite imagery. In this study, we propose a highly accurate image registration method using high-resolution satellite imagery from CAS500-1, KOMPSAT-3A, and KOMPSAT-3. To overcome the limitations of feature point detection, a ResShift-based super-resolution technique was applied to generate a dataset with higher resolution than the original data, maximizing the performance of the feature matching models. For deep learning-based feature point detection and matching models, SuperPoint, SuperGlue, LightGlue, and RoMa were utilized. Notably, the RoMa model demonstrated exceptional performance by recording over 2,300 correct matches on the super-resolved dataset. The results of this study are expected to contribute to effective image registration in various fields that utilize multi-temporal and multi-satellite imagery.

This work is supported by the Korea Agency for Infrastructure Technology Advancement(KAIA) grant funded by the Ministry of Land, Infrastructure and Transport (Grant RS-2022-00155763).

 

How to cite: Im, Y. and Lee, Y.: Image Registration of CAS500-1 and KOMPSAT-3/3A Satellite Images Using Deep Learning-Based Feature Matching and Super-Resolution Techniques, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6359, https://doi.org/10.5194/egusphere-egu25-6359, 2025.

EGU25-12032 | Posters on site | ESSI4.9

Between Heaven and Earth 

Katharina Schleidt and Stefan Jetschny

As the Copernicus program matures, ever more gridded data becomes available to researchers to incorporate into their studies. This data is partially raw satellite data, but an increasing amount of derived products are becoming available. In addition, data from various terrestrial sources is being aggregated to gridded formats, enabling integration with products derived from satellite data.

The technologies available for provision, sharing and processing of gridded data have traditionally been developed by the EO community, with functionality tailored towards the requirements of satellite data. Where these technologies have been applied to the more terrestrial products, both derived from satellite data as well as that generated from terrestrial sources, gaps become apparent in the metadata provided. 

These gaps pertain to concepts not required for satellite data, as they are either not relevant, or have clear default values. For example in ISO 19123-1:2023, while one can define if the value being provided pertains to the center of the grid cell or one of the corners (Pixel-in-center, pixel-in-corner), it is not possible to indicate that the value pertains to the entire area of the cell, as required for land cover or population grids.

A further gap becomes apparent regarding the Observable Property that is being conveyed by the provided data. When dealing with satellite data, the only Observable Property being provided tends to be radiance, the only additional metadata to be provided details the individual frequency bands. When dealing with terrestrial products there are almost infinite lists of Observable Properties for which data is collected or generated, clean indication of what exactly the data represents is essential.

In some cases such information is provided through the use of relevant extensions, e.g. the STAC raster extension, that foresees a link to a semantic resource defining what the data actually represents. However, often, this information is not provided in a structured form. The user must extract this information from textual documentation to understand what the data actually represents. 

Proper provision of Observable Property concepts with gridded data would greatly enhance both data discoverability and reuse, as essential concepts describing the data are cleanly exposed, not requiring the user to guess from titles or poorly defined keywords. Proper integration of Observable Property concepts in core metadata structures would greatly increase the FAIRness of provided data.

Further issues encountered in sharing gridded data from diverse sources have to do with the currently available standardized web services and APIs. OGC WCS has been shown to have integral errors providing data over time, while work on OGC API - Coverage has yet to be completed. The openEO API is an interesting alternative, but as this is a processing API, deployment of this API purely for data accessibility entails a great deal of unnecessary overhead.

In conclusion, in order to reap the potential that can be gained from the diverse gridded data products emerging from both terrestrial and satellite sources, there are still a number of issues to be resolved, both in their description and accessibility.

This work was enabled by the FAIRiCUBE EU Horizon Project.

How to cite: Schleidt, K. and Jetschny, S.: Between Heaven and Earth, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12032, https://doi.org/10.5194/egusphere-egu25-12032, 2025.

EGU25-13947 | Posters on site | ESSI4.9

High Altitude Platform Stations: A Novel Earth Observation Technique from the Stratosphere 

Daniel Price, Philipp Sueltrop, Mark Rocket, Matthew Fladeland, and Stefan Baumgartner

High Altitude Platform Stations (HAPS) are an emerging technology solution for Earth Observation that are beginning to reach fruition. Operating for weeks at a time in the lower stratosphere (~ 20 km, 65,000 ft), the solar-powered long endurance aircraft provide a transformative ability to monitor areas of interest at unprecedented temporal and spatial resolution. At these improved resolutions, HAPS provide a significant advantage over satellite-based sensors and have a broad range of scientific and operational applications. Large-scale deployment of HAPS technology will revolutionise Earth Sciences with direct benefit to any science question attempting to improve understanding of Earth-surface system processes. Key industry applications include environmental monitoring, precision agriculture, forestry, smart cities and atmospheric sounding. The technology could play a critical operational role in advancing maritime domain awareness and disaster response.

At Kea Aerospace we are currently conducting flight operations with our Mk1 Kea Atmos aircraft capable of stratospheric flight to an optimal altitude of ~50,000 ft. The Mk1 has a 12.5 m wingspan and can deploy a 2.5 kg payload, with a 200L x 200W x 300H mm volume. The average payload power consumption will influence the mission profile and thermal control requirements with power availability mission and payload specific. We aim to deploy optical hyperspectral, synthetic aperture radar and atmospheric sampling instrumentation with key scientific and industry partners including the National Aeronautics and Space Administration (NASA) and German Aerospace Center (DLR).

We present preliminary findings from our Mk1 flight test programme in New Zealand and an overview of our future aspirations and upcoming Mk2 stratospheric long endurance aircraft programme.

How to cite: Price, D., Sueltrop, P., Rocket, M., Fladeland, M., and Baumgartner, S.: High Altitude Platform Stations: A Novel Earth Observation Technique from the Stratosphere, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13947, https://doi.org/10.5194/egusphere-egu25-13947, 2025.

Accurate and efficient mapping of crop spatial distribution is crucial for agricultural monitoring, yield prediction, and environmental sustainability.  In this study, we developed a novel workflow, GEDI-Guided Crop Mapping Framework (GGCMF), for high-resolution mapping of corn and sorghum by integrating GEDI data, Sentinel-2 imagery, and machine learning classifiers within the Google Earth Engine (GEE) platform. The GGCMF workflow begins by utilizing historical CDL crop type maps to extract canopy height and vertical structural differences from GEDI L2A Vector data, which are processed within a newly developed GEE-compatible framework.  This ensures minimal geolocation errors and allows the accurate differentiation of high- and low-vegetation classes (e.g., corn + sorghum vs. other crops).  Subsequently, Sentinel-2 imagery is employed to capture unique phenological and spectral features, enabling the generation of high-quality training samples for the fine-scale differentiation of corn and sorghum.

This automated approach was applied to multiple years (2019–2022) and regions (China and the U.S.), assessing its transferability and robustness.  Validation of corn classification achieved an average overall accuracy (OA) of 0.91, with strong correlations to independent labels, published mapping products (R² = 0.98), and official statistics (R² = 0.96).  The current results for corn show that the GGCMF method is not only highly accurate but also robust across different temporal and spatial scales. The integration of GEDI and Sentinel-2 data within GEE offers a cost-effective and scalable solution for mapping structurally distinct crops.  By leveraging GEDI's canopy height data for automatic labeling and combining it with Sentinel-2's high-resolution imagery, GGCMF presents a novel, automated workflow for crop mapping.  This approach has significant potential for large-scale agricultural monitoring, providing timely and reliable data to support sustainable agricultural management.

How to cite: Li, Z.: Automated and Scalable Corn and Sorghum Mapping Across Diverse Regions Using GEDI and Time-Series Sentinel-2 Imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16846, https://doi.org/10.5194/egusphere-egu25-16846, 2025.

EGU25-17511 | Posters on site | ESSI4.9

Operational Results of GK-2A AMI Special Observations 

Hye-Won Kim and Sang Cherl Lee

KARI (Korea Aerospace Research Institute) is responsible for operating three geostationary satellites, including the GK-2A (Geo-KOMPSAT-2A), and plays a key role in ensuring the stable operation of these satellites. The institute contributes to the continuous acquisition of satellite imagery data, providing around-the-clock support for satellite operations and monitoring. GK-2A was launched on July 5, 2018, from the Guiana Space Centre in French Guiana. It is part of Korea's next-generation geostationary meteorological satellite program, designed to enhance weather forecasting capabilities. This satellite is equipped with advanced payloads, including the Advanced Meteorological Imager (AMI) and the Korea Space Environment Monitor (KSEM). The AMI is dedicated to observing atmospheric conditions in real time, providing high-resolution imagery for weather analysis and forecasting, while the KSEM is tasked with monitoring space weather phenomena, such as solar radiation and geomagnetic storms, which can impact satellite operations and communication systems. The AMI operates across multiple spectral bands, enabling detailed observations of clouds, precipitation, and other atmospheric phenomena. It covers a wide area, including the East Asia region, with a temporal resolution that allows for frequent imaging of the Earth’s atmosphere. One of the AMI observation modes, Local Area (LA), typically covers the Korean Peninsula. However, in the case of special observations, the AMI can perform LA observations, where it is capable of imaging any region within its Field of View (FOV), beyond the standard observation area. This flexibility enhances its capacity for targeted monitoring, making it particularly useful for high-priority events, localized weather phenomena and global disasters requiring rapid observations.
This paper presents an overview of the operational results since the launch of the GK-2A, with a particular focus on special observations conducted using the AMI. The results from the special observation operations during the normal operational period of GK-2A are expected to provide insights into the future direction for the development of special observation operations using the AMI. 

How to cite: Kim, H.-W. and Lee, S. C.: Operational Results of GK-2A AMI Special Observations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17511, https://doi.org/10.5194/egusphere-egu25-17511, 2025.

EGU25-18783 | ECS | Posters on site | ESSI4.9

Towards Sentinel-2-based monthly water occurrence mapping with the Dynamic World data suite 

Leo Helling, Barbara Belletti, Mathis Messager, Louis Rey, Hervé Parmentier, and Hervé Piégay

Global water occurrence data derived from satellite imagery provide critical insights into surface water dynamics, informing science and management of key issues like climate change, water scarcity, and biodiversity loss. The Landsat-based Global Surface Water (GSW) dataset (Pekel et al., 2016) has notably provided an important archive of the global surface water areas and its changes over time. However, its 30-m resolution limits its applicability for smaller river systems. Since the launch of the Copernicus Sentinel-2 (S2) program, higher-resolution imagery (10 m) at recurrence times of 5 days is available, but has not yet been fully exploited. The only large-scale, temporally explicit layer of water occurrence based on S2 was provided by Yang et al. (2020) for the French Metropolitan region, but is limited by noise from clouds, terrain shadows, and seasonal snow. 

The recently developed Dynamic World database (Brown et al., 2022) provides a probabilistic, pixel-scale land cover classification of S2 images updated globally in near-real time, potentially enabling computationally efficient, temporally continuous water mapping at high resolution. Here we evaluate DW’s water detection capabilities and propose a workflow for large-scale, monthly surface water occurrence mapping. Our approach integrates probabilistic and physical-based water classification, topographic filtering, and cloud masking to overcome limitations of GSW and existing Sentinel-2 applications. DW’s water probabilities were compared to spectral indices (NDWI, MNDWI) and combinations of these metrics were explored. We also assessed the potential for topographic data (FABDEM) and pixel-quality measures (CloudScore+) to reduce misclassification and allow the inclusion of more observations. The analysis is applied to the French Rhône-Mediterranean basin, a region chosen due to its diverse hydrological, climatic and geomorphological conditions. Verification is performed using a recently developed high-resolution annual land use product for mainland France (Manière, 2023) and results are compared to the GSW layer.

Preliminary results demonstrate that DW natively detects water in most areas well, but noise from shadow remains a challenge. Through combination with NDWI and further filtering with topographical data, significant classification improvements can be achieved. In addition, the pixel-based cloud-filtering with CloudScore+ enables the inclusion of more observations compared to previous methods. We implemented this approach on Google Earth Engine with a simple and efficient algorithm providing monthly water occurrence observations for a whole year. This scalable workflow holds the potential to address significant limitations of prior methods and facilitate large-scale surface water mapping at high resolution. The results are especially significant in areas where in-situ hydrological monitoring is scarce.

 

References

Brown, C. F., Brumby, S. P., Guzder-Williams, B., et al. (2022). Dynamic World, Near real-time global 10 m land use land cover mapping. Scientific Data, 9(1), Article 1. https://doi.org/10.1038/s41597-022-01307-4

Manière, L. (2023). Projet MAPD’O. https://bassinversant.org/wp-content/uploads/2023/03/presentation_mapdo.pdf

Pekel, J.-F., Cottam, A., Gorelick, N., & Belward, A. S. (2016). High-resolution mapping of global surface water and its long-term changes. Nature, 540(7633), Article 7633. https://doi.org/10.1038/nature20584

Yang, X., Qin, Q., Yésou, H., et al. (2020). Monthly estimation of the surface water extent in France at a 10-m resolution using Sentinel-2 data. Remote Sensing of Environment, 244, 111803. https://doi.org/10.1016/j.rse.2020.111803

How to cite: Helling, L., Belletti, B., Messager, M., Rey, L., Parmentier, H., and Piégay, H.: Towards Sentinel-2-based monthly water occurrence mapping with the Dynamic World data suite, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18783, https://doi.org/10.5194/egusphere-egu25-18783, 2025.

EGU25-19140 | ECS | Posters on site | ESSI4.9

When is a finer spatial resolution justified in remote sensing analysis? 

Yomna Eid and Edzer Pebesma

Remote sensing analysis is often used to provide supporting information for evidence-informed policy-making. Typically, such analysis presents results as classification maps, such as a land cover classification used to estimate deforestation areas in a region. For such analyses, where aggregated areal values of specific classes are the primary targets, a critical question arises: do the results significantly degrade when lower spatial resolution Earth Observation (EO) products are used instead of higher-resolution ones?

EO products like Dynamic World land use and land cover maps, produced at a high temporal and spatial resolution (5 days and 10m, respectively), are built on the widely held belief that higher resolutions inherently yield better results. However, with the exponential growth in data volumes and the computational demands of high-resolution workflows, it becomes increasingly important to determine where these resource-intensive approaches provide meaningful advantages — and where they do not — to balance computational efficiency with the need for accuracy in remote sensing workflows.

To address this question, we examine two case studies: deforestation in the Cerrado Biome of Brazil, and the imperviousness of sealed surfaces in Germany. Classification maps from each study are systematically downsampled from their native resolutions in steps up to 10 km spatial resolution. Using Ripley’s Equation1, numerically approximated with a Gaussian-Quadrature approach, we compute standard errors to assess the impact of spatial resolution on classification accuracy.

We report our findings on how the aggregated target values derived from lower-resolution data compare to those from higher-resolution inputs. We also try to identify the resolution thresholds beyond which the quality of the final product loses acceptable representation of the phenomena in the selected use cases.

1 See Eq. 3.4, page 23 in Ripley, B.D. (1981), “Spatial Sampling” in Spatial Statistics.

How to cite: Eid, Y. and Pebesma, E.: When is a finer spatial resolution justified in remote sensing analysis?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19140, https://doi.org/10.5194/egusphere-egu25-19140, 2025.

EGU25-19832 | Posters on site | ESSI4.9

A Features Reconstruction and Prediction Joint Learning Framework with Incomplete SITS for Agriculture Semantic Segmentation  

Yuze Wang, Mariana Belgiu, Aoran Hu, Rong Xiao, and Chao Tao

The dense Satellite Images Time Series (SITS) plays an important role in the agriculture semantic segmentation task. However, in real-world scenarios, cloud contamination and temporary sensor outages can lead to significant data missing in SITS, which declines the performance of models trained on ideal scenarios. A common approach is to reconstruct the complete SITS before the model’s prediction, where the reconstruction is independent of the prediction. This approach not only leads to the error accumulation from reconstruction to prediction, but also the detailed rebuilding of complete SITS may be redundant for the prediction. In this paper, we proposed a features reconstruction and prediction joint learning framework. The collaborative optimization of the two tasks aims to encourage the model to efficiently reconstruct complete features beneficial for prediction from incomplete SITS. Specifically, we simulate the data-missing scenarios with masks. The prediction task of masked data is supervised by labels. Meanwhile, by using the model that is well-trained on ideal scenarios as a teacher, we leverage its extracted temporal features from the data before masking as the target of the feature reconstruction task. The gradient flow of two tasks will be shared, which enables mutual supervision between them. Feature reconstruction prevents the model from acquiring incorrect reasoning ability caused by the shortest path problem during prediction, whereas prediction keeps reliability and reduces redundancy of reconstructed information. Furthermore, after training with the proposed framework, the model architecture remains unchanged and still maintains its robustness of complete SITS, which enhances the model's feasibility in practical applications. The experiments were conducted across multiple agricultural semantic segmentation datasets with incomplete SITS, sourced from Sentinel-2 and Planet satellites. We also validate its robustness for the common model architectures, and visualize the intermediate features to explore the mutual influence between the two tasks.

How to cite: Wang, Y., Belgiu, M., Hu, A., Xiao, R., and Tao, C.: A Features Reconstruction and Prediction Joint Learning Framework with Incomplete SITS for Agriculture Semantic Segmentation , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19832, https://doi.org/10.5194/egusphere-egu25-19832, 2025.

Semantic segmentation of cropland is critical for accurately extracting crop distribution from satellite remote sensing (RS) images. However, the dynamic temporal patterns caused by crop rotations and the heterogeneous spatial characteristics of cropland pose significant challenges for achieving high-precision segmentation. To tackle these issues, we propose a novel spatiotemporal feature-enhanced network (STFE) designed specifically for cropland segmentation in remote sensing time-series images (RSTI). The STFE network effectively integrates temporal and spatial features by introducing key innovations. First, we design an edge-guided spatial attention (EGSA) module to enhance spatial detail extraction, particularly for delineating ambiguous boundaries. Second, a progressive feature enhancement (PFE) strategy is developed to capture and fuse multi-scale features progressively across network layers. Third, for temporal feature extraction, we incorporate a differential awareness attention (DAA) module, built on ConvLSTM, to dynamically aggregate temporal information, enabling the model to better capture crop rotation patterns and temporal variations. Experimental results on three benchmark datasets—PASTIS, ZueriCrop, and DNETHOR—demonstrate the superior performance of STFE compared to state-of-the-art methods, achieving mean IoU improvements of 3.2% over the best-performing baseline. The model excels particularly in handling challenging scenarios such as irregular crop shapes and mixed cropping patterns. Its adaptability to complex and evolving agricultural landscapes provides a scalable and reliable solution for supporting sustainable farming practices and informed decision-making.

How to cite: chang, M. and li, S.: Cropland segmentation leveraging a synergistic edge enhancement and temporal difference-aware network with Sentinel-2 time-series imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20365, https://doi.org/10.5194/egusphere-egu25-20365, 2025.

EGU25-35 | ECS | Orals | ESSI4.10

How successive meteotsunami and storm activity disrupts saltmarsh vegetation. 

Clare Lewis, Jonathan Dale, Jess Neumann, Tim Smyth, and Hannah Cloke

Meteotsunami or meteorological tsunamis are globally occurring progressive shallow water waves with a period of between 2 to 120 minutes which result from an air-sea interaction. Meteotsunami are initiated by sudden pressure changes and wind stress from moving atmospheric systems. These waves are known to cause destruction to assets with injury and fatality to human life. Currently, there is no research into the impact upon ecological assets.

This presentation outlines the impact of two meteotsunami events (2016 and 2021) on an intertidal saltmarsh ecosystem in the southwestern UK. By utilizing satellite imagery and applying Normalized Difference Vegetation Index (NDVI) an assessment was carried out on vegetation before and after each event against a baseline 10-year mean. Results revealed that the 2016 meteotsunami resulted in a minimal impact upon vegetation, suggesting a potential resilience or adaptive response to a single episodic disturbance. In contrast, the 2021 event, compounded by two significant storms and multiple additional meteotsunami, led to a notable decline in NDVI values, indicating a likely short-term disruption to vegetation. Recovery appeared to be rapid (within one to three months.)

This comparative analysis underscores the complex interactions between meteotsunami events, climatic phenomena, and coastal vegetation dynamics, highlighting the necessity for ongoing monitoring and research to understand the resilience mechanisms of such ecosystems in the face of increasing climatic variability and extreme weather events.

 

How to cite: Lewis, C., Dale, J., Neumann, J., Smyth, T., and Cloke, H.: How successive meteotsunami and storm activity disrupts saltmarsh vegetation., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-35, https://doi.org/10.5194/egusphere-egu25-35, 2025.

Sea-level rise (SLR) driven by climate change has exacerbated coastal erosion, posing significant challenges for coastal management. Effective management necessitates robust tools to evaluate shoreline dynamics under varying climate scenarios, facilitating the identification of high-risk areas. However, the pixelated nature of coastlines and the limited scope of large-scale coastal projections under diverse climate conditions hinder comprehensive risk assessment. This study addresses these gaps by utilizing medium-resolution Landsat data integrated with a Convolutional Neural NetworkCNN - Random Forest-RF, enhanced by an activation function and five max-pooling, to process training predictors based on spectral indices MNDWI, NDWI, NDVI, GCVI, and SAVI, shorelines detection and demarcation. The analysis applies Bruun Rule to assess shoreline retreat relative to SLR along Pakistan's coast at five-year intervals from 2020 to 2050. SLR and SST are sourced from multiple satellite sensors, including AVHRR and SLSTR, and computed using CMEMS relative to a 2000–2023 baseline. Climate projections are derived from a multi-model ensemble of CMIP6 General Circulation Models (GCMs), spanning Shared Socioeconomic Pathways SSP1-2.6 to SSP5-8.5. The proposed CNN-RF model demonstrated high accuracy, achieving precision, recall, and F1 scores of 95.01%, 96.16%, and 96.91%. Results from historical regression rates, combined with SLR and SST projections, indicate widespread erosion in Indus Delta, with alarming retreat rates of -80.4 ±1.15 m/year between 2000 and 2010, corresponding to SLR values ranging from 0.015 to 0.085 m/year. From 2010 to 2023, SLR accelerated to 0.087–0.15 m/year, with SST increasing from 297.79 K to 300.3 K. Conversely, the Sandspit coast exhibited accretion, gaining 23.24 km² at rates of up to mean 49.45 ±1.16 m/year. Notable warming trends were observed, with central Arabian Sea SSTs exceeding 302.41 K, correlating strongly with SLR (R² = 0.40 by 2023). Under the high-emission scenario SSP5-8.5, projections for 2020–2025 indicate persistent erosion in the Indus Delta, with retreat rates of -25 to -60 m/year, while Gwadar Port up to 10 to 15 m/year. For 2025–2030 and 2030-2050 erosion in the Indus Delta, retreat rates up to -68 m/year and of -101 to -120 m/year, Sonmiani Aquifer may transition erosion up to mean -55.1 and -110 m/year). SST anomalies exhibit variability (0.3°C–0.8°C) and periodic spikes linked to climatic events, with annual increases of 0.02°C–0.05°C and a coefficient of variation of 12%–25%. Pearson’s correlation (R² = 0.6–0.8) suggests a positive relationship between SST and SLR, but highlighted variability, indicating areas for refinement. The impacts of the intrusion on the local coastal community are also analyzed with trends of communities’ migration. Our analysis revealed that erosion also results from reduced sediment flow linked to water infrastructures. Future policy and action plans should prioritize Integrated Coastal Zone Management frameworks (ICZMF), providing critical insights into erosion dynamics and addressing integrated nature-based solutions.

Keywords: Sea-level rise (SLR), Coastal Erosion, CNN-Random Forest (RF), Landsat, CMIP6, Integrated Coastal Zone Management frameworks (ICZMF)

How to cite: Aeman, H., Shu, H., and Nadeem, I.: Integrating Medium Resolution Satellite Data and CNN-RF Machine Learning for Shoreline Dynamics: Assessing Coastal Erosion and Accretion under Climate Change Scenarios, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-207, https://doi.org/10.5194/egusphere-egu25-207, 2025.

The Gulf of Khambhat, a 200 km stretch of coastline in Gujarat, India, is increasingly vulnerable to the impacts of rising sea levels, inundation, and erosion. The region is home to densely populated districts such as Bhavnagar, Surat, Bharuch, and Khambhat, as well as vital ports like Dahej, crucial for global trade and economic growth. However, urbanization, industrialization, and a growing population have placed additional pressure on the region's underground resources, making the soft sediments more prone to subsidence. This, coupled with the environmental pressures from climate change, significantly amplifies the area's vulnerability to coastal hazards. The land use and land cover (LULC) changes between 2017 and 2023 have shown an increase in built-up areas and a decline in vital ecosystems like mangroves. Between 2014 and 2017, approximately 28.66 square kilometers of high tidal mudflats were lost, which not only destroyed critical habitats but also exposed populated areas to tidal flooding. This accelerated erosion further threatens the stability of the coastline. According to the IPCC AR6, the sea level along the Gulf is projected to rise by 0.95 meters by 2100.  Tropical cyclones like Tauktae and Biparjoy, which caused significant damage in the region, may further intensify the risks of storm surges and flooding in the future. The combined effects of sea level rise (SLR), tropical cyclones, and vertical land motion (VLM) may further threaten the region’s biodiversity, health, and food security. In this context, this study aims to examine the combined effects of coastal subsidence and sea level rise on the coastal cities along the Gulf of Khambhat. Given the increasing frequency of cyclones in India, the study also assesses the risks of inundation and flooding due to SLR, storm surges, and land subsidence in the 21st century. The approach integrates scenario-based SLR projections from the IPCC AR6 (ranging from SSP1-1.9 to SSP5-8.5), vertical land motion rates, high-resolution Digital Elevation Models (DEMs), and historical storm surge data. The study uses C-band Sentinel-1 satellite data (92 SAR images) from March 2020 to June 2023, processed through the GMTSAR software with an advanced Small Baseline Subset (SBAS)-based Multi-Temporal Interferometric Synthetic Aperture Radar (MT-InSAR) technique. The analysis reveals a subsidence rate of over 5 mm/year in various areas of the Gulf, particularly in locations like Palsana, Volvad, Navetha, and Bhadbhut. Furthermore, the results suggest that if sea level rise continues as projected by the IPCC and if the subsidence rate persists, the inundated area will increase by approximately 1.57% by 2030, 4.70% by 2050, and 18.20% by 2100 under the worst-case scenario (SSP5-8.5). Additionally, a cyclone similar to Tauktae, with the worst 4-meter storm surge height, could further impact over 1,000 square kilometers of the Gulf region under the same scenario. Given these alarming projections, it is essential to develop comprehensive emergency response plans for flood-related disasters to mitigate the growing risks and protect both the environment and local communities.

How to cite: Sharma, S. and Ojha, C.: Coastal Subsidence and Inundation Risk in the Gulf of Khambhat, India: A Geospatial Perspective, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-385, https://doi.org/10.5194/egusphere-egu25-385, 2025.

Offshore rip currents are among the primary causes of drowning incidents for beachgoers. Maritime radar has demonstrated potential for monitoring rip currents. To investigate the characteristics of rip currents as captured in radar imagery, a radar-based monitoring station was established along the southwestern coast of Taiwan. This station acquires nearshore radar echo images every 20 minutes, and the observational experiments have been ongoing for over six months. Rip current features detected in radar images can be categorized into two types.

The first type is the offshore flow channel (channel rip) occurring within the surf zone. The highly irregular surface structures in the surf zone increase radar wave scattering intensity, resulting in strong electromagnetic echoes in radar imagery. Conversely, wave breaking within the offshore flow channel is often reduced compared to the surrounding areas, leading to weakened radar wave scattering.

The second type is the offshore rip head extending beyond the surf zone. Floating debris on the sea surface, influenced by the rip current, is transported offshore, forming a streak-like region. Compared to clean seawater, these floating materials generate stronger sea surface echoes. Additionally, interactions between the offshore-directed rip current and onshore-directed waves increase sea surface roughness, further enhancing radar backscatter intensity.

To better elucidate the rip current features observed in radar images, we conducted supplementary experiments during the radar monitoring period, including field surveys of bathymetry, aerial photography, and drifter experiments. Cross-validation of these diverse datasets aims to clarify the feasibility of microwave radar for detecting rip currents comprehensively.

How to cite: Wu, L.-C. and Lai, J.-W.: Microwave Radar Detection of Rip Currents: Observations and Characterization from a Coastal Monitoring Station in Southwestern Taiwan, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2112, https://doi.org/10.5194/egusphere-egu25-2112, 2025.

EGU25-2689 | ECS | Orals | ESSI4.10

Assessment of a generalized linear model for satellite-derived bathymetry in turbid waters using Sentinel-2 multi-temporal images 

Tea Isler, Xavier Monteys, Gema Casal, and Colman Gallagher

Climate change is reshaping the world’s coastlines through coupled dynamic processes. The increased importance of monitoring coastal changes over time can be partially addressed using satellite derived bathymetry (SDB), which is more cost effective than traditional methods and allows for monitoring capabilities. In this study we developed a two-step methodology aiming to improve shallow water depth estimates from multi-temporal Sentinel-2 satellite images. The pilot area lies in north-east Ireland in optically complex waters. A threshold criterion was applied to identify 10 suitable Sentinel-2 images over one year time (2021). Lyzenga and Stumpf empirical models were evaluated followed by the application of an empirical generalized linear model (GLM). The performance of atmospherically corrected composite images, created using a reducer function (mean and median), was also evaluated, and compared with the performance of single images. Validation results confirmed the outperformance of the GLM model compared to Lyzenga and Stumpf empirical models. The optimum combination of multi temporal images outperformed the single images regression scores, with a reduction of 45 % in RMSE and a MAE as low as 31 cm in the 0 to 10 m depth. The application of empirical models on the multi-temporal image analysis results in a reduction of error outliers. These results enhance the potential of SDB and Sentinel-2 data in a range of potential coastal monitoring applications, such as repetitive bathymetric changes, ecosystem mapping and environmental management.   

How to cite: Isler, T., Monteys, X., Casal, G., and Gallagher, C.: Assessment of a generalized linear model for satellite-derived bathymetry in turbid waters using Sentinel-2 multi-temporal images, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2689, https://doi.org/10.5194/egusphere-egu25-2689, 2025.

EGU25-2977 | ECS | Posters on site | ESSI4.10

Observation of Sediment Plume Dispersion around Ieodo Ocean Research Station in the East China Sea Using Satellites and UAVs 

Seong-Bin Hwang, Jong-Seok Lee, Sin-Young Kim, and Young-Heon Jo

Sediment plumes in marine environments significantly impact ecosystems by increasing turbidity, depleting bottom-water oxygen, and transporting pollutants. In general, there is a special plume, called Ieodo plume, which is a characteristic bent plume originating from the Ieodo seamount in the northern East China Sea. While satellite-based remote sensing is commonly used to study such phenomena, its spatiotemporal resolution is often insufficient for monitoring rapidly changing marine dynamics. Thus, it is still challenging to understand their specific behavior, dispersion patterns, and range of influence. This study investigates the behavior and dispersion of the Ieodo plume using integrated UAV (Unmanned Aerial Vehicle) and satellite observations. Continuous UAV-based hovering observations were conducted on the Ieodo Ocean Research Station, adjacent to the plume, utilizing optical and multispectral sensors. Optical sensors were employed to monitor flow at the plume's source, while surface currents derived from Optical Flow algorithm were combined with tide and wind data from real-time in situ observations at the research station to estimate plume dispersion range theoretically, using equations derived from plume dynamics. These theoretical predictions were validated against Sentinel-2 optical satellite imagery. Multispectral sensors were used to derive suspended sediment concentration (SSC) information within the plume based on remote sensing reflectance (Rrs). This study provides a comprehensive understanding of the initial characteristics and dispersion of the Ieodo plume based on theoretical and observational analysis. These results are expected to be applicable to predict plume dispersion caused by riverine outflows, seabed resource extraction, and dredging operations, thereby contributing to better management of such marine phenomena.

How to cite: Hwang, S.-B., Lee, J.-S., Kim, S.-Y., and Jo, Y.-H.: Observation of Sediment Plume Dispersion around Ieodo Ocean Research Station in the East China Sea Using Satellites and UAVs, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2977, https://doi.org/10.5194/egusphere-egu25-2977, 2025.

EGU25-3667 | ECS | Orals | ESSI4.10

A time-variable topo-bathymetry from coastal remote sensing observations 

Bene Aschenneller, Roelof Rietbroek, and Daphne van der Wal

Sea level rise is commonly associated with retreating shorelines. However, shoreline evolution is the result of the complex interaction between several groups of processes: Changes in inundation from changing water levels, vertical land motion and morphodynamics. Our goal is to quantify and separate the influence of these processes on the shoreline geometry by using remote sensing data. In a case study for the barrier island of Terschelling (the Netherlands), we found that between 1992 and 2022 morphodynamics had the largest impact on shoreline changes: Inundation by sea level rise, corrected for vertical land motion, accounted for -0.3 m/year shoreline retreat on average while the total average shoreline trend was -3.2 m/year.

These results are very site-specific and cannot be easily transferred to other places. The main limitation for upscaling this method lies in the availability of land elevation data. Local high-quality elevation datasets from airborne LiDAR or ship-based bathymetry are ideal but usually limited to countries which invest in regular observations campaigns. On the other hand, global Digital Elevation Models (DEMs) either lack the required vertical accuracy or horizontal resolution, they often cover only either the topography or the bathymetry, or they mix several data sources resulting in a mean elevation model spanning time periods of several years to decades.

Here we present a technique to derive a time-variable elevation grid that 1) can be applied globally, 2) has a high temporal resolution, 3) covers the intertidal area around the shoreline (foreshore and upper shoreface), and 4) has sufficient vertical accuracy and horizontal resolution. Additionally, we will address the question which accuracies are considered "sufficient" for certain problems.

To create such a time-variable topo-bathymetry model with yearly resolution for the years 1993-present, we combine existing global DEMs (e.g. DeltaDTM or CoastalDEM) with satellite remote sensing observations in a Kalman filter scheme. The observations are yearly 2.5D point clouds (x,y,h) of the intertidal zone that we generate by assigning sea surface heights from coastal altimetry to shoreline contours from optical remote sensing ("waterline method"). First, we incrementally update the global DEMs with these point clouds in a forward Kalman filter. Then, we use a backward smoother to derive the final elevation grid that best represents the topo-bathymetry at one point in time.

For validation, we apply this technique to sandy beaches in the Netherlands, Duck (USA) and Narrabeen (Australia), where high-quality elevation dataset are available. We hope to find that this method increases the accuracy of global DEMs and allows us to study temporal variations in coastal morphology and the role of sea level rise in data-sparse regions worldwide.

How to cite: Aschenneller, B., Rietbroek, R., and van der Wal, D.: A time-variable topo-bathymetry from coastal remote sensing observations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3667, https://doi.org/10.5194/egusphere-egu25-3667, 2025.

EGU25-5013 | Posters on site | ESSI4.10

An optimal approach for morphological changes of tidal flat using multi-satellite sensors 

Keunyong Kim, Jingyo Lee, Geun-Ho Kwak, and Joo Hyung

The tidal flats are important in an ecological and economical way, and continuous management is required because it shows very dynamic changes. In the Korean tidal flats, efficient management is more emphasized because the heavily economically active. In this study, we propose a way to create a topographic elevation and area theme map of the tidal flat using multiple satellites and to update it efficiently. The waterline method was used to generate a satellite based digital elevation model (DEM), and the topographic elevation was calculated using the tidal information at the time the satellite image was acquired. The exposure frequency of tidal flats was calculated through time-series images and compared and verified with the unmanned aerial vehicle-based DEM to present an optimal topographic elevation map generation method. For the satellite-based tidal flat area theme map, the tidal flat was classified using the supervised classification method, and compared and verified with the tidal flat area data provided by the Ministry of Oceans and Fisheries. The multi-satellite-based DEM of tidal flat could produce a precise theme with an error of about 21 cm with only 5 months of image collection, and even if the image collection period was longer, the accuracy was not significantly improved. In the case of the satellite-based tidal flat area, the accuracy was about 95% compared to the reference data, and it was analyzed that the tidal flat area, which was missing some surveys, could also be detected. Through the results of this study, it was confirmed that the satellite-based topographic elevation and area map production method can drastically shorten the update cycle while maintaining a level of accuracy similar to the current survey method.

How to cite: Kim, K., Lee, J., Kwak, G.-H., and Hyung, J.: An optimal approach for morphological changes of tidal flat using multi-satellite sensors, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5013, https://doi.org/10.5194/egusphere-egu25-5013, 2025.

EGU25-5796 | Posters on site | ESSI4.10

Integrated satellite and drone-based multispectral analysis for 40-Year shoreline reconstruction on the Southern Latium Coast 

Francesco Troiani, Giulia Iacobucci, Davide Torre, and Daniela Piacentini

Coastal zones are widely recognized as among the most dynamic and sensitive geomorphological systems, particularly in response to weather and climate conditions. Coastal erosion and deposition alternate cyclically, influenced by fluvial sediment transport, wave and tidal regimes, sea level rise, tectonics, coastal flooding, and anthropogenic pressures. With approximately 2.15 billion people residing in coastal areas - and projections indicating significant population growth in these zones - understanding shoreline morphodynamics is essential for cost-effective and sustainable management strategies.

Within the framework of Italy’s National Recovery and Resilience Plan (PNRR), funded by Next Generation EU, the extended partnership RETURN (multi-Risk sciEnce for resilienT commUnities undeR a changiNg climate) aims to strengthen research on environmental, natural, and anthropogenic risks associated with climate change. Specifically, the Diagonal Spoke (DS) 8, Science underpinning climate services for risk mitigation and adaptation, focuses on developing innovative models to forecast atmospheric, hydrological, and marine impact-oriented indicators, alongside assessing their uncertainties. In this context, shoreline position and morphology emerge as critical indicators for assessing the impacts of climate change on coastal regions. Italian coastline spans approximately 7,500 km, of which 943 km are currently eroding, and 970 km are prograding (ISPRA, 2023), based on comparisons of shorelines between 2006 and 2020. Reconstructing coastal dynamics in specific study areas is therefore pivotal for effective land management and provides a valuable tool for government agencies and stakeholders.

The southern coastal area of the Latium region (Central Italy) represents an ideal case study for investigating shoreline morphodynamics, with a coastline approximately 30 km long. This study utilizes multispectral and multi-mission satellite imagery from Landsat 4, 5, 8, and Sentinel-2, offering an unparalleled dataset for reconstructing coastal changes. The primary objectives of the research are: i) annual reconstruction of the instantaneous waterline, and ii) identification of erosional and depositional sectors with quantified rates. Using the Normalized Difference Water Index (NDWI), 40 instantaneous shorelines were reconstructed for the summer season from 1984 to 2024. The application of the Digital Shoreline Analysis System (DSAS) developed by the USGS revealed maximum shoreline regression rates of approximately 1 m/yr (1.07 m/yr and 1.2 m/yr, respectively LRR and WLR). Additionally, in winter 2024/2025 drone survey, conducted using a Matrice 350 RTK equipped with a multispectral MicaSense RedEdge-P camera, were integrated into the methodology to provide high-resolution and spatially detailed data on shoreline position and morphology, enhancing the accuracy of the reconstructed coastal dynamics and complementing the satellite-based analyses. Finally, the accuracy of the reconstructed shorelines was validated by comparing satellite-derived shorelines from 1998, 2005, and 2019 with ISPRA’s orthophoto-derived shorelines. The results demonstrate strong agreement, with RMSE of 14.44 m, 12.60 m, and 5.83 (1998, 2005 and 2019, respectively), falling within the uncertainty range of Landsat and Sentinel imagery. This study highlights the potential of multi-sensor remote sensing surveys and geospatial techniques in monitoring coastal dynamics, providing critical insights for climate adaptation and risk mitigation strategies in coastal regions.

How to cite: Troiani, F., Iacobucci, G., Torre, D., and Piacentini, D.: Integrated satellite and drone-based multispectral analysis for 40-Year shoreline reconstruction on the Southern Latium Coast, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5796, https://doi.org/10.5194/egusphere-egu25-5796, 2025.

EGU25-6098 | ECS | Posters on site | ESSI4.10

Satellite-derived shoreline evolution in Inhambane province (Mozambique) using high-resolution imagery 

Eva Pavo-Fernández, Carlos Loureiro, Gorka Solana, Vicente Gracia, Abrange Mavimbele, and Manel Grifoll

Satellite imagery is crucial for studying shoreline evolution due to its ability to provide consistent, high-resolution, and large-scale data over time (Gomes da Silva et al., 2024); and it plays a crucial role in countries with limited coastal information sources. The aim of this study is to explore the use of satellite imagery to investigate shoreline evolution at three different locations in Mozambique: Morrungulo Beach, Barra Beach, and Tofo Beach in Inhambane Province. These three locations are characterized as exposed and mesotidal beaches and were selected as representative of the typical coastal archetype in the south coast of Mozambique. This study uses satellite imagery provided by Planet Labs, which is explored with the open-source code CoastSat.PlanetScope toolkit to map and analyze in detail shoreline changes in the study sites (Doherty et al., 2022). PlanetScope satellite imagery has a spatial resolution of approximately 3 meters and almost daily temporal resolution, allowing for detailed observation of shoreline features. For the automated extraction of the shoreline, CoastSat.PlanetScope takes into account the beach slope and tide to provide shoreline positions along user-defined transects, determined using a water index and pixel thresholding. The temporal scope of the satellite imagery utilized in this study extends from July 2016 to June 2024, using one image per month, offering a comprehensive dataset for examining monthly to multiannual shoreline dynamics. Shoreline positions have been evaluated using data total of 101 shorelines for Barra Beach, 94 shorelines for Morrungulo Beach, and 108 shorelines for Tofo Beach. Through this analysis, it was also possible to determine the shoreline impacts of tropical cyclones that made landfall in the region. Barra Beach revealed a strong erosion rate of 3.7 m/year as calculated using the End Point Rate (EPR) method, which measures the net shoreline change over time, and a moderate erosion rate of 0.5 m/year based on the Linear Regression (LR) method, suggesting relative stability in shoreline position when more shoreline positions are considered. Morrungulo Beach presented an accretion rate of 1.7 m/yr based on EPR, but evidenced an erosion rate of 0.4 m/yr with LR. Tofo Beach presented more consistent erosion, with a rate of 1.8 m/yr for EPR and 0.7 m/yr for LR. The analysis of shoreline changes across the three selected beaches in Mozambique highlights distinct patterns of erosion and accretion over the study period. Barra Beach demonstrated considerable differences in erosion rate according to the method, while Morrungulo Beach exhibited a mix of accretion and minor erosion, depending on the analysis method used. Conversely, Tofo Beach showed consistent erosion. These findings highlight the need to carefully consider shoreline change metrics, selecting those that better represent the coastal processes of interest to ensure site-specific management strategies along Mozambique’s coastline. This study has been funded by DOORS project (H2020 – 101000518 – DOORS), and co-funded by the FI AGAUR grant (2022 FI_B 00897).

How to cite: Pavo-Fernández, E., Loureiro, C., Solana, G., Gracia, V., Mavimbele, A., and Grifoll, M.: Satellite-derived shoreline evolution in Inhambane province (Mozambique) using high-resolution imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6098, https://doi.org/10.5194/egusphere-egu25-6098, 2025.

Continuous sampling and analysis of data from the Atlantic sector of the Southern Ocean is key to monitoring rapid, stochastic ecosystem changes in the region. Antarctic krill (hereafter krill), the subject of this research, is particularly prone to regional warming, with a southward contraction of its habitat forecast. While coastal regions of the Antarctic Peninsula, South Orkney Islands and South Georgia are relatively well surveyed using trawl and acoustic survey methods, offshore environments represent significant data gaps. Many important stages in krill life cycle take place offshore including spawning and grazing, and moreover transit between the Antarctic Peninsula and South Georgia entails over 200+ days of advective oceanic (mainly passive) transport with Antarctic Circumpolar Current and its associated fronts. To fill in this significant data gap and infer patterns in temporal and spatial offshore distribution patterns necessitates the integration of diverse data sources, including primary historical observations of krill abundance from surveys and fishing activity as well as secondary observations from remotely sensed environmental variables.

Ideally, we could detect krill directly using hyperspectral imaging to measure the concentration of astaxanthin pigments in surface waters (Basedow et al. 2019). However, given such methods are still in development, we utilize Species distribution models (SDM) to infer spatiotemporal krill distributions. SDMs are models that relate abundance/ occurrence of species with environmental data for a given set of sample locations (Elith and Leathwick 2009).  In this research we use multivariable regression methods to build SDMs to predict krill abundance in relation to both static (geographical area, bathymetry) and dynamic (SST) environmental features, and numerical density of krill as a target variable. We explore the accuracy of several nonlinear methods including Random Forest and Boosted Regression Trees, through comparisons of model accuracy (R2 values, standardized RMSE values and so on) and cross-validation. We then compare these predictions to eddy statistics calculated from satellite altimetry data, and phytoplankton concentrations derived from ocean colour data, with both products accessed through the Copernicus Marine Service. In this way, we will use SDMs for spatiotemporal predictions and use these mapped predictions to explain important relationships e.g. krill density as a function of eddy size.

References:

Elith, Jane, and John R. Leathwick. 2009. Species Distribution Models: Ecological Explanation and Prediction Across Space and Time. Annual Review of Ecology, Evolution, and Systematics 40 (1): 677–97. doi: https://doi.org/10.1146/annurev.ecolsys.110308.120159

Basedow, S.L., McKee, D., Lefering, I. et al. Remote sensing of zooplankton swarms. Sci Rep 9, 686 (2019). doi: https://doi.org/10.1038/s41598-018-37129-x

How to cite: Kelly, C., Daae, R., Ellingsen, I., and Omholt Alver, M.: Filling Data Gaps in the Southern Ocean: Fusion of Remote Sensing Observations with Historic Krill Data to Explain Coastal and Offshore Variability in Krill Abundance, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6296, https://doi.org/10.5194/egusphere-egu25-6296, 2025.

EGU25-6763 | ECS | Orals | ESSI4.10

Inconsistency of resolution in bathymetry mapping may lead to misconception of coastal resilience to climate change 

Bo Miao, Peter Arlinghaus, Corinna Schrum, and Wenyan Zhang

As science and technology continue to advance, the accuracy of coastal and ocean bathymetry mapping continues to improve. Bathymetric mapping of coastal zones usually integrates products from multiple instruments for optical sensing (satellite, LiDAR) and/or acoustic sensing (single beam, multibeam and sidescan sonars) that are of varying accuracy and spatial resolution. Merging of these data from different sources may lead to spatial and temporal inconsistencies in the joint bathymetric data and inhibits their use for reliable assessment of coastal resilience to climate change such as sea level rise. This particularly requires caution since the rate of sea level change is typically on the order of a few mm yr-1, which is much smaller than the accuracy of bathymetric data, e.g. the accuracy ranges from the order of a few cm for LiDAR and multibeam eco sounding data to a few tens of cm for satellite data. In this study, we first demonstrate a problem, which is often overlooked in existing literature, in using coastal bathymetric data derived from state-of-the-art techniques for assessing coastal resilience to sea level rise. Using the Germen Wadden Sea as example, we found that the inconsistency of spatial resolution in the bathymetry mapping, when merged into a uniform gridded dataset, could result in a false trend in the change of the mean elevation of tidal basins, leading to a misconception of coastal resilience to sea level rise. We developed an analytical method to identify inconsistency in gridded bathymetry dataset that can be applied worldwide. Based on the identified inconsistency, we propose two solutions to minimise the associated effect. Our methods are broadly applicable to reduce the error in coastal bathymetry mapping and improve quantitative assessment of coastal resilience to climate change.

How to cite: Miao, B., Arlinghaus, P., Schrum, C., and Zhang, W.: Inconsistency of resolution in bathymetry mapping may lead to misconception of coastal resilience to climate change, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6763, https://doi.org/10.5194/egusphere-egu25-6763, 2025.

EGU25-8104 | ECS | Posters on site | ESSI4.10

Assessing Sand Deposit Dynamics at the Island of Langeoog, Germany by Means of Multi-Sensor Remote Sensing Data 

Julia Holzner, Sandro Martinis, and Simon Plank

Langeoog is one of the East Frisian Islands located in the German Wadden Sea. The island's coastal morphodynamics are strongly influenced by environmental factors such as tides and currents. As natural sediment supplies cannot always compensate for coastal erosion of sand deposits, coastal protection measures are crucial for preserving the island. Langeoog is a special case in this context, as it is the only island in the barrier island chain that is mostly managed without the use of coastal protection structures such as groynes or revetments, due to its prevailing current and sedimentation processes. The island’s settlement and infrastructure are surrounded by a protective dune and adjacent sandy beach areas to the west and north. Conservation measures to preserve the dune are only necessary in the northern part, in front of the Pirolatal, where sand replenishments are carried out regularly to counteract ongoing beach erosion by restocking the sand deposits in front of the protective dune. To initiate necessary measures and estimate the required sediment volumes, knowledge about the development of this beach section is essential for local coastal protection authorities.

In this study, we investigate the suitability of optical multi-sensor remote sensing data to analyse changes in the sand deposits and their effects on the condition of the protective dune in front of Pirolatal on Langeoog Island from 2018 to 2023. For this purpose, we processed high-resolution (HR) and medium-resolution (MR) optical satellite data, applying index-based threshold methods to estimate several proxies of coastal dynamics, such as the instantaneous waterline, the location and state of the protective dune, and the extent of permanently dry sand areas under regular tidal conditions. We compare the results to elevation data to assess the potential of 2D remote sensing data for monitoring this coastal section. The results show that the state of the beach and the height of the dune’s break-off are strongly influenced by accretion events (sand replenishments) and ongoing erosion, particularly during storm surges in the winter season. The condition of the sand deposit is also crucial for determining the position of the instantaneous waterline.

This study demonstrates the benefit of a multi-sensor optical satellite data approach to support coastal monitoring and applied coastal protection efforts.

How to cite: Holzner, J., Martinis, S., and Plank, S.: Assessing Sand Deposit Dynamics at the Island of Langeoog, Germany by Means of Multi-Sensor Remote Sensing Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8104, https://doi.org/10.5194/egusphere-egu25-8104, 2025.

EGU25-8120 | ECS | Posters on site | ESSI4.10

Comparison of Aquaculture Facilities with Sentinel-1 Data for Change Detection 

Yunjee Kim and HyunSoo Choi

Aquaculture facilities are not only an important component of fisheries and local economies, but they also have significant economic and environmental impacts that require efficient and sustainable management. Additionally, knowing the exact locations of aquaculture facilities is crucial for operating ships in nearshore areas, as their presence significantly affects the safe navigation of vessels. However, current aquaculture facility data suffer from slow update cycles due to reliance on field surveys and data processing, and their low spatial resolution does not meet the accuracy requirements in the field. To address this issue, this study takes a first step toward updating aquaculture facility data in quasi-real time using satellite imagery. Specifically, we evaluated the spatial agreement between detected aquaculture facility data and existing data based on Sentinel-1 satellite imagery. While many previous studies on aquaculture facility detection have utilized optical satellites, this study aims to verify the detectability of aquaculture facilities using SAR (Synthetic Aperture Radar) imagery, which can be acquired regardless of weather conditions or time of day. The aquaculture facility data provided by the Korea Hydrographic and Oceanographic Agency is available in both polygon and point formats, with the last update date being December 19, 2024. Accordingly, we analyzed Sentinel-1 data acquired around the same time (December 20, 2024) and compared it with the polygon data. Our analysis revealed significant discrepancies between the two datasets. These findings highlight the need to update current aquaculture facility data and suggest that satellite imagery, with its ability to regularly cover broad areas, could be employed to improve the accuracy and timeliness of aquaculture data updates. This confirms the potential value and utility of satellite imagery as an effective tool for managing aquaculture facilities.

How to cite: Kim, Y. and Choi, H.: Comparison of Aquaculture Facilities with Sentinel-1 Data for Change Detection, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8120, https://doi.org/10.5194/egusphere-egu25-8120, 2025.

Shallow water bathymetry is vital for understanding coastal ecosystems, managing marine resources, and monitoring environmental changes. However, global mapping remains challenging due to the limited penetration of optical and near-infrared light in water, which is rapidly absorbed and scattered by suspended particles and water molecules. Other electromagnetic frequencies, such as microwaves, do not penetrate deeply enough, rendering photogrammetric methods ineffective for underwater mapping.

High-accuracy methods like airborne LiDAR, sonar, and ICESat-2 (a spaceborne altimetric LiDAR) provide detailed bathymetric measurements but are limited by sparse spatial coverage and infrequent revisits. This study combines the strengths of airborne LiDAR and ICESat-2 data to train Machine Learning models for bathymetry extraction from Sentinel-2 multispectral imagery. Sentinel-2 offers global coverage, 10-meter resolution, and a ~5-day revisit cycle, presenting a scalable solution for large-scale mapping. Atmospheric corrections were applied to Sentinel-2 data, and ICESat-2 data were adjusted for tidal and refraction effects. Using Machine Learning models, we evaluate whether smaller ICESat-2-derived training datasets can achieve comparable accuracy to those trained on airborne LiDAR data, which provide a more comprehensive depth range.

In the past, correlations between the logarithm of Sentinel-2 blue-green band ratios versus depth has been widely used in bathymetric studies. We seek to improve prediction accuracy from optical imagery by incorporating other nonlinear relationships and leveraging additional spectral bands, allowing for more robust modeling across varying environmental and water conditions.

Our research underscores the complementary strengths and limitations of ICESat-2 and airborne LiDAR for bathymetric modeling and highlights the potential of Sentinel-2 for global, repeatable bathymetry. Achieving accurate and frequent mapping could revolutionize coastal monitoring, enabling applications such as disaster impact assessments and change detection after events like oceanic landslides, volcanic eruptions or earthquakes.

Keywords: Shallow water bathymetry, ICESat-2 ATL03, Airborne LiDAR, Sentinel-2, Random Forest, Coastal mapping

How to cite: Hsu, H. J. and Moortgat, J.: Enhancing Shallow Water Bathymetry Using Machine Learning with ICESat-2, Airborne LiDAR, and Sentinel-2 Imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10185, https://doi.org/10.5194/egusphere-egu25-10185, 2025.

EGU25-10452 | ECS | Orals | ESSI4.10

Machine Learning Approaches for Tidal Data Interpolation in Satellite-Derived Bathymetry Applications 

Mario Luiz Mascagni, Antonio Henrique da Fontoura Klein, Anita Maria da Rocha Fernandes, Dennis Kerr Coelho, Andrigo Borba dos Santos, and Laís Pool

Satellite-derived bathymetry (SDB) has been developed since the 1970s and has grown exponentially with the popularization of remote sensing technologies. Over the years, several authors have proposed various methods to perform bathymetric inversion from the information contained in the pixels of satellite images, aiming to improve the accuracy and reliability of non-direct methods for estimating depth data in shallow waters. 

Despite the potential of remote sensing-based algorithms and global models to monitor multiple parameters of the planet's surface, few studies correlate SDB with water level in satellite images, obtained for the same region under different tidal conditions. Most recent efforts are limited to cluster analyses, separating the images into high-tide and low-tide groups to perform SDB with empirical models in a segmented approach, adjusting the linear coefficients of the regression models, partly for high-tide conditions and partly for low-tide conditions. The present study seeks to integrate tidal variation data with SDB techniques through Machine Learning (ML), particularly through the input channels of a Convolutional Neural Network (CNN). 

Previous research employing a simpler ML model, the Multi-Layer Perceptron (MLP), in Babitonga Bay, a microtidal region situated along the southern coast of Santa Catarina, Brazil, was compared to empirical SDB models that rely on the linear interaction of electromagnetic spectrum bands with the water column. The findings demonstrated that the nonlinear inferences generated by deep neural networks can enhance the accuracy of SDB data by more than 100% in optically complex environments, influenced by high concentrations of Colored Dissolved Organic Matter (CDOM) and Suspended Particulate Matter (SPM), such as Babitonga Bay. 

The application of more complex neural networks, such as CNN combined with additional input layers incorporating tidal data, has great potential for enhancing the performance of SDB, since CNN models utilize kernels that analyze multiple pixels surrounding a target point, enabling a more robust and context-aware approach, unlike MLP models, which infer depth on a pixel-by-pixel basis. The introduction of tide level variables as input channels in these deep learning neural networks makes these models suitable for universal application across micro-, meso-, and macrotidal environments. 

The CNN model applied to Babitonga Bay yielded substantial improvements in SDB accuracy, reducing the mean absolute error (MAE) from 2.9 m (traditional SDB methods) and 1.3 m (MLP) to 0.1 m. These results were obtained using field data collected in 2018 through single-beam echo sounder surveys for training, testing, and validation for both cases, the traditional empirical SBD models, and the machine learning models (MLP and CNN).

How to cite: Mascagni, M. L., Klein, A. H. D. F., Fernandes, A. M. D. R., Coelho, D. K., Santos, A. B. D., and Pool, L.: Machine Learning Approaches for Tidal Data Interpolation in Satellite-Derived Bathymetry Applications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10452, https://doi.org/10.5194/egusphere-egu25-10452, 2025.

EGU25-10754 | ECS | Orals | ESSI4.10

A new method for evaluating satellite-derived waterline detection in macrotidal beaches with complex intertidal morphology 

Mª Carmen Millán Roldán, Albert Gallego Jiménez, Paula Gomes da Silva, Josep E. Pardo-Pascual, Jesús Palomar, Carlos Cabezas-Rabadán, Erica Pellón, and Jara Martínez Sánchez

Satellite-Derived Waterlines (SDWs) have become highly valuable assets in coastal studies due to their extensive data availability, offering temporal and spatial resolutions of up to 5 days and 10 m, respectively. Numerous tools for SDW extraction have been developed, being widely used in microtidal beaches with high reliability. However, macrotidal environments present significant challenges due to their large intertidal extensions. The dynamic nature and complex morphology of these areas frequently lead to inaccuracies in shoreline detection by existing tools. Furthermore, the large volume of SDWs makes identifying errors challenging. Understanding misdetection conditions is key to automating error flagging and improving efficiency.

This study aims at improving the understanding of waterline (mis)detection by identifying the environmental conditions that influence incorrect identification of the sand-water interface. SDWs extracted using the SHOREX tool, developed by the Geo-Environmental Cartography and Remote Sensing Group from the Universitat Politècnica de València, were analyzed in Salinas (144 SDWs) and El Puntal (141 SDWs), two macrotidal beaches in northern Spain with complex intertidal topography.

The analysis was undertaken with the aim of taking a step back to understand what beach features are identified as waterline by currently available tools. This involved a detailed visual inspection of SDWs compared to their corresponding RGB imagery, conducted by a validated operator. The beaches were discretized into equally spaced transects, and for each SDW, the operator classified the detected feature in each transect as one of the following: Waterline (sand-water interface), Maximum High Tide Level (dry-wet sand interface), Intertidal Water (boundary of accumulated water in the intertidal zone), Intertidal Morphological Features (dry-wet sand interface due to intertidal bars), Backshore elements, or Clouds. Three analyses were derived: (1) the percentage of transects classified as each indicator per SDW, (2) the confidence level perceived by the operator for each indicator, and (3) the correlation between met-oceanic variables (e.g., wave height, peak period, storm surge, astronomical tide, and tidal stage) and the percentage of Waterline identification per SDW.

The results revealed a strong positive correlation (R=0.56) between the percentage of transects classified as waterline (ideal identification) and a variable combining tidal level and phase (flood/ebb). Better detections during high tides likely occurred due to drier intertidal sand, while wet sand during ebb tides led to detection problems. However, a lack of representation of the highest tidal states was observed in the satellite time series. Wave parameters (Hs and Tp) showed weaker inverse correlations to the percentage of waterline detection (R=−0.15 and −0.28, respectively), likely due to increased sand saturation during the rundown phase of energetic waves. High vertex count correlated positively with waterline identification (R=0.51), indicating improved detection with greater shoreline detail, while strong negative correlation with SDW sinuosity (R=−0.51) suggested misdetection due to complex intertidal features.

This new approach advances understanding of SDW detection in macrotidal beaches, paving the way for improving detection methodologies. Ongoing work includes assessing additional SDW detection tools, extending analyses to diverse beach types (depending on hydro-morphodynamic conditions), and developing methods for automatic error flagging in each environment.

How to cite: Millán Roldán, M. C., Gallego Jiménez, A., Gomes da Silva, P., Pardo-Pascual, J. E., Palomar, J., Cabezas-Rabadán, C., Pellón, E., and Martínez Sánchez, J.: A new method for evaluating satellite-derived waterline detection in macrotidal beaches with complex intertidal morphology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10754, https://doi.org/10.5194/egusphere-egu25-10754, 2025.

Satellite-derived bathymetry is an essential tool for mapping shallow coastal areas with complex reef structures where ship-based surveys are either unsafe or inefficient. The focus of this study is to present a baseline study of the seabed morphology of a coastal region (0-130 m water depth) to support before-and-after analyses of an artificial reef deployment site under the OCEAN CITIZEN project.

OCEAN CITIZEN aims to develop a sustainable and innovative protocol for coastal restoration and biodiversity conservation. Adapted to specific ecozones, yet replicable across regions, this protocol emphasises the expansion of Marine Protected Areas (MPAs), the creation of ecological corridors to support ecosystem interactions, the restoration of biodiversity, the enhancement of blue carbon through innovative techniques, and the establishment of self-sustaining economic models for long-term sustainability.

In this context, the study investigates seafloor morphology using optical and acoustic methods to assess the suitability of satellite-derived data for mapping coastal habitats. Satellite data from Maxar's WorldView-2 satellite sensor were compared with ship-based data acquired with a Norbit iWBMS multibeam system (190 - 400 kHz) to assess their performance in water depths ranging from 10 to 20 m. Ground truthing was carried out using underwater video surveys to validate substrate, classification and biological observations.

Preliminary results show that satellite-derived data effectively capture broad-scale seafloor morphology, with contours closely matching multibeam data. However, small-scale and complex reef structures could only be resolved by ship-based surveys. Comparisons of seafloor reflectance (optical) and backscatter (acoustic) showed different sensitivities: as expected, backscatter distinguished sandy areas between hard outcropping substrate, whereas satellite reflectance is also sensitive to variations in substrate brightness. These differences highlight the need to be aware of the complementary nature of the two methods and their potential to provide additional insight into coastal restoration planning.

How to cite: Schönke, M., Jensen, M., Lohrberg, A., Feldens, P., and Schneider von Deimling, J.: Comparison between satellite derived and ship based seafloor characteristic's in areas with complex seafloor morphology - A Baseline study for artificial reef deployment, Tenerife Island, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12115, https://doi.org/10.5194/egusphere-egu25-12115, 2025.

EGU25-13317 | Orals | ESSI4.10

Use of high resolution multispectral Unmanned Aerial vehicle (UAV) imagery to retrieval nearshore bathymetry using photogrammetric and spectral techniques in the Cantabrian Sea in Spain. 

Javier Sánchez-Espeso, Gabriel Bellido, Ignacio García-Utrilla, Etienne Faugére, Beatriz Pérez-Díaz, Mirian Jiménez, and Sonia Castanedo

The scientific, economic and social interest in coastal environments requires the continuous and sufficiently accurate determination of bathymetries of these areas, in particular of estuaries and beaches, at least at shallow depths of around 20 m.

The usual techniques for their determination, which combine the use of space geodesy (GNSS) and multibeam echosounder techniques, are a very accurate methodology, but also very costly in economic terms. The methodologies associated with the use of UAVs, which employ a wide variety of sensors, from RGB optical, thermal or multispectral cameras to Lidar Detection and ranging (LiDAR), are providing spatial information and images of high metric and thematic quality, with very short capture times for large extensions, at significantly lower costs than previous techniques.

Typically, photogrammetric techniques, in particular Structure for Motion (Sfm), have been used for the orientation process of the photogrammetric model, which obtain successful results in emerged areas, but have many limitations, or are simply impossible to apply, in the determination of bathymetries, due fundamentally to two aspects, key to the conventional photogrammetric process. Firstly, due to the difficulty in identifying homologous points that allow Bundle Block Adjustment, due to air-water refraction, and secondly, and equally important, due to the reflective behavior of the sea surface itself.

To overcome the indicated barriers, the first results obtained using a multispectral sensor with 10 bands on board a UAV, ranging from 444 to 842 nanometres (nm), highlighting 4 bands in the blue and green ranges (444, 475, 531 and 560 nm), are presented. With the images obtained, and by applying spectral techniques used in satellite-derived bathymetry, previously normalized radiometrically and mosaicked to the sea surface, we have proceeded to determine the sea depth, in different conditions of turbidity and clarity that can be considered globally unfavorable and that characterize the Cantabrian Sea in the North of Spain.

Essel, B.; Bolger, M.; McDonald, J.; Cahalane, C. Developing a Theoretical Assessment Method for an Assisted Direct Georeferencing Approach to Improve Accuracy when Mapping over Water: The Concept, Potential and Limitations. In Proceedings of the ISPRS 12th International Symposium on Mobile Mapping Technology (MMT), Padua, Italy, 24–26 May 2023.

Román, A.; Heredia, S.; Windle, A.E.; Tovar-Sánchez, A.; Navarro, G. Enhancing Georeferencing and Mosaicking Techniques over Water Surfaces with High-Resolution Unmanned Aerial Vehicle (UAV) Imagery. Remote Sens. 2024, 16, 290. https://doi.org/10.3390/rs16020290.

Windle, A.E.; Silsbe, G.M. Evaluation of Unoccupied Aircraft System (UAS) Remote Sensing Reflectance Retrievals for Water Quality Monitoring in Coastal Waters. Front. Environ. Sci. 2021, 9, 674247.

How to cite: Sánchez-Espeso, J., Bellido, G., García-Utrilla, I., Faugére, E., Pérez-Díaz, B., Jiménez, M., and Castanedo, S.: Use of high resolution multispectral Unmanned Aerial vehicle (UAV) imagery to retrieval nearshore bathymetry using photogrammetric and spectral techniques in the Cantabrian Sea in Spain., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13317, https://doi.org/10.5194/egusphere-egu25-13317, 2025.

EGU25-13902 | ECS | Orals | ESSI4.10

Enhanced monitoring of coastal change: a comprehensive validation framework for satellite imagery 

Yeray Castillo Campo, Xavier Monteys, Gema Casal, and Conor Cahalane

Coastal change assessments have significant socioeconomic, environmental, and infrastructure implications due to the extensive impacts of climate change, such as rising sea levels, the increasing frequency and intensity of storms, as well as the consequences of human intervention.  Satellite products have been used to monitor the coast at relatively high resolution (30 m) since the 1970s through the Landsat program. However, the arrival of the EU Copernicus/Sentinels in 2014 introduced a marked increase in coastal applications thanks to the improved spatial and temporal resolutions. The research presented in this study explores the derivation of waterlines from Sentinel 2 imagery and the creation of a novel holistic approach to a validation framework. Specifically, this study aims to: a) explore the inherent waterline errors against reference datasets and begin to establish the overall uncertainty in deriving waterlines from optical satellite imagery; and b) assess the potential of these results and their suitability for coastal change applications. The results indicate an average positional error of approximately 4 meters for Sentinel images in coastal regions by evaluating the Sentinel-2 satellite images with distinct features visible in aerial orthophotography. Subsequently, the horizontal and vertical inaccuracies of the satellite-derived waterlines (SDWL) were further determined by using a GNSS line as a reference dataset. The horizontal assessment was conducted by calculating the average distance between the SDWLs and the GNSS reference lines across eighteen Sentinel-2 images corresponding to the years 2021, 2022, 2023 and 2024. These were analysed, showing a median displacement of 15 meters, and indicating an offshore trend for the satellite-derived waterlines. The vertical assessment, or height error, was computed by comparing the average height of SDWLs (as determined by the average tide gauge heights) with the reference dataset height (as measured by GNSS), resulting in a mean absolute error of 6 cm. The vertical results indicate that the SDWLs’ heights, as measured by the local tide gauges, align well with in situ local height measurements. The results of this study will aid in identifying temporal and spatial scales and resolutions at which Earth Observation products are suitable for coastal management. The initial stages of a validation framework are presented to assess the quality and applicability of satellite-derived waterlines for coastal change monitoring based on specific user requirements. Identifying the sources of error and improving uncertainty models for satellite-derived products enables better decision-making in coastal management. These analyses will demonstrate whether the outcomes remain consistent among satellite images or change according to local environmental conditions. Increasing end-user confidence in the rates of change obtained from available satellite products can provide crucial information in study areas, and at space-time resolutions previously unattainable.

How to cite: Castillo Campo, Y., Monteys, X., Casal, G., and Cahalane, C.: Enhanced monitoring of coastal change: a comprehensive validation framework for satellite imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13902, https://doi.org/10.5194/egusphere-egu25-13902, 2025.

The problem of coastline erosion is of global concern. Acquisition and processing of useful earth observation data about coastal changes is crucial to accurate change monitoring [1]. With the availability of sophisticated machine learning techniques, it is possible to accurately detect and analyze patterns of changes in coastal regions. One important aspect here is the explainability of the machine learning model used to predict changes and the possibility to incorporate human expertise in the process of detection [2]. In this research, we use an explainable artificial intelligence model to discover data patterns in Sentinel-2 time-series images to describe changes over a 7-year study period. Time-series imagery was acquired every month from January 2018 to September 2023, covering 4,694 cloud-free locations along the North Sea and Baltic Sea coastlines, each spanning 5 km x 5 km. These locations were selected using farthest point sampling to ensure representative coverage. The imagery was further divided into smaller scenes of 1.28 km x 1.28 km, and active learning techniques were employed to minimize labeling efforts. We have used Latent Dirichlet Allocation (LDA), a Bayesian generative model recently established as explainable model [1]. Being a probabilistic model, LDA is able to output certainty score for its predictions. We use the LDA as an unsupervised explainable model to create interpretable intermediate visual outcomes that support model explainability, while certainty scores of each prediction enhances trust. These interpretable outcomes are used by the domain expert to assess quality of the outcomes. Two kinds of visualizations are produced: 1) visual topic maps -LDA retrieved visual topics depicting latent data patterns, often perceived by humans as visual objects 2) change class maps and change signature maps - maps showing which land cover classes (e.g wave-breaking zones, dry sand, inter-tidal area, vegetation) have gone through most changes ( we produce histograms showing percentage of change per class per year, and also over the whole study period ); change signatures describe the nature of change in every class.  We conclude the research by validating our results by domain experts.

This work is part of Helmholtz Autocoast project.

Keywords: Explainable AI, Coastal Change Monitoring, Sentinel-2 time-series, Visualizations

 

References:

  • Fejjari, G. Valentino, J. A. Briffa and S. D'Amico, "Detection and Monitoring of Maltese Shoreline Changes using Sentinel-2 Imagery," 2023 IEEE International Workshop on Metrology for the Sea; Learning to Measure Sea Health Parameters (MetroSea), La Valletta, Malta, 2023, pp. 52-56, doi: 10.1109/MetroSea58055.2023.10317486.
  • Karmakar, C. O. Dumitru, G. Schwarz and M. Datcu, "Feature-Free Explainable Data Mining in SAR Images Using Latent Dirichlet Allocation," in IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 14, pp. 676-689, 2021, doi: 10.1109/JSTARS.2020.3039012.
  • Karmakar, C.O. Dumitru, N. Hughes and M. Datcu, "A Visualization Framework for Unsupervised Analysis of Latent Structures in SAR Image Time Series", IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 16, pp. 5355-5373, 2023.

How to cite: karmakar, C.: Explainable Unsupervised Model for Coastline Change Monitoring with Sentinel-2 Time Series, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16180, https://doi.org/10.5194/egusphere-egu25-16180, 2025.

EGU25-17895 | ECS | Posters on site | ESSI4.10

An automated shoreline detection method using PlanetScope satellite imagery  

Alfred Hewetson, James Lawrence, and Ioannis Karmpadakis

Multispectral satellite images can survey the surf zone through discretizing the land-sea interface, at a known water level, to monitor recession and accretion rates along the coastline. This shoreline detection method can be enhanced by utilizing the daily return frequency of PlanetScope data, allowing a higher temporal resolution of the observed shorelines. Similar shoreline detection tools, such as CoastSat(Doherty et al., 2022; Vos et al., 2019), discretize the land-sea interface by thresholding the image using a single index, such as NDWI (normalized difference water index) (McFeeters, 1996) and contouring the image at this threshold. Presented here is an alternative approach. In using several multilayer perceptrons (MLP) acting together, each pixel’s probability of being classed as land or sea is calculated. The final shoreline contour is then probabilistically defined whithout the use of manual threshold. The advantage of this method is that it allows for spatial variability within satellite bands, for regions of shadow and geographical features, to still be correctly discretized. It also allows for further use case beyond just sandy beaches, due to the implementation of multiple indices allowing identification of different classes that could be interfacing with the sea. Characteristically, apart from the usual NDWI and NDVI index, we use the RGB and IR bands as well as 24 further band relationships for a total set of 28 indices to train the MLPs. The root mean squared error (RMSE), the distance between the derived shoreline and a height contour relative to the instantaneous water-level, of this method tested at Seaford UK for cloud cover <90% is ~7m. 

NDWI=Green−IRGreen+IR">NDWI=Green−IRGreen+IRNDWI=Green−IRGreen+IR

 

NDVI=IR−RedIR+Red">NDVI=IR−RedIR+RedNDVI=IR−RedIR+Red

 

Doherty, Y., Harley, M. D., Vos, K., & Splinter, K. D. (2022). A Python toolkit to monitor sandy shoreline change using high-resolution PlanetScope cubesats. Environmental Modelling and Software, 157. https://doi.org/10.1016/J.ENVSOFT.2022.105512 

McFeeters, S. K. (1996). The use of the Normalized Difference Water Index (NDWI) in the delineation of open water features. International Journal of Remote Sensing, 17(7), 1425–1432. https://doi.org/10.1080/01431169608948714 

Vos, K., Splinter, K. D., Harley, M. D., Simmons, J. A., & Turner, I. L. (2019). CoastSat: A Google Earth Engine-enabled Python toolkit to extract shorelines from publicly available satellite imagery. Environmental Modelling & Software, 122, 104528. https://doi.org/10.1016/J.ENVSOFT.2019.104528 

 

How to cite: Hewetson, A., Lawrence, J., and Karmpadakis, I.: An automated shoreline detection method using PlanetScope satellite imagery , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17895, https://doi.org/10.5194/egusphere-egu25-17895, 2025.

EGU25-18173 | Posters on site | ESSI4.10

Coastal benthic habitat monitoring using Copernicus and contributing missions 

Luca Cicala, Marzia Cianflone, Marco De Mizio, and Simonetta Fraschetti

Monitoring of seagrasses and macroalgae is important both for surveillance of habitat conservation and for quantifying the effects of anthropogenic pressure on marine ecosystems. Satellite remote sensing, allows for global and local scale environmental monitoring with low cost and high revisit time. The availability of agency satellites, such as Sentinel-2, and commercial satellites, such as the Planetscope constellation, makes continuous and long-term monitoring of the underwater vegetation possible. However, the interposition between the underwater vegetation and the marine surface of the water column significantly limits the possibility of carrying out satellite monitoring, which is therefore suitable for shallow coastal areas but not for the open sea.  Furthermore, in order to properly and detailly interpret the nature of the monitored vegetation, for example in terms of species, it is necessary to compare satellite data with sea truth.

In this work, some strategies are proposed to delimit areas of marine vegetation and to compare them with the sea truth in order to monitor ecosystems continuously and in the long term, possibly starting from an initial accurate on field assessment. The use of free agency data (with lower spatial resolution) and commercial data (with higher resolution) is combined in such a way as to contain the costs of data acquisition. Furthermore, data obtained from the Copernicus Marine Service are used, together with bathymetry data, to estimate the effects of the water column on the reflectance of underwater vegetation. Multi-temporal analysis approaches are proposed to identify possible changes in vegetation covers that can trigger acquisition campaigns at sea, to directly verify the detected anomalies. The proposed approaches, as mentioned, exploit the availability and large geographical coverage of satellite data, without renouncing the use of (more expensive) resources on field when strictly necessary.

How to cite: Cicala, L., Cianflone, M., De Mizio, M., and Fraschetti, S.: Coastal benthic habitat monitoring using Copernicus and contributing missions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18173, https://doi.org/10.5194/egusphere-egu25-18173, 2025.

EGU25-18784 | ECS | Posters on site | ESSI4.10

UAV monitoring for assessing beach litter pollution and coastal morphodynamics: case studies from the Molise region (central Adriatic coast, Italy) 

Grazia Dilauro, Ludovica Di Renzo, Giorgio Anfuso, Gianluigi Di Paola, Angela Rizzo, and Carmen Maria Rosskopf

Beach litter (BL) poses a significant threat to coastal ecosystems and marine biodiversity. Monitoring its density, composition, and distribution is crucial to develop effective management strategies and support the sustainable use of the beach environments. Traditional in situ visual surveys allow detailed identification and classification of BL items but are often limited in spatial coverage and require significant time investment. Recent advancements, including the use of Unmanned Aerial Vehicles (UAV), have enabled more efficient assessments of coastal litter and related geomorphological features.

This study investigates the characteristics and distribution of BL and its relationship with coastal morphodynamics along two unmanaged beaches located along the Molise coast (southern Italy): Petacciato and Ramitelli beaches. These sites were selected based on their distinct morphodynamic characteristics, and their free beach status.

The monitoring methodology considered international guidelines [1, 2]. UAV surveys were conducted before and after a significant storm event to assess its impact not only on beach morphology but also on BL distribution. Flights were carried out with a Yuneec Typhoon H520 hexacopter using a flight height of 15 meters [2]. High-resolution orthophotos were analysed to manually identify BL larger than 2.5 cm and classify them according to the Joint List for Marine Macrolitter Monitoring [2]. Items were categorized by material type, size, and weight, and unusual objects were documented in detail. Shoreline and dune foot variations along with morphological changes of the beach were also quantified to evaluate the role of coastal processes in BL dispersion and accumulation patterns.

First results reveal significant differences in BL density and composition between the two study sites, but with plastic materials dominating both the collected items, consistent with broader Mediterranean trends [3]. The post-storm survey highlighted the role of weather events in redistributing litter, particularly at the southern limit of the Ramitelli beach which is in contact with the jetty of the Saccione River mouth that drives the accumulation of beach litter on the adjacent shoreline. This study underscores the importance of integrating UAV-based monitoring with geomorphological analyses to better understand the interplay between coastal dynamics and BL distribution. The monitoring of these relationships can provide essential data to support and improve coastal management and design targeted beach clean-up strategies.

 

 

Key Words

Remote Sensing, Coastal monitoring, Coastal geomorphology, UAV images, Visual assessment, Litter beach analysis.

 

[1] Vlachogianni T. (2017) - Methodology for Monitoring Marine Litter on Beaches. Macro-Debris (> 2.5 cm). DeFishGear, 1-16.

[2] Fleet D., Vlachogianni T. & Hanke G. [Eds] (2021) - A joint list of litter categories for marine macrolitter monitoring. JRC Technical Reports, Publications Office of the European Union, Luxembourg, 30348, 52.

[3] Rizzo A., Sozio A., Anfuso G., La Salandra M., Sasso C. (2022) – The use of UAV images to assess preliminary relationships between spatial litter distribution and beach morphodynamic trends: the case study of Torre Guaceto beach (Apulia Region, Southern Italy). Geogr. Fis. Dinam. Quat. 45 (2022). 237-250.

How to cite: Dilauro, G., Di Renzo, L., Anfuso, G., Di Paola, G., Rizzo, A., and Rosskopf, C. M.: UAV monitoring for assessing beach litter pollution and coastal morphodynamics: case studies from the Molise region (central Adriatic coast, Italy), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18784, https://doi.org/10.5194/egusphere-egu25-18784, 2025.

 Surface reflectance of remote sensing datasets contribute to various fields such as natural resources management (Liang et al., 2024), agricultural practices (Liu et al., 2020; Stratoulias et al., 2017), ecological monitoring (Liang et al., 2024), and climate studies (Liu et al., 2020), providing critical information about Earth's surface conditions and resources. Nonetheless, inaccuracies in the raw remote surface reflectance data, resulting from both internal sensor anomalies (Hu et al., 2012) and external atmospheric effects (Dash et al., 2011; Vermote et al., 2016), reveal that correction of these datasets is essential. Moreover, Surface Reflectance datasets of coastal and inland waters are significantly affected by cloud coverage (Wang & Chen, 2024) introducing noise (Qing et al., 2021) into the imagery and shadows. This study introduces a methodology to correct and fill in missing data from multispectral Level 2 Surface Reflectance daily time-series, by identifying logical errors and implementing Principal Component Analysis. The study successfully results in continuous two-decade surface reflectance dataset to assure its reliability and utility across various applications.

References
Dash, P., Walker, N. D., Mishra, D. R., Hu, C., Pinckney, J. L., & D’Sa, E. J. (2011). Estimation of cyanobacterial pigments in a freshwater lake using OCM satellite data. Remote Sensing of Environment, 115(12), 3409–3423. https://doi.org/10.1016/j.rse.2011.08.004
Hu, C., Lee, Z., & Franz, B. (2012). Chlorophyll a algorithms for oligotrophic oceans: A novel approach based on three-band reflectance difference. Journal of Geophysical Research: Oceans, 117(1), 1–25. https://doi.org/10.1029/2011JC007395
Liang, S., Li, Y., Wei, H., Dong, L., Zhang, J., & Xiao, C. (2024). Research on Hyperspectral Surface Reflectance Dataset of Typical Ore Concentration Area in Hami Remote Sensing Test Field. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, 10(1), 137–144. https://doi.org/10.5194/isprs-annals-X-1-2024-137-2024
Liu, J. L., Cheng, F. Y., Munger, W., Jiang, P., Whitby, T. G., Chen, S. Y., Ji, W. W., & Man, X. L. (2020). Precipitation extremes influence patterns and partitioning of evapotranspiration and transpiration in a deciduous boreal larch forest. Agricultural and Forest Meteorology, 287(January), 107936. https://doi.org/10.1016/j.agrformet.2020.107936
Qing, S., Cui, T., Lai, Q., Bao, Y., Diao, R., Yue, Y., & Hao, Y. (2021). Improving remote sensing retrieval of water clarity in complex coastal and inland waters with modified absorption estimation and optical water classification using Sentinel-2 MSI. International Journal of Applied Earth Observation and Geoinformation, 102, 102377. https://doi.org/10.1016/j.jag.2021.102377
Stratoulias, D., Tolpekin, V., de By, R. A., Zurita-Milla, R., Retsios, V., Bijker, W., Hasan, M. A., & Vermote, E. (2017). A workflow for automated satellite image processing: From raw VHSR data to object-based spectral information for smallholder agriculture. Remote Sensing, 9(10). https://doi.org/10.3390/rs9101048
Vermote, E., Justice, C., Claverie, M., & Franch, B. (2016). Preliminary analysis of the performance of the Landsat 8/OLI land surface reflectance product. Remote Sensing of Environment, 185, 46–56. https://doi.org/10.1016/j.rse.2016.04.008
Wang, J., & Chen, X. (2024). A new approach to quantify chlorophyll-a over inland water targets based on multi-source remote sensing data. Science of the Total Environment, 906(July 2023), 167631. https://doi.org/10.1016/j.scitotenv.2023.167631

How to cite: biliani, I. and Zacharias, I.: Satellite Surface Reflectance correction and completion methodology by using Principal Component Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20830, https://doi.org/10.5194/egusphere-egu25-20830, 2025.

EGU25-294 | Orals | ESSI4.11

Enhancing Forest Change Detection Using Self-Supervised Learning with Multi-Source EO Data 

Ridvan Kuzu, Antony Zappacosta, Oleg Antropov, and Octavian Dumitru

This study presents advancements in forest change detection by leveraging self-supervised learning (SSL) methods with multi-source and multi-temporal Earth Observation (EO) data. Transitioning from traditional bi-temporal approaches, the developed methodology incorporates multi-temporal analysis and multimodal data fusion using Sentinel-1, Sentinel-2, and PALSAR-2 imagery. Key innovations include mapping the magnitude of forest changes rather than binary classifications, enabling nuanced assessment of disturbance severity.

Experiments demonstrate the effectiveness of SSL-pretrained backbones, such as ResNet architectures, in extracting features for change detection. The integration of multi-temporal Sentinel-1 time series further improved the reliability and accuracy of disturbance tracking over time. These advancements show the potential of SSL to enhance forest change monitoring, providing scalable solutions for continuous and precise assessment of forest dynamics.

How to cite: Kuzu, R., Zappacosta, A., Antropov, O., and Dumitru, O.: Enhancing Forest Change Detection Using Self-Supervised Learning with Multi-Source EO Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-294, https://doi.org/10.5194/egusphere-egu25-294, 2025.

Forest fragmentation disrupts habitat continuity, reshapes ecosystem processes, and threatens biodiversity. Effective conservation efforts in fragmented landscapes rely on precise monitoring of these changes. This study leverages remote sensing through vegetation indices to evaluate forest health and detect fragmentation-induced alterations over time. Focusing on the Tuchola Forest in Poland, an area increasingly affected by windstorms, we analyzed Sentinel-2 imagery from 2016 to 2024 using 19 vegetation indices. Machine learning classifiers—Extra Trees, Random Forest, and LightGBM—were employed to assess which indices best capture fragmentation stress. The Extra Trees classifier outperformed the others in accuracy and generalization, identifying NDWI and GNDVI as the most effective indicators. These indices were particularly responsive to shifts in vegetation water content and canopy density linked to fragmentation. Our findings underscore the utility of targeted vegetation indices for precise ecological monitoring and inform conservation strategies in fragmented forests.

How to cite: Dutt, S. and Kunz, M.: Uncovering Fragmentation Patterns: Optimal Vegetation Indices for Monitoring the Tuchola Forest Ecosystem, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-353, https://doi.org/10.5194/egusphere-egu25-353, 2025.

EGU25-670 | ECS | Orals | ESSI4.11

Integrating High-Resolution Thermal Mapping and Greenhouse Gas Emission Analysis for Climate Resilience in Urban, Peri-Urban and Rural Areas 

Naji El Beyrouthy, Mario Al Sayah, Rita Der Sarkissian, and Rachid Nedjai

Monitoring urban, peri-urban, and rural temperatures, along with greenhouse gas (GHG) emissions, is crucial for understanding local climate dynamics, especially in rapidly urbanizing areas. This study leverages advanced remote sensing techniques and environmental analysis to enhance high-resolution Land Surface Temperature (LST) mapping. It further investigates the relationship between LST and methane (CH₄) emissions - a significant driver of climate change - and their combined impact on Urban Heat Island (UHI) effects.

Leveraging multispectral atmospherically corrected imagery from LANDSAT 8-9 and SENTINEL-2 satellites, spectral harmonization techniques and Convolutional Neural Network (CNN)-based super-resolution models were applied to improve the spatial resolution and accuracy of LST calculation. These methods are further refined through the integration of key environmental indices, including soil characteristics, land cover, Normalized Difference Vegetation Index (NDVI), Normalized Difference Built-up Index (NDBI), and Normalized Difference Water Index (NDWI), which capture land use characteristics and their impact on thermal variations. The resultant LST at 1m was statistically validated against meteorological datasets by calculating Root Mean Squared Error and Mean Absolute Error, showing errors consistently below 2°C, with 75% of the values within 1°C. Making use of the accurate LST readings, air temperature (Ta) was derived using polynomial regression models, ultimately resulting in LST-derived air temperature maps with R² values exceeding 0.75.

Building upon this high-resolution thermal mapping, the study examines how agricultural zones are influenced by urban thermal dynamics exacerbated by GHG emissions creating a negative feedback loop where increased temperatures further impact agricultural practices and lead to additional GHG emissions. Seasonal and phenological variations in CH₄ emissions from major crops cultivated in the Loiret region including wheat, were analyzed. Results reveal that land use, crop phenology and soil characteristics significantly modulate LST, influencing both the intensity and distribution of urban heat anomalies. Moreover, the thermal contributions of these areas are analyzed within the context of their dual role. On one hand, these areas can act as potential moderators of UHIs by providing vegetative cover and cooling effects. On the other hand, they contribute to regional methane fluxes due to agricultural practices. This dual role highlights the complexity of peri-urban and rural zones, as they can simultaneously alleviate and exacerbate environmental challenges.

The presented framework can be considered as a contribution to bridging the gap between remote sensing advancements and climate science by providing actionable insights into the interactions between urban and rural thermal dynamics. The methodology not only offers a scalable approach for improving LST and Ta monitoring in data-sparse regions but also highlights the implications of land management practices for mitigating urban heat and reducing GHG emissions. By combining cutting-edge data processing techniques with environmental analysis, the study underscores the importance of integrating thermal mapping with greenhouse gas emission assessments to inform sustainable planning and climate adaptation strategies. In conclusion, this study contributes to the broader understanding of urban-rural thermal interdependencies and their role in shaping regional climate resilience, while also aiming to develop a new approach that leverages remote sensing to GHG emissions across wide areas.

How to cite: El Beyrouthy, N., Al Sayah, M., Der Sarkissian, R., and Nedjai, R.: Integrating High-Resolution Thermal Mapping and Greenhouse Gas Emission Analysis for Climate Resilience in Urban, Peri-Urban and Rural Areas, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-670, https://doi.org/10.5194/egusphere-egu25-670, 2025.

EGU25-1883 | ECS | Posters on site | ESSI4.11

Development of a Remote Crop Quality Sensor: Advancing Carotenoid Assessment with Raman Spectroscopy 

Miri Park, Annette Somborn, Dennis Schlehuber, and Volkmar Keuter

The accurate evaluation of crop quality is vital for sustainable agriculture and optimized production. Raman spectroscopy, renowned for its insensitivity to water interference and its ability to deliver molecular-specific information, presents significant potential as a remote sensing technology. This study explores the feasibility of adapting advanced Raman spectroscopy as a remote crop quality sensor for the precise assessment of carotenoids. Carotenoids were chosen due to their dual role as key stress indicators in crops and their well-established antioxidant benefits for human health.

To explore carotenoid variability, Arabidopsis thaliana and Spinacia oleracea were analyzed. Raman spectroscopy measurements were performed on two leaves per plant using a 785 nm laser. For the carotenoid quantification, Linear Discriminant Analysis (LDA) was adapted. The spectra were processed through smoothing, background removal, and normalization, followed by modification with an amplifying factor. This study evaluated the impact of these processing methods, particularly the application of the amplifying factor, on the accuracy of the model. High-Performance Liquid Chromatography (HPLC) was employed as the reference method for validation. Three-quarters of the samples were used to construct the model, while the remaining one-quarter was reserved for validation. As a result, the model utilizing spectra modified with the amplifying factor in most cases achieved higher validation accuracy compared to models based on unmodified spectra.

This study introduces a novel Raman spectroscopy-based remote sensing approach for crop quality assessment, establishing an enhanced model for interpreting spectral data. By enabling precise detection of stress-induced changes in plant chemical composition, including carotenoids, this technique paves the way for scalable, real-time monitoring through Raman-equipped machinery or drones, advancing sustainable agriculture practices.

How to cite: Park, M., Somborn, A., Schlehuber, D., and Keuter, V.: Development of a Remote Crop Quality Sensor: Advancing Carotenoid Assessment with Raman Spectroscopy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1883, https://doi.org/10.5194/egusphere-egu25-1883, 2025.

EGU25-1993 | Orals | ESSI4.11

Leveraging Satellite Earth Observation for Detecting Bloom Shifts and Phenological Patterns in California’s Almond Orchards 

Tarin Paz-Kagan, Oren Lauterman, Fadi Kizel, Maciej A. Zwieniecki2, Jessica Orozco, and Or Sperling

Given the impact of climate change on deciduous crop yields, our research focuses on leveraging earth observation remote sensing to accurately detect flowering periods in almond orchards and evaluate a climate-based dormancy model for predicting flowering times. This study addresses the challenge of monitoring almond flowering phenology by employing automated crop mapping techniques to support phenology monitoring across California's Central Valley. Using Sentinel-2 (S2) multispectral satellite imagery, we compare its effectiveness with the carbohydrate-temperature (C-T) dormancy model. The study area encompasses approximately 30,000 almond orchards, precisely identified using the Almond Industry Map. We utilized time-series analyses of the Enhanced Bloom Index (EBI) and the Normalized Difference Vegetation Index (NDVI) to quantify bloom periods and intensity and determine peak bloom times. Leveraging around 4,000 S2 tiles, enhanced vegetation indices, and in situ time-lapse camera data collected from 2019 to 2022, we developed a robust methodology for accurately identifying peak bloom periods. This process created a comprehensive phenological dataset, which was standardized and interpolated to daily resolution for improved time-series analysis. Our approach achieved a mean absolute error (MAE) of just 1.9 days in detecting peak bloom, demonstrating the accuracy of satellite-based phenological monitoring. This underscores both the advantages and limitations of remote sensing technologies in agricultural phenology. The dataset was then used to validate projections from the climate-based carbohydrate-temperature (C-T) dormancy model, offering valuable insights and supporting the refinement of this mechanistic approach. The study revealed significant spatial and temporal patterns in flowering phenology, emphasizing the role of regional climatic conditions in influencing crop development. Results highlight the potential of remote sensing and satellite imagery to detect the start, peak, and end of bloom in almond orchards with high precision, generate valuable phenological datasets, monitor patterns at both regional and field scales, and assess the reliability of dormancy models. This research has critical implications for improving agricultural practices and supporting decision-making in the almond industry. By advancing phenological monitoring techniques, our study presents a scalable and innovative approach to managing perennial crops in the face of climate change.

How to cite: Paz-Kagan, T., Lauterman, O., Kizel, F., Zwieniecki2, M. A., Orozco, J., and Sperling, O.: Leveraging Satellite Earth Observation for Detecting Bloom Shifts and Phenological Patterns in California’s Almond Orchards, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1993, https://doi.org/10.5194/egusphere-egu25-1993, 2025.

Selecting the appropriate unmanned aerial vehicle flight height is beneficial for increasing the monitoring efficiency. We firstly used an unmanned aerial vehicle to explore the scale effect on monitoring rice aboveground biomass. The results confirmed the feasibility of using vegetation indices and textures from hyperspectral images to improve the estimations at different spatial resolutions. The monitoring accuracy of combining vegetation indices and textures was the highest, and exhibited a decreasing trend as the spatial resolution decreased with the greatest accuracy appearing at 13 cm. Two new concepts were proposed: “appropriate monitoring scale domain” to define the range of spatial resolution where the monitoring accuracy was less affected by scale effect, and “appropriate monitoring scale threshold” to define the spatial resolution where accuracy dropped noticeably. The appropriate monitoring scale domains varied at different growth stages and the appropriate monitoring scale thresholds of using vegetation indices and textures were lower than those using textures: 39 cm, 52 cm, and 65 cm at the pre-heading, post-heading, and entire growth stages, respectively when using textures, and 52 cm, 65 cm, and 78 cm at the corresponding growth stages when combining vegetation indices and textures. In terms of aboveground biomass level, growth stage and error value, the relatively lower aboveground biomass levels, earlier growth stages of the multi-temporal models, and overestimations were more likely to yield notable accuracy changes when the spatial resolution converted to lower level on both sides of appropriate monitoring scale threshold. Vegetation indices containing red-edge or near-infrared bands were effective for estimation. Yellow/green band textures and vegetation indices containing green bands with near-infrared/red-edge bands also obtained inspiring performances. MEA was indispensable in estimation while more diverse textures were incorporated into the models of the entire growth stages and models established at lower spatial resolutions. These findings are essential for understanding the scale effect in estimating rice aboveground biomass, facilitating efficient monitoring at field scale.

How to cite: Xu, T., Wang, F., and Shi, Z.: Multi-scale monitoring of rice aboveground biomass by combining spectral and textural information from UAV hyperspectral images, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2093, https://doi.org/10.5194/egusphere-egu25-2093, 2025.

EGU25-2120 | ECS | Orals | ESSI4.11

A new framework for mapping time series rubber plantation in Southeast Asia 

Yaoliang Chen and Hongfeng Xu

Accurate yield estimation and appropriate planting management policies for rubber plantations require their precise information on spatiotemporal change data. Previous studies on mapping of rubber plantations did not employ the dynamic rubber phenology features and had difficulty in obtaining historical samples. Here we attempted to develop a new mapping framework through taking historical sample migration, dynamic phenology, and change detection variables into the classification procedure. An automatic sample migration algorithm was first proposed to generate historical samples. Then, two new variable types, dynamic phenology indices and change detection variables, were developed. Another four commonly used variable types -spectral bands, yearly composite spectral indices, terrains, and textures were also extracted. Five combinations of variable types were designed to explore key variable types. Subsequently, the framework with recommended variable types was applied at an experimental site in China and was finally evaluated to two test sites in Myanmar and Thailand for examining its transferability. Results showed that the average overall accuracy of historically migrated samples reached over 97% at the experimental site. Dynamic phenology indices and change detection variables were found as two crucial variable types for rubber plantations mapping. The average rubber plantations mapping accuracy during 2003-2022 reached 93.68%. Transferring the proposed framework to two test sites confirmed the independent roles of change detection variables and dynamic phenology indices. Their average rubber plantations mapping accuracy during 2003-2022 reached 94.34% and 93.73%, respectively. Good spatial consistency between the classified maps and Google Earth images was observed, displaying clear boundaries between rubber plantations and farmland, evergreen broadleaf forest, and shrub. Overall, the proposed framework has great potential for time series rubber plantations mapping in Southeast Asia.

How to cite: Chen, Y. and Xu, H.: A new framework for mapping time series rubber plantation in Southeast Asia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2120, https://doi.org/10.5194/egusphere-egu25-2120, 2025.

The snow cover occurrence index (SCOI), deffned as the ratio of the number of times that a pixel is classiffed as snow to the number of times that the pixel is observed in optical remote sensing data over a given year, can effectively mitigate the inffuence of clouds and holds great potential for extracting the annual snow duration and glacier extent in mountainous regions. The SCOI of the Qinghai–Tibet plateau (QTP) is calculated and analyzed on the basis of Landsat images from 1985 to 2021. The results indicate the following: 1) the evaluation based on station snow depth reveals that the SCOI is stable when the number of combined years reaches 5; 2) the SCOI has a strong correlation with snow cover days (SCD) determined from Moderate Resolution Imaging Spectroradiometer (MODIS) snow cover products; and 3) the SCOI has good potential for glacier extraction and exhibits a high level of consistency with glacier boundary survey data. Overall, owing to the higher spatial resolution and longer duration of the Landsat-based SCOI, it can accurately describe the distribution characteristics and changes in snow cover and glaciers in complex mountainous areas. 

How to cite: Wang, X.: A Novel Snow Cover Occurrence Index (SCOI) for the Dynamics of Snow Duration and Glacier Extent in Mountainous Regions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2205, https://doi.org/10.5194/egusphere-egu25-2205, 2025.

EGU25-2520 | ECS | Orals | ESSI4.11

Linking Citrus Fruit Cracking Intensity to Plant Water Status: Insights from UAV-Derived Metrics Validated by Ground-Based Data 

Moshe (Vladislav) Dubinin, Michael Morozov, Avi Sadka, and Tarin Paz-Kagan

Citrus fruit cracking, a physical failure of the peel, causes yield losses of 10% to 35%, peaking during October-November. Water status of the tree and water flow into the fruit influence this phenomenon. with excessive irrigation during critical fruit development stages exacerbates cracking. As part of the EU-Horizon CrackSense project, this study is aimed to link citrus tree plant water status (PWS) to fruit cracking, emphasizing how deficit irrigation can reduce yield loss due to cracking. Using UAV and eco-physiological measurements, we developed models to predict PWS and its relationship with cracking intensity early in the season. The study, conducted in 2023-2024 in a commercial orchard near Kfar Chabad, Israel, tested four irrigation treatments: control, defined as the standard irrigation, two deficits irrigations regimes (50% of control) early and late in the season, and excessive irrigation (150% of control) throughout the season. Ground-based measurements included fruit and trunk diameter, stem water potential (SWP), stomatal conductance, plant area index (PAI), and growth rate (TG). UAV flights integrated multispectral, thermal, and LiDAR sensors to capture spatial-temporal variability in PWS. Canopy metrics, such as height, volume, LiDAR-based PAI, and spectral and thermal indices, were incorporated into PWS models. Results revealed significant differences in TG, SWP, and stomatal conductance for 50% of early and late deficit irrigation treatments compared to other treatments. Random forest models demonstrated strong predictive performance for SWP (R² > 0.77) and TG (R² > 0.76). LiDAR-derived PA correlated highly with field optical measurements (R² = 0.92), yield (R² = 0.67), and cracked fruit percentages (R² > 0.50). This study underscores the importance of precise irrigation management in reducing fruit cracking. It highlights the potential of remote sensing systems for predicting cracking and managing water status at the tree level. The developed models equip farmers with tools to apply controlled water stress, minimizing cracking and improving yield.

How to cite: Dubinin, M. (., Morozov, M., Sadka, A., and Paz-Kagan, T.: Linking Citrus Fruit Cracking Intensity to Plant Water Status: Insights from UAV-Derived Metrics Validated by Ground-Based Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2520, https://doi.org/10.5194/egusphere-egu25-2520, 2025.

EGU25-2532 | Orals | ESSI4.11

Cross-Year Crop Mapping with Thermal Calendar from Optical Satellite Image Time Series 

Mehmet Ozgur Turkoglu and Helge Aasen

Traditional approaches for crop type classification from optical satellite images typically evaluate algorithms using training and test datasets from the same year and based on calendar days. However, this experimental setup is not practical for real-world applications due to (i) year-to-year variations in crop growth caused by climate, which limit generalization, and (ii) the inability to apply a model to the current year if trained on current-year data. This work addresses these challenges by introducing a cross-year experimental setting and incorporating thermal calendars into our deep learning model. Specifically, we train an attention-based deep learning model on the 2021 Swiss crop dataset, validate it in 2022, and test it in 2023. Thermal calendars, derived from accumulated daily average temperatures, align crop growth with thermal time instead of calendar time, addressing temporal shifts caused by climatic variations. Our results demonstrate that integrating thermal calendars improves performance compared to baseline using standard calendar encodings, achieving better generalization across years and showcasing the potential for large-scale operational crop classification.

How to cite: Turkoglu, M. O. and Aasen, H.: Cross-Year Crop Mapping with Thermal Calendar from Optical Satellite Image Time Series, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2532, https://doi.org/10.5194/egusphere-egu25-2532, 2025.

EGU25-2602 | ECS | Orals | ESSI4.11

LiDAR-based indices and machine learning efforts to model biophysical estimations of corn (Zea mays L.) 

K. Colton Flynn, Gurjinder Baath, Bala Ram Sapkota, and Douglas R. Smith

Light Detection and Ranging (LiDAR) in precision agriculture is gaining traction as the technology becomes both accessible and affordable, particularly for assessing biophysical characteristics of vegetation. This study investigates the potential of unmanned aerial vehicle (UAV)-based LiDAR data for modeling Leaf Area Index (LAI), a key indicator of crop health and productivity. We explore laser penetration indices to model LAI and compare these results with machine learning models using various LiDAR return types (e.g., ground, vegetation, first, last). In both approaches, in-situ LAI measurements obtained with a LiCOR LAI-2000 were used as ground truth. The study was conducted over two years with a multi-date planting of corn (Zea mays L.) in Temple, TX. Our findings indicate that LiDAR-based methods, both through penetration indices and machine learning, hold promise for accurately modeling LAI and other biophysical crop traits in precision agriculture.

How to cite: Flynn, K. C., Baath, G., Sapkota, B. R., and Smith, D. R.: LiDAR-based indices and machine learning efforts to model biophysical estimations of corn (Zea mays L.), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2602, https://doi.org/10.5194/egusphere-egu25-2602, 2025.

Bananas are the tropical fruit with the largest global cultivation area, sales volume, and 
international trade. China is the world's second-largest producer and consumer of bananas. 
Rapid and accurate acquisition of banana planting range and spatial distribution information 
is crucial for promoting the sustainable development of the banana industry in China. 
Currently, research on banana classification and identification faces challenges such as 
insufficient mechanistic understanding, poor generalizability, and difficulties in large-scale 
application. Additionally, banana cultivation areas are often located in regions with cloudy 
and rainy climates, limiting the acquisition of optical imagery. To address this, this study 
constructs a banana identification model based on phenological characteristics: (1) Sentinel
1/2 imagery is utilized to obtain time series curves of banana spectral and scattering features, 
followed by interpolation and filtering of the time series data; (2)A phenological index based 
on optical and scattering features is developed according to banana phenological 
characteristics. By combining SAR with the index, the model's mechanistic understanding is 
enhanced while alleviating the challenges posed by cloud cover in tropical and subtropical 
regions; (3)Using the constructed phenological index alongside banana spectral, texture, and 
temporal features, a classification model is trained for banana identification in the study area. 
This banana forest identification model and the developed phenological index aim to resolve 
current issues in banana classification and provide theoretical and practical support for large
scale banana extraction and the study of tropical and subtropical economic crops.

How to cite: wang, Z.: Banana plantation identification using remote sensing data in tropical and subtropical regions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2709, https://doi.org/10.5194/egusphere-egu25-2709, 2025.

EGU25-4248 | Orals | ESSI4.11

Characterisation and Calibration of Low-Cost IoT Monitoring Systems for Extreme Environmental Conditions  

Laura Mihai, Cristina Toma, Razvan Mihalcea, Karolina Sakowska, Loris Vescovo, Luca Belelli Marchesini, Valerio Coppola, Francesco Renzi, and Riccardo Valentini

Monitoring forests in hard-to-reach locations and under extreme climatic conditions requires reliable, long-term data collection systems. Low-cost devices are increasingly being developed for this purpose; however, deploying these systems without thorough characterisation and calibration can compromise data quality. This work emphasises the importance of fully characterising and calibrating such systems prior to installation to ensure accuracy and reliability over extended periods. This study was conducted as part of the RemoTrees project, which aims to develop a unique IoT tree monitoring system equipped with satellite communication and designed to withstand extreme environmental conditions. A set of the alpha version prototypes, developed within the project, was evaluated in this work. The evaluation focused mainly on a set of low-cost environmental monitoring devices equipped with radiometric sensors measurements. The key performance parameters were assessed, including signal-to-noise ratio (SNR), irradiance sensor detector nonlinearity, sensitivity to temperature variations, and angular response influenced by the diffusive optics. Each parameter was analysed to determine system performance under close to real-world conditions, using both laboratory and in situ validation setups. Key findings revealed that without proper optics used the accuracy of irradiance measurements are significantly influenced. Improvements on the system design and on calibration procedures were implemented to address these issues, improving the overall accuracy and stability of the systems. By addressing these challenges, the systems demonstrated enhanced robustness and suitability for long-term environmental monitoring in extreme conditions. This study underscores the necessity of rigorous pre-deployment testing and calibration for low-cost monitoring devices, particularly when deployed in challenging environments. The findings contribute to advancing the development and deployment of cost-effective technologies for environmental monitoring, enabling more sustainable and accessible data collection practices in forests under extreme climatic conditions.

How to cite: Mihai, L., Toma, C., Mihalcea, R., Sakowska, K., Vescovo, L., Marchesini, L. B., Coppola, V., Renzi, F., and Valentini, R.: Characterisation and Calibration of Low-Cost IoT Monitoring Systems for Extreme Environmental Conditions , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4248, https://doi.org/10.5194/egusphere-egu25-4248, 2025.

EGU25-4513 | ECS | Orals | ESSI4.11

A Machine Learning-based Surrogate Model for Optimization of Cropping Systems in Denmark 

Meshach Ojo Aderele, Edwin Haas, Klaus Butterbach-Bahl, and Jaber Rahimi

Process-based agricultural system models (PBMs) are pivotal tools for evaluating the environmental impacts of agricultural practices. However, their large-scale application is constrained by significant computational demands, extensive time requirements, and data availability. These challenges hinder policymakers and land managers in implementing sustainable agricultural practices at scales meaningful for decision-making. Recent advancements in machine learning (ML) offer a promising solution by providing computationally efficient alternatives, yet the lack of interpretability regarding agro-environmental processes remains a critical barrier.

In this study, we address this challenge by developing a machine learning-based surrogate model for LandscapeDNDC (LDNDC) framework. The surrogate model predicts key agro-environmental variables, including yield, nitrous oxide (N2O) emissions, nitrate leaching (NO3-), and soil organic carbon (SOC), at a national scale for Denmark. Synthetic data were generated using a factorial design based on observed crop practices in Denmark, utilizing field-level data collected across six Danish catchments between 2013 and 2019 as part of the National Monitoring Program for Water Environment and Nature (NOVANA; LOOP-program). This approach incorporated crop rotations as well as spatially disaggregated information on soils and weather, resulting in a dataset comprising approximately 2 billion rows. To enhance the dataset's versatility and account for potential future scenarios, factors like manure amount and synthetic fertilizer amount were extrapolated beyond its current observed ranges. The synthetic dataset was subsequently simulated using the LDNDC modelling framework, and the resulting outputs were employed to train a variety of machine learning algorithms utilizing multi-task learning, optimizing predictions for multiple agro-environmental variables of interest.

Our results demonstrate that the ML-based surrogate model not only significantly reduces computational cost and processing time but also enables the exploration of multiple cropping scenarios with greater efficiency. This approach facilitates rapid scenario testing and optimization, making it accessible to policymakers and farmers without the constraints imposed by traditional PBM frameworks. We propose this methodology as a scalable and practical tool for advancing sustainable agricultural decision-making.

How to cite: Aderele, M. O., Haas, E., Butterbach-Bahl, K., and Rahimi, J.: A Machine Learning-based Surrogate Model for Optimization of Cropping Systems in Denmark, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4513, https://doi.org/10.5194/egusphere-egu25-4513, 2025.

EGU25-4726 | ECS | Posters on site | ESSI4.11

Combining geometric-optical and spectral invariants theories for modeling canopy fluorescence anisotropy 

Yachang He, Yelu Zeng, and Dalei Hao

The spectral invariants theory (p-theory) has received much attention in the field of quantitative remote sensing over the past few decades and has been adopted for modeling of canopy solar-induced chlorophyll fluorescence (SIF). However, the spectral invariant properties (SIP) in simple analytical formulas have not been applied for modeling canopy fluorescence anisotropy primarily because they are parameterized in terms of leaf total emissions and scatterings, which precludes the differentiation between forward and backward leaf SIF emissions. In this study, we have developed the canopy-SIP SIF model by combining geometric-optical (GO) theory to account for asymmetric leaf SIF forward and backward emissions at the first-order scattering and by modeling multiple scattering based on the p-theory, thus avoiding the dependence on radiative transfer models. The applicability of the model simulations especially over 3D heterogeneous canopies was improved by incorporating canopy structure through multi-angular clumping index, and by modeling single scattering from the four components of the scene in view according to the GO approach. The results show good consistency with both the state-of-the-art SIF models and multi-angular field SIF observations over grass and chickpea canopies. The coefficient of determination (R²) between the simulated SIF and field measurements was 0.75 (red) and 0.74 (far-red) for chickpea, and 0.65 (both red and far-red) for grass. The average relative error was approximately 3% for 1D homogeneous scenes when comparing the canopy-SIP SIF model simulations to the SCOPE model simulations, and around 4% for the 3D heterogeneous scene when comparing to the LESS model simulations. The results indicate that the proposed approach for separating asymmetric leaf SIF emissions is a robust way to keep a balance between satisfactory simulation accuracy and efficiency. Model simulations suggest that neglecting the leaf SIF asymmetry can lead to an underestimation of canopy red SIF by 16.1% to 43.4% for various canopy structures. This study presents a simple but efficient analytical approach for canopy fluorescence modeling, with potential for large-scale canopy fluorescence simulations.

How to cite: He, Y., Zeng, Y., and Hao, D.: Combining geometric-optical and spectral invariants theories for modeling canopy fluorescence anisotropy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4726, https://doi.org/10.5194/egusphere-egu25-4726, 2025.

EGU25-4910 | ECS | Orals | ESSI4.11

Mapping 10-m monoculture and intercropped maize of Kenya with phenology knowledge and Sentinel-2 data 

Yang Chen, Lijun Zuo, Xianhu Wei, Xiao Wang, and Jinyong Xu

In East Africa, lack of agriculture inputs and unstable climates lead to 50% yield gaps, making intercropping—the planting of more than one crop in the same parcel of land—a common agricultural management practice among smallholder farmers to improve land-use efficiency and reduce risks. In Kenya, where maize is the staple food, maize is often intercropped with beans, legumes, and potatoes. Despite its widespread, agricultural statistics on intercropping are currently sparse, and remote sensing approaches for large-scale crop monocultures are often unsuitable for intercropping monitoring. Mapping intercropping at national scale is extremely challenging because of heterogeneous landscapes, lack of cloud-free satellite imagery, and the scarcity of high-quality ground-based situ data in these regions. This study addressed these challenges using a phenology-assisted automated mapping framework on Google Earth Engine (GEE) to create 10m-resolution maps of monoculture and intercropped maize across Kenya for the long and short rainy seasons of 2023.
First, we computed 10-day median composites of Sentinel-2 optical reflectance data for each pixel in the region to build monoculture/intercropped/non-maize Random Forest (RF) classifiers. Several thousand crop ground labels were collected during field surveys in 2023, including monoculture maize (mono-maize), intercropped maize (in-maize), and other crops (e.g., wheat, rice, coffee, tea, sugarcane, potatoes, beans, etc.). To address the limited availability of intercropped maize samples, a novel phenology-based approach was implemented. Maize was first differentiated from other crops by analyzing TCARI and OSAVI during the vegetative phase and ARI during maturity. Additionally, lower greenness and moisture levels in intercropped systems, which have larger planting width and more short-term crops, were detected using the SWIR1/NDVI ratio, effectively distinguishing mono-maize from in-maize. Automatically derived monoculture/intercropped maize samples and 40% of ground samples were used for training, while the remaining ground data were used for accuracy assessment. 
For the long rainy season, the overall accuracy (OA) was 0.88, with an F1-score of 0.87 for mono-maize and 0.78 for in-maize. For the short rainy season, OA dropped to 0.85, with F1-scores of 0.82 for mono-maize and 0.72 for in-maize. Misclassification primarily arose from phenological similarities between mono-maize and in-maize and increased planting of other crops with similar patterns during the short rainy season. Results revealed that 854,432 hectares of mono-maize were concentrated in the Western region and Rift Valley plateau during the long rainy season, while 1,061,701 hectares of in-maize were widely distributed across the region, particularly near Mount Kenya and the Eastern region. In the short rainy season, reduced and erratic precipitation led to decreased maize planting, with more farmers opting for intercropped systems and short-term crops to reduce risks of crop failure. 
We are convinced that this study is a crucial first step to demonstrate the potential of Sentinel-2 data and phenology-based automated mapping for large-scale monitoring of intercropping, providing critical insights for agricultural monitoring in sub-Saharan Africa. It serves as a foundation for developing a regional archive of monoculture and intercropped crop systems and addressing key agricultural challenges across the region.

How to cite: Chen, Y., Zuo, L., Wei, X., Wang, X., and Xu, J.: Mapping 10-m monoculture and intercropped maize of Kenya with phenology knowledge and Sentinel-2 data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4910, https://doi.org/10.5194/egusphere-egu25-4910, 2025.

A study conducted in a northern Jordanian arid Mediterranean grassland between 2017 and 2021 examined the relationship between remotely sensed Normalized Difference Vegetation Index (NDVI) and modeled standing crop biomass. The research sought to determine the utility of high-resolution (10-meter) Sentinel-2 imagery, coupled with the PHYGROW model, for biomass estimation in this challenging environment, and to assess the potential of NDVI as a cost-effective alternative to traditional ground-based methods. Data were aggregated into 10-day intervals for temporal analysis. Results indicated a significant positive correlation (p < 0.001) between NDVI and standing crop (kg/ha), described by the linear model: Standing crop = 60.40 + 3567.56 × NDVI (R² = 0.52). This finding suggests that NDVI offers a reliable and time effective approach to biomass estimation in such settings.

The strong positive correlation between NDVI and standing crop highlights the potential of remote sensing for large-scale rangeland health monitoring. Tracking NDVI changes over time provides insight into vegetation responses to climate, grazing, and conservation efforts. This understanding supports decision-making for sustainable grazing, water management, and conservation strategies. Future research should validate these findings on larger scales and explore integrating NDVI with other data, like soil moisture, to refine predictive models and improve accuracy. The study advocates adopting NDVI-based monitoring in arid rangeland management.

How to cite: Alhamad, M. N.: Integrating Sentinel-2 Imagery and PHYGROW Model for Biomass Estimation in Arid Rangelands of Northern Jordan, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5178, https://doi.org/10.5194/egusphere-egu25-5178, 2025.

EGU25-5892 | ECS | Orals | ESSI4.11

Assessment of bud flush and damage in young Norway Spruce trees through airborne high-resolution multispectral images 

Louisa Eurich, Sara López Fernández, Malin Elfstrand, María Rosario García-Gil, Jonas Bohlin, and Eva Lindberg

Scandinavia is facing climate changes with a predicted increase in mean temperature of 2-4°C. For Swedish forests to be adapted to this challenge, the Swedish tree breeding program aims to select trees that are adapted to different biotic and abiotic conditions. Information on spring phenology, damage and vitality are important variables in the Norway spruce selection process. Traditionally, the data is gathered through manual assessment of each tree, which requires significant resources and limits the number and frequency of variables that can be measured. As an alternative, Remote Sensing is a promising technology to evaluate bud flush and vitality in conifers, offering the advantage of scoring more trees in a shorter time with fewer resources while obtaining data for several time points during the vegetation season, and its use of algorithms to measure variables reduces the risk of human error.

This project aims to develop methods that can be used within the breeding program by collecting information on spring phenology, damage and vitality using high-resolution multispectral drone images of young Norway spruce trees. Data were collected during spring 2023 and 2024. Bud flush is estimated from the spectral values of the tree crowns using manual assessment of the flush in a subset of the trees as training data. The high-resolution multispectral images will also be used to assess the damage and vitality of the new shoots. To ensure capturing the bud flush at a high temporal resolution, images were taken before the vegetation season and up to twice weekly during the period with the most rapid flush. In the final step, the spatial pattern within the study sites will be analyzed and connected to damage and vitality of the young Norway Spruce trees.

 

How to cite: Eurich, L., López Fernández, S., Elfstrand, M., García-Gil, M. R., Bohlin, J., and Lindberg, E.: Assessment of bud flush and damage in young Norway Spruce trees through airborne high-resolution multispectral images, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5892, https://doi.org/10.5194/egusphere-egu25-5892, 2025.

EGU25-5987 | ECS | Posters on site | ESSI4.11

Soil Moisture Retrieval Over Agricultural Fields Using Synthetic Aperture Radar (SAR) Data 

Nguyen-Thanh Son, Chi-Farn Chen, Cheng-Ru Chen, Yi-Ting Zhang, Shu-Ling Chen, and Shih-Hsiang Chen

Soil moisture is vital for agricultural fields as it determines water availability for crops, directly affecting plant growth and productivity. It regulates nutrient uptake, root development, and microbial activity, ensuring efficient use of fertilizers and soil resources. Proper soil moisture levels prevent water stress, reduce crop failure risks, and optimize water irrigation efficiency. Accurate soil moisture monitoring supports sustainable farming practices, helps mitigate drought impacts, and enhances climate resilience. By maintaining optimal soil moisture, farmers can improve resource use, boost crop yields, and promote long-term agricultural sustainability. This study aims to develop an approach for retrieving soil moisture from Sentinel-1 A Synthetic Aperture Radar (SAR) data. The SAR data were processed for the 2024 dry season using a triangle-based approach in the Mekong Delta, Vietnam, following three main steps: (1) data preprocessing to convert raw radar backscatter values into the sigma naught (σ₀) backscatter coefficient in decibels (dB). This involves radiometric calibration, noise removal, and logarithmic scaling to enhances data interpretability, allowing for better comparisons across different radar acquisitions and improved analysis accuracy, (2) soil moisture retrieval by means of a triangle-based method developed based on the dual-polarization modes of the vertical transmit and vertical receive polarization (VV) and vertical transmit and horizontal receive polarization (VH). This method employs the triangular feature space created by using change in VV backscatter coefficients and the radar vegetation index (RVI), in which RVI helps distinguish vegetation effects while VV backscatter provides information on soil moisture. Combining both parameters thus allows for more precise moisture estimation even in complex environments, and (3) error verification. The results of soil moisture retrieval compared with the reference data showed moderate positive correlation, with the values of correlation coefficient (r) greater than 0.5 and the root mean square error (RMSE) smaller than 0.05, respectively. The lower soil moisture levels were especially observed in coastal areas, where higher evaporation rates, saline intrusion, and limited rainfall contribute to drier soils. These conditions create challenges for agriculture in coastal regions, as crops are more susceptible to drought stress and water shortages. Consequently, managing soil moisture becomes crucial for maintaining crop productivity and ensuring sustainable farming in coastal provinces. Eventually, soil moisture data was spatially aggregated with cropping areas to improve management practices in the region, allowing precise monitoring of soil conditions relative to specific crops and enabling tailored irrigation and water management strategies. This approach, leveraging dual-polarization SAR data with aid of the triangle-based method, could enhance soil moisture monitoring in agriculture and is completely transferable to other regions across the globe for soil moisture monitoring.

How to cite: Son, N.-T., Chen, C.-F., Chen, C.-R., Zhang, Y.-T., Chen, S.-L., and Chen, S.-H.: Soil Moisture Retrieval Over Agricultural Fields Using Synthetic Aperture Radar (SAR) Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5987, https://doi.org/10.5194/egusphere-egu25-5987, 2025.

EGU25-6071 | Orals | ESSI4.11

Sensor Spatial Planning Methodology for Optimal Coverage and Data Accuracy in Agricultural Parcels 

Efthymios Papachristos, Marios Vlachos, and Angelos Amditis

Accurate sensor placement is critical in precision agriculture to collect high-resolution data essential for effective monitoring and decision-making. This study presents a comprehensive methodology for optimizing the spatial placement of sensors, focusing on determining the number of sensors needed and their optimal positions to ensure data quality and adequate area coverage. This methodology addresses the challenges posed by terrain restrictions, cost constraints, and data resolution needs. It is versatile, supporting in-situ monitoring, UAV-based sensing, and soil sampling for applications such as soil health analysis and soil organic carbon prediction models.

In many Research and Innovation Labs (RILs), the resolution of Earth Observation (EO) data, such as Sentinel-5 imagery with a resolution of 5×3 km, is often insufficient for the specific needs of agricultural parcels. To complement EO data, additional information must be gathered using in-situ sensors or UAVs. These additional data collection methods can provide higher resolution and more diverse data types, which are crucial for localized agricultural applications. However, the placement of sensors significantly impacts the quality and adequacy of the collected data. Dense sensor deployment across an entire area is often infeasible due to terrain challenges, budgetary limits, and the specific nature of the data being collected.

The methodology developed to address these challenges combines convex optimization, soft clustering, and cost-minimization techniques. The process begins by analyzing the statistical properties of the dataset, such as maximizing variance and maintaining the mean value, to ensure comprehensive data representation. This approach identifies key locations within the parcel that can adequately describe distributed values, reducing the need for excessive sensor deployment while maintaining data integrity.

For areas with existing spatial maps or datasets, the methodology applies weighted subsampling and soft clustering to identify optimal sensor locations. Weighted distributions prioritize critical areas for data collection, ensuring that key zones receive sufficient coverage. In cases where spatial maps are unavailable, an in-house cost-minimization algorithm guides the placement of sensors or UAVs. This algorithm incorporates factors such as terrain, accessibility, and installation costs to balance logistical constraints with data coverage requirements.

This methodology is compatible with diverse data sources, including EO data, hyperfield data, and in-situ sensor data from IoT networks. For instance, it can leverage data from soil moisture monitoring systems. Additionally, the methodology can guide soil sampling strategies for soil health assessment and serve as input for soil organic carbon prediction models. Its adaptability allows it to meet the needs of various agricultural monitoring applications, ranging from broad-scale field evaluations to detailed soil property studies.

Moreover, it enhances data quality by ensuring optimal sensor placement that captures maximum variability within the monitored area and it reduces costs and improves efficiency by minimizing the number of sensors needed. The approach is scalable and flexible, accommodating parcels of varying sizes and adapting to different data collection requirements and its integration with multiple data sources provides a comprehensive and cost-effective solution for advancing precision agriculture and sustainable resource management.

Acknowledgement:

This research has been funded by European Union’s Horizon Europe research and innovation programme under ScaleAgData project (Grant Agreement No. 101086355).

How to cite: Papachristos, E., Vlachos, M., and Amditis, A.: Sensor Spatial Planning Methodology for Optimal Coverage and Data Accuracy in Agricultural Parcels, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6071, https://doi.org/10.5194/egusphere-egu25-6071, 2025.

EGU25-6085 | Orals | ESSI4.11

Integrating UAV Multispectral Data into a Combined Crop-Radiative Transfer Model for Processing Tomatoes Using a Particle Filter 

Amit Weinman, Nitzan Malachy, Raphael Linker, and Offer Rozenstein

The proliferation of remote sensing (RS) data and advancements in mechanistic crop modeling and data assimilation techniques necessitate a framework that digitally represents cropping systems and their spectral properties. Such a framework would enable crop growth simulation, scenario testing, and timely prediction updates using RS data.

In this study, we develop a comprehensive coupling scheme that links a crop model (DSSAT-CROPGRO) with a radiative transfer model (RTMo module in SCOPE). This integration allows for the utilization of reflectance data from all measured spectral bands during data assimilation (DA) into the crop model.

We apply this coupled crop-radiative-transfer model in a DA experiment using a novel particle filter scheme. The assimilated data consists of observed reflectance measurements obtained by a multispectral camera mounted on an unmanned aerial vehicle (UAV). Using multispectral data with a high spatial resolution for analyzing a row crop required a dedicated analysis to fit model simulations to measurements. The suggested DA scheme was implemented in an irrigation and fertilization trial with processing tomatoes to evaluate its effectiveness.

The results showed that applying the DA scheme improved the NRMSE of the Leaf Area Index (LAI) from 59% to 41.8% and yield from 63.6% to 35.4%. The DA scheme performed best when the treatment that included the most severe stress was excluded from weight calculation, resulting in NRMSE of 34.1% and 15.5% for LAI and yield, respectively. After showing promising results, the suggested data assimilation scheme should be tested in large-scale, commercial fields using space-borne RS data to examine its applicability in various scenarios.

How to cite: Weinman, A., Malachy, N., Linker, R., and Rozenstein, O.: Integrating UAV Multispectral Data into a Combined Crop-Radiative Transfer Model for Processing Tomatoes Using a Particle Filter, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6085, https://doi.org/10.5194/egusphere-egu25-6085, 2025.

EGU25-6149 | ECS | Posters on site | ESSI4.11

UAV-based disease and pest detection using AI: Time to reconsider our approach? 

Eline Eeckhout, Pieter Spanoghe, and Wouter Maes

Rapid advancements in technology, particularly the rise of artificial intelligence (AI) and the integration of uncrewed aerial vehicles (UAVs) equipped with RGB, multi- and hyperspectral sensors, have boosted agricultural research on crop disease detection. This has led to a surge in studies exploring high-technology approaches to detecting crop diseases. While numerous studies have demonstrated high accuracy in detecting specific diseases or pests in crops, concerns arise regarding their reproducibility and generalisability.

We conducted a meta-analysis of over 100 research papers to examine how models are trained and validated, with a focus on how datasets for training, validation and testing were handled. In principle, a model can only be considered robust and widely applicable if it performs well on an entirely new dataset, i.e., a dataset it wasn’t specifically trained one. Otherwise, AI models risk overfitting to specific datasets or fields, potentially detecting signals that are not universal or not related to the targeted pest or disease. This issue arises when datasets are randomly split in training, validation and test subsets.

Our analysis revealed significant limitation in current practices. Nearly half of the reviewed papers relied on a single dataset (one single field, one single flight) for both model training and validation. About one-quarter of the studies used data from a single field with repeated flights during the same growing season. Only another quarter utilized datasets from multiple fields; however, the majority of these studies still used a random split for training and testing, meaning their models were not evaluated on independent datasets. In addition, a handful of studies using RGB data, applied transfer learning, with models pretrained on public (non-UAV) datasets and then applied to UAV datasets.

Overall, only 10% of the reviewed papers validated their models on fully independent datasets, i.e, using transfer learning or using an independent (untrained) separate field to test the model. We found that particularly models constructed with multispectral or hyperspectral data did not use independent datasets. On top of that, none of the studies explicitly tested whether their models were pest- or disease-specific, i.e., whether the models were sensitive only to the pest or disease they were trained to detect.

These findings highlight a critical limitation in the robustness and scalability of current AI-approaches to crop disease detection with UAVs. To address this, we call on researchers to include independent test datasets in their studies, and urge journals and reviewers to prioritize this criterion during evaluations. Additionally, we advocate for the public sharing of datasets to enable the development of robust and generalisable methods.

How to cite: Eeckhout, E., Spanoghe, P., and Maes, W.: UAV-based disease and pest detection using AI: Time to reconsider our approach?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6149, https://doi.org/10.5194/egusphere-egu25-6149, 2025.

EGU25-6243 | ECS | Posters on site | ESSI4.11

Multisource data analysis at the catchment scale to quantify and map sustainable agricultural management practices 

Maria S. Vesterdal, Tommy Dalgaard, and René Gislum

Natural environments face substantial challenges from human activities related to food, feed, and energy production. Unsustainable nutrient management is a key issue, with excess nutrients leaching into the groundwater cycle or escaping intended cropland through other pollution pathways ending up in the atmosphere or in nearby coastal systems. This nutrient loss depletes soil health, contributes to the climate crisis and impacts water quality, especially when combined with intensive farming practices lacking conservation efforts. Innovative mitigation actions, such as the Nature-based Solutions framework, designed to enhance water quality and advance sustainability in agricultural management, require thorough assessment and monitoring to encourage stakeholder participation in these strategies. Conducting research to explore the extent of their effects is thus essential, with a deeper understanding of the nutrient cycle playing a pivotal role in achieving these goals.

With the cumulatively increasing availability of remote sensing data sources and advancements in machine learning technologies, automating monitoring and assessment efforts has become a hot and important topic. The challenge is to construct transparent and transferable models capable of working with real-time data to accurately predict crop types, crop status or other desired features. The primary goal of this study is to investigate how an automated multisource data analysis approach, with a focus on remotely sensed data, can support the quantification and mapping of sustainability efforts in agricultural crop management while enhancing the understanding of nutrient flow within large-scale agricultural catchments. Centered on the Hjarbæk Fjord in Denmark, the study also aims to assess the transferability of its models across different sites in Europe. This research is part of a broader project investigating the potential of integrating permanent grasslands into crop rotations as a Nature-based Solution in the catchments surrounding Hjarbæk Fjord. The project aims to develop a decision support tool to guide the planning and optimization of grassland implementation in terms of extend and location. This tool is designed to maximize benefits across various parameters, including the number of stakeholders impacted, economic considerations, crop yield, biodiversity, and other critical factors. The output of the current study, involving the training of a deep learning model to predict cropland trends related to grassland implementation, can in turn be integrated as input for the described decision support tool.

This is an explorative study that relies on the availability of accurate ground truth data to train and validate a deep learning model, providing insights into trends associated with the implementation of sustainable management strategies. A key challenge lies in acquiring knowledge of and access to comprehensive datasets that capture relevant parameters, such as actual yield values, quantitative values of nutrients in different stages of the growth season and different nutrient pools within the cropland environment, accurate accounts of management actions and other contributors to the nutrient cycle. Additional challenges involve preprocessing satellite data to establish a robust pipeline for the automated collection of satellite imagery, ensuring a coherent time series. This includes addressing temporal and spatial data gaps through extrapolated estimations to create a consistent dataset.

How to cite: S. Vesterdal, M., Dalgaard, T., and Gislum, R.: Multisource data analysis at the catchment scale to quantify and map sustainable agricultural management practices, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6243, https://doi.org/10.5194/egusphere-egu25-6243, 2025.

EGU25-6285 | ECS | Posters on site | ESSI4.11

Can SuperDove Multispectral Satellite Data Optimize Citrus Orchard Monitoring? 

Lamia Rahali, Salvatore Pratico, and Giuseppe Modica

The increasing global demand for food and the pressing need for sustainable agricultural practices have made technological innovations essential in modern agriculture. Satellite imagery, as a cornerstone of precision agriculture (PA), provides valuable tools for monitoring crops and optimizing resource management. This study evaluates the potential of PlanetScope’s (PS) advanced 8-band multispectral sensor (SuperDove) for citrus orchard monitoring. The primary objectives are to investigate the effectiveness of PS data in assessing orchard health and dynamics and to explore its utility in detecting spatial variability within citrus orchards. The methodology involves preprocessing SuperDove data to derive key vegetation indices (VIs), such as NDVI, SAVI, and EVI, which are widely used to gain insights into the vigor and condition of citrus orchards. To assess the reliability and practicality of PS data, the study includes a comparison with free and open-source alternatives, such as Sentinel-2. This research emphasizes the importance of integrating high-resolution satellite imagery into citrus orchard management practices. While still in the early stages, the study aims to provide insights into how advanced satellite data can support sustainable agriculture.

How to cite: Rahali, L., Pratico, S., and Modica, G.: Can SuperDove Multispectral Satellite Data Optimize Citrus Orchard Monitoring?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6285, https://doi.org/10.5194/egusphere-egu25-6285, 2025.

EGU25-7430 | Orals | ESSI4.11

Improving satellite-based actual evapotranspiration estimations using data from local weather stations  

Offer Rozenstein, Jessey Kwame Dickson, and Josef Tanny

Evapotranspiration (ET) is crucial for water resource management, agricultural planning, and understanding land-atmosphere interactions. Numerous approaches are available for estimating ET at various spatial and temporal scales, including ground-based measurements, mechanistic models, and remote sensing. In this study, we aimed to enhance the accuracy and applicability of the Sentinel for Evapotranspiration (Sen-ET) plugin for estimating ET in diverse field crops in Israel. The primary objectives were to validate the Sen-ET method using eddy covariance (EC) measurements across various seasons and crop types, improve Sen-ET estimates by incorporating local weather station data, and illustrate the influence of weather station distance from measurement sites on Sen-ET accuracy.

The research was conducted across eight test sites in Israel, including fields with spring wheat, potato, cotton, and tomato. In applying Sen-ET model, we utilized high-resolution Sentinel-2 and Sentinel-3 imagery, along with ERA-5 meteorological data and local weather station inputs. The ET estimations by Sen-ET involved preprocessing satellite data, resampling meteorological data, and using a Two Source Energy Balance model to derive daily ET values. These estimates were compared against EC measurements.

The results demonstrated that incorporating local weather station data significantly improved the accuracy of the Sen-ET estimates, with most sites showing a substantial reduction in root mean square error (RMSE) of daily ET compared to the standard Sen-ET method. For example, at one of the wheat sites, the RMSE was reduced from 0.60 mm to 0.14 mm day-1. On the other hand, one of the tomato sites showed a slight deterioration, with an increase of 0.01 mm day-1 in RMSE when data from a weather station 7 km away was used. However, when a closer weather station at 1.17 km was used, the RMSE was reduced by 0.34 mm day-1, thus demonstrating the importance of employing representative weather data in the model.

This study underscores the contribution of localized meteorological data in refining satellite-based ET models and provides a robust approach for precise ET estimation in agricultural landscapes. The findings have significant implications for improving water resource management and irrigation practices.

How to cite: Rozenstein, O., Kwame Dickson, J., and Tanny, J.: Improving satellite-based actual evapotranspiration estimations using data from local weather stations , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7430, https://doi.org/10.5194/egusphere-egu25-7430, 2025.

Abstract: The universal contamination of arable land with potentially toxic elements (PTEs) poses a severe threat to food security and jeopardizes worldwide efforts to meet the United Nations Sustainable Development Goals (SDGs). How to obtain information on PTEs in regional agricultural soils more reliably is a priority problem to be solved. Multispectral satellite remote sensing, with its advantages of high spatial and temporal resolution, broad coverage, and low cost, offers the potential to acquire distribution information of PTEs over large areas. However, owing to the complexity of soil environments and the insufficiency of spectral information, the mechanism for retrieving concentrations of soil PTEs via multispectral satellites is not yet clear, and the accuracy needs to be improved. In this study, we aimed to assess whether employing a fusion of spectral information and environmental covariates, results in more accurate predictions of PTEs, specifically chromium (Cr) and mercury (Hg), in croplands than does employing spectral information alone. Three machine learning algorithms—kernel-based support vector machine (SVM), neural network-based back propagation neural network (BPNN), and tree-based extreme gradient boosting (XGBoost)—were developed to retrieve soil Cr and Hg concentrations. The results showed that the fusion of spectral information and environmental covariates combined with the XGBoost model performed best in retrieving both Cr and Hg concentrations with coefficient of determination (R2) values of 0.73 and 0.74, respectively. Environmental covariates are important variables for determining Cr and Hg concentrations in agricultural soils, but the ability to retrieve these element concentrations by utilizing multispectral information alone is limited. High Cr concentrations occurred in central towns and southern hilly mountains. High Hg concentrations were detected in the northwestern region and southern hilly mountains. The potential of fusing multispectral data and environmental variables to precisely retrieve soil PTE concentrations can serve as a reference for agricultural information monitoring worldwide.

Keywords: Potentially toxic elements; Sentinel-2; Environmental covariates; Machine learning; Farmland

How to cite: Zha, Y.: Retrieval of chromium and mercury concentrations in agricultural soils: Using spectral information, environmental covariates, or a fusion of both?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7525, https://doi.org/10.5194/egusphere-egu25-7525, 2025.

EGU25-7922 | Posters on site | ESSI4.11

Machine Learning-Based Rice Disease Diagnosis Through Joint Utilization of Satellite, Drone, and Weather Data 

Jae-Hyun Ryu, Kyung-Do Lee, Young-ah Jeon, Geun-Ho Kwak, Soo-Jin Lee, and Lak-Yeong Choi

Remote sensing and machine learning techniques enable precise diagnosis of crop growth anomalies, providing an effective means to mitigate production losses caused by disease outbreaks while supporting sustainable agricultural management. This study aims to detect rice diseases using satellite, drone, and weather data in a timely manner. A random forest model for rice disease detection was developed using drone imagery collected in 2023 year, where disease-damaged pixels were classified through K-means clustering, and the corresponding damaged areas were used for rice paddy disease classification model training. This model has been applied to agricultural fields in 2024 year as follows. First, Sentinel-1 and Sentinel-2 satellite data were utilized to classify paddy rice fields, with irrigated areas identified through the normalized difference vegetation index, land surface water index, and VV polarization. Second, the risk of rice disease occurrence was calculated based on air temperature, relative humidity, and precipitation. These variables represent weather conditions that can cause crop diseases. Third, drone measurements were conducted to monitor the abnormal growth of paddy rice when the risk score increased. Fourth, the location of disease outbreaks was detected using the random forest model, which uses surface reflectance at blue, green, red, red-edge, and near-infrared wavelengths as input data. Subsequently, drone spraying operations were carried out to reduce crop damage caused by the identified diseases. These results highlight the potential of agricultural management using remote sensing techniques.

Acknowledgments: This research was funded by RDA, grant number RS-2022-RD010059.

How to cite: Ryu, J.-H., Lee, K.-D., Jeon, Y., Kwak, G.-H., Lee, S.-J., and Choi, L.-Y.: Machine Learning-Based Rice Disease Diagnosis Through Joint Utilization of Satellite, Drone, and Weather Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7922, https://doi.org/10.5194/egusphere-egu25-7922, 2025.

EGU25-7950 | ECS | Orals | ESSI4.11

Detecting irrigation amount from the integration of remote sensing data in the soil water model  

Fatemeh Khamseh and Mohammad Danesh-Yazdi

Agriculture is one of the primary consumers of freshwater globally. However, precise data on irrigation water use (IWU) at the regional scale is often lacking, which hampers the development of effective water management plans. This information gap is particularly crucial in water-stressed regions, resulting in significant resource waste. Remote sensing datasets offer a valuable opportunity to monitor irrigation patterns over extended periods at a regional scale. Since irrigation affects both soil moisture (SM) and actual evapotranspiration (ET), increases in SM and ET values following irrigation events can be leveraged to frequently retrieve IWU from remotely sensed data. In this regard, we first developed an irrigation-free soil water model in the root zone to simulate SM dynamics during non-growing periods. We then computed the residuals between the modeled SM and the 9 km root zone SM retrieved from SMAP L3, as well as the residuals between the modeled ET and both 30-m OpenET and 500-m PML, to estimate IWU. We used annual IWU data from Arizona State, USA, in 2017 to examine model performance. The simulated SM by our soil water model showed strong agreement with SMAP, evidenced by R2 = 0.68 and RMSE = 0.015 [mm3/mm3]. The estimated IWU using OpenET closely aligned with benchmark data, showing a bias of -17%. However, IWU retrieved by PML led to a much higher bias of -56%, indicating the deficiency of course-resolution ET products in capturing irrigation signals. We further found that over 97 % of the estimated IWU was attributed to ET rather than SM residuals, which is due to SMAP’s low spatial resolution, which limits its ability to resolve farm-scale irrigation volumes.

 

How to cite: Khamseh, F. and Danesh-Yazdi, M.: Detecting irrigation amount from the integration of remote sensing data in the soil water model , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7950, https://doi.org/10.5194/egusphere-egu25-7950, 2025.

EGU25-8008 | ECS | Posters on site | ESSI4.11

Modeling and managing erosion in arid ravines using high-resolution satellite imagery 

Amir Mor-Mussery, Eli Zaady, and Lior Blank

Abstract

Ravines in arid lands are affected by various soil erosion processes caused by inconsistent rainfall regimes, flooding patterns, and anthropogenic interventions. These effects are expressed in the geomorphological and vegetation patterns of the ravine's land segments. To address these changes, a study was planned with the following objectives: [1] Modeling the effects of ravine erosion processes on its land-segments vegetation using high-resolution satellite imagery; [2] Suggesting analysis schemes based on remote sensing to suit land management practices for the ravine parts.  The study site is located in Migda Ravine, Northern Negev, between Gerar and Patish ephemeral streams. Due to the loess soil and extreme arid conditions, the area suffers from soil erosion and land incision. Using imaging from PlanetScope® satellite constellation (spatial resolution: 3m pixel-1, temporal: Image per 3 days, and spectral: Red-Green-Blue-Near Infra-Red bands) between 2017 and 2024, from January to August each year, NDVI median and quartiles ranges of the ravine land segments were calculated and normalized against a stabilized reference plot. Thirteen erosion processes were defined, and classified into ravine surrounding areas, banks, and ephemeral stream water flow. The findings indicate erosional processes that dramatically decreased the Normalized Fresh Vegetation Reflectance (NFVR)in 2019, with a lighter decrease in 2020. Some erosion processes were characterized by a subsequent NFVR increase after the soil erosion event, while others, such as subsurface erosion, showed a continuous NFVR decrease. Stream plots were characterized by soil deposition, which resulted in vegetation flocculation. Using vegetation change patterns, NDVI normalization, and multi-year temporal analysis can aid in formulating land management practices for the ravine land segments and predicting long-term erosional patterns.

How to cite: Mor-Mussery, A., Zaady, E., and Blank, L.: Modeling and managing erosion in arid ravines using high-resolution satellite imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8008, https://doi.org/10.5194/egusphere-egu25-8008, 2025.

EGU25-8077 | Posters on site | ESSI4.11

Development of Cropping Pattern Product Using Sentinel-2 Satellite Data 

Lak-Yeong Choi, Jae-Hyun Ryu, Ho-Yong Ahn, Soo-Jin Lee, Geun-Ho Kwak, Young-Ah Jeon, and Kyung-Do Lee

Understanding cropland utilization is essential for improving agricultural productivity and efficiently managing cropland resources. Analyzing region-specific cropping systems enables the establishment of sustainable agricultural policies tailored to environmental conditions. However, conducting field surveys over extensive agricultural areas presents significant challenges. Satellite data for agricultural monitoring provides continuous and large-scale information for cropland. The purpose of this study is to develop a cropping pattern product for annual crops using satellite data. The study area is ‘Gimje-si’ in the Republic of Korea. Sentinel-2 Level-2 data was acquired from 2022 to 2024. The normalized difference vegetation index (NDVI) was calculated after eliminating cloud and contaminated pixels, and then the monthly mean NDVI was computed. Cropland was extracted using a farmland boundary map in vector file format. Types of cropping patterns were classified into single and sequential (e.g., double, triple) cropping, and non-cultivated land, based on the number of peaks in the time-series NDVI data. The threshold for NDVI peaks was set to 0.4, and the minimum distance between NDVI peaks was set to 3. The final product was generated in vector format and includes monthly NDVI values, cropping patterns, and peak month information for each field. The annual map for 3 years showed changes in cropping patterns. These products were useful for detecting changes in cropland and confirming whether it was being cultivated. There was an increasing trend in the number of fields with sequential cropping from 2022 to 2024. Our results help comprehend the use and change of cropland spatiotemporally.

Acknowledgments: This research was funded by RDA, grant number PJ01676802.

How to cite: Choi, L.-Y., Ryu, J.-H., Ahn, H.-Y., Lee, S.-J., Kwak, G.-H., Jeon, Y.-A., and Lee, K.-D.: Development of Cropping Pattern Product Using Sentinel-2 Satellite Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8077, https://doi.org/10.5194/egusphere-egu25-8077, 2025.

Accurate cultivated land parcels (CLPs) information is essential for precision agriculture. Deep learning methods have shown great potential in CLPs delineation but face challenges in detection accuracy, generalization capability, and parcel optimization quality. This study addresses these challenges by developing a high-generalization multi-task detection network coupled with a specialized parcel optimization step. Our detection network integrates boundary and region tasks and design distinct decoders for each task, employing performance-enhancing modules as well as more balanced training strategies to achieve both accurate semantic recognition and fine-grained boundary depiction. To improve network's ability to train more generalized models, our study identifies the variations in image hue, landscape surroundings, and boundary granularity as the key factors contributing to generalization degradation and employ color space augmentation and attention mechanisms on spatial and hierarchy to enhance the generalization. Additionally, the parcel optimization step repairs long-distance boundary breaks and performs object-level fusion of delineated regions and boundaries, resulting in more independent and regular CLP results. Our method was trained and validated on GaoFen-1 images from four diverse regions in China, demonstrating high delineation accuracy. It also maintained stable spatiotemporal generalization across different times and regions. Comprehensive ablation and comparative experiments confirmed the rationale behind our model improvements and demonstrated our method's effectiveness over existing single-task models (SegNet, MPSPNet, DeeplabV3+, U-Net, ResU-Net, R2U-Net), and recent multi-task models (ResUNet-a, BSiNet, SEANet). 

How to cite: Zhu, Y. and Pan, Y.: A deep learning method for cultivated land parcels (CLPs) delineation from high-resolution remote sensing images with high-generalization capability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8391, https://doi.org/10.5194/egusphere-egu25-8391, 2025.

EGU25-8765 | ECS | Orals | ESSI4.11

Using Remote Sensing Spectral Image Dynamics for early prediction of biotic stress in wheat: lessons from Armenia and southern Russia 

Igor Sereda, Andrey Medvedev, Grigor Ayvazyan, and Shushanik Asmaryan

Winter and spring wheat are among key agricultural crops in the Republic of Armenia, and represent a significant share of grain production. However, their yield is threatened to substantially decline due to the negative impact of various biotic factors, including weeds and phytopathogens such as rust, powdery mildew, and tan spot. Remote sensing methods, particularly multitemporal dynamics of plant spectral imagery, offer opportunities for early detection and monitoring of these diseases. Early identification allows for timely management interventions to stabilize crop conditions, preserve yields, and enable mapping of problem areas before scheduled applications, allowing more effectively application of herbicides and fungicides.

Hyperspectral spectrometry of winter wheat crops under increased pathogen stress, together with control plots without increased pathogen stress, were studied in experimental fields in southern Russia (Krasnodar Krai) between 2017-2023. The results show that the temporal dynamics in reflectance during the spring-summer growth period of winter wheat likely indicate disease levels, where the period between stem elongation and heading was identified as crucial. A series of high-frequency spectral measurements (every 2–3 days) allowed the classification of areas with infected and healthy plants (accuracy of 70–88%) but also reasonably accurate predictions of the maximum development stage of various pathogens (R² = 0.48–0.55) 10–12 days before peak development. Moreover, these patterns were confirmed using data from ground-based spectrometry, UAVs, and satellite imagery.

Additionally, this methodology was tested on spring wheat fields in the Republic of Armenia (Aragatsotn, Nerkin Sasnashen) in 2024. Using a series of multitemporal UAV surveys, the fields were divided into zones based on the temporal behavior of spectral imagery that successfully identifies zones of weed emergence and negative consequences of agronomic errors. However, identification of more sensitive spectral regions with pathogen hotspots was hindered by the high heterogeneity of the fields.

Based on these methodologies, we defined the optimal dates for initiating phytosanitary monitoring for different regions in Armenia. This part of the investigation shows that zoning territories by the timing of the phenophase "stem elongation" with an error <10 days is crucial for the start of intensive spectral monitoring, and can be achieved by combining NDVI data with meteorological and topographical parameters.

Altogether, the results demonstrate the early diagnosis of biotic stress in plants is feasible using spectral data and can improve decision-making for field treatments in the long term. The early detection of biotic stress in plants enhances the potential of precision agriculture, as time is a crucial factor in addressing these challenges. Furthermore, the described methods have shown the capability to be scaled from local experiments, as is currently the case in most studies, to a regional scale.

How to cite: Sereda, I., Medvedev, A., Ayvazyan, G., and Asmaryan, S.: Using Remote Sensing Spectral Image Dynamics for early prediction of biotic stress in wheat: lessons from Armenia and southern Russia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8765, https://doi.org/10.5194/egusphere-egu25-8765, 2025.

EGU25-9499 | Posters on site | ESSI4.11

Spectral-based monitoring methods to optimise precision irrigation in maize 

Attila Nagy, Andrea Szabó, Gift Siphiwe Nxumalo, Erika Budayné Bódi, and János Tamás

Precision irrigation is one of the fundamental areas of modern agriculture that aims to manage water use more efficiently and sustainably. Continuous monitoring of crop status is essential for the optimisation of irrigation systems, in which spectral-based monitoring methods play a key role. These methods use the spectral properties of the light reflected or absorbed by plants to determine vegetation indices, soil moisture and other plant life parameters. Measurements in the optical and infrared (IR) wavelengths are particularly important as these wavelengths are sensitive to the biochemical and physical properties of plants, such as chlorophyll content, nitrogen levels and water content.

The primary aim of the study is to expand the area of remote sensing in agricultural monitoring using laboratory, field scale proximal sensors, field an UAV imaging by creating a new rapid non-invasive approach for predicting crop health and water demand using spectral data. The study seeks to close the gap where chlorophyll estimations are generally not plant-specific by offering an integrated and refined approach to improve reliability and accessibility in chlorophyll estimation. Besides Integrating VI and thermal imaging with UAV technology can be used in precision agriculture in a number of areas, such as crop monitoring, yield forecasting and optimisation of irrigation water allocation. Furthermore, using several VIs were found to be optimal in crop coefficient estimation, so as to more precise calculation of crop evapotranspiration The ultimate result is giving new approaches to farmers and agricultural stakeholders for more precise and dependable tools for measuring crop evapotranspiration, crop health while promoting sustainability, efficiency, and scalability in irrigation practices.

The research presented in the article was carried out within the framework of the Széchenyi Plan Plus program, with support from the RRF 2.3.1 21 2022 00008 project. This research was supported by the János Bolyai Research Scholarship of the Hungarian Academy of Sciences

 

How to cite: Nagy, A., Szabó, A., Nxumalo, G. S., Budayné Bódi, E., and Tamás, J.: Spectral-based monitoring methods to optimise precision irrigation in maize, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9499, https://doi.org/10.5194/egusphere-egu25-9499, 2025.

EGU25-9634 | ECS | Orals | ESSI4.11

Assessing generalization of deep learning models for crop classification under climatic variability in Denmark 

Muhammad Rizwan Asif, Mehdi Rafiei, Rasmus Nyholm Jørgensen, Michael Nørremark, and Nima Teimouri

This study explores the impact of climatic variability on the generalization capabilities of a deep learning model for pixel-level crop classification using multi-temporal Sentinel-1 SAR data in Denmark. With agriculture accounting for 61% of Denmark’s land area, accurate and timely crop mapping is essential for providing insights into crop distribution, offering valuable information to advisors and authorities to support large-scale agricultural management, and address challenges posed by changing climatic conditions.

Our study leverages a novel deep learning architecture that combines a 3-D U-Net with a conv-LSTM module to effectively capture both spatial and temporal dependencies in crop growth patterns. We consider 14 crop types over an eight-year period (2017–2024) and growth season (May to August), with ground truth data derived from Denmark’s Land Parcel Identification System (LPIS). Our analyses reveal that climatic variables such as precipitation, temperature, and humidity significantly influence model performance across years. Notably, extreme years like 2018 (characterized by drought and high solar radiation) and 2024 (marked by record precipitation) challenge the model’s ability to generalize effectively. By correlating inter-annual model accuracy trends with climatic data, the study demonstrates the necessity of incorporating environmental context into AI-driven agricultural monitoring systems.

We also evaluate the benefits of training the model on multi-year datasets to enhance robustness against climatic variability. Our findings reveal that leveraging temporal diversity improves model performance but highlights persistent difficulties in generalizing to outlier years with extreme climate conditions. While training on multi-year datasets helps capture a broader range of crop phenological variations, the results underscore that this approach alone is not sufficient, and underscores the importance of integrating auxiliary data, such as local climatic variables, to enable models to better adapt to evolving crop growth patterns influenced by changing environmental conditions.

This work represents one of the most comprehensive evaluations of deep learning for crop classification, spanning eight years and covering over 1.5 million hectares of agricultural land. By linking model performance to climatic variability, this study provides critical insights for improving the generalization capabilities of deep learning models in precision agriculture. These findings not only pave the way for enhanced crop monitoring under diverse climatic scenarios but also emphasize the potential of integrating climate-resilient AI technologies to address global agricultural and environmental challenges.

How to cite: Asif, M. R., Rafiei, M., Jørgensen, R. N., Nørremark, M., and Teimouri, N.: Assessing generalization of deep learning models for crop classification under climatic variability in Denmark, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9634, https://doi.org/10.5194/egusphere-egu25-9634, 2025.

EGU25-9672 | ECS | Orals | ESSI4.11

Barley Yield Estimation Using Regression Models and Spatial Pattern Analysis 

Faten Ksantini, Miguel Quemada, Andrés F. Almeida-Ñauñay, Ernesto Sanz, and Ana M. Tarquis

Precision agriculture (PA) has emerged as a key strategy for optimizing agricultural production. Using data-driven technologies such as sensors and satellite imagery, PA improves the efficiency of agricultural processes. Accurate crop yield estimation is an essential component of PA. An important aspect of yield estimation within PA is the ability to assess and map spatial variations in yield in an agricultural field. Understanding these spatial patterns enables more precise management decisions and targeted interventions.

Therefore, this study aimed to develop two regression approaches, multiple linear regression (MLR) and random forest regression (RFR), to estimate crop yield using sixteen input variables with a 6 m resolution. These variables were obtained using different sensors, reflecting the soil and crop spatial variability. The estimation performance of the studied approaches was assessed using the coefficient of determination (R²), showing very satisfactory results (R² > 0.85) for both approaches.

The spatial distribution of barley yield was assessed, focusing on identifying areas of high and low productivity within the field. RFR demonstrated its ability to capture yield patterns. By incorporating spatial factors, RFR effectively modelled the varying yield potential in the crop field.

 

Keywords—precision agriculture, multiple linear regression, random forest regression, spatial pattern, barley

 

Acknowledgments: Financed by Ministerio de Ciencia e Innovación, Spain (PID2021-124041OB-C22)

 

How to cite: Ksantini, F., Quemada, M., Almeida-Ñauñay, A. F., Sanz, E., and Tarquis, A. M.: Barley Yield Estimation Using Regression Models and Spatial Pattern Analysis, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9672, https://doi.org/10.5194/egusphere-egu25-9672, 2025.

EGU25-9872 | ECS | Orals | ESSI4.11

Remote Sensing-based Wheat Area and Yield Estimation: Insights from Uttarakhand, India 

Priya Singh and Kritika Kothari

India is one of the world's leading exporters of wheat grain, making monitoring its growth and yield one of the country's top economic priorities. This study aimed to develop a methodology for delineating wheat cultivation areas and estimating wheat yields using Landsat 8 (30 m spatial resolution) data for the Nainital and Udham Singh Nagar districts of Uttarakhand, India. The cultivated wheat fields were identified using a supervised classification-based Random Forest (RF) algorithm during the growing season from November 2020 to April 2021. To characterize the wheat class, a total of 239 and 226 wheat points, along with 201 and 166 non-wheat geometry points, based on NDVI time series were allotted for Nainital and Udham Singh Nagar districts, respectively. The calculated wheat area was found to be 778.94 sq. km and 209.48 sq. km, compared to the actual reported areas by the Agriculture Department, Government of Uttarakhand of 1059.61 sq. km and 212.78 sq. km for Udham Singh Nagar and Nainital, respectively. The RF algorithm showed an underestimation for both districts, achieving a kappa coefficient of 0.97, producer accuracy of 0.97, user accuracy of 0.96, and overall accuracy of 0.98 for the Nainital district. For the Udham Singh Nagar district, the kappa coefficient was 0.89, with producer accuracy of 0.89, user accuracy of 0.93, and overall accuracy of 0.93. The study also utilized weather data along with Landsat 8 imagery as inputs for the Carnegie-Ames-Stanford Approach (CASA) to estimate wheat yields and get spatial wheat yield maps. The estimated mean yields were 3.73 t ha⁻¹ and 3.37 t ha⁻¹, whereas the actual mean yields were 3.82 t ha⁻¹ and 4.45 t ha⁻¹ for Nainital and Udham Singh Nagar districts, respectively. The study demonstrates the potential of combining remote sensing and supervised classification techniques for reliable wheat yield estimation in data-scarce regions, which can be a promising tool for agricultural policy and decision-making.

Keywords: Crop classification, Landsat 8, random forest, wheat, spatial yield map 

How to cite: Singh, P. and Kothari, K.: Remote Sensing-based Wheat Area and Yield Estimation: Insights from Uttarakhand, India, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9872, https://doi.org/10.5194/egusphere-egu25-9872, 2025.

EGU25-10478 | ECS | Posters on site | ESSI4.11

Correlation between NDVI and soil sensor data collected by UAV 

Andrea Szabó, Erika Budayné Bódi, Ademola Blessing Blessing, Sándor Kun, Éva Nikolett Kiss, János Tamás, and Attila Nagy

The development of UAVs and the reduction in the weight of payload-bearing devices is making remote sensing of crops possible. This technology is cheaper, more time-efficient and produces higher resolution images in a non-destructive way. Another important feature of drone imagery is its ability to monitor crops on a regular basis. The raw data collected by drones can be integrated into models for analysis and further corrective measures can be created to improve crop yields. Drones are capable of assessing soil conditions, assisting in irrigation, fertilizer application and monitoring crop health. The Normalized Difference Vegetation Index (NDVI) was used to quantify the greenness of vegetation to assess changes in vegetation density and health. When near-infrared light reaches the leaves of a healthy plant it is reflected back into the atmosphere, as the amount of chlorophyll produced by the plant decreases, less near-infrared radiation is reflected back. The result can then be used to assess the overall health of the plant. The values are calculated for each pixel of your map, giving you an index in the range -1 to 1.

 

4 sampling points (A-D) were selected in the sample area Nyírbator, Hungary. Soil moisture and soil temperature probes were deployed at three depths in the points and data were downloaded during bi-weekly sampling and measurements. The vegetation monitoring of the irrigated and non-irrigated area was carried out by taking NDVI images every 2 weeks using UAV remote sensing. During the NDVI processing of the irrigated area, only the first half of the area was captured for the initial images, at the beginning of the vegetation. NDVI images were processed in Pix4D and ArcGIS Pro software. In ArcGIS Pro, the minimum, maximum, mean and standard deviation values for the study area were observed and subsequently evaluated separately point by point using a zonal statistics algorithm.

 

In the study area, a larger temperature variation is observed for the deployed soil probes at a depth of 10 cm, which underlines the sensitivity of the surface temperature to environmental conditions. With increasing depth, a gradual decrease in temperature is observed, indicating the influence of soil properties on heat retention and dissipation. Consistently fluctuating moisture levels near the surface (at a depth of 10 cm) were observed in response to precipitation or irrigation events. The fluctuation of the curves gradually decreases with increasing depth. At all depth levels, a more consistent linear gradient is observed, reflecting the prolonged drought conditions in the soil. This observation is consistent with the low mean NDVI values observed simultaneously in the same zone. The data show that the irrigated area tends to have higher average NDVI values than the non-irrigated area, which has significantly lower values.

 

 

 

 

This research was supported by the János Bolyai Research Scholarship of the Hungarian Academy of Sciences. The research presented in the article was carried out within the framework of the Széchenyi Plan Plus program with the support of the RRF 2.3.1 21 2022 00008 project.

How to cite: Szabó, A., Budayné Bódi, E., Blessing, A. B., Kun, S., Kiss, É. N., Tamás, J., and Nagy, A.: Correlation between NDVI and soil sensor data collected by UAV, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10478, https://doi.org/10.5194/egusphere-egu25-10478, 2025.

EGU25-11964 | Orals | ESSI4.11

What influences alpine pasture productivity? Exploring the relation among topography, climate, and biomass using remote sensing. 

Fabio Oriani, Helge Aasen, Manuel Schneider, and Pierluigi Calanca

Mountain pastures are a biodiversity rich and heterogeneous ecosystem of the Alps influenced by a complex topography and a variable climate. Understanding the impact of these factors on pasture productivity is of primary importance for forage production and ecosystem preservation.

We present here a regional analysis covering the alpine pastures in the Grisons Canton (eastern Switzerland, 1997 sq. km), for which we developed a collection of high-resolution (10-m) annual growth indicators based on the Enhanced Vegetation Index (EVI) derived from Sentinel-2 images, from 2016 to 2024. We correlate our growth maps to a 1-km gridded climate dataset (Meteoswiss) and a 10-m digital elevation model (Swisstopo) to understand which weather factors - rainfall, temperature, or radiation - influence the most the growing season and from which period of the year. In addition, we explore the variability of these dependencies in space, in relation to elevation and derived topographic descriptors, e.g. slope or valley orientation.

This analysis shades light on the climate dynamics impacting the most the growing season in conjunction to a complex local topography. The results can be used to identify vulnerability levels along the elevation profile, influenced by soil depth and valley orientation, where growth varies the most from year to year in function of annual weather variations. In these zones, pasture management will need extra flexibility measures and real-time monitoring to adapt to annual fluctuations of a future climate change.

How to cite: Oriani, F., Aasen, H., Schneider, M., and Calanca, P.: What influences alpine pasture productivity? Exploring the relation among topography, climate, and biomass using remote sensing., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11964, https://doi.org/10.5194/egusphere-egu25-11964, 2025.

Understanding and monitoring crop growth is crucial for addressing global food security challenges and promoting sustainable agricultural practices. Traditional methods of observing crop traits in plot experiments are labor-intensive, limiting their spatial and temporal resolution. While conventional satellite platforms like Sentinel-2 and Landsat have proven valuable for large-scale agricultural monitoring, their spatial resolutions and temporal gaps are insufficient for time series of small experimental plots. Recent advancements, such as PlanetLabs’ SuperDove constellation, provide an alternative by offering daily imagery at a 3 m resolution, making them suitable for small-scale plot-level analysis. Despite their high spatial detail, these images face challenges related to radiometric stability, spatial co-registration accuracy, and quality masks, which must be resolved for effective small-scale monitoring. Addressing these limitations, this research investigates the use of PlanetScope data to estimate canopy cover (CC) and leaf area index (LAI) in plot experiments. High-resolution Unmanned Aerial System (UAS) RGB imagery was used as a reference to estimate early-stage CC. By applying a machine learning-based segmentation technique, we distinguished foliage from background pixels. This segmentation enabled us to integrate UAS-derived CC estimates with 8-band multispectral imagery from PlanetLabs’ SuperDove constellation. After improving the radiometric stability and spatial accuracy of the satellite imagery, we used the multispectral data along with UAS-derived canopy cover estimates as inputs to identify the most sensitive satellite-derived vegetation indices (VIs) for estimating CC during the early growth stages. In conjunction with LAI, we generated model-based time-series growth curves covering all phenological stages. The method was validated on experimental plots in northern Switzerland, with varying soil compaction and fertilization treatments. The study demonstrates successful segmentation of high-resolution UAS-based RGB imagery, providing a robust baseline for validating satellite-derived data and training novel retrieval methods for canopy cover. Comparative analyses identify vegetation indices from PlanetScope imagery that correlate with early crop growth. This research highlights the potential of high-resolution satellite data for generating time-series growth curves, offering a valuable tool for improving crop management and optimizing resource use across diverse farming systems.

How to cite: Boos, T. and Aasen, H.: Using High-Resolution Satellite Data to Estimate Canopy Cover and Leaf Area Index in Plot Experiments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12456, https://doi.org/10.5194/egusphere-egu25-12456, 2025.

EGU25-12606 | ECS | Orals | ESSI4.11

Comparison of zonation approaches by means of remote sensing vegetation indices for agricultural applications  

Gunay Hasanli, Sadra Emamalizadeh, Riccardo Mazzoleni, and Gabriele Baroni

Remote sensing vegetation indices play a vital role in agricultural zoning by providing detailed insights into crop health, productivity, and environmental conditions. They enable researchers and professionals to monitor environmental changes, urban expansion, and natural events with exceptional accuracy and precision. This progress has been fueled by major technological developments in satellite sensors, data processing algorithms, and analytical methods, enabling the capture of more detailed information and increased observation frequency across expansive regions. Despite these excellent opportunities, numerous image processing techniques have been suggested, each customized for particular applications, datasets, and user needs, yet no widely recognized standard methods have been established. This absence of standardization creates difficulties of interoperability, reproducibility, and consistency in analytical results. Researchers and practitioners frequently encounter challenges choosing the most suitable methods, since the effectiveness of these techniques can fluctuate based on factors like spatial resolution, temporal frequency, and the type of landscape under examination. As a result, there is an increasing demand for the creation of thorough guidelines and uniform procedures that can facilitate the use of remote sensing instruments while ensuring dependable and comparable outcomes across various studies and fields. In this research, we analyze zonation outcomes obtained from remote sensing images captured at different times, using several vegetation indices and applying various clustering techniques. The objective is to evaluate how time-related changes, the selection of vegetation indices, and the use of different clustering methods affect the precision and dependability of land classification. Through the examination of these combination performance, this comparative examination underscores both the advantages and drawbacks of each approach while offering important insights for improving classification methods in varied and changing environments.

How to cite: Hasanli, G., Emamalizadeh, S., Mazzoleni, R., and Baroni, G.: Comparison of zonation approaches by means of remote sensing vegetation indices for agricultural applications , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12606, https://doi.org/10.5194/egusphere-egu25-12606, 2025.

EGU25-13239 | Posters on site | ESSI4.11

LED-induced chlorophyll fluorescence during heat and drought stress as assessed in a microcosm experiment on sunflower 

Szilvia Fóti, Ádám Mészáros, Islam Guettala, Evelin Péli, Krisztina Pintér, Zoltán Nagy, and János Balogh

Like sun-induced fluorescence (SIF), LED-induced fluorescence (LEDIF) became frequently used to establish and analyze leaf- and canopy-level stress responses. Different plant phenotypes (trees, understory shrubs, crops, vineyards, etc.) were subjected to, in most of the studies, blue LED illumination during the night or in darkened boxes for assessing either the entire broad-band (650-850 nm) spectrum of LEDIF or one of the wavelength bands of the red (~ 690 nm) and far-red (~ 740 nm) peak emissions. It seems to be however less common to apply close to “white” LED lighting, mixed from different wavelength ranges all below 650 nm (to overcome spectral overlap of red excitation and emission) as a light source. Moreover, stress manipulation in microcosm experiments is also scarce within studies while detecting LEDIF signal changes.

In our study, we established a microcosm experiment with four treatments on sunflowers: well-watered – no heat stressed, well-watered – heat stressed, water-stressed – no heat stressed, and water-stressed - heat stressed. The plants were gradually exposed to the treatments during the two months of the experiment between October and December 2024. We captured reflectance and the broad-band fluorescence spectra above the canopy with a VIS-NIR spectrometer facing downwards toward the canopy between the LED panels. We followed the response of the plants to the imposed stress by weekly/bi-weekly measurements and analyzed the changes in the shapes of the curves. We also captured the canopy architecture with side-view photos and leaf area growth with top-view photos. There was a clear increase in the LEDIF signal during the canopy development, and then a heterogeneous response depending on the treatment.

How to cite: Fóti, S., Mészáros, Á., Guettala, I., Péli, E., Pintér, K., Nagy, Z., and Balogh, J.: LED-induced chlorophyll fluorescence during heat and drought stress as assessed in a microcosm experiment on sunflower, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13239, https://doi.org/10.5194/egusphere-egu25-13239, 2025.

Accurate and timely seasonal yield predictions before harvest are becoming ever more relevant due to increasing pressure on the agricultural sector under climate change. Especially for agricultural planning, logistics, and food markets, seasonal predictions are of significant importance in the context of food security and price stability.

A novel approach to enhance early-season yield forecasts at the regional scale will be presented. Earth observation (EO) data from the Copernicus Sentinel-3 satellite are able to trace spatio-temporal vegetation dynamics (e.g., crop phenological status, crop growth, photosynthesis via FAPAR, or chlorophyll indices) in near real-time. By deriving daily satellite composites and combining these data with physical modelling using the Lund-Potsdam-Jena managed Land (LPJmL) dynamic global vegetation model (DGVM) in a newly developed assimilation process, enhanced yield forecasts can be achieved. There are currently no interfaces for continuous assimilation of EO data for the LPJmL model, thus, approaches such as parameter forcing and ensemble methods allowing for continuous parameter optimization during the course of the growing season are presented and compared conceptually to improve the LPJmL model for seasonal yield predictions. Existing methods for model parameter calibration and optimization with EO data using machine learning are applied to agricultural areas in the study area.

While these results focus on the study area of Bavaria, southern Germany, the approach is scalable also on national or European scale. For demonstration purposes, the year 2018 – a comparably dry year – was chosen due to the availability of detailed land use data. LPJmL was designed for global simulations, hence, a regional downscaling is necessary for its application at the regional scale.

Integrating different remote sensing data sources enables a more detailed picture of plant growth, which will allow a regional early warning system for food security and farmer’s turnover in the future. The combination of process- and data-based approaches is likely to improve accuracy and lag time.

How to cite: Jörges, C., Hank, T., and Fader, M.: Chances and Challenges of Data Assimilation for Seasonal Yield Predictions Using Sentinel-3 Satellite Data and the Agro-Ecosystem Model LPJmL, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13318, https://doi.org/10.5194/egusphere-egu25-13318, 2025.

EGU25-14174 | Posters on site | ESSI4.11

Mapping Soil Organic Carbon Dynamics in Taiwan’s Agricultural Land Using Field and Remote Sensing Data. 

Miguel Conrado Valdez Vasquez, Chi-Farn Chen, Jien-Hui Syu, and Liang-Chien Chen

Soil organic carbon (SOC) stocks represent the second-largest natural carbon reservoir globally, surpassed only by the oceans. SOC plays a vital role in maintaining ecosystem health, offering numerous benefits such as enhancing soil structure, increasing nutrient availability, and boosting water retention capacity. Beyond its ecological significance, SOC is integral to climate change mitigation, given its ability to sequester atmospheric carbon dioxide effectively. Additionally, SOC contributes to improving the physical, chemical, and biological properties of soil, making it indispensable for sustainable land management. Taiwan, an island in the western Pacific Ocean, spans an area of approximately 35,800 square kilometers. Shaped like a tobacco leaf, the island extends 400 kilometers in length and 150 kilometers at its widest point. Taiwan’s landscape is characterized by a Central Mountain Range running north to south, steep slopes, and geologically fragile formations. In recent decades, Taiwan has experienced significant changes in land use and land cover, particularly in urban areas where cropland and forest land on city outskirts have been replaced by infrastructure development. These transformations have directly impacted SOC levels across the island, underscoring the need for accurate mapping to estimate SOC stocks and assess soil functionality, particularly in agricultural regions. Traditional ground sampling methods for estimating SOC, though precise, are often costly and labor-intensive. To address these limitations, alternative approaches, such as remote sensing, offer cost-effective solutions. Among various predictive modeling techniques, machine learning algorithms like Random Forest (RF) have emerged as highly effective tools for SOC estimation. RF models excel due to their ability to minimize correlation among individual decision trees and provide reliable error estimates, ensuring robust predictions.

In this study, we combined field sampling data (2010–2021) with remote sensing, topographic, and climatic datasets to estimate SOC stocks in the topsoil layer (0–30 cm) of Taiwan’s agricultural areas. Using the RF algorithm, we initially employed 23 explanatory variables and subsequently refined the model by eliminating less significant predictors, reducing the final set to 12 variables. The refined model demonstrated strong predictive accuracy, with R² values exceeding 0.70 for agriculture land in Taiwan. Our findings revealed spatial variations in SOC levels, with mountainous regions exhibiting higher SOC stocks compared to suburban and low-lying agricultural areas, where values were notably lower. SOC levels for agricultural lands ranged from a maximum of 7.14 kg/m² to a minimum of 2.55 kg/m², with an average value of 3.43 kg/m². Agricultural practices incorporating agroforestry techniques showed relatively higher SOC stocks, emphasizing the role of sustainable practices in enhancing soil carbon storage. The results of this study hold significant implications for long-term monitoring of SOC in Taiwan and provide a crucial reference for policymakers aiming to develop effective carbon sequestration strategies. By integrating field data with advanced modeling and remote sensing technologies, this research contributes to a deeper understanding of SOC dynamics and supports efforts to promote sustainable land management and climate resilience.

How to cite: Valdez Vasquez, M. C., Chen, C.-F., Syu, J.-H., and Chen, L.-C.: Mapping Soil Organic Carbon Dynamics in Taiwan’s Agricultural Land Using Field and Remote Sensing Data., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14174, https://doi.org/10.5194/egusphere-egu25-14174, 2025.

Air pollution, particularly surface ozone, has become a significant threat to agriculture in China, severely impacting the productivity of essential staple crops like winter wheat. However, the spatiotemporal variability of ozone concentrations and its interactions with other environmental factors—such as temperature and droughts—remain inadequately understood regarding their impact on agricultural productivity. To address this gap in knowledge, this study integrates multi-source remote sensing data with advanced statistical analysis and machine learning techniques to quantitatively examine the spatiotemporal variation of ozone pollution and its interactions with climate change and other environmental factors on winter wheat productivity.

The study first employs the Geographically and Temporally Weighted Regression (GTWR) model, utilizing high-resolution remote sensing data from 2013 to 2019, to assess the spatiotemporal response of winter wheat productivity to ozone pollution. To further investigate the interactions between ozone and other environmental factors, an interpretable machine learning framework is applied, specifically using the eXtreme Gradient Boosting (XGBoost) algorithm augmented by SHapley Additive exPlanations (SHAP). Additionally, a structural equation model is developed to elucidate the underlying mechanisms of these interactions. The results indicate that the negative impact of surface ozone on winter wheat has intensified annually, with significant spatial variation. Particularly in high-pollution areas, such as eastern Henan and northern Anhui provinces, the effects of ozone on winter wheat are most pronounced. Furthermore, the study reveals that the impact of ozone on winter wheat productivity varies across different growth stages, with the most severe effects observed during the later stages in May. Additionally, the research reveals the complex interactions between ozone and other environmental factors, such as temperature and aerosol concentration. Notably, the harmful effects of ozone are exacerbated under conditions of high aerosol concentration and elevated temperatures. Interestingly, drought conditions were found to partially mitigate the negative impact of ozone on productivity.

This study provides a systematic and actionable analytical framework for quantitatively evaluating the effects of ozone pollution and its interactions with climate change and other environmental factors on crop productivity. The findings underscore the need for targeted agricultural measures and pollution control strategies, particularly in high-pollution regions and during critical growth stages. These results provide theoretical support for sustainable agricultural development and climate adaptation management. Furthermore, the study contributes valuable insights into the application of remote sensing technology for large-scale agricultural monitoring, thereby enhancing the management efficiency and adaptive capacity of agricultural ecosystems in response to environmental challenges.

How to cite: Du, C.: Evaluating Air Pollution Impacts on Agricultural Productivity in China: Insights from Remote Sensing Data and Geospatial Modeling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14198, https://doi.org/10.5194/egusphere-egu25-14198, 2025.

EGU25-14688 | Posters on site | ESSI4.11

Monitoring Methane Emissions from Rice Paddies in Middle Taiwan Using Remote Sensing Data. 

Cheng-Ru Chen, Chi-Farn Chen, Nguyen-Thanh Son, Liang-Chien Chen, Tsang-Sen Liu, and Yao-Cheng Kuo

Methane (CH₄) emissions from paddy rice fields significantly contribute to greenhouse gas emissions and global climate change. In Taiwan, rice cultivation occupies approximately 20% of agricultural land. This study utilizes Sentinel-2 and Sentinel-5P satellite data to monitor methane emissions from these fields. The research follows four key steps: 1) classifying rice cropping areas; 2) detecting the phenological stages of rice; 3) correlating spatial and temporal data with rice cultivation and methane emissions; and 4) validating the results with in-situ data. The preliminary findings identify methane emission hotspots during the rice-growing seasons, revealing substantial temporal variability linked to agricultural practices such as water management, organic matter application, and rice phenology. Peak emissions occur during the early to mid-growing stages. The adoption of satellite data for monitoring emissions offers a cost-effective and scalable alternative to traditional methods, which are often labor-intensive and geographically limited. The research can also enhance the sustainable agricultural management strategies for achieving local greenhouse gas reduction targets.

How to cite: Chen, C.-R., Chen, C.-F., Son, N.-T., Chen, L.-C., Liu, T.-S., and Kuo, Y.-C.: Monitoring Methane Emissions from Rice Paddies in Middle Taiwan Using Remote Sensing Data., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14688, https://doi.org/10.5194/egusphere-egu25-14688, 2025.

EGU25-14999 | Orals | ESSI4.11

Early prediction of within-field variability wheat productive potential using Sentinel2 satellite data. 

Elena Pareja-Serrano, José González-Piqueras, and André Chanzy

Assessing agricultural production in the context of climate change is a global concern. In the recent decades, variable rate technology (VRT) for agricultural machinery has made it possible to adjust fertiliser rates on-the-go, allowing the within-field crop management. In this context, in order to select the most effective management practices, it is essential to identify the driving factors that determine yield variability, mapping the spatial distribution of these driving factors and to determine the local yield variability potential.

Mapping the homogeneous within-field areas of yield potential is used to define management zones. Remote sensing data provide a practical means of delineating these zones. The crop biophysical variable, cumulative evapotranspiration (ETccum), derived from NDVI time series and climate data, was analysed to evaluate its ability to estimate yield. In the semi-arid conditions of the Spanish Central Plateau, wheat ETccum maps were correlated with yield maps by non-linear regression with an R2 of 88%. ETccum serves as an effective proxy for yield estimation and the statistical analysis to determine the level of homogeneity within the field, the driving factors that determine yield variability, and mapping the spatial distribution of these driving factors. Nevertheless, the observed saturation effect in the biophysical variable highlights limitations that require further analysis.

Additionally, during the wheat season, expected potential yields can fluctuate in response to actual weather conditions. Consequently, updating yield predictions early in the season is critical for informed irrigation and fertilisation management decisions. The ability of ETccum to forecast yields at early phenological stages, such as flag leaf and flowering—key stages for yield formation—is examined. Finally, the stability of spatial variability patterns, compared to those derived from ETccum at maturity, is analysed as an indicator of the spatial distribution of yield drivers.

Acknowledgments: this work was supported by the research project NSBOIL (Horizon, GA 101091246).

How to cite: Pareja-Serrano, E., González-Piqueras, J., and Chanzy, A.: Early prediction of within-field variability wheat productive potential using Sentinel2 satellite data., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14999, https://doi.org/10.5194/egusphere-egu25-14999, 2025.

Effectively tracking drought effects using satellite data can be conducted by combining atmospheric data with additional information of vegetation indices (VIs) from optical data. While VIs detect drought when plant damage is often irreversible, information about the plant physiological status can help detect drought effects much earlier. Remotely-sensed solar-induced chlorophyll fluorescence (SIF), emitted directly from the photosynthetic apparatus (Drusch et al., 2017), provides such information.  When abiotic stress occurs due to an increased dissipation of thermal energy through the process of non-photochemical quenching (NPQ), the fluorescence yield is decreased, which can be measured as SIF (Berger et al., 2022, Damm et al., 2018).

Top of canopy (TOC) SIF is available from Sentinel-5P’s TROPOMI sensor since 2018 (Guanter et al., 2021, Köhler et al., 2018). This data, however, is affected by incoming radiation and canopy structure. These effects need to be removed In order to calculate the fluorescence yield in form of the quantum efficiency at leaf level (ΦF), which provides the pure information on the actual physiological status of the plant. Equation (1) uses the vegetation index NIRv (NDVI*NIR (Badgley et al., 2017)) to serve as a combined proxy of the fraction of absorbed photosynthetically active radiation (fAPAR) and the fluorescence escape probability (fesc) (Dechant et al. 2020, Liu et al. 2023). Both SIF data at 743 nm and the reflectance used to calculate the NIRv come from TROPOMI, while the photosynthetically active radiation (PAR) is provided by MODIS.

ΦF = π*SIF743canopy/(NIRv*PAR) (1)

This study presents a new multi-year (2018-2023) ΦF dataset at 0.05° resolution covering Germany with daily temporal resolution. To assess ΦF’s potential as an early drought stress indicator for agricultural and forest ecosystems, it is compared to the anomaly of subsurface water storage (sss), which serves a reference parameter for plant water availability generated by combining the hydrological model PARFLOW and common land model (CLM) (Belleflamme et al., 2023). ΦF and sss anomaly data were split into periods of prolonged negative sss anomaly indicating drought events (cross-referenced as watch/warning periods using the Combined Drought Indicator (European Commission)). Cross-correlation coefficients for different time lags were calculated to compare the datasets. The data was spatially aggregated daily and temporally averaged using a two-day rolling average.

Results show that cross-correlation coefficients for ΦF and sss anomaly are highest at a 2-day lag, dropping again after 3 days, indicating that ΦF follows the negative sss anomaly trend with a 2-day delay in both agricultural and forest ecosystems. Non-normalized canopy SIF and vegetation indices (NIRv, NDVI) showed no pattern and low cross-correlation coefficients during the observed periods. Our findings prove that ΦF has the ability to detect insufficient plant water availability and thus can be used for early drought stress detection in agricultural and forest ecosystems. The comparison of the capabilities of ΦF and TOC SIF to track short-term changes in subsurface water storage illustrates that a proper downscaling and normalization of canopy SIF is essential to use SIF satellite measurements for the early detection of drought events.

How to cite: Herrera, D., Rascher, U., Belleflamme, A., and Siegmann, B.: On the Potential of a Novel Satellite-Based Time-Series of Normalized Far-Red Solar-Induced Fluorescence to Track Short-Term Changes in Subsurface Water Storage, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15038, https://doi.org/10.5194/egusphere-egu25-15038, 2025.

EGU25-15653 | ECS | Posters on site | ESSI4.11

Comparing Different Unmixing Methods for weed detection and identification 

Inbal Ronay, Ran Nisim Lati, and Fadi Kizel

Herbicides are extensively used for weed management worldwide. However, their use is a significant cause 
of environmental pollution and human health problems. Efficient Site-Specific weed management (SSWM) 
practice attempts to reduce herbicide use and its negative impacts by adjusting herbicide application based 
on weed composition and coverage. Such an application requires high-resolution data in spatial and spectral 
domains, which is not always available. Consequently, Mixed pixels are likely to exist, creating a challenge 
to generate accurate weed maps. In this regard, Spectral Mixture Analysis (SMA) can mitigate this challenge
by exploiting subpixel information. This study assesses the potential benefits of four SMA methods for 
estimating weed coverage of different botanical groups. We examined four methods- Constrained Least 
Squares Unmixing (FCLSU), Sparse Unmixing via variable Splitting and Augmented Lagrangian (SUnSAL), 
Sparse Unmixing via variable Splitting and Augmented Lagrangian and Total variation (SUnSAL-TV) and 
the Vectorized Code Projected Gradient Descent Unmixing (VPGDU). Each suggests a distinct advantage 
for spectral unmixing. We used controlled hyperspectral and multispectral field datasets to compare the four 
methods. The controlled data included weed species characterized by distinct botanical groups, while the 
field dataset included a corn field with weeds at varying densities. We assessed the performance of the 
different methods in estimating weed coverage and composition at various spatial resolutions. Our results
demonstrated the advantages of the total variation regularization of SUnSAL-TV and the superiority of the 
SAM-based method, VPGDU, over other approaches. VPGDU was the best-performing method, with MAE 
values consistently lower than 8.6% at all resolutions, underscoring the advantage of its objective function 
in unmixing weed botanical groups and the significant effect of illumination on the results. This result was 
also consistent in the field data as VPGDU yielded the lowest MAE of 11.95%,

How to cite: Ronay, I., Lati, R. N., and Kizel, F.: Comparing Different Unmixing Methods for weed detection and identification, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15653, https://doi.org/10.5194/egusphere-egu25-15653, 2025.

EGU25-15760 | ECS | Orals | ESSI4.11

Generally applicable method for unsupervised weed detection in row crops using UAV-based high-resolution RGB imagery 

Ambroos Van Poucke, Jan Verwaeren, and Wouter Maes

Advancements in sensing technology and in machine and deep learning have expanded UAV remote sensing applications in agriculture. Most of these applications rely on supervised techniques, but generalization remains a critical and underexplored challenge. Agricultural datasets often exhibit variability across fields, sensors, crops and growth stages. While models such as convolutional neural networks (CNNs) perform well when trained on millions of samples, this approach is impractical with UAV-based agricultural data. This suggests that a location-specific, unsupervised approach might be more effective.

This study proposes a generally applicable method to map weed densities in row crops using high resolution RGB UAV data. The workflow first starts with a vegetation masking based on the Excess Green index, followed by a novel row detection model that separates intra- and interrow vegetation. Pseudo-labels generated from this step are used to train the CNN segmentation model Deeplabv3.

The method was applied on 12 maize datasets collected across multiple locations in Belgium, at different growth stages, and using three different UAV cameras, leading to ranges in ground sampling distance (GSD). The model was also applied on a public sugar beet dataset, PhenoBench, covering 3 dates was used to validate the model. Model performance was evaluated against manually annotated ground truth segmentation maps from each field (n = 50).

Semantic segmentation of crops achieved consistent mean Intersection over Union (IoU) values, exceeding 0.7 (F1-score > 0.89). Weed detection performance was relatively low in very early growth stages (IoU>0.4, F1-score > 0.6) due to limited plant sizes, but improved as weeds grew, with IoU reaching 0.63 (F1-score = 0.83) in later stages. The model was equally performant on maize and on sugar beet.

Despite these early-season limitations, the lower weed detection accuracy had minimal impact on field-level weed density maps, which are primarily used for relative density comparisons to guide site-specific herbicide applications. Regression analyses of predicted crop and weed areas against ground truth annotations showed strong linear relationships. Early-season datasets exhibited slight underestimates of weed area, whereas later-season datasets demonstrated a near-perfect 1:1 relationship (R² > 0.80). GSD proved to be a reciprocal indicator of accuracy, with the highest accuracy at GSDs below 1mm/pixel. GSD above 3 mm/pixel showed a rapid decrease in accuracy.

Overall, the proposed approach effectively generates accurate field-level weed density maps, offering a robust tool for precision weed management in agriculture.

How to cite: Van Poucke, A., Verwaeren, J., and Maes, W.: Generally applicable method for unsupervised weed detection in row crops using UAV-based high-resolution RGB imagery, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15760, https://doi.org/10.5194/egusphere-egu25-15760, 2025.

EGU25-16547 | ECS | Orals | ESSI4.11

Knowledge-encoded deep fusion for yield estimation under extreme climate stress 

Xingguo Xiong, Renhai Zhong, Qiyu Tian, Ioannis Athanasiadis, and Tao Lin

Accurately modeling the impacts of climate stress on crop growth and yield is crucial for ensuring food security. Data-driven models are increasingly utilized for yield estimation because they can learn effective crop growth features from vast amounts of remote sensing and meteorological data. However, extreme climate stress conditions have few yield labels available for these models to modeling the interaction in crop responses. The response of crops to extreme climate stress often exhibits varied delays which are captured in remote sensing observations. In this study, we explicitly encode the time lag effect quantified by remote sensing and climate stress indicators into a two-stream fusion framework for estimating crop yield under extreme climate stress. Each stream employs a pyramid structure that progressively aggregates remote sensing and climate time series into feature embeddings. A time-lag-encoded cross attention mechanism fuses feature embeddings between the two streams, while phenology-sensitivity-guided linear attention is applied on top of the pyramid structures for processing ultimate time-lag encoded features. The proposed model is evaluated across nine Midwestern states within the US Corn Belt at the county level from 2006 to 2012, simulating climate stress situations with fewer samples. End-of-season results demonstrate that the knowledge-encoded two-stream model (RMSE=1.17 Mg ha-1) outperforms both the feature-stacking-based two-stream model (RMSE=1.43 Mg ha-1) and random forest (RMSE=1.68 Mg ha-1) under extreme climate stress. The improved estimation performance indicates that knowledge-encoded data fusion is more effective than merely stacking multi-source input data. In-season results reveal that our model proficiently captures extreme events and effectively predicts yield 8 weeks in advance. The time-lag knowledge could be extended to other forms of climate stress. Also, cross attention enables integration with additional data sources to enhance the interaction modeling of complex biomass accumulation and yield formation.

How to cite: Xiong, X., Zhong, R., Tian, Q., Athanasiadis, I., and Lin, T.: Knowledge-encoded deep fusion for yield estimation under extreme climate stress, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16547, https://doi.org/10.5194/egusphere-egu25-16547, 2025.

EGU25-16843 | Orals | ESSI4.11

Explainable Machine Learning for Forest Fire Detection with Remote Sensing for Effective Rescue Planning 

Octavian Dumitru, Chandrabali Karmakar, and Shivam Goyal

In the present decade, forest fires have become more common than ever [1]. Efficient strategies to cope with fire situations, and/damage assessments need efficient automatic forest fire detection model. In this research, we propose an unsupervised eXplainable machine learning model to assess the severity of forest fire with remote sensing data. The model, namely, Latent Dirichlet Allocation is a Bayesian Generative model, is capable of generating interpretable visualizations. LDA uncertainty quantifiable and explainable [2]. We do not need labelled data to train the model. Other usefulness of the model is that it is simple to combine any kind of input data (for example, UAV images, wind speed information). In the scope of this contribution, we use Sentinel-2 spectral bands to extract information to compute indices indicating severity of fire [1]. Uncertainty of each prediction of the model is computed to ascertain robustness of the model. As a use case, we have chosen the recent forest fire incident at Los Angeles, USA [6].

The methodological approach is as the following:

1) we acquire pre-fire, post-fire Seintinel-2 images, 2) compute three indices : Normalized Difference Vegetation Index (NDVI), Normalized Burn Ratio (NBR), and Burned Area Index for Sentinel (BAIS) based on state of the art literature and generate index maps, 3) compute difference between the pre-fire and post-fire index maps, 4) apply the unsupervised xAI LDA model to retrieve semantic classes in pre-fire and post-fire Sentinel-2 band images, general corresponding classification maps and plot a binary class-to-class change map,  5) Analyze the maps with visual tool to find the most affected semantic classes (e.g., dense vegetations, urban areas etc.) and produce a data-driven estimation of per-class changes due to fire [7].

In future, we plan to fuse other data sources (e.g., wind speed information [5]) to help practical applications.

Reference:  

[1] Lasaponara, A. M. Proto, A. Aromando, G. Cardettini, V. Varela and M. Danese, "On the Mapping of Burned Areas and Burn Severity Using Self Organizing Map and Sentinel-2 Data," in IEEE Geoscience and Remote Sensing Letters, vol. 17, no. 5, pp. 854-858, May 2020, doi: 10.1109/LGRS.2019.2934503.

[2] Karmakar, C. O. Dumitru, G. Schwarz and M. Datcu, "Feature-Free Explainable Data Mining in SAR Images Using Latent Dirichlet Allocation," in IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 14, pp. 676-689, 2021, doi: 10.1109/JSTARS.2020.3039012.

[3] California Wildfires Live Updates: 24 Dead in L.A. as Dangerous Winds Threaten Fire Growth - The New York Times

[4] Sentinel-2 mission. Available online: https://sentinel.esa.int/web/sentinel/missions/sentinel-2

[5] Global Wind Atlas. Available online: https://globalwindatlas.info/en/about/dataset

[6] ESA news based on Sentinel-2. Available online: https://www.esa.int/ESA_Multimedia/Missions/Sentinel-2/(offset)/100/(sortBy)/published/(result_type)/images

[7] Karmakar, C.O. Dumitru, N. Hughes and M. Datcu, "A Visualization Framework for Unsupervised Analysis of Latent Structures in SAR Image Time Series", IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 16, pp. 5355-5373, 2023.

How to cite: Dumitru, O., Karmakar, C., and Goyal, S.: Explainable Machine Learning for Forest Fire Detection with Remote Sensing for Effective Rescue Planning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16843, https://doi.org/10.5194/egusphere-egu25-16843, 2025.

EGU25-17615 | ECS | Orals | ESSI4.11

Automated detection of tuta absoluta (Meyrik) lesions on tomato plants using artificial intelligence 

Andrés Felipe Almeida-Ñauñay, Ernesto Sanz, Juan José Martín-Sotoca, Ruben Moratiel, Esther Hernández-Montes, and Ana M. Tarquis

The invasive tomato pest Tuta absoluta (Meyrik) poses a significant threat to global agriculture, often resulting in severe yield losses if not detected and managed early. This study investigates the application of artificial intelligence (AI) to develop an automated system for detecting T. absoluta (Meyrik) lesions on tomato plants. Leveraging open-source computational tools such as Google Colab, the research aims to provide an accessible and efficient solution through computational experiments, without requiring field trials.

A curated dataset of tomato plant images is prepared for training and evaluation. The YOLO (You Only Look Once) model is utilized for its proven effectiveness in small-object detection tasks, making it an ideal choice for identifying pest lesions. Model performance is assessed using metrics such as mean Average Precision (mAP), precision, recall, and F1-score, ensuring robust and reliable results across varying conditions. Prior research has highlighted the success of similar AI-based approaches in agricultural pest detection, achieving high accuracy while supporting sustainable farming practices  

This work emphasises leveraging multi-source data and advanced modelling approaches to enhance agricultural sustainability. By integrating sensing data and AI techniques, the study supports improved Integrated Pest Management (IPM) strategies, offering a scalable and environmentally friendly solution for pest monitoring in tomato production. Furthermore, the approach demonstrates how AI-driven insights from remote sensing can contribute to the broader goals of ecosystem productivity and nature-based solutions for climate change mitigation.

Acknowledgements: The authors acknowledge the support of the Project “LIFE23-CCA-ES-LIFE ACCLIMATE: Cultivating Resilience: Climate Change Adaptation Strategies for Greenhouses to Enhance Yield and Resource Efficiency from the Programme for the Environment and Climate Action (LIFE-EU) (project number: 101157315).

How to cite: Almeida-Ñauñay, A. F., Sanz, E., Martín-Sotoca, J. J., Moratiel, R., Hernández-Montes, E., and Tarquis, A. M.: Automated detection of tuta absoluta (Meyrik) lesions on tomato plants using artificial intelligence, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17615, https://doi.org/10.5194/egusphere-egu25-17615, 2025.

EGU25-17735 | ECS | Orals | ESSI4.11

A multi-sensor remote sensing approach to monitor illegal charcoal production sites in Somalia’s forests 

Luca de Guttry, Iqro Abdi Olow, Paolo Paron, Michele Bolognesi, Ugo Leonardi, Laura Stendardi, Giovanni Argenti, Marco Moriondo, and Camilla Dibari

Illegal charcoal production, by means of indiscriminate logging activities, poses significant threats to the stability of the drylands’ ecosystem in the Somali territory. In addition, the revenues from the charcoal trade often serve further illegal activities, exacerbating the already complex socio-political context of the country. In this work, we investigated the application of freely available multi-sensor remote sensing products (Sentinel-1 and Sentinel-2) and machine learning techniques to detect the presence of charcoal production sites (i.e., kilns) over large areas. Exploiting Google Earth Engine and open-source tools, we were able to develop a binary classification of kilns’ presence-absence for the years 2019, 2020, and 2021 in a remote area (approximately 32000 km2) north-west of Mogadishu, Somalia. Concerning the workflow, we first computed median images, spanning the first three months of each year, composed of numerous optical, SAR (Synthetic Aperture Radar), and combined vegetation indices. Images were then subtracted between consecutive years and a Support Vector Classification (SVC) algorithm was trained and validated on the indices’ values extracted from those. As a reference dataset, we employed known kilns’ locations from a preceding study by FAO-SWALIM, where photointerpretation of very high resolution images was used to individuate the appearance of illegal charcoal kilns. The evaluation of the classifications showed that our approach has great capabilities for the automatic individuation and the monitoring of illegal charcoal production sites, with R2 values and accuracy metrics ranging between 0.80-0.88 for the three considered years (2019, 2020, 2021). Moreover, mappings of the predicted presence-absence of kilns (at 10 m spatial resolution) were produced starting from the trained SVC model, giving a spatial representation of the phenomenon and allowing an assessment of the most impacted areas. In conclusion, our results represent a significant advancement in monitoring illegal charcoal production activities in Somalia, offering a reliable and transferable methodology based on accessible satellite imagery and tools.

How to cite: de Guttry, L., Abdi Olow, I., Paron, P., Bolognesi, M., Leonardi, U., Stendardi, L., Argenti, G., Moriondo, M., and Dibari, C.: A multi-sensor remote sensing approach to monitor illegal charcoal production sites in Somalia’s forests, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17735, https://doi.org/10.5194/egusphere-egu25-17735, 2025.

EGU25-17743 | ECS | Orals | ESSI4.11

Remote sensing applied to phenology monitoring in vineyards: testing through field observations 

Eduardo Jiménez-Jiménez, Guillermo Muñoz-Gómez, Beatriz Lara, Federico Fernández-González, and Rosa Pérez-Badia

In this paper we study the relationship between vegetative phenology obtained from satellite-derived vegetation indices (VIs) and vegetative and floral phenology based on field observations. The work was conducted during 2023 and 2024 in vineyards belonging to the Designation of Origin Uclés, located in the west of Cuenca province (Castilla-La Mancha region, central Spain). The field work was carried out in seven plots that are frequently ploughed and lack cover crops and green covers. All plots grow under similar conditions and the maximum distance between plots is less than 2 kilometers. Phenological sampling was carried out weekly on 20 grapevines per plot, using the BBCH scale.

Different VIs (NDVI, EVI, SAVI and SAVI2) were calculated using Google Earth Engine (GEE) and Sentinel-2 data, but EVI was selected due to its greater amplitude in the index curves. The R package Phenofit was used to clean the data, curve fitting and extraction of phenology metrics. For curve fitting, the Elmore method was used, and for phenology metrics extraction, the Threshold, Inflection and Gu methods from the Phenofit package were applied. Although Inflection and Gu differ in their approach, they both divide the curve into four phenological metrics: greenup, when index starts to growth; maturity, when the index value remains stable; senescence, when it decreases; and dormancy, when it stops decreasing and remains at a low value. Threshold considers only greenup and dormancy.

The results show that greenup is associated with the inflorescence development. This phase starts in a similar day of the year (DOY) in all plots and in the two studied years. Maturity, marked by Inflection and Gu methods, occurs between flowering and fruit development stages, that is, between DOY 140–198. The senescence period is marked between fruit development and leaf discoloration (178–310 DOY), and despite its amplitude, 75% of the observations place senescence between the final stages of the fruit and leaf discoloration. Finally, dormancy occurs between leaf discoloration and the leaf fall which is correct but usually it is marked excessively late.

Phenological metrics derived from Vegetation Indices (VIs) such as greenup (potentially related to inflorescence development), senescence (potentially related to leaf discoloration), and dormancy (potentially related to leaf discoloration and fall) can be linked to the grapevine cycle on the BBCH scale. However, more studies are needed to accurately link field phenological observations with satellite-derived vegetation indices.

This work has been funded by the Junta de Comunidades de Castilla-La Mancha (JCCM) through the project SBPLY/ 21-180501-000172 and by the University of Castilla-La Mancha (UCLM) through the project 2022-GRIN-34507. EJJ thanks to the Investigo Program for a contract co-financed by the European Social Fund Plus.

How to cite: Jiménez-Jiménez, E., Muñoz-Gómez, G., Lara, B., Fernández-González, F., and Pérez-Badia, R.: Remote sensing applied to phenology monitoring in vineyards: testing through field observations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17743, https://doi.org/10.5194/egusphere-egu25-17743, 2025.

EGU25-18129 | ECS | Orals | ESSI4.11

Temporal and Spatial Analysis of Critical Field Points Using High-Resolution Soil Water Content Estimation Employing Remote Sensing and Deep-Learning 

Mehdi Rafiei, Muhammad Rizwan Asif, Michael Nørremark, and Claus Aage Grøn Sørensen

This study presents a novel deep-learning approach for estimating Soil Water Content (SWC) with high spatial resolution across multiple soil depths. Additionally, the study identifies critical field points based on their drying-out times analyzed by SWC estimations over extended periods. Understanding potential critical points regarding SWC allows operators of heavy agricultural equipment to gain insight into the field's traits and prevent excessive soil compaction. Additionally, this information can support more strategic and efficient harvesting plans by accounting for the impact of varying drying patterns on crop growth and soil strength to not only minimize soil degradation but also maximize yield production, offering a more productive and sustainable crop production.

In this regard, our proposed method offers a practical approach to integrating diverse data types, including:

  • Spatial data: remote sensing data (Synthetic Aperture Radar (SAR) and vegetation index), land elevation, and soil profiles at various depths (soil content and bulk density).
  • Temporal data: historical weather information (precipitation, temperature, wind, and global radiation).
  • Contextual data: date, groundwater level, and crop type.

Previous machine learning and numerical models primarily used temporal and contextual data alongside point-based parameter values as inputs. In contrast, we incorporated spatial information instead of point values, allowing the model to capture better the surrounding influences—such as elevation, water flow, and vegetation shadows—on SWC.

To be able to estimate the SWC using the comprehensive analysis of spatial, temporal, and relevant contextual factors, these inputs are processed by a novel multi-model deep learning framework comprising:

  • U-Net to capture spatial features and the impacts of 2D image data.
  • Temporal Convolutional Network (TCN) to extract temporal dependencies from weather data.
  • Feed-Forward Network (FFN) to model the influence of contextual inputs.

Our model is trained and validated using ground truth data from site measurements in the HOBE dataset. These measurements are conducted at 30 locations within the Skjern River Catchment in Western Denmark, with each data sample containing SWC at different depths: surface, 20cm, and 50cm. By utilizing data collected between 2014 and 2018 from point 1.09 in the HOBE dataset, we demonstrated that the proposed model achieved a mean absolute error (MAE) of 0.0207. For comparison, a numerical model (Daisy) and a machine learning approach that did not account for spatial context produced higher MAEs of 0.0382 and 0.0269, respectively.

Subsequently, the developed model is employed to estimate SWC over extended periods and identify critical points within fields. To achieve this, we collaborated with several farmers who manually classified their field maps into regular, late-drying, and critical parts. The distinction between the latter two categories is crucial, as our observations revealed that "not every wet point is a critical point." The collected temporal SWC data is analyzed with land elevation to differentiate between these two classes. This aspect of the study remains under investigation, and further research is being conducted to refine the classification process and validate its effectiveness.

How to cite: Rafiei, M., Asif, M. R., Nørremark, M., and Sørensen, C. A. G.: Temporal and Spatial Analysis of Critical Field Points Using High-Resolution Soil Water Content Estimation Employing Remote Sensing and Deep-Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18129, https://doi.org/10.5194/egusphere-egu25-18129, 2025.

EGU25-18174 | ECS | Orals | ESSI4.11

Innovative Approaches to Carbon Stock Assessment in Agroecosystems: The Potential of TLS 

Martina Leoni, Maria Vincenza Chiriacò, Simona Castaldi, and Riccardo Valentini

The European Union’s Carbon Removal Certification Framework (CRCF) establishes robust quality standards and transparent monitoring, reporting, and verification (MRV) systems to ensure the credibility of carbon removal initiatives. Reliable MRV systems are critical for maintaining the environmental integrity of European carbon farming efforts and building stakeholder confidence in carbon accounting and reporting. Achieving these objectives requires the integration of innovative technologies with traditional methods to enhance accuracy and scalability carbon stock estimations.

Within this framework, growing attention is being directed toward methodologies for estimating carbon stocks across various pools in agroecosystems. While soil carbon estimation methods are well-established, the estimation of above-ground biomass (AGB) in agroforestry systems remains underexplored. Significant challenges in this domain include the difficulty of conducting destructive sampling in productive agricultural systems, the lack of species-specific allometric equations for woody crops, and the variability in tree structure introduced by pruning and other anthropogenic interventions.

This study applies terrestrial laser scanning (TLS) in a plum (Prunus domestica L.) orchard to address these challenges and perform non-destructive sampling of AGB for carbon stock assessment. The research employs quantitative structure modeling (QSM) to estimate tree volume and AGB with high precision, demonstrating TLS's ability to overcome limitations associated with destructive sampling, offering a scalable and repeatable approach for accurate biomass estimation in agroforestry systems. Furthermore, the study highlights the role of agroforestry in carbon sequestration efforts.

The findings highlight TLS as a valuable tool for improving the precision and reliability of carbon accounting in agroforestry systems. Its ability to provide accurate, non-destructive AGB estimates supports the effective implementation of the CRCF and advances the EU’s climate goals. Moreover, the scalability and adaptability of TLS make it a promising addition to MRV frameworks, offering stakeholders practical solutions for enhancing carbon removal initiatives.

How to cite: Leoni, M., Chiriacò, M. V., Castaldi, S., and Valentini, R.: Innovative Approaches to Carbon Stock Assessment in Agroecosystems: The Potential of TLS, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18174, https://doi.org/10.5194/egusphere-egu25-18174, 2025.

EGU25-18417 | ECS | Orals | ESSI4.11

An Intercomparison of Two Satellite-Based Hyperspectral Imagery (PRISMA & EnMAP) for Agricultural Mapping: Potential of these sensors to produce hyperspectral time-series essential for tracking crop phenology and enhancing crop type mapping 

Mohamed Bourriz, Ahmed Laamrani, Ali El-Battay, Hicham Hajji, Nadir Elbouanani, Hamd Ait Abdelali, François Bourzeix, Abdelhakim Amazirh, and Abdelghani Chehbouni

In recent decades, space-borne hyperspectral sensors have demonstrated significant potential for agricultural monitoring by providing rich spectral information, improved feasibility, and cost-effectiveness compared to multispectral satellite imagery. In this study, we investigated the consistency of two hyperspectral satellite sensors, PRISMA and EnMAP, for agricultural mapping during the 2025 growing season in the Meknes region: one of the most fertile and productive areas for cereals and vegetables at the national level of Morocco. The primary objective was to conduct a comparative analysis of the two datasets and perform a binary classification (crop vs. no-crop) to support land use monitoring, inform decision-making, and enable advanced crop type mapping.

Our methodology included a correlation analysis of reflectance values across the visible to near-infrared (VNIR) and shortwave infrared (SWIR) ranges, as well as the evaluation of NDVI indices using two methods: band averaging and hyperspectral NDVI (hNDVI). Classification was performed using three machine learning algorithms—Support Vector Machine (SVM), Random Forest (RF), and CatBoost—based on 16 optimal hyperspectral narrow-bands (i.e., 427,  535, 567, 714, 775, 805, 839, 913, 977, 1175, 1246, 1295, 1717, 2077, 2191, 2343 nm) from PRISMA and EnMAP that best capture the variability of vegetation biophysical and biochemical characteristics.

Results demonstrated high Pearson correlation coefficients between the two sensors, with r=0.93 in the VNIR and r=0.91 in the SWIR ranges. NDVI comparison also showed strong consistency results, with correlations of r=0.84 using the hNDVI method and r=0.85 using band averaging. The utilization of optimal hyperspectral narrow-bands achieved superior classification accuracies of 99.95% with PRISMA and 99.65% with EnMAP, with SVM outperforming other algorithms, followed by RF and CatBoost. Moreover, an Explainable Artificial Intelligence (XAI) based analysis indicated that bands in the NIR and SWIR regions were the most critical features driving these high classification performances.

These findings highlight the consistency and complementarity of PRISMA and EnMAP for agricultural monitoring. They also demonstrate the potential of these sensors to produce hyperspectral time-series essential for tracking crop phenology and enhancing crop type mapping, thereby overcoming the constraints posed by limited revisit intervals in current imaging spectroscopy missions.

How to cite: Bourriz, M., Laamrani, A., El-Battay, A., Hajji, H., Elbouanani, N., Ait Abdelali, H., Bourzeix, F., Amazirh, A., and Chehbouni, A.: An Intercomparison of Two Satellite-Based Hyperspectral Imagery (PRISMA & EnMAP) for Agricultural Mapping: Potential of these sensors to produce hyperspectral time-series essential for tracking crop phenology and enhancing crop type mapping, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18417, https://doi.org/10.5194/egusphere-egu25-18417, 2025.

EGU25-18748 | ECS | Posters on site | ESSI4.11

Species Distribution Models: Application to the Identification of Populations and Potential Distribution Areas of the Forage Plant Bituminaria bituminosa 

Javier San Martin Loren, Jesús Fernandez Habas, and Pilar Fernandez Rebollo

The species Bituminaria bituminosa (L.) C.H. STIRT has been studied over the past two decades to be integrated as a forage crop in agro-silvo-pastoral systems due to its nutritional qualities and low water requirements (<200 mm). These efforts have led to the development of new varieties using genotypes from the Canary Islands. These varieties are expected to be utilized in mixed or monoculture systems, leveraging their drought tolerance to extend the availability of high-quality feed, thus reducing costs during the forage shortages of the summer season. The ability of Bituminaria to fulfill this role in Mediterranean basin farms will largely depend on its adaptation to environmental conditions.

This study aims to explore the circum-Mediterranean distribution of Bituminaria using Species Distribution Models (SDMs) and 33,132 occurrences from the GBIF platform on natural populations of the species. Bioclimatic, edaphic, geomorphological, and satellite-derived variables were used in model development through the biomod2 package in R, achieving ensemble model metrics with a mean True Skill Statistic (TSS) of 0.78. Eight clusters have been proposed to group occurrences based on the most important variables identified in the ensemble model, which also aids in identifying isolated populations or localized scenarios that may serve as a foundation for breeding programs aimed at improving specific traits. These results will contribute to a deeper understanding of the ecology, phenotypic plasticity, population dynamics, movement patterns, and evolutionary processes within the genus Bituminaria.

How to cite: San Martin Loren, J., Fernandez Habas, J., and Fernandez Rebollo, P.: Species Distribution Models: Application to the Identification of Populations and Potential Distribution Areas of the Forage Plant Bituminaria bituminosa, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18748, https://doi.org/10.5194/egusphere-egu25-18748, 2025.

Scalably sensing/estimating local information of environment, management, and crop at the field level is the first step of a System-of-Systemssolution to quantify field-level agroecosystem dynamics (Guan et al., Earth-Science Reviews, 2023). This sensing effort involves two major and inherently connected tasks: (1) ground truth collection, and (2) cross-scale sensing. Agricultural ground truth is scarce and expensive to collect, however, the need for ground truth data is non-negotiable and should be a major investment with public funding. We have developed cross-scale sensingapproaches to scale-up ground truthcollection to large scales. In this talk, we will review our recent progress in using "cross-scale sensing" to accurately estimate critical variables of agroecosystem dynamics, covering management practices (e.g. tillage practice, crop rotation, cover crop adoption, irrigation), environmental conditions (e.g. soil properties), and crop traits and conditions (e.g. LAI, Vmax, phtosynthesis, crop yield). We will also identify current challenges and future opportunities to further advance remote sensing for sustainable and precision agriculture. 

How to cite: Guan, K.: Recent progress in remote sensing for advancing sustainable and precision agriculture, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19065, https://doi.org/10.5194/egusphere-egu25-19065, 2025.

EGU25-19406 | ECS | Posters on site | ESSI4.11

Assessment of Retention Basin Potential Using Active Remote Sensing 

Dávid Pásztor, Attila Nagy, Zsolt Fehér, and János Tamás

The increasing frequency of drought periods and the intensification of precipitation distribution extremes in Central Europe, particularly in eastern Hungary, pose significant challenges for water resource management. The Great Hungarian Plain (Alföld) experiences an annual precipitation deficit of 150–250 mm, exacerbating the adverse effects of drought. The Eastern Main Canal (Keleti-főcsatorna) plays a crucial role in water supply, transporting 300–400 million m³ of water annually as part of the Civaqua program. This initiative aims to channel water from the Tisza River to the Tócó stream, ensuring sustainable water supply for the region and maintaining critical water levels in local reservoirs, including the Vezér Street Retention Basin. The basin serves not only water retention and flood control purposes but also provides recreational opportunities for the local community.
This study aims to evaluate strategies for maximizing the capacity and efficiency of retention basins by optimizing the water supply from the Tisza River and the Eastern Main Canal, particularly during drought periods. Additionally, the research explores the potential of basin retention for the storage of precipitation and excess water within the basin and surrounding landscapes. Such retention solutions contribute to efficient water resource management, mitigating drought impacts and enhancing the long-term sustainability of water management practices.
The research employed active remote sensing technologies, including the Apache 3 unmanned surface vessel equipped with a monobeam sonar, providing depth measurement accuracy within 1% of the measured depth. For terrestrial surveys, the Stonex X120GO SLAM Laser Scanner was utilized, delivering millimeter-level precision in 3D mapping. The integration of these technologies enabled the development of detailed basin models, capturing both underwater and aboveground features of the retention basin. The primary focus was the Vezér Street Retention Basin, which serves flood control, water retention, and recreational functions in the Debrecen area.
The lowest point of the Vezér Street Retention Basin is at an elevation of 110.65 m above Baltic Sea level, while the highest point of the basin crown is 114.39 m, resulting in a maximum depth of 3.74 m. The basin’s total storage capacity, when fully saturated, is 39,213.59 m³, with a water surface area of 16,354.93 m². At the average water level of 113.69 m, the basin holds approximately 28,253.2 m³ of water, with a water surface area of 15,000.08 m². During the summer, under conditions of 20°C, average atmospheric pressure, and humidity, evaporation rates reach 3 mm/day/m², resulting in a daily water loss of 45,000.24 mm/day. The aquatic biodiversity of the basin is characterized by the presence of Typha species, which serve as critical ecological indicators.
The preliminary findings highlight that active remote sensing methods, such as sonar and the Stonex X120GO SLAM Laser Scanner, provide reliable tools for maximizing basin capacity and developing efficient water retention strategies.

 

The research presented in the article was carried out within the framework of the Széchenyi Plan Plus program with the support of the RRF 2.3.1 21 2022 00008 project.

How to cite: Pásztor, D., Nagy, A., Fehér, Z., and Tamás, J.: Assessment of Retention Basin Potential Using Active Remote Sensing, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19406, https://doi.org/10.5194/egusphere-egu25-19406, 2025.

EGU25-19560 | ECS | Orals | ESSI4.11

Assessment of phenology of winter wheat using Sentinel 2 multispectral data for varying sowing dates  

Hitesh Upreti, Chinthamaneni Sriyodh, and Manoj Yadav

Wheat is one most widely grown and consumed crops globally. Region-wise, the north Indian plains are one of the largest producers of wheat in the world. However, there remains a substantial variation in the sowing dates and thus the phenology of wheat grown in the area owing to variation in cropping patterns, soil type and agricultural practices. In this study, field data including the extent of wheat crops along with their sowing and harvest dates were collected in the Gautam Buddha Nagar district of Uttar Pradesh, India during the 2022-23 crop season. The study region is then classified into croplands and further into wheat and non-wheat areas using the random forest classifier in the Google Earth Engine. On the basis of the sowing dates, the study region is divided into early sowing (sowing date before 10 November 2022) and late sowing (sowing date after 25 November 2022) areas. The phenology of the wheat agricultural fields is analyzed using the normalized difference vegetation index (NDVI) derived using the Sentinel 2 surface reflectance data product available in the Google Earth Engine. Results showed that the early sowing wheat has the largest period (6 to 7 weeks) in which canopy cover was near maximum. The same period for late sown wheat was found as 4 to 5 weeks for late sown wheat. In general, the peak vegetation density for the crop season decreased as the sowing time of the wheat was delayed. The average value of peak normalized difference red-edge index (NDRE) was found as 0.67 (in second week of February 2023) and 0.62 (in first week of March 2023) for the early and late sown wheat, respectively. The lengths of the crop seasons of the early and late sown wheat were found as 140 and 120 days, respectively. The findings of the present study can be extrapolated to understand the phenology as well as the yield patterns of the wheat in one largest wheat producing regions in the world.

How to cite: Upreti, H., Sriyodh, C., and Yadav, M.: Assessment of phenology of winter wheat using Sentinel 2 multispectral data for varying sowing dates , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19560, https://doi.org/10.5194/egusphere-egu25-19560, 2025.

EGU25-19779 | ECS | Posters on site | ESSI4.11

Integrating Orthophotos and Field Data for Precision Vineyard Yield Prediction: A Case Study of Tempranillo Grapevines 

Maite Novellón, Sara Lacalle, Ana María Tarquis, and Pilar Baeza

Anticipating the response of grapevines to environmental variability is crucial for opti-mizing field management practices. This study explores the interaction between vines and their habitat across the growing cycle to inform more effective vineyard management. The research was conducted at the "Alhambra" plot in Ciudad Real (38.8089720, -3.0705830), which spans approximately 6 hectares of irrigated Tempranillo (Vitis vinifera L.) vines. Vine spacing is 3.05x1.54 m², and the training system is a double guyot pruned, vertical shoot positioning. The study utilizes data collected over 2024.


Within the plot, three replicates of 30 plants each were sampled. Measurements were taken from consecutive rows, 15 plants each. At the phenological stage of separated clus-ters, the number of clusters was recorded, while berry weight and the number of berries per cluster were assessed at veraison and harvest. Yield partitioning was determined at harvest. Additional parameters were also measured, including total soluble solids, surface area, pruning and shoot weight.


A custom script was developed to analyze the orthophotos of the vineyard to quantify the trellis length occupied by vines, excluding gaps where vines were missing. This method enables precise calculation of the vine-covered productive area. By combining these or-thophoto analyses with field-estimated yields per linear meter of vine, the study could provide accurate vineyard yield predictions. The accuracy and effectiveness of this inte-grated methodology are thoroughly evaluated.


Acknowledgements BigPrediData

How to cite: Novellón, M., Lacalle, S., Tarquis, A. M., and Baeza, P.: Integrating Orthophotos and Field Data for Precision Vineyard Yield Prediction: A Case Study of Tempranillo Grapevines, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19779, https://doi.org/10.5194/egusphere-egu25-19779, 2025.

Andean communities in central Peru play a key role in the conservation of vicuñas (Vicugna vicugna), a protected species that depends on puna grass and flooded vegetation for food and access to water throughout the year. This study focuses on seven communities of Lucanas in Ayacucho, a dry mountainous region of Peru, emphasizing the need for accurate information to monitor resources in a context of climate change and support community decision-making. In this research, based on Google Earth Engine (GEE), we evaluated the performance of classification algorithms using Sentinel-1 (S1) and Sentinel-2 (S2) image data for rangelands classification. The process used ground-based and image-based points to train and validate the models, a filter to minimize spatial autocorrelation between training and validation sets; and spectral separability measurements using the Jeffries-Matusita (JM) distance, all of steps allowed an adequate discrimination and representation of the classes. Additionally, we used 64 feature variables (including vegetation, texture, topographic, snow, water, minerals, radar features) and applied Cloud Score+, quality assessment (QA) processor in S2 image collection, to improve classification accuracy. Random Forest (RF) algorithm achieved an overall accuracy (OA) of 92% and a Kappa coefficient of 0.908 outperforming the Support Vector Machine (SVM) algorithm, which obtained an OA of 90.9% and a Kappa coefficient of 0.895. The results show that, in the semi-captivity sectors, 1,777.5 hectares of puna grass and 319.1 hectares of flooded vegetation were identified, while in wild management areas 5,431.1 hectares of puna grass and 843.8 hectares of flooded vegetation were recorded. These findings highlight the importance of integrating remote sensing tools and machine learning algorithms to generate key information in the management of natural resources in communities.

How to cite: Ochoa, J., Juarez, H., Sotomayor, D., and De Haan, S.: Mapping Rangeland Vegetation Using Sentinel-1 and Sentinel-2 Imagery with Machine Learning: A Case Study of Vicuña Conservation in the Central Andes of Perú, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21573, https://doi.org/10.5194/egusphere-egu25-21573, 2025.

EGU25-21604 | ECS | Orals | ESSI4.11

Leaf Area Index and Leaf Chlorophyll Content estimation from hyperspectral imaging using SCOPE model inversion 

Chiara Rivosecchi, Aya Amar, Paola A. Deligios, Eline Eeckhout, Matteo Francioni, Geert Haesaert, Luigi Ledda, Adriano Mancini, and Wouter H. Maes

Leaf Area Index (LAI) and Leaf Chlorophyll Content (LCC) are key vegetation indices for modeling energy and mass exchange between the atmosphere and land surfaces and can therefore be utilized for yield prediction. Consequently, suitable methods have been developed to retrieve LAI and LCC from remotely sensed data. Among these, the inversion of Radiative Transfer Models stands out as a promising approach, as it addresses the issue of limited transferability and minimizes the need for extensive field measurements also accounting for crop variability.

The objective of this study is to assess the applicability of the Soil Canopy Observation of Photochemistry and Energy Fluxes (SCOPE) model for estimating LAI and LCC of potato (Solanum tuberosum L.) using time series of hyperspectral images captured by an uncrewed aerial vehicle. A field experiment was conducted in Belgium from June to October 2024, involving two potato varieties, early and late, subjected to two nitrogen fertilization levels and six different biostimulants. Throughout the crop growth cycle, hyperspectral UAV images were captured biweekly using the Specim AFX10 camera. On the same days, in situ measurements of LAI and LCC were performed. LAI and LCC were estimated using a look-up table (LUT) approach based on the inversion of the SCOPE model. A cost function (norm2 distance) was employed to sort the LUT and identify a set of spectra that minimized the distance between measured reflectance and simulated reflectance in the LUT. The estimated LAI and LCC values were then compared with their corresponding in situ measurements.

Preliminary results indicate that simulated LAI and LCC showed potential for use in designing models to predict measured LAI and LCC (R2=0.26 and R2=0.30, respectively, p<0.001). In conclusion, simulated LAI and LCC correlated well with measured values for the late variety at the beginning of the crop cycle. Future work will focus on extending the analysis to cover the entire season, incorporating remote sensing observations into the parametrization of a crop growth model for yield predictions.

How to cite: Rivosecchi, C., Amar, A., Deligios, P. A., Eeckhout, E., Francioni, M., Haesaert, G., Ledda, L., Mancini, A., and Maes, W. H.: Leaf Area Index and Leaf Chlorophyll Content estimation from hyperspectral imaging using SCOPE model inversion, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21604, https://doi.org/10.5194/egusphere-egu25-21604, 2025.

EGU25-21750 | ECS | Posters on site | ESSI4.11 | Highlight

PANGEOS COST action: Uncertainty propagation in remote sensing  

Egor Prikaziuk, Gary Llewellyn, Laura Mihai, Agnieszka Bialek, Andreas Hueni, Mike Werfeli, Jose Luis Gomez-Dans, Jochem Verrelst, Jose Luis Garcia-Soria, Joseph Fennell, Dessislava Ganeva, and Shawn Carlisle Kefauver

 

Pan-European Network of Green Deal Agriculture and Forestry Earth Observation Science (PANGEOS) funded by the European Cooperation in Science and Technology (COST) organisation brings together researchers to share their expertise and bring up a new generation of scientists. In October 2024 PANGEOS conducted an intensive 5-day summer school where more than 20 participants learnt how to propagate uncertainty of spectral measurements to uncertainty in higher-level products. The training material in the form of Python Jupyter notebooks is publicly available on GitHub https://github.com/pangeos-cost/uq-training.

This presentation is going to highlight the steps of uncertainty propagation from ground measurements through vegetation indices and retrieved plant traits towards higher-level model estimates, like gross primary productivity and evapotranspiration. All three pathways of retrieval uncertainty estimation, regression-based (vegetation indices), radiative transfer model-based and hybrid, are demonstrated. In addition, challenges of uncertainty propagation through satellite imagery are discussed in a separate block.

Finally, a highlight of current and future activities of the PANGEOS COST action will be given.

Acknowledgement

This abstract is supported by the EU COST (European Cooperation in Science and Technology) Action CA22136 “Pan-European Network of Green Deal Agriculture and Forestry Earth Observation Science” (PANGEOS).

How to cite: Prikaziuk, E., Llewellyn, G., Mihai, L., Bialek, A., Hueni, A., Werfeli, M., Gomez-Dans, J. L., Verrelst, J., Garcia-Soria, J. L., Fennell, J., Ganeva, D., and Kefauver, S. C.: PANGEOS COST action: Uncertainty propagation in remote sensing , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21750, https://doi.org/10.5194/egusphere-egu25-21750, 2025.

Sea surface temperature anomalies (SSTAs) over the North Atlantic (NA) have a significant impact on the weather and climate in both local and remote regions. This study first evaluated the seasonal prediction skill of NA SSTA using the North American multi-model ensemble and found that its performance is limited across various regions and seasons. Therefore, this study constructs models based on the long short-term memory (LSTM) network machine learning method to improve the seasonal prediction of NA SSTA. Results show that the seasonal prediction skill can be significantly improved by LSTM models since they show higher capability to capture nonlinear processes such as the impact of El Nin ̃o-Southern Oscillation on NA SSTA. This study shows the great potential of the LSTM model on the seasonal prediction of NA SSTA and provides new clues to improve the seasonal predictions of SSTA in other regions.

How to cite: Yan, X. and Tang, Y.: Seasonal prediction of North Atlantic sea surface temperature anomalies using the LSTM machine learning method , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-153, https://doi.org/10.5194/egusphere-egu25-153, 2025.

EGU25-3747 | Orals | CL4.6

Bridging paleoclimate and seasonal climate prediction: The case of European summer climate 

Martin Wegmann and Stefan Brönnimann

Understanding monthly-to-annual climate variability is essential for improving climate forecast products as well as adapting to future climate extremes. Previous studies show, that European summer climate, including temperature and precipitation extremes, is modulated by hemispheric large-scale circulation patterns, which themselves are connected to Earth system components such as sea surface temperature across temporal scales. Nevertheless, it remains unclear as to how stationary these teleconnections are and if their predictive power is potent across multiple centuries and background climates. By combining d18O isotopes from a European tree ring network with independent paleo-climate reanalyses, we highlight precursors and atmospheric dynamics behind European summer climate over the last 400 years.

We further present evidence that centennial ensemble seasonal climate forecasts capture the causality of the atmospheric
dynamics behind these teleconnections in the 20th century. Our results suggest that tropical sea surface temperature anomalies trigger specific precipitation and diabatic heating patterns which are dynamically connected to extratropical Rossby wave trains and the formation of a circumglobal teleconnection pattern weeks later.

How to cite: Wegmann, M. and Brönnimann, S.: Bridging paleoclimate and seasonal climate prediction: The case of European summer climate, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3747, https://doi.org/10.5194/egusphere-egu25-3747, 2025.

EGU25-3839 | ECS | Posters on site | CL4.6

Causal Links Between North Atlantic SSTs and Summer East Atlantic Pattern Predictability: Implications for Seasonal Forecasting 

Julianna Carvalho Oliveira, Giorgia Di Capua, Leonard F. Borchert, Reik V. Donner, and Johanna Baehr

We use causal effect networks to assess the influence of spring North Atlantic sea surface temperatures (NA-SSTs) on summer East Atlantic (EA) pattern predictability during 1908–2008. In the ERA-20C reanalysis, a robust causal link is identified for 1958–2008, where the spring meridional SST gradient causes a 0.2 standard deviation change in the summer EA. Additionally, the spring meridional SST index has an estimated negative causal effect (~−0.2) on summer 2m air temperatures over northwestern Europe. However, both links are absent when analysing the full period and are sensitive to interannual variability.

Analysis of the Max Planck Institute Earth System Model in mixed resolution (MPI-ESM-MR) shows that historical simulations fail to reproduce the observed causal links, while initialised ensembles occasionally capture them but underestimate their strength. Predictive skill assessments conditioned on these causal links indicate limited overall impact but suggest potential local improvements for European summer climate forecasts. These findings underscore the value of causal approaches for refining seasonal predictability.

How to cite: Carvalho Oliveira, J., Di Capua, G., Borchert, L. F., Donner, R. V., and Baehr, J.: Causal Links Between North Atlantic SSTs and Summer East Atlantic Pattern Predictability: Implications for Seasonal Forecasting, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3839, https://doi.org/10.5194/egusphere-egu25-3839, 2025.

EGU25-5880 | Orals | CL4.6

Intermittency of seasonal forecast skill for the wintertime North Atlantic Oscillation and East Atlantic Pattern  

Laura Baker, Len Shaffrey, Antje Weisheimer, and Stephanie Johnson

The wintertime North Atlantic Oscillation (NAO) and East Atlantic Pattern (EA) are the two leading modes of North Atlantic pressure variability and have a substantial impact on winter weather in Europe. The year-to-year contributions to multi-model seasonal forecast skill in the Copernicus C3S ensemble of seven prediction systems are assessed for the wintertime NAO and EA, and well-forecast and poorly-forecast years are identified. Years with high NAO predictability are associated with substantial tropical forcing, generally from the El Niño Southern Oscillation (ENSO), while poor forecasts of the NAO occur when ENSO forcing is weak. Well-forecast EA winters also generally occurred when there was substantial tropical forcing, although the relationship was less robust than for the NAO. These results support previous findings of the impacts of tropical forcing on the North Atlantic and show this is important from a multi-model seasonal forecasting perspective.

How to cite: Baker, L., Shaffrey, L., Weisheimer, A., and Johnson, S.: Intermittency of seasonal forecast skill for the wintertime North Atlantic Oscillation and East Atlantic Pattern , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5880, https://doi.org/10.5194/egusphere-egu25-5880, 2025.

EGU25-6006 | ECS | Orals | CL4.6

 Investigating the sensitivity of 20th century seasonal hindcasts to tropospheric aerosol forcing 

Matthew Wright, Antje Weisheimer, Tim Woollings, Retish Senan, and Timothy Stockdale

Previous studies have identified multi-decadal variations in the skill of winter seasonal forecasts of large-scale climate indices, including ENSO, the PNA, and NAO. Forecast skill is significantly lower in the middle of the 20th century (1940—1960) than at the start or end of the century. We hypothesise that tropospheric aerosol forcing, which is spatially and temporally heterogeneous and poorly constrained in the hindcasts used in previous studies, contributes to this low skill mid-century period.

This study assesses the sensitivity of ECMWF’s state-of-the-art seasonal forecasting model to tropospheric aerosol forcing, using a newly developed aerosol forcing dataset based on CEDS emissions data. We analyse DJF hindcasts initialised every November from 1925—2010, each with 21 ensemble members. For each year, we run hindcasts with ‘best guess’, doubled, and halved aerosol forcing (perturbing both anthropogenic and natural aerosols). All experiments exhibit similar multi-decadal variability in skill for large-scale climate indices. Aerosol forcing has no significant impact on forecast skill but some impacts on mean biases, suggesting other factors drive the mid-century skill minimum.

Aerosol forcing has large regional impacts. Increasing aerosol forcing leads to cooler 2m temperature and SSTs globally, with amplified cooling in regions with large aerosol forcings, such as northern India and North Africa. Dynamical responses include an ‘anti-monsoon’ circulation over Africa, with a weakening of the trade winds and Atlantic Walker circulation, and local southwards shift of the ITCZ. The magnitude of the response increases when ocean initial conditions are perturbed to represent the cumulative impact of aerosol forcing, suggesting that coupling enhances the atmospheric response.

These results highlight the model’s sensitivity to tropospheric aerosols, with large differences in bias and mean state after four months, despite limited impact on skill. The circulation changes over Africa warrant further investigation, with implications for future aerosol scenarios. Planned experiments will explore the impact in summer and quantify the timescale of the response to aerosols.

How to cite: Wright, M., Weisheimer, A., Woollings, T., Senan, R., and Stockdale, T.:  Investigating the sensitivity of 20th century seasonal hindcasts to tropospheric aerosol forcing, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6006, https://doi.org/10.5194/egusphere-egu25-6006, 2025.

This study shows a close relationship between winter Arctic sea ice concentration (WASIC) anomalies in the Barents-Greenland Seas and the subsequent autumn Indian Ocean Dipole (IOD) based on the observational analysis and numerical simulations. Particularly, more (less) WASIC in the Barents-Greenland Seas tends to lead to a positive (negative) IOD in the following autumn. Above-normal WASIC in the Barents-Greenland Seas results in reduction of the upward turbulent heat flux and induces tropospheric cooling over the Arctic. This tropospheric cooling triggers an atmospheric teleconnection extending from the Eurasian Arctic to the subtropical North Pacific. Numerical experiments with both the linear barotropic model and atmospheric general circulation model can well capture the atmospheric teleconnection associated with the WASIC anomalies. The subtropical atmospheric anomalies generated by the WASIC anomalies then result in subtropical sea surface temperature (SST) warming, which sustains and expands southward to the equatorial central Pacific during the following summer via a wind-evaporation-SST feedback. The resulting equatorial central Pacific SST warming anomalies induce local atmospheric heating and trigger an anomalous Walker circulation with descending motion and low-level anomalous southeasterly winds over the southeastern tropical Indian Ocean. These anomalous southeasterly winds trigger positive air-sea interaction in the tropical Indian Ocean and contribute to the development of the IOD. The close connection of the WASIC anomalies with the subsequent IOD and the underlying physical processes can be reproduced by the coupled climate models participated in the CMIP6. These results indicate that the condition of WASIC is a potential effective precursor of IOD events.

How to cite: Xin, C.: Influence of winter Arctic sea ice anomalies on the following autumn Indian Ocean Dipole development, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6176, https://doi.org/10.5194/egusphere-egu25-6176, 2025.

EGU25-7163 | Orals | CL4.6

Robust decadal predictability of cold surge frequency in Taiwan and East Asia through teleconnection of North Atlantic Oscillation 

Wan-Ling Tseng, Yi-Chi Wang, Ying-Ting Chen, Yi-Hui Wang, Huang-Hsiung Hsu, and Chi-Cherng Hong

This study investigates the decadal predictability of cold surge frequency (CSF) in East Asia, including Korea, Japan, and Taiwan, through the lens of the North Atlantic Oscillation (NAO) index. The findings suggest that extreme events such as cold surges can be predicted on decadal timescales when the teleconnection mechanism is robustly established. The study revisits and consolidates the dynamical mechanisms underlying wave propagation and the teleconnection between the NAO and the East Asian trough, highlighting their role in creating a winter environment conducive to cold surges in Taiwan and East Asia. The study demonstrates the skill of climate models in capturing the NAO's decadal variability, and develops a statistical-dynamical hybrid approach. This method integrates decadal prediction datasets with a statistical model to enhance the prediction of extreme cold surge occurrences on a multi-annual timescale. The results of the study underscore the scientific significance of merging climate dynamical mechanisms with decadal prediction systems for extreme events, and introduce a hybrid framework that combines numerical decadal climate predictions with statistical regression models. This addresses the challenges posed by biases in climate prediction models and advances the capability to predict regional extreme events such as cold surges.

How to cite: Tseng, W.-L., Wang, Y.-C., Chen, Y.-T., Wang, Y.-H., Hsu, H.-H., and Hong, C.-C.: Robust decadal predictability of cold surge frequency in Taiwan and East Asia through teleconnection of North Atlantic Oscillation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7163, https://doi.org/10.5194/egusphere-egu25-7163, 2025.

EGU25-8693 | ECS | Orals | CL4.6

Decadal prediction for the European Energy Sector 

Benjamin Hutchins, David Brayshaw, Len Shaffrey, Hazel Thornton, and Doug Smith

The timescale of decadal climate predictions, from a year-ahead up to a decade, is an important planning horizon for stakeholders in the energy sector. With power systems transitioning towards a greater share of renewables, these systems become more vulnerable to the impacts of both climate variability and climate change. As decadal predictions sample both the internal variability of the climate and the externally forced response, these forecasts can provide useful information for the upcoming decade. 

There are two main ways in which decadal predictions can benefit the energy sector. Firstly, they can be used to try to predict how a variable of interest, such as average temperature, may evolve over the coming year or decade. Secondly, a large ensemble of decadal predictions can be aggregated into a large synthetic event set to explore physically plausible extremes, such as winter wind droughts. 

We find predictive skill at decadal timescales for surface variables over Europe during both winter (ONDJFM) and summer (AMJJAS). Although this skill is patchy, there are regions of relevance to the energy sector, such as over the UK for temperature, where this skill emerges. We find significant skill when using pattern-based (e.g., NAO) approaches to make predictions of European energy indicators during the extended winter, including Northern Europe offshore wind generation, Spanish solar generation, and Scandinavian precipitation. For predicting UK electricity demand, we find significant skill when directly using the model predictions of surface temperature. Our results highlight the potential for operational decadal predictions for the energy system, with potential benefits for both the planning and operation of the future power system. 

How to cite: Hutchins, B., Brayshaw, D., Shaffrey, L., Thornton, H., and Smith, D.: Decadal prediction for the European Energy Sector, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8693, https://doi.org/10.5194/egusphere-egu25-8693, 2025.

EGU25-8904 | Orals | CL4.6

On the predictive skill for warm spells in Germany across seasons  

Fabiana Castino, Tobias Geiger, Alexander Pasternack, Andreas Paxian, Clementine Dalelane, and Frank Kreienkamp

Intense warm spells, such as heatwaves, can significantly impact human health, the environment, and socio-economic systems. Although heatwaves are typically associated with summer, the occurrence of warm spells during cold seasons can also have profound effects on various sectors. While some effects, such as reduced cold-related mortality, can be considered beneficial, the long-term consequences, e.g. on ecosystems, forests, and agriculture, are concerning. Warm spells during the cold seasons can alter the natural dormancy cycles of plants, causing premature sprouting, flowering, or growth and negatively affecting crop yield and quality. In addition, cold season warm spells can reduce snow accumulation in mountainous regions, potentially affecting downstream water availability. As climate change drives increases in the frequency, intensity, and duration of warm spells, their impacts are becoming more severe and far-reaching. This makes predicting such events a key priority for climate science and risk management.

Climate forecast models offer the potential to predict extreme events like warm spells weeks to months in advance, becoming increasingly relevant for decision-making across various socio-economic sectors. This study examines the predictive skill of the downscaled German Climate Forecast System Version 2.1 (GCFS2.1) for warm spells in Germany on a seasonal scale, encompassing both warm seasons (spring and summer) and cold seasons (autumn and winter).  The analysis relies on hindcast data from the 1991-2020 base period, statistically downscaled to 5 km resolution. It evaluates multiple extreme temperature climate indices, as for example the Warm Spells Duration index, and applies various statistical metrics to assess the predictive skill. The findings reveal high heterogeneity in the ability of the (downscaled) GCFS2.1 to forecast warm spells across seasons, with higher predictive skill during the cold seasons but more limited for the warm seasons.

How to cite: Castino, F., Geiger, T., Pasternack, A., Paxian, A., Dalelane, C., and Kreienkamp, F.: On the predictive skill for warm spells in Germany across seasons , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8904, https://doi.org/10.5194/egusphere-egu25-8904, 2025.

EGU25-8980 | ECS | Orals | CL4.6

Predicting North Atlantic Temperature Trends with the Analogue Method using the MPI CMIP6 Grand Ensemble 

Lara Heyl, Sebastian Brune, and Johanna Baehr

The analogue method is a powerful and efficient tool for climate predictions, particularly in regions like the North Atlantic, where impacts of climate change have been relatively modest. While climate projections effectively estimate global mean surface temperature trends over a century, decadal trends in the North Atlantic diverge from the global trend. Here, we leverage on the similar evolution of analogous patterns on a decadal time scale by comparing SST patterns in observed data with patterns from an existing simulation ensemble. We apply this method to ten-year SST trend reconstructions in the North Atlantic using the MPI CMIP6 grand ensemble. In addition, we assess the impact of volcanic eruptions on the quality of the SST trend reconstruction for the time period 1960-2019. We also provide a prediction for 2020–2029. We find that the analogue method delivers high correlation of SST trend reconstructions with observed trends for the MPI CMIP6 grand ensemble. Volcanic influence can be accounted for by trimming the time series to those times unaffected by volcanic eruptions, which results in a higher correlation. Our results suggest that the decadal predictions of SST trends might also be achieved without the need for new, computationally expensive simulations.

How to cite: Heyl, L., Brune, S., and Baehr, J.: Predicting North Atlantic Temperature Trends with the Analogue Method using the MPI CMIP6 Grand Ensemble, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8980, https://doi.org/10.5194/egusphere-egu25-8980, 2025.

EGU25-9006 | Posters on site | CL4.6

Is the winter mean NAO white noise? Models and observations. 

Bo Christiansen and Shuting Yang

The NAO is a dominant mode of variability in the Northern Hemisphere with strong impacts on temperature, precipitation, and storminess. The predictive skill of the NAO on annual to decadal scales is therefore an important topic, which is often studied using, e.g., (initialized) climate models. The temporal structure is closely related to the predictability, and on inter-annual time scales the observed NAO is frequently described to have power at 2-7 years and sometimes with a distinct peak around 7 or 8 years.  However, the observational record is brief, and such estimations have high uncertainty.

Here, we present a thorough study to answer the questions: is the winter mean NAO different from white noise and is the observed NAO different from the NAO in historical experiments with contemporary climate models (CMIP6)? To this end we use a range of statistical tools in both the temporal and spectral domain: Power-spectra, wavelet-spectra, autoregressive models, and various well-known time-series statistics.

Overall, we find little evidence to reject that the NAO is white noise. For observations, the peak in the power-spectrum at 8 years is, taken individually, significant in the period after 1950 but not before. However, considering the complete spectrum, significant peaks will often occur at some frequency, even for white noise.  The large CMIP6 multi-model ensemble is statistically very similar to an ensemble of similar size of white noise, e.g., the ensemble averages of the power spectrum and the wavelet spectra are completely flat.  Furthermore, for both observations and the model ensemble the tests based on autoregressive modelling and time-series statistics do not reject the null-hypothesis of white noise.

How to cite: Christiansen, B. and Yang, S.: Is the winter mean NAO white noise? Models and observations., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9006, https://doi.org/10.5194/egusphere-egu25-9006, 2025.

EGU25-10305 | ECS | Posters on site | CL4.6

Towards improved forecast initialisations with an observation-informed ocean grid 

Marlene Klockmann, Kai Logemann, Sebastian Brune, and Johanna Baehr

For climate forecasts it is crucial to initialise the ocean state from observations because they rely on the memory of the ocean. If, however, the initialised ocean state is far away from the model’s own preferred mean state, predictive skill will suffer due to model drift. We are testing whether an ocean grid with variable resolution - designed to represent sparse and well-observed regions with appropriate resolution - has advantages over an ordinary grid with uniform resolution. The locally high resolution could lead to an improved mean ocean state through a better representation of mesoscale processes. The observation-informed grid will allow for high-resolution data assimilation in well-observed areas, which will potentially lead to improved initial conditions and predictive skill.  

We developed such a grid for the ocean component of the coupled ICON model designed for seamless predictions (ICON-XPP). The grid resolution varies from 40 to 10km, depending on the observation density in the EN4 database from 1960 to 2023. The local refinement in well-observed areas leads to a better representation of ocean features such as fronts and western boundary currents. We assess the effect of these improvements on the mean climate state by comparing to a reference simulation with a uniform 20km ocean resolution. 

 

How to cite: Klockmann, M., Logemann, K., Brune, S., and Baehr, J.: Towards improved forecast initialisations with an observation-informed ocean grid, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10305, https://doi.org/10.5194/egusphere-egu25-10305, 2025.

EGU25-10747 | Posters on site | CL4.6

Ocean–atmosphere feedbacks key to NAO decadal predictability 

Panos J. Athanasiadis, Casey Patrizio, Doug M. Smith, and Dario Nicolì

Recent studies using initialised large-ensemble re-forecasts have shown that the North Atlantic Oscillation (NAO) exhibits significant decadal predictability, which is of great importance to society given the significant climate anomalies that accompany the NAO. However, the key physical processes underlying this predictability, including the role of ocean–atmosphere interactions, have not yet been pinned down. Also, a critical deficiency in the representation of the associated predictable signal by climate models has been identified in recent studies (the signal-to-noise problem), still lacking an explanation.

In this study, the decadal prediction skill for the NAO and the interactions of the associated atmospheric circulation anomalies with the underlying ocean are assessed using retrospective forecasts from eight decadal prediction systems and observation-based data. We find considerable spread in the NAO skill across these systems and critically, that this is linked to differences in the representation of ocean–NAO interactions across the systems. Evidence is presented that the NAO skill depends on a direct positive feedback between subpolar sea surface temperature anomalies and the NAO, which varies in strength across the prediction systems, yet may still be too weak even in the most skillful systems compared to the observational estimate. This positive feedback is opposed by a delayed negative feedback between the NAO and the ocean circulation that also contributes to disparities in the NAO skill across systems. Our findings therefore suggest that North Atlantic ocean–atmosphere interactions are central to NAO decadal predictability. Finally, it is suggested that errors in the representation of these interactions may be contributing significantly to the signal-to-noise problem.

How to cite: Athanasiadis, P. J., Patrizio, C., Smith, D. M., and Nicolì, D.: Ocean–atmosphere feedbacks key to NAO decadal predictability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10747, https://doi.org/10.5194/egusphere-egu25-10747, 2025.

EGU25-10815 | Posters on site | CL4.6

Planktonic foraminifera as a tool of past seasonality reconstruction 

Zhoufei Yu, Baohua Li, and Shuai Zhang

Seasonal changes in seawater temperature leave large imprints on the stable oxygen isotope composition (δ18O) of planktonic foraminiferal tests, based on which the past seasonal changes can be reconstructed. However, there are still problems needed to be figured out in regard to this new method, to improve the reliability of seasonality reconstruction. For example, the selected foraminiferal species, the used size fraction, and the sample area. As a result, by analyzing planktonic foraminiferal test δ18O from the sediment trap samples deployed in the South China Sea, we found that foraminiferal seasonal δ18O signal is strongly distorted (amplified or damped) by seasonal variations in their habitat depth, particularly for the species living in low latitude. Furthermore, Globigerinoides ruber of 300-355 um can record the most comprehensive seawater seasonality information. This study provides strong support to the reconstruction of past seawater seasonal temperature by using individual planktonic foraminifera.

How to cite: Yu, Z., Li, B., and Zhang, S.: Planktonic foraminifera as a tool of past seasonality reconstruction, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10815, https://doi.org/10.5194/egusphere-egu25-10815, 2025.

EGU25-11024 | ECS | Orals | CL4.6

Skill assessment of a multi-system ensemble of initialized 20-year predictions 

Dario Nicolì, Sebastiano Roncoroni, Wolfgang A. Mueller, Holger Pohlmann, Sebastian Brune, Markus Donat, Rashed Mahmood, Steve Yeager, William J. Merryfield, Reinel Sospedra-Alfonso, and Panos J. Athanasiadis

Decadal predictions have advanced greatly in recent years: not only have they become operational worldwide and have been demonstrated to be skillful in various aspects of climate variability, including predicting changes in the atmospheric circulation and in the occurrence of extremes several years ahead, but —as such— they are also being used increasingly in climate services. Climate adaptation and policy making, however, also require climate predictions that go beyond the 10-year horizon. For climate information beyond 10 years into the future, uninitialized climate projections, which completely miss any predictability stemming from internal variability, have been the only available product. Trying to account for this lack of information in climate projections regarding any predictable components of internal variability, methods to constrain climate projections using information from large ensembles of initialized decadal predictions have been developed and have been shown to reduce the uncertainty and increase the skill of climate projections, even beyond the 10-year horizon. The demonstrated benefits of such indirect methods to account for predictable internal variability indicate that the latter remains significant beyond the 10-year limit of decadal predictions. Hence, directly harnessing this predictability through running initialized 20-year predictions emerges as a strategic endeavour.
In this study a novel, multi-system ensemble of initialized extended-decadal predictions is assessed. These predictions consist of a grand ensemble of 71 members derived from 6 forecast systems. They are initialized every 5 years from 1960 onward and run ahead for 20 years. Our analysis uses an elaborate drift- and bias-correction method that accounts for the correct representation of trends. Importantly, we show significant skill against observations for a number of variables (fields and indices), even in the second decade of the forecasts. The origin of such predictability is discussed together with the limitations of these 20-year predictions. The respective experimental protocol was defined in the framework of the ASPECT EU project and has been proposed as a tier-2 Decadal Climate Prediction Project (DCPP) protocol for the Coupled Model Intercomparison Project phase 7 (CMIP7).

How to cite: Nicolì, D., Roncoroni, S., Mueller, W. A., Pohlmann, H., Brune, S., Donat, M., Mahmood, R., Yeager, S., Merryfield, W. J., Sospedra-Alfonso, R., and Athanasiadis, P. J.: Skill assessment of a multi-system ensemble of initialized 20-year predictions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11024, https://doi.org/10.5194/egusphere-egu25-11024, 2025.

EGU25-11166 | ECS | Orals | CL4.6

Multidecadal variability of the ENSO teleconnection to Europe in early-winter and implications for seasonal forecasting 

Pablo Fernández-Castillo, Teresa Losada, Belén Rodríguez-Fonseca, Diego García-Maroto, Elsa Mohino, and Luis Durán

El Niño-Southern Oscillation (ENSO) is the leading mode of global climate variability. Through its associated teleconnections, ENSO can impact the climate of numerous regions worldwide at seasonal timescales, highlighting its role as the main source of seasonal predictability. Numerous studies have demonstrated a significant influence of ENSO on the climate of the Euro-Atlantic sector, but the impacts and mechanisms of the teleconnection in early-winter (November-December) remain unclear. Besides, in early-winter, ENSO teleconnections involve tropospheric pathways, which may change in response to different background states of the ocean. Thus, a crucial research question to address is whether the early-winter teleconnection to the Euro-Atlantic sector has changed under the different background states of sea surface temperature (SST) over the Pacific Ocean. 

 

This work aims to analyse the ENSO early-winter teleconnection to the Euro-Atlantic sector from a nonstationary perspective. Specifically, the teleconnection is analysed under different background states of SST over the Pacific Ocean, related to changes in the phase of the Pacific Decadal Oscillation (PDO). Using observational and reanalysis datasets for the period 1950-2022, results reveal that the tropospheric pathways of the teleconnection change under the different Pacific SST background states, leading to distinct responses of the North Atlantic atmospheric circulation to ENSO. We also confirm that these distinct responses in the North Atlantic entail significantly different impacts of ENSO on the surface climate across Europe, particularly on surface air temperature. Furthermore, the teleconnection is analysed in the SEAS5 state-of-the-art dynamical seasonal prediction model. The analysis within the model is also conducted from a nonstationary perspective, and aims to determine whether the model successfully reproduces a shift in the teleconnection in the late 1990s identified in reanalysis and observations. Results show that the model accurately captures the spatial pattern of the teleconnection impacts across Europe after the late 1990s, but not before. In turn, significant changes in the skill of seasonal forecasts are observed between before and after the late 1990s. However, skill after the late 1990s is just moderate due to a significant underestimation of the teleconnection impacts. 

 

The results of this study shed light on the nonstationary behaviour of the early-winter teleconnection to the Euro-Atlantic sector and have important implications on seasonal predictability in Europe. Particularly, the nonstationarity of the teleconnection gives rise to the emergence of windows of opportunity for seasonal forecasting, in which forecast skill may be greater than initially expected from a stationary analysis.

How to cite: Fernández-Castillo, P., Losada, T., Rodríguez-Fonseca, B., García-Maroto, D., Mohino, E., and Durán, L.: Multidecadal variability of the ENSO teleconnection to Europe in early-winter and implications for seasonal forecasting, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11166, https://doi.org/10.5194/egusphere-egu25-11166, 2025.

EGU25-11511 | Orals | CL4.6

Constraining near-term climate projections by combining observations with decadal predictions 

Rémy Bonnet, Julien Boé, and Emilia Sanchez

Reducing the uncertainty associated with internal climate variability over the coming decades is crucial, as this time frame aligns with the strategic planning needs of stakeholders in climate-vulnerable sectors. Three sources of information are available: non-initialized ensembles of climate projections, initialized decadal predictions, and observations. Non-initialized ensembles of climate projections span seamlessly from the historical period to the end of the 21st century, encompassing the full range of uncertainty linked to internal climate variability. Initialized decadal predictions aim to reduce uncertainty from internal climate variability by initializing model simulations with observed oceanic states, phasing the simulated and observed climate variability modes. However, they are usually limited to 5 to 10 years, with small added value after a few years, and they are also subject to drift due to the shock from the initialization. Finally, we can also use observations that can provide information to constrain the climate evolution over the next decades. Providing the best climate information at regional scale over the next decades is therefore challenging. Previous methods addressed this challenge by using information from either the observations or the decadal predictions to constrain uninitialized projections. In this study, we propose a new method to make use of the different sources of information available to provide relevant information about near-term climate change with reduced uncertainty related to internal climate variability. First, we select a sub-ensemble of non-initialized climate simulations based on their similarity to observed predictors with multi-decadal signal potential over Europe, such as Atlantic multi-decadal variability (AMV). Then, we further refine this sub-ensemble of trajectories by selecting a subset based on its consistency with decadal predictions. We present a case study focused on predicting near-term future surface temperatures over Europe. To evaluate the effectiveness of this method in providing reliable climate information, we conduct a retrospective analysis over the historical period.

How to cite: Bonnet, R., Boé, J., and Sanchez, E.: Constraining near-term climate projections by combining observations with decadal predictions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11511, https://doi.org/10.5194/egusphere-egu25-11511, 2025.

EGU25-12107 | Orals | CL4.6

Overcoming the spring predictability barrier with a supermodel 

Noel Keenlyside, Tarkeshwar Singh, Ping-Gin Chiu, Francois Counillon, and Francine Schevenhoven

Climate models suffer from long-standing biases that degrade climate prediction skills. While radically increasing resolution offers promise, we are still many years away from being able to perform operational climate predictions with models that can explicitly resolve the most important physical processes. Here we demonstrate that supermodelling can enhance climate predictions through better using the current generation of models. A supermodel connects different models interactively so that their systematic errors compensate. It differs from the standard non-interactive multi-model ensembles, which combines model outputs a-posteriori. We have developed an ocean-connected Earth System model using NorESM, CESM, and MPIESM in their CMIP5 versions. The model radically improves the simulation of tropical climate, strongly reducing SST and double ITCZ biases. We perform seasonal predictions for the period 1990-2020, initialized through (EnOI) data assimilation of SST. We have performed one forecast per season but are currently extending the ensemble size to ten members. The supermodel shows marked improvement in prediction skill for forecasts started before boreal spring, significantly overcoming the spring predictability barrier. Initial investigation indicates the skill enhancement is connected to better simulation of ocean-atmosphere interaction during the first part of the year, which also leads to improved initial conditions. Our results indicate the importance of better representing the signal-to-noise in the western and central Pacific during boreal spring.

How to cite: Keenlyside, N., Singh, T., Chiu, P.-G., Counillon, F., and Schevenhoven, F.: Overcoming the spring predictability barrier with a supermodel, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12107, https://doi.org/10.5194/egusphere-egu25-12107, 2025.

EGU25-12143 | Posters on site | CL4.6

Probabilistic climate outcomes from prediction aggregation 

Robin Lamboll, Sofia Palazzo Corner, and Moritz Schwarz

Currently, much of the literature around the Paris Agreement, Paris Compliance and manging the transition to net zero requires heavy use of integrated assessment models (IAMs). IAMs provide economic projections of future emissions, conditional on idealised scenarios. However, for most adaptation and cost-benefit analysis, policymakers require predictions, which IAMs do not even attempt to provide. How can we use aggregated estimates of emissions and resulting climate change to give probability distributions of climate impacts? We outline why human computation likely out-performs other prediction methods and present a flexible method to collect intended predictions from a variety of people to effectively estimate future emissions, temperatures and climate impacts via prediction aggregation platforms. These can subsequently be used to inform estimates of climate impacts. It can also highlight deficiencies in the IAM scenarios literature and indicate relative probabilities of scenarios. We estimate all-uncertainty temperatures in 2050 and outline extensions of the work.

How to cite: Lamboll, R., Palazzo Corner, S., and Schwarz, M.: Probabilistic climate outcomes from prediction aggregation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12143, https://doi.org/10.5194/egusphere-egu25-12143, 2025.

EGU25-12247 | ECS | Orals | CL4.6

Forecasting monthly-to-seasonal sea surface temperatures and marine heatwaves with graph neural networks and diffusion methods 

Varvara Vetrova, Ding Ning, Karin Bryan, and Yun Sing Koh

Knowing future sea surface temperature (SST) patterns play a crucial role not only in industries such as fisheries, shipping and tourism but also in conservation of marine species . For example, DNA of endangered species can be sampled prior to anticipated marine heatwaves to preserve marine biodiversity. Overall, availability of SST forecasts allows to mitigate potential adverse impacts of extreme events such as marine heatwaves. 

There is a strong interest in accurate forecasts of SST and their anomalies on various time scales. The commonly used approaches include physics-based models and machine learning (ML) methods. The first approach is computationally intensive and limited to shorter time scales. While several attempts have been made by the community to adapt ML models to SST forecasts several challenges still remain. These challenges include improving accuracy for longer lead SST anomaly forecasts. 

Here we present an integrated deep-learning based approach to the problem of SST anomalies and MHW forecasting. On one hand, we capitalise both on inherent climate data structure and recent advances in the field of geometric deep learning. We base our approach on a flexible architecture of graph neural networks, well suited for representing teleconnections. From another hand, we adapt the diffusion method to increase lead time of the forecasts.  Our integrated approach allows marine heatwave forecasts up to six months in advance.

How to cite: Vetrova, V., Ning, D., Bryan, K., and Koh, Y. S.: Forecasting monthly-to-seasonal sea surface temperatures and marine heatwaves with graph neural networks and diffusion methods, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12247, https://doi.org/10.5194/egusphere-egu25-12247, 2025.

The expansion of and increasing dependency on renewable energy that exploit climate variables, such as wind and precipitation, are highly sensitive to climate variability and weather extremes. Climate Futures is a Center of Research-based Innovation that aims to “co-produce new and innovative solutions for predicting and managing climate risks from sub-seasonal-to-seasonal (S2S) and seasonal-to-decadal (S2D) time scales with a cluster of partners in climate- and weather-sensitive sectors, including the renewable energy sector, through long-term cooperation between businesses, public organizations and research groups.

The aim of the cross-sectoral collaboration is for renewable energy companies to integrate improved climate predictions into their decision making. The long-term implications are a more resilient energy sector and stable power production. Examples of ongoing projects within the center include (1) using large ensemble climate model simulations to estimate near-future changes in precipitation variability, and (2) estimating future wind power production and variability using state-of-the-art decadal climate predictions. These results are important for future wind- and hydropower operations and infrastructure planning.

How to cite: Svendsen, L.: Climate services for and with the renewable energy sector in Norway, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12574, https://doi.org/10.5194/egusphere-egu25-12574, 2025.

EGU25-13076 | Posters on site | CL4.6

Usage of seasonal forecasts in Tropical Cyclone risk models 

Rudy Mustafa, Ulysse Naepels, Hugo Rakotoarimanga, Rémi Meynadier, and Clément Houdard

Tropical cyclones (TCs) pose significant risks to lives, infrastructure and economies, especially in coastal areas.

AXA has been developing stochastic natural hazard models (also called natural catastrophe or NatCat models) to quantify the impact of events such as TCs on its portfolios. However, NatCat models tend to model the average annual risk for a given peril. NatCat models do not consider the present state of the atmosphere and therefore are not conditioned with respect to the current tropical cyclone season.

Information about the TC risk in the upcoming weeks or months of a season could be crucial for an insurer, especially regarding its reinsurance coverage, but also for better risk mitigation through reinforced and more efficient prevention systems.

Previous studies have demonstrated that ensemble seasonal forecasts have skill in predicting TC occurrence several weeks in advance. We explore the ability of ensemble seasonal forecasts to provided skilled information on the general activity of the season to come for various lead-times (number of occurrences, number of landfalls, ACE…) and how can NatCat models be adapted to provide a more dynamic vision of the TC risk.

How to cite: Mustafa, R., Naepels, U., Rakotoarimanga, H., Meynadier, R., and Houdard, C.: Usage of seasonal forecasts in Tropical Cyclone risk models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13076, https://doi.org/10.5194/egusphere-egu25-13076, 2025.

EGU25-13668 | Orals | CL4.6

Forecasting the annual CO2 rise at Mauna Loa 

Richard Betts, Chris Jones, Jeff Knight, John Kennedy, Ralph Keeling, Yuming Jin, James Pope, and Caroline Sandford

For the last 9 years, the Met Office has issued forecasts of the annual increment in atmospheric carbon dioxide measured at Mauna Loa, accounting for both anthropogenic emissions and the effects of El Niño Southern Oscillation (ENSO) on natural carbon sinks and sources. The first forecast was produced when the 2015-2016 El Niño was emerging, and correctly predicted the largest annual CO2 increment on record at the time. In most years, the inclusion of ENSO provides a more skilful forecast than just considering emissions alone, except for 2022-2023 when La Niña conditions in late 2022 were followed by an early emergence of El Niño conditions in the second quarter of 2023. The impacts of interannual differences in emissions on the CO2 rise are usually smaller than those of ENSO variability, except in 2020 when the emergence of an unexpected large drop in global emissions due to societal responses to the COVID-19 pandemic required the forecast to be re-issued with a new estimate of the annual profile of emissions. Our forecast methodology also provides a simple means of tracking the changes in anthropogenic contributions to the annual atmospheric CO2 rise against policy-relevant scenarios. The Met Office forecast for 2023-2024 predicted a relatively large annual CO2 rise, but the observed rise was even larger, with exceptional wildfires in the Americas a likely contributor to the additional increase. Even without the effects of El Niño and other climatic influences on carbon sinks, the human-driven rise in CO2 in 2023-2024 would have been too fast to remain compatible with IPCC AR6 scenarios that limit global warming to 1.5°C with little or no overshoot. While the 2024-2025 rise is predicted to be smaller than 2023-2024, it will still be above these 1.5°C scenarios.

How to cite: Betts, R., Jones, C., Knight, J., Kennedy, J., Keeling, R., Jin, Y., Pope, J., and Sandford, C.: Forecasting the annual CO2 rise at Mauna Loa, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13668, https://doi.org/10.5194/egusphere-egu25-13668, 2025.

EGU25-13771 | Posters on site | CL4.6

Seasonal forecasting of East African short rains 

Giovanni Liguori, Agumase Kindie Tefera, William Cabos, and Antonio Navarra

The variability of East African Short Rains (October-December) has profound socioeconomic and environmental impacts on the region, making accurate seasonal rainfall predictions essential. We evaluated the predictability of East African short rains using model ensembles from the multi-system seasonal retrospective forecasts from the Copernicus Climate Change Service (C3S). We assess the prediction skill for 1- to 5-month lead times using forecasts initialized in September for each year from 1993 to 2016. Although most models exhibit significant mean rainfall biases, they generally show skill in predicting OND (October-December) precipitation anomalies across much of East Africa. However, skill is low or absent in some northern and western parts of the focus area. Along the East African coasts near Somalia and over parts of the western Indian Ocean, models demonstrate skill throughout the late winter (up to December-February), likely due to the persistence of sea surface temperature anomalies in the western Indian Ocean. Years when models consistently outperform persistence forecasts typically align with the mature phases of El Niño Southern Oscillation (ENSO) and/or Indian Ocean Dipole (IOD). This latter mode, when tracked using the Dipole Mode Index, is generally able to predict the sign of the rainfall anomaly in all models. Despite East Africa's proximity to the west pole of the IOD, the correlation between short rains and IOD maximizes when both east and west are considered. This finding confirms previous studies based on observational datasets, which indicate that broader-scale IOD variability associated with changes in the Walker Circulation, rather than local SST fluctuations, is the primary driver behind East African rainfall.     

How to cite: Liguori, G., Tefera, A. K., Cabos, W., and Navarra, A.: Seasonal forecasting of East African short rains, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13771, https://doi.org/10.5194/egusphere-egu25-13771, 2025.

EGU25-13847 | ECS | Posters on site | CL4.6

Decadal Predictions with Diffusion Models: Combining Machine Learning and Earth System Modelling 

Simon Lentz, Johanna Baehr, Christopher Kadow, Johannes Meuer, Felix Oertel, and Bijan Fallah

In the past years, decadal prediction systems have started to fill the gap between seasonal forecasts and long-term climate projections. Despite huge progress in predictive skill and decadal predictions outperforming climate projections in almost all forecast tasks, decadal predictions still possess large rooms for improvement. Machine learning based forecast systems have already outperformed traditional weather forecast systems in recent years. Similarly, machine learning has successfully transformed or assisted in data assimilation or climate data reconstruction tasks. Despite its success in the climate sciences, machine learning methods have not yet been successfully integrated in decadal prediction systems.

Combining machine learning and numerical modeling, we attempt to produce decadal climate predictions utilizing Diffusion Models, essentially probabilistic neural networks. We use such a neural network to predict global 2m-air temperatures by training it on the historical MPI-ESM-LR Grand Ensemble and finetuning it on the MPI-ESM-LR decadal predictions and on ERA5 reanalyses. The resulting predictions are qualitatively comparable to the standard MPI-ESM-LR decadal prediction system, surpassing their predictive skill for leadyears 1 and 2. With diffusion models still new to climate predictions, we expect this result to stand only at the beginning of further machine learning integration into climate predictions in general and decadal predictions in particular.

How to cite: Lentz, S., Baehr, J., Kadow, C., Meuer, J., Oertel, F., and Fallah, B.: Decadal Predictions with Diffusion Models: Combining Machine Learning and Earth System Modelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13847, https://doi.org/10.5194/egusphere-egu25-13847, 2025.

EGU25-15772 | Orals | CL4.6

A perfect-model perspective on the signal-to-noise paradox in initialized decadal climate predictions 

Markus G. Donat, Rashed Mahmood, Francisco J. Doblas-Reyes, and Etienne Tourigny

Initialized climate predictions are skillful in predicting regional climate conditions in several parts of the globe, but also suffer from different issues arising from imperfect initializations and inconsistencies between the model and the real world climate and processes. In particular, a so-called signal-to-noise paradox has been identified in recent years. This ‘paradox’ implies that the models can predict observations with higher skill than they predict themselves, despite some physical inconsistencies between modeled and real world climate. This is often interpreted as an indicator of model deficiencies.

Here we present a perfect-model decadal prediction experiment, where the predictions have been initialized using climate states from the model's own transient simulation. This experiment therefore avoids issues related to model inconsistencies, initialization shock and the climate drift that affect real-world initialized climate predictions. We find that the perfect-model decadal predictions are highly skillful in predicting the near-surface air temperature and sea level pressure of the reference run on decadal timescales. Interestingly, we also find signal-to-noise issues, meaning that the perfect-model reference run is predicted with higher skill than any of the initialized prediction members with the same model. This suggests that the signal-to-noise paradox may not be due just to model deficiencies in representing the observed climate in initialized predictions, but other issues that affect the statistical properties of the predictions. We illustrate that this signal-to-noise problem is related to analysis practices that concatenate time series from different discontinuous initialized simulations, which introduces inconsistencies compared to the continuous transient climate realizations and the observations. In particular, the concatenation of predictions initialized independently into a single time series breaks the auto-correlation of the time series.

How to cite: Donat, M. G., Mahmood, R., Doblas-Reyes, F. J., and Tourigny, E.: A perfect-model perspective on the signal-to-noise paradox in initialized decadal climate predictions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15772, https://doi.org/10.5194/egusphere-egu25-15772, 2025.

EGU25-18643 | Orals | CL4.6

Extending the Lead Time for European Winterstorm Activity Predictions 

Gregor C. Leckebusch, Kelvin S. Ng, Ryan Sriver, Lisa Degenhardt, Eleanor Barrie, and Elisa Spreitzer

The most dangerous and costly meteorological hazards in Europe are extreme extra-tropical cyclones and associated windstorms (EUWS) in winter. Recent studies have shown that seasonal prediction systems can skilfully predict the seasonal frequency of EUWS with a one-month lead time using November initialisations. Given that many seasonal prediction systems produce seasonal forecasts at the start of each month, this raises the question whether pre-November initialised seasonal forecasts could provide usable information in predicting seasonal activity of EUWS.

In this study, we will present preliminary results of an approach aimed at extending the predictive horizon of seasonal EUWS activity. While the direct outputs of the pre-November initialised seasonal predictions of EUWS do not have the sufficient skill, skilful predictions of seasonal EUWS activity can be obtained by an approach that utilises the information of the upper ocean mean potential temperature from seasonal prediction systems. Based on our approach, skilful predictions of seasonal EUWS activity becomes possible as early as October.

How to cite: Leckebusch, G. C., Ng, K. S., Sriver, R., Degenhardt, L., Barrie, E., and Spreitzer, E.: Extending the Lead Time for European Winterstorm Activity Predictions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18643, https://doi.org/10.5194/egusphere-egu25-18643, 2025.

Long-range winter predictions over the Euro-Atlantic sector have demonstrated significant skill but suffer from systematic signal-to-noise errors. Here, we examine sources of early winter seasonal predictability in across state-of-the-art seasonal forecasting systems. As in previous studies, these systems demonstrate skill in the hindcasts of the large-scale atmospheric circulation in early winter, associated with the East Atlantic pattern. The predictability is strongly tied to the ENSO teleconnection to the North Atlantic, though the systems' response to ENSO is systematically too weak. The hindcasts of the East Atlantic index exhibit a substantial signal-to-noise errors, with the systems' predicted signal generally being smaller than would be expected for the observed level of skill, though there is substantial spread across systems. The signal-to-noise errors are found to be strongly linked to the strength of the ENSO teleconnection in the systems, those with a weaker teleconnection exhibit a larger signal-to-noise problem. The dependency on modelled ENSO teleconnection strength closely follows a simple scaling relationship derived from a toy model. Further analysis reveals that the strength of the ENSO teleconnection in the systems is linked to climatological biases in the behaviour of the North Atlantic jet. 

How to cite: O'Reilly, C.: Signal-to-noise errors in early winter Euro-Atlantic predictions linked to weak ENSO teleconnections and pervasive North Atlantic jet biases, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18821, https://doi.org/10.5194/egusphere-egu25-18821, 2025.

EGU25-21570 | ECS | Posters on site | CL4.6

Predicting climate indicators at the decadal scale using a hybrid prediction system: application to SUEZ water management plans over France 

Joanne Couallier, Ramdane Alkama, Charlotte Sakarovitch, and Didier Swingedouw

As climate change reshapes hydrological cycles, workers in water management face unprecedented challenges in ensuring resource availability, mitigating flood risks, and maintaining resilient infrastructure. Nowadays, water utilities and authorities rely on long-term climate projections to plan for challenges extending through the end of the century. However, critical gaps persist in actionable information for shorter timescales, such as the decadal scale, which better aligns with political and operational decision-making. In this context, decadal climate predictions can be pivotal to address the needs of the water management sector and develop efficient climate services. However, their added values as compared to projections remained limited up to now.
To better understand user requirements, we collaborate with various teams from SUEZ, a company specializing in water management. Through interviews, we have identified the demand for specific indicators based on climate variables (e.g., precipitation, temperature) and corresponding spatio-temporal scales. Building on this understanding, we also develop in IPSL-EPOC decadal prediction team a new hybrid approach to improve our forecasts. This approach includes identifying a climate index (e.g., NAO, WEPA) derived from Sea Level Pressure (SLP) that correlates with the climate variable of interest. Using all the available decadal climate predictions from the DCPP project, we evaluate the predictability of this index, which is usually high for NAO and WEPA. This index is then employed to subsample a few of member CMIP6 climate projections that are in phase with the prediction of the DCPP ensemble. This latter step allows to inflate the amplitude of the predictable signal, resolving the limitation coming from the signal-to-noise paradox. It is also allowing to perform a proper statistical downscaling, used to refine these forecasts, ensuring their usability for identified needs. The resulting forecasts are designed to integrate seamlessly into SUEZ’s water sector models.
Preliminary work has identified diverse parameters of interest for water management, such as daily precipitation (resource availability forecasting), extreme precipitation events at fine temporal resolution (Combined Sewer Overflows modeling), and the number of very cold or very hot days (linked to risks of water mains and service lines failures, respectively). Early findings also suggest that, for the average precipitation over France, the WEPA index exhibits the largest correlations, unlike the NAO, which has greater influence for other European regions. The production of forecasts is currently underway, and their performance regarding the initially identified parameters will be presented.

How to cite: Couallier, J., Alkama, R., Sakarovitch, C., and Swingedouw, D.: Predicting climate indicators at the decadal scale using a hybrid prediction system: application to SUEZ water management plans over France, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21570, https://doi.org/10.5194/egusphere-egu25-21570, 2025.

EGU25-2316 | Posters on site | GI1.3

Development of an intelligent recirculating water system for land-based sea cucumber aquaculture 

Kuo-Hua Chien, Wen-Shun Huang, Jinn-Chyi     Chen , and Xiangfei   Ren 

  Presently, three fundamental methods are employed for sea cucumber aquaculture: pond culture, dam culture, and submarine seedling culture. However, these methods are susceptible to environmental water quality degradation due to factors such as sea cucumber feces, excessive feed, and climate change, which can impede sea cucumber growth and affect yields.

In order to address the issues outlined above, this study presents the intelligent circulating water system (ICWS), which is composed of a composite low-energy physical liquid-solid separator and a multi-mixed biofilter. A detailed description of these components is provided below.

  • Composite Low Energy Physical Liquid-Solid Separator

The liquid-solid separator uses minimal energy because of its innovative composite type. It extracts the contaminant source from the aquatic environment, reducing biofilter bed load and energy demand.

  • Multi-mixed Biofilter

The configuration of hybrid arrangement structures increases the specific surface area of the biofilter, leading to a reduction in its volume. The structure controls flow rate, hydraulic residence time, and hydraulic loading, which can be used to regulate temperature, salinity, pH, dissolved oxygen, and ammonia levels. This ensures the provision of high-quality water that meets the needs of sea cucumbers.

The innovative low-energy-consuming water recycling system outlined in this project has the theoretical potential to achieve complete water recycling without the necessity of replenishing the source water. This scenario presents a mutually beneficial opportunity for the sustainable utilization of Earth's water resources and the realm of commercial aquaculture, exhibiting no inherent incompatibility.

How to cite: Chien, K.-H., Huang, W.-S., Chen , J.-C.  .  ., and Ren , X.  .: Development of an intelligent recirculating water system for land-based sea cucumber aquaculture, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2316, https://doi.org/10.5194/egusphere-egu25-2316, 2025.

EGU25-4974 | ECS | Posters on site | GI1.3

Improving Discharge Measurement in Unmeasured Zones of ADCPs 

Jongmin Kim and Dongsu Kim

The Acoustic Doppler Current Profiler (ADCP) is one of the most commonly used instruments for measuring river discharge by utilizing the Doppler effect of acoustic waves. However, its reliance on a single transducer introduces certain limitations. During the transition between transmission and reception of the acoustic signal, the returning signal cannot be captured, resulting in an inability to measure discharge near the sensor.

Additionally, side-lobe interference generated by acoustic waves reflects off the riverbed and contaminates measurements near the bottom. To mitigate this, discharge data within 5% of the water depth from the bottom are typically excluded from results. Furthermore, in shallow areas where the unmeasured regions near the sensor and near the bottom overlap, discharge cannot be accurately measured.

To address these gaps, discharge in the unmeasured regions of ADCP measurements is typically extrapolated using data from the measured sections or calculated using empirical equations. In this study, a method to improve the measurement accuracy in the unmeasured regions of the ADCP was developed and evaluated.

 

Acknowledgements 

This work was supported by Korea Environment Industry & Technology Institute(KEITI) through Research and development on the technology for securing the water resouces stability in response to future change Program, funded by Korea Ministry of Environment(MOE)(RS-2024-00336020)

How to cite: Kim, J. and Kim, D.: Improving Discharge Measurement in Unmeasured Zones of ADCPs, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4974, https://doi.org/10.5194/egusphere-egu25-4974, 2025.

EGU25-9193 | Posters on site | GI1.3

Monitoring for  sustainable and inclusive urban areas 

Francesco Soldovieri, Vincenzo Cuomo, and Jean Dumoulin

Urban areas need to rethink their policies to strengthen their capacities to prepare for and respond to hazards and become more resilient, intelligent and inclusive. In this context, one of the objectives is to ensure the resilience of their services and systems against multi-hazard scenarios, where the effect of local hazards combines with global challenges such as climate change and pandemics. Moreover, the concept of inclusiveness is becoming crucial, as highlighted during COVID, which showed that the most vulnerable population is the one living in sparsely and densely populated areas, where the level of social and physical services is often inadequate [1].

In this context, one possible response to this need is the development of monitoring and surveillance approaches [2]. The present contribution will focus on three aspects

The first is that resilience must be addressed as a whole, since services and networks are interconnected and interdependent (e.g. health system, transport, energy and water distribution, air quality, protection from extreme weather events, etc.). The main consequence of these interconnections is that the complete collapse of services (blackout) may become a realistic possibility.

The second aspect is that resilience can only be achieved in the presence of continuous and detailed monitoring of both the structures/infrastructure/services and the territory on which they insist, and that without such a monitoring it is impossible to correctly define the interventions to be carried out and their priorization.

The third aspect concerns the development of new monitoring systems based on Earth observation, positioning, navigation, and ICT technologies that exploit the citizen as a sensor and the so-called ‘non-sensors’, i.e. sensors that provide useful information for monitoring even if they are not designed for this purpose. All this ‘sensory’ data must be integrated to obtain a complete and reliable awareness of the scenario; hence the need to process and systematize large amounts of information that can only be processed by AI and HPC.

 

[1] V. Cuomo F. Soldovieri F. Bourquin, N. -E. El Faouzi, J. Dumoulin. The necessities and the perspectives of the monitoring/surveillance systems for multi-risk scenarios of urban areas including COVID-19 pandemic. Proceedings of the TIEMS Annual Conference, 18-20 November 2020, Paris, France, ISBN: 978-94-90297-19-0, vol. 27

[2] Cuomo V., Soldovieri F., Ponzo F.C., Ditommaso R. (2018). A holistic approach to long-term SHM of transport infrastructures. The International Emergency Management Society (TIEMS) Newsletter 33, pp. 67-84.

How to cite: Soldovieri, F., Cuomo, V., and Dumoulin, J.: Monitoring for  sustainable and inclusive urban areas, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9193, https://doi.org/10.5194/egusphere-egu25-9193, 2025.

EGU25-10582 | Orals | GI1.3

Near real-time, water quality event monitoring in small rivers, in the context of increasing frequency and intensity of hydrodynamic events due to climate change. 

Lisa Cronin, Cian M. Taylor, Ciprian Briciu-Burghina, Fiona Regan, and Frances E. Lucy

Freshwater quality continues to decline despite the adoption of the Water Framework Directive (WFD) almost twenty five years ago with the recovery of water quality in Europe plateauing since the 2010s (Haase et al., 2010).  Pollution from diffuse sources, particularly from agriculture remains a key challenge to restoring water quality to at least ‘good status’ under the WFD (EEA, 2018) compounded by water quality declines due to increased frequency and intensity of hydrodynamic events (van Vliet et al., 2023). 

Assigning accurate WFD classes and detecting changing trends in water quality have been challenging where traditional low frequency monitoring approaches have been implemented (Skeffington et al., 2015).  Higher monitoring frequency and spatial coverage is required to effectively identify improvements in water quality (Westerhoff et al., 2022) particularly when detecting changes over shorter time periods (Mcdowell et al., 2012).  High frequency monitoring is required to identify temporal water quality changes linked to rainfall driven pollutant transfer from land to waters (Métadier and Bertrand-Krajewski, 2012) with monitoring over multiple events required to capture the variability in pollutant concentrations and pollutant loads across events (Kozak et al., 2019).  Furthermore, 50% of surface waterbodies in the EU are impacted by multiple pressures (EEA, 2018), with increased urbanisation requiring a more complex, multi-pollutant approach to assessing impacts on river quality (Strokal et al., 2021).

The aim of this research was to identify if rainfall driven transient pollution events were occurring at two monitoring stations in a river catchment, and if continuous instream monitoring of turbidity and other water quality parameters could be used to capture changes in water quality and potential instances of such events.  One of the objectives was to identify if continuous monitoring could create a site-specific water quality profile that could be used to identify early warning indicators of rainfall driven or other transient pollution events. 

Results from this study indicate that changes in water quality are happening during rainfall events and that turbidity alongside other parameters can be used to track such events, trigger alarms when a probable event is occurring and automatically activate more intense monitoring during these events.  The integrated monitoring approach adopted allows for the tracking of water quality changes across temporal and spatial scales for multiple pollutants and allows for temporal fluctuations, and variation in pollutant loads during hydrodynamic events to be determined. 

The significant advantages of this approach are it’s suitability for remote deployments with no requirement for permanent infrastructure, the use of site specific water quality profiles to identify potential water quality events at individual sites and to activate further monitoring if required, the ability to tailor the monitoring for pollutant screening or more specific pollutants of concern, and the cost effectiveness of moving the integrated monitoring station between different water bodies. 

How to cite: Cronin, L., Taylor, C. M., Briciu-Burghina, C., Regan, F., and Lucy, F. E.: Near real-time, water quality event monitoring in small rivers, in the context of increasing frequency and intensity of hydrodynamic events due to climate change., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10582, https://doi.org/10.5194/egusphere-egu25-10582, 2025.

EGU25-10725 | ECS | Posters on site | GI1.3

Advanced Ultrasound Techniques for Investigating Air-Water Two-Phase Flow: An Experimental Approach 

Juan Calderon, Max Dormann, Till Branß, Martin Balcewicz, Jochen Aberle, and Erik Saenger

The König-Project, funded by the German state, is a long-term project focused primarily on developing a multi-scale wave measurement laboratory to improve flow measurements in an industrial context. Part of this project researches the propagation of ultrasound waves inside a moving fluid. A wide variety of flow scenarios are considered, and new methods for ultrasonic flow measurement can be developed and optimized. One experimental scenario includes the determination of volume fraction and drop size distributions of air dispersed in water using ultrasonic waves.

For this purpose, a modular system is used as an initiative to integrate manufacturer-independent measurement components with open-source software for the acquisition and processing of ultrasound signals. The modular system equipment consists of a multichannel system, which allows the positioning of several transceivers to send and receive ultrasonic waves from different directions along the experimental zone of interest. The concentration of dispersed air in water will be determined by measuring the reduced transit time caused by the added compressibility of the air phase.

Characterizing multiphase flows using other techniques can be time-consuming and the accuracy can fall short as the complexity of the fluid grows. The use of ultrasound to characterize fluid flows has many advantages such: as a non-invasive method that doesn’t alter the fluid path, real-time data acquisition, and high-temporal resolution, it is cost-effective and can be used on opaque fluids. Therefore this technique is gaining more attention in several industrial applications, including oil and gas, hydrogen, and geothermal energy generation. The results of this investigation will be validated and compared with the output of a numerical simulation, in which the boundary conditions and the flow characteristics will be similar to the experimental setup.

 

How to cite: Calderon, J., Dormann, M., Branß, T., Balcewicz, M., Aberle, J., and Saenger, E.: Advanced Ultrasound Techniques for Investigating Air-Water Two-Phase Flow: An Experimental Approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10725, https://doi.org/10.5194/egusphere-egu25-10725, 2025.

EGU25-10832 | ECS | Posters on site | GI1.3

Numerical Study to Determine Water-Air Dispersion with Ultrasound Waves 

Max Dormann, Juan Calderon, Claudia Finger, Martin Balcewicz, and Erik H. Saenger

The König-project is funded by the German state with the aim to develop a calibrated and virtual measurement laboratory to enhance methods based on ultrasound measurements that find application in the determination of flow velocity or particle movement. By comparing the results of controlled laboratory and real-world experiments with numerical simulations, the understanding of the interaction between ultrasonic waves and fluid flow is intended to be improved. The amount of scatterers within a fractured medium directly affects  the effective velocity of elastic waves. Thus we investigate, if the effects found in solid media can be transferred to fluids. We ran a series of numerical experiments, simulating ultrasound transmission measurements for multiple concentrations of bubbles of varying diameter dissolved in a stationary water layer. For the simulation of elastic wave propagation, we used a rotated staggered finite-difference scheme. We investigate the relation between the effective wave speed and the bubble concentration and compare those to results of laboratory experiments. Future research will then expand to moving fluid-gas mixtures.

How to cite: Dormann, M., Calderon, J., Finger, C., Balcewicz, M., and Saenger, E. H.: Numerical Study to Determine Water-Air Dispersion with Ultrasound Waves, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10832, https://doi.org/10.5194/egusphere-egu25-10832, 2025.

EGU25-10867 | Posters on site | GI1.3

Radar Altimetry Reveals the Smoothness of the Surface: the Case of Salar de Uyuni, Bolivia 

Francesco De Biasio, Stefano Vignudelli, Ron Abileah, and Paula Pacheco Mollinedo

Salar de Uyuni is a salt desert in Bolivia, spanning approximately 10,000 km2. During the wet season a thin layer of rainfall water covers the salt flats, making its surface mirror-like and earning it the title of “the largest natural mirror in the world”. The surface reflects the sky like a mirror, and attracts tourists who document this effect only from its outer perimeter. No evidence is documented in the interior, accessible only during the dry season. The only frequent observations of the Salar surface are from satellites, particularly altimetric radars, which are specifically designed to measure topography. Originally developed to measure sea level [1], they have recently been used, with a different metrics, to describe how emitted radar pulses are reflected by the surface, measuring the intensity of the reflected echo, and thus the Radar Cross Section (RCS) of the surface [2]: higher RCSs correspond to smoother surfaces. RCS was initially estimated in [1] with a an approximate method. Later EUMETSAT shared a better estimate by solving the radar equation with satellite parameters that were previously unknown to us [3]. In this study we used Sentinel-3A and 3B RCS measurements over the Salar flats, along six ground tracks, to describe for the first time the evolution of the Salar surface smoothness in space and time. A field campaign (16th - 20th of February 2024) was also conducted to validate the interpretation of radar measurements during the Sentinel-3A overpass on the track 167. At the field site, in a water depth of 1.8 cm (horizontal wind 4.5-3.4 ms-1), we measured a null vertical surface displacement to within ±0.5 mm, which classifies the surface as electromagnetically smooth at the radar frequency. The RCS values near the site were around 120 dBsm, as expected for radar return from a smooth surface. Three peaks are observed on the statistical distribution of the RCS: 87 (dry), 101 (intermediate) and 120 dBsm (wet season).The wet season, characterized by values above 101 dBsm, begins in December, peaking from late January to early March. February thus ensures the highest chance to observe mirror-like effects. Rainfall climatology from Uyuni city meteorological station reflects such statistics. The spatial and temporal evolution of RCS over the Salar, however, do not describe this place like a uniform mirror at the radar frequency, and so it is unlikely to observe such effect at shorter wavelengths, contrary to what is believed in the literature. Finally, satellites can help tourism stakeholders in programming the most enjoyable experience for travellers.

[1] Vignudelli et all. 10.1007/s10712-019-09569-1

[2] Abileah and Vignudelli, 10.1016/J.Rse.2021.112580

[3] Dinardo and Lucas, EUM/RSP/TEN/23/1376566

How to cite: De Biasio, F., Vignudelli, S., Abileah, R., and Pacheco Mollinedo, P.: Radar Altimetry Reveals the Smoothness of the Surface: the Case of Salar de Uyuni, Bolivia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10867, https://doi.org/10.5194/egusphere-egu25-10867, 2025.

EGU25-12996 | Orals | GI1.3

Integrating Remote Sensing Technique with 3D Numerical Modelling for Enhanced Maintenance of Critical Infrastructure in Landslide-Prone Areas 

Diana Salciarini, Alice Vitaletti, Erica Cernuto, and Filippo Ubertini

Landslides, alongside earthquakes and floods, are among primary natural phenomena that are responsible for significant social and economic losses. Their impact poses an increasing threat worldwide, particularly in marginal and degraded contexts, affecting urban areas, infrastructures, environmental, historical, and cultural heritage, and, in severe cases, resulting in human casualties. In recent years, the number of infrastructure collapses or severe structural damages due to landslide movements has risen significantly, hindering the functionality of infrastructures, and highlighting the urgent need to deeply understand their interactions. Landslides can endanger roads, bridges, and railways, compromising the accessibility and inclusivity, and exacerbating social and economic exclusion in affected areas. Critical infrastructures are often located in challenging areas, where the susceptibility to landslides and natural hazards is significantly elevated. These sites demand advanced monitoring technologies to ensure infrastructure safety and mitigate the social and economic impacts of landslides. This study explores an innovative approach that integrates Interferometric Synthetic Aperture Radar (InSAR) data with numerical Finite Element Modelling (FEM) to address these challenges. The proposed method was applied to a case study involving a partial interaction between a slow-kinematic landslide, documented in the Inventory of Landslide Phenomena in Italy (IFFI), and a bridge along a highway section in the Liguria Region. Leveraging high-resolution satellite-based data from the Copernicus European Ground Motion Service (EGMS), the InSAR analysis provided spatial and temporal monitoring of ground displacements. Satellite remote sensing offers a wide spatial and temporal coverage over multiple regions, enabling for the detection of extensive or hard-to-access areas with millimetric precision in deformation velocity, ensuring high efficiency at a favourable cost-benefit ratio. However, while InSAR analysis can precisely measure ground motions, it lacks the ability to provide insights into the physical mechanisms under varying loading conditions. To address this limitation, FEM modelling was used to simulate the three-dimensional landslide mechanical behaviour under hydraulic loading, offering a deeper understanding of the slope stability and infrastructure deformations. InSAR data post-processing enabled the estimation of transverse and vertical components of the actual displacement vector, aligning with the observed landslide deformations and facilitating the numerical model validation. Simultaneously, FEM results highlighted significant displacements downstream of the landslide area, indicating a slope stability close to the limit equilibrium condition. Quantitative analysis also revealed relevant deformations at the base of bridge piers located within the landslide, caused by horizontal forces impacting the foundations. The integration of InSAR observations and FEM calculations demonstrated consistency in the identified movement, validating the efficacy of the combined method in identifying critical zones in landslide-prone regions. This study highlights how advanced remote sensing technologies, when coupled with numerical simulations, can enhance the monitoring and maintenance of critical infrastructure, particularly in marginal or extensive contexts. By identifying vulnerable areas and supporting the maintenance strategies, this methodology can contribute to hydrogeological risk management and promote inclusivity in regions where social and economic disparities exacerbate natural hazards impacts.

How to cite: Salciarini, D., Vitaletti, A., Cernuto, E., and Ubertini, F.: Integrating Remote Sensing Technique with 3D Numerical Modelling for Enhanced Maintenance of Critical Infrastructure in Landslide-Prone Areas, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12996, https://doi.org/10.5194/egusphere-egu25-12996, 2025.

EGU25-13576 | Orals | GI1.3

An Approach for IoT-Based Smart Sensors Placement in Urban Water Networks Under Natural Hazards 

Bahram Malekmohammadi, Mehdi Rahimi, Reza Kerachian, Vijay P. Singh, Roger A. Falconer, Roohollah Noori, and Farhad Bahmanpouri

Natural hazards such as floods, storms, and earthquakes present significant threats to urban infrastructures, particularly water supply and distribution networks. These events can severely impact the quality and quantity of water resources, leading to serious consequences for public health and social security. Factors such as unplanned urban development and non-compliance with engineering standards further increase the vulnerability of these systems. Recent advancements in technologies like the Internet of Things (IoT) and Artificial Intelligence (AI) have enabled real-time monitoring and data analysis of these critical infrastructures. IoT-based smart sensors capture essential information, including flow rate, water quality, corrosion, leakage, and pipeline ruptures. These data are processed using machine learning and deep learning algorithms to identify anomalies. Such systems can enhance monitoring capabilities and support effective decision-making in crisis situations. This study explores key criteria for selecting optimal locations for sensor deployment. These criteria include connection points, infrastructure accessibility, water quality, natural hazard risks, and historical incident data. For example, evaluating the location of connection points and their impact on water flow and distribution can help identify optimal routes, reducing costs and response times. Easy access to infrastructure facilitates sensor installation and maintenance, improving system efficiency. Monitoring water quality at various points in the distribution network is also critical to identifying sensitive locations and ensuring water safety. Additionally, identifying areas prone to natural hazards helps prioritize vulnerable regions for monitoring and improve system resilience. Historical data on anomalies and past incidents provide patterns that highlight risk-prone areas and help refine monitoring strategies. Based on these criteria, a multi-criteria decision-making approach is applied to propose the most effective locations for sensor placement. This method suggests prioritizing locations that have the highest impact and accessibility. These recommendations aim to enhance system efficiency and improve response capabilities during emergencies.

Ketwords: Smart Infrastructures, Internet of Things, MCDM, Artificial Intelligence, Natural Hazards

How to cite: Malekmohammadi, B., Rahimi, M., Kerachian, R., Singh, V. P., Falconer, R. A., Noori, R., and Bahmanpouri, F.: An Approach for IoT-Based Smart Sensors Placement in Urban Water Networks Under Natural Hazards, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13576, https://doi.org/10.5194/egusphere-egu25-13576, 2025.

EGU25-17346 | Orals | GI1.3

Drone based radar technologies for wide rural areas resources exploration: potentialities and challenges 

Ilaria Catapano, Giuseppe Esposito, Gianluca Gennarelli, and Francesco Soldovieri

Rural areas, i.e. areas with a low population density and a small number of anthropogenic environments, represent a significant resource in the pursuit of a green and sustainable development of the European Community [1]. This development involves not only the ecological and balanced use of agriculture and forestry resources, but also policies devoted to environmental protection and monitoring. In this context, drone-based technologies offer valuable opportunities because they facilitate the effective and non-invasive surveillance and monitoring of wide and inaccessible places. These technologies, indeed, allow surface and subsurface explorations, while concomitantly reducing the financial and logistical demands associated with investigation missions.

The present contribution is focused on Unmanned Aerial Vehicle (UAV)-Ground Penetrating Radar (GPR) technology and the potential of UAV-GPR technological solutions in subsurface prospecting [2]. The discussion encompasses the collection and processing of data, emphasising the efficacy and sustainability of the technology. The contribution will address the development of guidelines for the design of the flight grid and the formulation of an effective imaging strategy that can account for deviations in motion relative to the nominal trajectory.

[1] Bizottság, E. (2024). The long-term vision for the EU’s rural areas: key achievements and ways forward. Report from the Commission to the European Parliament, the Council, the European Economic and Social Committee and the Committee of the Regions, Brüsszel, The long-term vision for the EU’s rural areas: key achievements and ways forward, Report from the Commission [Letöltve: 2024.06. 20.].

[2] Noviello, C., Gennarelli, G., Esposito, G., Ludeno, G., Fasano, G., Capozzoli, L., Soldovieri, F., & Catapano, I. (2022). An Overview on Down-Looking UAV-Based GPR Systems. Remote Sensing, 14(14), 3245. https://doi.org/10.3390/rs14143245

Acknowledgements: The communication has been funded by EU - Next Generation EU Mission 4 “Education and Research” - Component 2: “From research to business” - Investment 3.1: “Fund for the realisation of an integrated system of research and innovation infrastructures” - Project IR0000032 – ITINERIS - Italian Integrated Environmental Research Infrastructures System - CUP B53C22002150006.

The authors acknowledge the Research Infrastructures participating in the ITINERIS project with their Italian nodes: ACTRIS, ANAEE, ATLaS, CeTRA, DANUBIUS, DISSCO, e-LTER, ECORD, EMPHASIS, EMSO, EUFAR ,Euro-Argo, EuroFleets, Geoscience, IBISBA, ICOS, JERICO, LIFEWATCH, LNS, N/R Laura Bassi, SIOS, SMINO.

How to cite: Catapano, I., Esposito, G., Gennarelli, G., and Soldovieri, F.: Drone based radar technologies for wide rural areas resources exploration: potentialities and challenges, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17346, https://doi.org/10.5194/egusphere-egu25-17346, 2025.

EGU25-18196 | Posters on site | GI1.3

SEAWATCH Project: A year of advancements in Short-Range K-Band Radar for Coastal Monitoring 

Giovanni Ludeno, Pasquale Contestabile, Diego Vicinanza, Matteo Antuono, Caludio Lugni, Ilaria Catapano, Giuseppe Esposito, Carlo Noviello, Francesco Soldovieri, and Gianluca Gennarelli

Coastal regions are crucial for human settlements and economic development. However, their distinctive environmental characteristics, particularly in deltas, bays, and gulfs, render them highly vulnerable to threats such as erosion phenomena and pollution. The effective management of these areas depends on the accurate predictions of wave dynamics and their interactions with the shoreline and seabed. Reliable forecasts require numerical wave propagation models to be initialized with precise data and detailed bathymetric representations, and their accuracy depends on calibration operations using high-quality sea state observations.

Sea state data are typically collected through in-situ sensors, such as buoys and drifters, or remote sensing devices, including radars and video-monitoring systems [1]. Remote sensing technologies are often preferred due to their ability to provide both spatial and temporal information. Among these, ground-based radar systems like High-Frequency and X-band radars have proven effective in retrieving wave spectra and coastal sea state information. However, these systems face notable limitations, including difficulties in acquiring data near the shoreline. Additionally, they are bulky, heavy, and cumbersome, which complicates the deployment stage.

To address these challenges, the Italian PRIN-PNRR 2022 Project SEAWATCH—Short-Range K-Band Wave Radar System Close to the Coast—was launched on November 30, 2023. SEAWATCH focuses on developing an innovative, portable, short-range K-band radar prototype specifically designed for sea state monitoring in nearshore zones. Thanks to its compact size, lightweight design, and low power requirements, the system enables flexible, on-demand surveys, meeting critical safety and environmental management needs in harbors and coastal zones.

This communication outlines the key activities and initial results achieved during the first year of the SEAWATCH project. This last is organized into six milestones, supported by a robust collaboration between research units to ensure efficient knowledge sharing and steady progress. Preliminary here results shown highlight the radar prototype potential to overcome traditional limitations, offering enhanced spatial resolution and real-time monitoring capabilities near the coastline [2]-[4].

Future efforts will focus on further refining the radar prototype and validating its performance across diverse coastal environments.

 

References:

  • P. Neill, M. Reza Hashemi, Chapter 7 - In Situ and Remote Methods for Resource Characterization, Editor(s): Simon P. Neill, M. Reza Hashemi, In E-Business Solutions, Fundamentals of Ocean Renewable Energy, Academic Press, 2018, Pages 157-191.
  • Afolabi, L. A., et al. (2025). Underestimation of Wave Energy from ERA5 Datasets: Back Analysis and Calibration in the Central Tyrrhenian Sea. Energies, 18(1), 3.
  • Ludeno, G., Antuono, M., Soldovieri, F., & Gennarelli, G. (2024). A Feasibility Study of Nearshore Bathymetry Estimation via Short-Range K-Band MIMO Radar. Remote Sensing16 (2), 261.
  • Ludeno, G.; Esposito, G.; Lugni, C.; Soldovieri, F.; Gennarelli, G. A Deep Learning Strategy for the Retrieval of Sea Wave Spectra from Marine Radar Data.  Mar. Sci. Eng.202412, 1609.

 

Acknowledgment: This work was supported and funded by the European Union—NextGenerationEU PNRR Missione 4 “Istruzione e Ricerca”—Componente C2 Investimento 1.1, “Fondo per il Programma Nazionale di Ricerca e PRIN—SEAWATCH—Short-rangE K-bAnd Wave rAdar sysTem Close to tHe coast CUP B53D23023940001, and partially funded by the research project STRIVE—La scienza per le transizioni industriali, verde, energetica CUP B53C22010110001.

How to cite: Ludeno, G., Contestabile, P., Vicinanza, D., Antuono, M., Lugni, C., Catapano, I., Esposito, G., Noviello, C., Soldovieri, F., and Gennarelli, G.: SEAWATCH Project: A year of advancements in Short-Range K-Band Radar for Coastal Monitoring, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18196, https://doi.org/10.5194/egusphere-egu25-18196, 2025.

EGU25-18356 | ECS | Posters on site | GI1.3

Applicability of cheap and lightweight magnetic sensors to geophysical exploration 

Filippo Accomando and Giovanni Florio

 In recent years, there was a notable technological advancement in geophysical sensors. In the case of magnetometry, several sensors were used having the common feature to be miniaturized and lightweight, thus idoneous to be carried by UAV in drone-borne magnetometric surveys. Moreover, such sensors have the common feature to be very cheap, so that it is in principle very easy to have the resources to combine two or three of them to form gradiometers. Nonetheless, another common feature is that their sensitivity ranges from 0.1 to about 200 nT, thus not comparable to that of alkali vapor, standard flux-gate or even proton magnetometers. However, their low-cost, small volume and weight remain as very interesting features of these sensors. In this communication, we want to explore the range of applications of small tri-axial magnetometers commonly used for attitude determination in several devices. We compare the results of ground-based surveys performed with conventional geophysical instruments with those obtained using these sensors.

 

How to cite: Accomando, F. and Florio, G.: Applicability of cheap and lightweight magnetic sensors to geophysical exploration, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18356, https://doi.org/10.5194/egusphere-egu25-18356, 2025.

EGU25-18926 | ECS | Orals | GI1.3

Remote Sensing for Volcanic Eruptions and Earthquake Emergency Management Strategies in Developing Countries 

Tesfaye Tessema, Elias Lewi, and Fabio Tosti

Volcanic eruptions and earthquakes present significant challenges to developing countries, where limited monitoring infrastructure restricts effective risk mitigation efforts. Satellite remote sensing observations offer essential information, including surface deformation and thermal anomalies, for hazard assessment, early warning, and emergency response. These satellite-based observations enable comprehensive spatial and temporal monitoring, utilising both publicly available medium-resolution and commercial high-resolution datasets. Over the past decade, Sentinel radar and optical observations have been employed in areas with limited in-situ measurement capabilities[1]. Nonetheless, the utilisation of these datasets in developing countries is frequently hampered by insufficient computational and analytical resources.

This study examines the role of remote sensing in strengthening disaster risk management within resource-constrained contexts. We propose a collaborative framework that utilises satellite remote sensing data processing Centres in developed countries to assist developing nations in analysing pre-, during, and post-crisis events. Moreover, we advocate for engaging with space agencies to enhance satellite tasking during crisis observation, thereby improving our understanding of the event’s driving mechanisms. We highlight the critical role of remote sensing through a case study of recent seismic and volcanic activity in the Main Ethiopian Rift, specifically between the Fentale and Dofen volcanoes[2]. While national seismic and geodetic networks provide data on large and medium-magnitude earthquakes and significant deformations, they cannot detect low-magnitude precursory events or local deformations due to their proximity to volcanic centres. Furthermore, the installation of temporary monitoring facilities is often constrained by various limitations. Remote sensing bridges this gap by offering detailed data to support local research, inform timely decision-making, and strengthen crisis management. The crises have impacted under-resourced regions, the primary import-export corridor, and nearby urban centres, including Addis Ababa, where rapid urbanisation has raised safety concerns. This study underscores the necessity of integrated remote sensing solutions and international collaboration to enhance resilience and mitigate risks in disaster-prone areas.

Keywords: Sentinel, Main Ethiopian Rift, Fentale Volcano, Developing Countries, Emergency Management

 

Acknowledgements

The Authors would like to express their sincere thanks and gratitude to the following trusts, charities, organisations and individuals for their generosity in supporting this project: Lord Faringdon Charitable Trust, The Schroder Foundation, Cazenove Charitable Trust, Ernest Cook Trust, Sir Henry Keswick, Ian Bond, P. F. Charitable Trust, Prospect Investment Management Limited, The Adrian Swire Charitable Trust, The John Swire 1989 Charitable Trust, The Sackler Trust, The Tanlaw Foundation, and The Wyfold Charitable Trust.

 

References

[1] Tessema, T. T., Biggs, J., Lewi, E., & Ayele, A. (2020). Evidence for active rhyolitic dike intrusion in the northern Main Ethiopian Rift from the 2015 Fentale seismic swarm. Geochemistry, Geophysics, Geosystems, 21, e2019GC008550. https://doi.org/10.1029/2019GC008550

[2] Derek Keir, Alessandro La Rosa, Carolina Pagli, et al. (2024). The 2024 Fentale Diking Episode in a Slow Extending Continental Rift. ESS Open Archive DOI: 10.22541/au.172979388.80164210/v1

How to cite: Tessema, T., Lewi, E., and Tosti, F.: Remote Sensing for Volcanic Eruptions and Earthquake Emergency Management Strategies in Developing Countries, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18926, https://doi.org/10.5194/egusphere-egu25-18926, 2025.

EGU25-19565 | ECS | Orals | GI1.3

Innovative Geothermal Mining through Membrane Technologies 

Bruno Marco Inzillo, Sergio Santoro, Efrem Curcio, and Salvatore Straface

Critical raw materials (CRMs) are crucial for technological advancements and the global energy transition, especially in sectors such as renewable energy, electronics, and electric mobility. The sustainable and secure management of these materials is increasingly important. Geothermal springs represent a promising source of CRMs, offering valuable materials such as lithium, magnesium, strontium, and boron in addition to clean energy. Depending on where they come from geologically, geothermal springs can have lithium levels that are at least 10 times higher than seawater (0.18 mg/L) and about the same as salt lakes (0.04–3 g/L). The moderate Mg2+/Li+ molar ratio (~35) also shows that the two elements might be better separated, which would allow for more Mg2+ recovery. This study introduces a novel method for the recovery of CRMs from geothermal brines, combining Reverse Osmosis (RO), Nanofiltration (NF), and Membrane Distillation (MD) for efficient separation of water and valuable materials. The experiments are conducted using a synthetic laboratory-reproduced geothermal spring solution, which accurately replicates the pH, temperature, and ionic composition typical of natural geothermal waters. This experimental approach ensures that the results reflect real-world conditions, which is critical for evaluating the feasibility and scalability of the proposed method. The process begins with RO and NF to concentrate the brine and selectively separate multivalent ions (e.g., Mg) from monovalent ions (e.g., Li), leveraging differences in ionic valence. Following this, MD is applied to reduce brine volume and minimize thermal energy consumption, thereby optimizing both water recovery and the concentration of CRMs. A key innovation of this work is the exploitation of the elevated temperature of geothermal brines (> 35°C), which allows the use of MD with minimal external heating. This significantly reduces energy requirements and operational costs. The process minimizes Specific Thermal Energy Consumption (STEC), highlighting its efficiency and sustainability. This method not only enhances the recovery of lithium and magnesium from geothermal springs, but it also offers a cleaner, more sustainable approach to CRM extraction by utilizing renewable geothermal heat.

How to cite: Inzillo, B. M., Santoro, S., Curcio, E., and Straface, S.: Innovative Geothermal Mining through Membrane Technologies, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19565, https://doi.org/10.5194/egusphere-egu25-19565, 2025.

EGU25-19593 | ECS | Orals | GI1.3

Enhancing Hydrological Models with Remote Sensing: A Review of Products, Techniques, and Uncertainties 

Soufiane Taia, Yassine Ait Brahim, Mohammed Hssaisoune, Andrea Scozzari, and Bouabid El Mansouri

Distributed hydrological models are crucial for flood prediction, drought analysis, and water resource monitoring. They are typically calibrated using streamflow observations at the watershed outflow to determine the best parameter values within their common ranges. These models are then applied to analyze management and climate scenarios. However, accurately representing hydrological complexities is challenging due to limited knowledge, data availability, and imprecise measurements. Uncertainties in these models arise from parameters, model structure, calibration processes, and data, especially in regions with scarce data. Consequently, hydrological models require extensive hydro-meteorological data for calibration and validation, which can be costly and time-consuming. Recently, remote sensing techniques advanced hydrological modeling by providing regular sampling of essential variables like precipitation, soil moisture, and evapotranspiration. However, thanks to technological advancements, numerous global and regional remote seeing products for the same variable have become freely available. These products vary in their algorithms, approaches, spatial and temporal resolutions, leading to diverse datasets for the same variable. Therefore, different products can perform differently in terms of parameter estimation, model robustness, and water balance predictions within the same area. However, each product may introduce biases or uncertainties, necessitating modelers to assess their performance and carefully choose the most suitable product for their study objectives. This research reviews commonly used remotely sensed products and the techniques and approaches for integrating them into distributed and semi-distributed hydrological models. Additionally, this review examines the uncertainties associated with different existing products and their performance within hydrological models.

How to cite: Taia, S., Ait Brahim, Y., Hssaisoune, M., Scozzari, A., and El Mansouri, B.: Enhancing Hydrological Models with Remote Sensing: A Review of Products, Techniques, and Uncertainties, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19593, https://doi.org/10.5194/egusphere-egu25-19593, 2025.

EGU25-19717 | ECS | Orals | GI1.3

Advancing Community-Based Water Quality Monitoring through Low-Cost Open-Source Optical Sensors and Data Integration 

Riccardo Cirrone, Amedeo Boldrini, Alessio Polvani, Xinyu Liu, and Steven Loiselle

To meet European (WFD) and International objectives (SDGs), there is a growing demand for water quality data with elevated spatial and temporal resolution. This has been an ongoing process, achieved by integrating data from governmental agencies with community-based monitoring initiatives (crowdsensing). Community-based monitoring has proven effective in addressing information gaps in managing and monitoring aquatic ecosystems, particularly in small rivers that often lack agency monitoring. However, there are still challenges regarding the reliability of such data. To fill this gap, there is an urgent need to develop affordable, reliable, and open-source instrumentation for water quality monitoring. These instruments should also comply with the recent European guidelines on the use of toxic substances in technology development.

This study presents the development and validation of a RoHS directive-compliant, open-source, low-cost optical sensor for detecting nitrates and phosphates in community-based monitoring initiatives. The sensor setup takes advantage of light-emitting diodes (LED) as light sources and a commercial ambient light detector. A second light sensor positioned at a 90° angle is employed for scattering correction. All components are managed by a Raspberry Pi Zero W microcomputer and housed in a custom 3D-printed poly(lactic acid) case. The device enables data collection, including GPS coordinates, with results stored offline or transmitted in real-time through Wi-Fi. The sensor’s analytical performance was evaluated in both laboratory and field conditions using reference materials and river samples. Results demonstrated accurate and repeatable measurements which were shown to increase resolution and precision compared to standard colorimetric methods. To promote accessibility and replication, the 3D-box CAD model, software, and usage guidelines are freely available online.

How to cite: Cirrone, R., Boldrini, A., Polvani, A., Liu, X., and Loiselle, S.: Advancing Community-Based Water Quality Monitoring through Low-Cost Open-Source Optical Sensors and Data Integration, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19717, https://doi.org/10.5194/egusphere-egu25-19717, 2025.

EGU25-20136 | ECS | Posters on site | GI1.3

Towards the Integration of GPR and Magnetic Data for the Study of Urban and Rural Areas 

Francesco Mercogliano, Andrea Barone, Andrea Vitale, Giuseppe Esposito, Pietro Tizzani, and Ilaria Catapano

Among Non-Destructive Testing (NDT) methods, Ground Penetrating Radar (GPR) and magnetic surveys are among the most widely used techniques for various applications, including geo-environmental, archaeological, geotechnical, and engineering purposes. Their success is attributed to factors such as cost-efficiency, versatility, and data collection capabilities. Additionally, both methods enable the detection of buried targets through their respective magnetic and electromagnetic properties. Integrating the results from these two methodologies can yield excellent outcomes for an in-depth analysis of the investigated environment and significantly enhance the detection capabilities for anomaly sources.

This study presents preliminary results on the integration of simulated GPR and magnetometric data for a representative scenario. Advanced imaging techniques, including the Depth from Extreme Points (DEXP) method for magnetic data and the microwave tomography approach for GPR data, were applied to produce an initial high-resolution visualization of the simulated target.

Building on these results, an arithmetic integration approach was used to merge the two datasets into a single image, enhancing the interpretation of the anomaly source, including its morphology, position, and depth.

These preliminary results demonstrate the potential of this workflow, based on the arithmetic integration of these datasets, to provide more accurate and detailed subsurface models. This approach paves the way for real-world applications, and further developments aim to refine it for broader geophysical purposes.

Acknowledgments: the project ITINERIS "Italian Integrated Environmental Research Infrastructure Systems" (IR0000032), which funded the research

How to cite: Mercogliano, F., Barone, A., Vitale, A., Esposito, G., Tizzani, P., and Catapano, I.: Towards the Integration of GPR and Magnetic Data for the Study of Urban and Rural Areas, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20136, https://doi.org/10.5194/egusphere-egu25-20136, 2025.

Microwave links from cellular communication networks have been proposed as an opportunistic source of precipitation data more than two decades ago. The first scientific studies demonstrating the potential of this ground-based remote sensing technique, in particular for areas around the world lacking dedicated rainfall observation networks, were published more than 15 years ago. Since then, a small but dedicated community of scientists and engineers working at universities, national meteorological services, engineering firms, mobile network operators and telecommunication equipment manufacturers has been making significant progress in turning this promise into a reality. In the meantime, numerous papers and reports have been published, conference presentations have been given and courses have been delivered. However, real-time access to high-resolution rainfall information from commercial microwave link networks over large continental areas is still a dream. How far have we come after more than 20 years of research and development? What does the future have in stall for the hydrological and meteorological communities? What should be done to turn this dream into a reality? Finally, which other hydrometeorologically relevant variables could potentially be retrieved using received signal levels from commercial microwave links? This sollicited presentation will attempt to provide some preliminary answers to these questions.

How to cite: Uijlenhoet, R.: Hydrometeorological Monitoring using Microwave Links from Cellular Communication Networks: Opportunities and Challenges, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20151, https://doi.org/10.5194/egusphere-egu25-20151, 2025.

EGU25-20810 | Posters on site | GI1.3

Building a Smart Dendrometer: Calibration and Field Deployment of a linear magnetic driven IoT Sensor for Real-Time Radial Growth Assessment 

Luca Belelli Marchesini, Jim Yates, Francesco Renzi, and Riccardo Valentini

Technological advancements in forest digitization have revolutionized real-time monitoring of tree ecophysiological processes. Direct measurement sensors, such as dendrometers, sap flow sensors, and spectrometers, enable high-resolution insights into tree function and growth. Here, we present a novel dendrometer designed to monitor radial stem increment using a Hall effect-based linear magnetic encoder system integrated into an IoT-enabled platform.

The dendrometer employs a commercially available linear magnetic encoder chip (AMS OSRAM GmbH) that operates without physical contact, ensuring low power consumption and long-term monitoring suitability. Key design components include a linear arm, sensor housing, rail, magnetic tape, and chip braces. Calibration was conducted using a stepper motor for linear movements at 0.1 mm increments, capturing 100 data points per step in four replicates. Regression analysis demonstrated high accuracy, with an R² of 0.99 and an RMSE of 0.05 mm. Temperature sensitivity tests (0–40°C) revealed minimal impact on sensor performance.

Field tests over one growing season involved four dendrometers installed on specimens of spruce (Picea abies (L.) H.Karst)) and silver fir (Abies alba Mill.). Seasonal radial growth patterns captured by the devices aligned closely with established static UMS D1 diameter belt measurements, demonstrating their capacity to detect both long-term trends and short-term diel stem oscillations.

This study highlights the potential of an IoT-driven dendrometer for capturing high-resolution radial growth data, offering insights into tree physiology and forest responses to environmental changes. Future development should focus on enhancing measurement precision through design optimization and improved access to power width modulation components in the AS3511 chip. This dendrometer represents a promising tool for advancing forest monitoring and understanding the impacts of climate change on tree growth dynamics.

How to cite: Belelli Marchesini, L., Yates, J., Renzi, F., and Valentini, R.: Building a Smart Dendrometer: Calibration and Field Deployment of a linear magnetic driven IoT Sensor for Real-Time Radial Growth Assessment, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20810, https://doi.org/10.5194/egusphere-egu25-20810, 2025.

EGU25-20884 | Posters on site | GI1.3

A Cost-Effective Automatic Chamber for Permanent CH4 and N2O Assessments inWetland Environments 

Milan Shay Kretzschmar, Maren Dubbert, Mathias Hoffmann, Milos Bielcik, Joana Bergmann, and David Dubbert

Wetland ecosystems exhibit large spatial and temporal variability in terms of greenhouse gas (GHG) fluxes, necessitating new technologies to ensure that they are well-monitored. Both manual and automated chamber-based approaches are currently costly and thus limited either in spatial or temporal resolution. Following on from Wang et al. (2022), we propose a new, inexpensive autochamber (TraceCatch) for long-term outdoor installation. Costs for one unit are less than 800€ in total, making it affordable and scalable for long-term ecological research, also in lower income countries such as the global south. The system is based on gathering gas samples over two weeks into four gas bags on a high-frequency sampling schedule. TraceCatch is controlled using an Arduino Uno, connected to a peristaltic pump for sampling of chamber headspace air as well as a number of sensors for air temperature and humidity (SHT-41), air pressure (BMP280), and CO2 concentrations (K30FR; 0–5,000 ppm, 30 ppm resolution). The latter are used to track the sealing condition of the chamber. We validated the system using defined injection amounts of technical gas (100% CO2). In addition, the system was applied to measure GHG fluxes from three wetland cores placed inside three ecotrons (UGT EcoLab flex, manufactured by Umwelt-Geräte-Technik GmbH, Germany). Gas samples were collected 4 times a day for 2 weeks during a 1 hour chamber closure time at t0, t20, t40, t60 and subsequently analyzed using gas chromatography (Nexis GC-2030, manufactured by Shimadzu Corporation, Japan). Average GHG fluxes determined over the two-week period were then compared to single measurements obtained using multi-gas sensors (LI-COR LI-7820 and LI-7810 analyzers, manufactured by LI-COR Biosciences, USA). If adopted, the system’s low cost, scale and robustness for permanent field deployments could help improve wetland GHG monitoring, offering a cost-efficient and practical alternative to traditional methods for global-scale biogeochemical cycle assessments.

How to cite: Kretzschmar, M. S., Dubbert, M., Hoffmann, M., Bielcik, M., Bergmann, J., and Dubbert, D.: A Cost-Effective Automatic Chamber for Permanent CH4 and N2O Assessments inWetland Environments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20884, https://doi.org/10.5194/egusphere-egu25-20884, 2025.

EGU25-21862 | Posters on site | GI1.3

TreeTalker Cyber: A Multi-Sensor, Low-Cost IoT Platform for Real-Time Monitoring of Tree Ecophysiology 

Francesco Renzi, Jim Yates, Valerio Coppola, Salvatore Riggi, Maria Vincenza Chiriacò, and Riccardo Valentini

The Earth is a complex ecosystem and each element is strictly linked with the others. It is often required to collect data on multiple aspects variously related to the main phenomenon in order to understand its mechanism. Moreover, the increasing use of machine learning algorithms requires the creation of new, reliable and extensive dataset in order to obtain significant results. The increasing demand for accurate, real-time monitoring of tree ecophysiological parameters presents challenges in developing affordable and efficient technologies, in particular in difficult environments such as mountains. TreeTalker Cyber, an innovative IoT platform, addresses these needs by integrating multiple sensors into a single, cost-effective device capable of measuring radial growth, radiation intensity below the canopy across 26 spectral bands, sap-flow, microclimate data, and trunk inclination. This presentation explores its capabilities, practical applications, and potential to transform forest monitoring globally. The use of a single platform to collect all the aforementioned parameter greatly reduces the cost of the equipment per collected parameter providing at the same time all main information required to evaluate the status of a tree, improving the maintenance of the network at the same time. The device is equipped with an NB-IoT or LoRaWAN transmission module to transmit collected data and make them available remotely. A comprehensive description of the platform and real field data are presented along with the technologies used for data transmission and storage with their strength and weaknesses. The OGC SensorThings API is also briefly described along with FROST (FRaunhofer Opensource SensorThings-Server) as an alternative to efficiently store IoT data and make them compliant with the FAIR principles, making them usable by both scientific and public communities. The creation of a dataset of trees ecophysiological parameters will help deepening the knowledge and understanding of forests around the world. TreeTalker Cyber lays the groundwork for advancing forestry research, providing fine-scale data as ground truth for forestry models and a starting point for future scenarios predictions, in particular when based on machine learning algorithms.

How to cite: Renzi, F., Yates, J., Coppola, V., Riggi, S., Chiriacò, M. V., and Valentini, R.: TreeTalker Cyber: A Multi-Sensor, Low-Cost IoT Platform for Real-Time Monitoring of Tree Ecophysiology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-21862, https://doi.org/10.5194/egusphere-egu25-21862, 2025.

EGU25-810 | ECS | PICO | HS6.9

Mapping field-scale crop water stress for wheat using satellite remote sensing data by formulating lower baseline using a novel approach 

Manoj Yadav, Likhit Muni Narakala, Sriyodh Chinthamaneni, and Hitesh Upreti

The quantification of crop water stress is very crucial for efficient irrigation water management and sustainable agriculture. The empirically derived crop water stress index (CWSI) is a widely used method for quantifying the crop water status. However, developing lower baseline is a prerequisite for estimating the crop water stress using the empirical approach. Traditionally, the lower baseline is formulated by taking in-situ observations of a well-watered crop canopy using infrared radiometers. In this study, a novel methodology is formulated for estimating the lower baseline using land surface temperature (LST) and normalized difference vegetation index (NDVI) for the wheat crops using Landsat-8, Landsat-9 and Sentinel-2 satellite data. This study is conducted during the 2021-22 and 2022-23 wheat crop seasons, covering approximately 630 acres of agricultural fields, managed by local farmers in the western part of Uttar Pradesh, India. The entire analysis is conducted on Google Earth Engine.  Initially, multi-temporal image classification is performed, employing the synergetic use of Sentinel-2 and machine learning algorithms, to distinguish the wheat and non-wheat fields. The manually collected ground truth data are used to train and test the random forest model. Subsequently, the candidate pixels are selected based on the maximum NDVI range, from (NDVImax - 0.1) to NDVImax, which represents dense and healthy wheat patches. These candidate pixels are further refined by selecting the pixels having less than 10th percentile of the LST values, which account for relatively higher evapotranspiration. The lower baseline is derived using LST values of the refined candidate pixels along with concurrent air temperature (Ta) and relative humidity measurements recorded by an automatic weather station.  Finally, CWSI is mapped for the study area using the empirical approach.

Classification accuracy of 96% and 95% was achieved for the classification of wheat and non-wheat fields during the 2021-22 and 2022-23 seasons, respectively, with corresponding Kappa coefficients of 0.85 and 0.80. For the classified wheat pixels, the lower baseline equation formulated by the proposed methodology are (LST – Ta) = -1.864VPD + 1.325 for 2021-22 season and (LST – Ta) = -4.92VPD + 3.14 for 2022-23 season, where VPD is vapour pressure deficit. The fixed upper baseline of (LST – Ta) = 4°C is taken for empirically deriving and mapping CWSI for both seasons. The minimum and maximum values of the CWSI ranged from 0 to 0.89 during the 2021-22 season and from 0 to 0.78 during the 2022-23 season. The 2021-22 cropping season observed increased CWSI values as compared to 2022-23, primarily due to the heatwave that occurred in the study area from during the latter part of the 2021-22crop season. Significant spatial and temporal variability is obtained in the CWSI values within the study area.  The results suggest that the proposed methodology can be effectively used for mapping crop water stress at field scale without the requirement of tedious in-situ canopy temperature observations.

How to cite: Yadav, M., Narakala, L. M., Chinthamaneni, S., and Upreti, H.: Mapping field-scale crop water stress for wheat using satellite remote sensing data by formulating lower baseline using a novel approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-810, https://doi.org/10.5194/egusphere-egu25-810, 2025.

EGU25-8175 | ECS | PICO | HS6.9

Interferometric radar satellite and in-situ well time-series reveal groundwater extraction rate changes in urban and rural Afghanistan 

Najibullah Kakar, Sabrina Metzger, Tilo Schöne, Mahdi Motagh, Hamidullah Waizy, Nasir Ahmad Nasrat, Milan Lazecky, Falk Amelung, and Bodo Bookhagen

Population growth, climate change, and a lack of infrastructure have increased water demand and groundwater exploitation in urban and rural Afghanistan, resulting in significant ground subsidence in various regions. 

Using Sentinel-1 radar-interferometric time-series data based on over 7-years (2015-2022), we assess country-wide Afghan subsidence rates for groundwater levels, precipitation, and changes in irrigation practices. Urban Kabul city and the growing agricultural sector of rural Ghazni provinces are of particular focus. In Kabul city, we compare spatiotemporal subsidence patterns to water table heights and precipitation amounts. In Ghazni, we monitored the transition from ancient to modern irrigation techniques by mapping solar-panel arrays as a proxy for electrical water pumping and evaluating the vegetation index as a proxy for agricultural activity.

Several provinces in Afghanistan such as Kabul, Ghazni, Helmand, Farah, Baghlan, and Kunduz exhibit significant subsidence of more than ~5 ± 0.1 cm/yr. In Kabul, ground subsidence is most pronounced in the city center with a 6-yr total of 31.2 ± 0.5 cm, but it’s the peripheral wells of the Kabul basin that exhibit the highest water-table drops, where aquifers are also thinner and wells are deeper. In Ghazni, a 7-yr total of 77.8 ± 0.5 cm ground subsidence was recorded. Before 2018 barren lands were transformed into farmland throughout the province, and traditional irrigation such as Kariz networks were replaced by electrical water pumps to tap groundwater, which enabled the conversion of barren land into farmland and marked the acceleration of ground subsidence after 2018. In addition severe droughts in 2020 and 2021 further exacerbated groundwater depletion, leading to m-wide and km-long desiccation cracks that appeared in the area with the highest irrigation volume and ground subsidence.

How to cite: Kakar, N., Metzger, S., Schöne, T., Motagh, M., Waizy, H., Nasrat, N. A., Lazecky, M., Amelung, F., and Bookhagen, B.: Interferometric radar satellite and in-situ well time-series reveal groundwater extraction rate changes in urban and rural Afghanistan, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8175, https://doi.org/10.5194/egusphere-egu25-8175, 2025.

EGU25-8550 | ECS | PICO | HS6.9

Regional Agricultural water productivity Monitoring for Climate Change Adaptation 

Angura Louis, Fehér Zsolt, Tamás János, and Nagy Attila

Agricultural Water scarcity, amplified by climate change, poses a great challenge to global agricultural productivity and sustainability. This study explores a new indicator to monitor regional crop water productivity in agricultural systems. Using a combination of Moderate Resolution Imaging Spectroradiometer (MODIS) satellite data, and ground observations, we assess spatiotemporal trends in water productivity across an agricultural production regional scale .The water productivity indicator (CWPSM ) was computed as a ratio of normalized difference vegetation index (NDVI) to volumetric soil moisture content at 30cm and 60cm soil depths respectively and compared against a benchmark water productivity indictor ( CWPEC) computed as a ratio of Gross primary productivity (GPP)  to Evapotranspiration (ET). Our research findings highlight a consistent strong positive correlation and alignment of CWPSM at 30 cm, CWPSM at 60 cm and CWPEC trends over time with however CWPSM at 60 cm demonstrating superior accuracy and reliability compared to CWPSM at 30 cm as a proxy for CWPEC. The results highlight the importance of ensuring that water reaches deeper layers to at least 60 cm depth during irrigation due to the stability of soil moisture, observed at this depth.

By providing actionable insights, the study contributes to achieving sustainable development goals of climate action, ending hunger and underscoring the importance of monitoring crop water productivity in addressing water management challenges in agricultural production

This research was funded by Szechenyi Plan Plus Program under the RRF 2.3.1 21 2022 00008 project. We gratefully acknowledge their tremendous support and contributions to the research.

How to cite: Louis, A., Zsolt, F., János, T., and Attila, N.: Regional Agricultural water productivity Monitoring for Climate Change Adaptation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8550, https://doi.org/10.5194/egusphere-egu25-8550, 2025.

EGU25-11344 | ECS | PICO | HS6.9

Improved floodplain modelling with FABDEM and Sentinel-2 earth observations in the middle valley of the Senegal River 

Issa Leye, Andrew Ogilvie, Soussou Sambou, and Didier Martin

In the alluvial plains of large rivers, the study of flood dynamics is essential to appreciate water resource variations and preserve associated ecosystem services, in particular biodiversity, groundwater recharge and flood-recession agriculture. Hydraulic modelling provides valuable opportunities to simulate the dynamics of surface water flows but are challenged by the very flat topography and the sparse field observations, especially in Africa. By combining advances in earth observations (Digital Elevation Models and Sentinel-2 surface water areas), field observations (stage, flow gauging, river profiles) and hydraulic modelling (HEC-RAS), we aim to improve the understanding of surface water dynamics in the Senegal River floodplain. In this region, flood-recession agriculture is a complementary activity to irrigated agriculture and plays an important role in the subsistence of local populations.

Recent, open-access DEMs (AW3D, COPDEM, FABDEM, NASADEM, SRTM, TanDEM) were compared against field observations revealing the superior performance of FABDEM (RMSE = 0.58). FABDEM was then pre-processed to recondition the elevations of the river bed based on field river profiles. The HEC-RAS model was calibrated to simulate the flow propagation from the Bakel to Diama over the period 2017-2020 and to accurately map flood-prone areas detected on Sentinel-2 imagery at the scale of individual depressions and the whole floodplain. Results show that the model reproduces flood dynamics with good accuracy, with KGE on water levels reaching 0.78 at Bakel and 0.65 at Matam gauging stations. The model also enabled the 2D representation of flooded areas, providing the first accurate representation of inundated areas in this floodplain, and their variations under climatic and dam construction scenarios. The excellent performance obtained with FABDEM highlights the enhanced opportunities it extends to develop hydraulic models of complex, poorly gauged floodplains, and support the management of water resources.

Key words: Floodplain, HEC-RAS, remote sensing, hydraulic modelling, Senegal River, Middle valley.

How to cite: Leye, I., Ogilvie, A., Sambou, S., and Martin, D.: Improved floodplain modelling with FABDEM and Sentinel-2 earth observations in the middle valley of the Senegal River, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11344, https://doi.org/10.5194/egusphere-egu25-11344, 2025.

EGU25-11981 | ECS | PICO | HS6.9

Data-Driven Modelling of Lake Water Quality Response to Catchment Dynamics 

Zhenyu Tan, Stefan Simis, and Mark Warren

The water quality of lakes and reservoirs is influenced by atmospheric and land-use pressures, requiring actionable insights for effective management. Given the unique nature of each water body, data-driven modelling provides a practical solution for identifying sensitivities to these pressures, circumventing the complexity of hydrological-biogeochemical models. Remote sensing technologies offer consistent, multi-temporal, and multi-scale water quality monitoring, while global weather forecasting models enable predictions of key environmental parameters. Integrating these datasets facilitates a systematic examination of catchment-to-lake dynamics.

This study introduces a unified approach to modelling relationships between satellite-derived water quality metrics, such as Chlorophyll-a (Chl-a) concentration and turbidity, and meteorological drivers influencing catchments. Using multivariate autoregressive models, we aim to identify the influence of environmental factors on water quality variations, and  determine which sub-basins exert the greatest influence on lake dynamics. This approach supports short-term predictions of water quality changes. Ultimately, we anticipate that the data-driven models can be used to predict short-term water quality changes

The study focuses on small and medium-sized lakes and reservoirs in the United Kingdom, using Sentinel-2 MSI observations for high-resolution water quality datasets. ERA5-Land hourly reanalysis data provided meteorological variables influencing water quality, including wind, lake mixed-layer temperature, solar radiation, precipitation, and runoff. Both datasets were aggregated into five-day time series to address observation intervals caused by orbital patterns and cloud cover. Aggregated data were normalized and stabilized to account for variable magnitudes before being input into autoregressive models.

Vector Autoregression (VAR) was used to assess long-term environmental influences on water quality, leveraging Impulse Response Function (IRF) and Forecast Error Variance Decomposition (FEVD). The reliance of VAR models on historical data enabled analysis of prolonged effects, with optimal four-time lags. In contrast, Autoregressive Integrated Moving Average with Explanatory Variables (ARIMAX) incorporated contemporary meteorological inputs, allowing for short-term impact analysis. ARIMAX models also enabled near-term water quality predictions using forecasted meteorological variables. At the sub-basin level, models were evaluated using the Fréchet distance, which quantifies the similarity between time-series curves. By comparing Fréchet distances across sub-basins, the relative contributions of each sub-basin to lake water quality variations were determined.

Our findings suggest that: 1) VAR models explained the temporal variability in lake water quality variables with a strong fitness (R2 > 0.82 for Chl-a and R2 > 0.69 for turbidity); 2) VAR models relied heavily on the lake water quality inputs from priors with optimal four time lags. The first lag contributed the most, with a mean weight of 0.61 (σ = 0.45) for Chl-a concentration and 0.71 (σ = 0.46) for turbidity; 3) Catchment drivers exhibited weights up to 2.3% at the second time horizon, with their influence increasing over time, while the contribution from water quality observations decreased; 4) ARIMAX models demonstrated high accuracy in simulating lake water quality variations (R2 > 0.83 for Chl-a and R2 > 0.68 for turbidity), showing promise for future water quality predictions.

How to cite: Tan, Z., Simis, S., and Warren, M.: Data-Driven Modelling of Lake Water Quality Response to Catchment Dynamics, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11981, https://doi.org/10.5194/egusphere-egu25-11981, 2025.

In a global warming and climate change context, populations all over the world are impacted by an increasing number of hydrological crisis (flood events, droughts, ...), mainly related to the lack of knowledge and monitoring of the surrounding water bodies. In Europe, flood risk accounts for 46% of the extreme hazards recorded over the last 5 years and current events confirm these figures. Although the main rivers are properly monitored, a wide set of small rivers contributing to flood events are not monitored at all. There is a clear lack of river basins monitoring regarding the rapid increase of extreme events. Moreover, hydrological surveys are currently insured by heterogeneous means from a country to another and even inside a country, from a region to another. It results in a high-cost level to deploy a robust, relevant, and efficient monitoring of all watercourses at risks. Therefore, there is a real need for affordable, flexible, and innovative solutions for measuring and monitoring hydrological areas to address climate change and flood risk within the water big cycle. 

 VorteX.io aims to provide an innovative and intelligent service for monitoring hydrological surfaces, using easy-to-install and fixed in-situ instruments, based on compact light-weight device inspired from satellite technology: the micro-stations. From a technical point of view, vorteX-io micro stations are designed like remote sensing nanosatellites that do not fly, but are installed above watercourses (i.e., under a bridge). Onboard remote sensing instruments (lidar, thermal and multispectral camera, GNSS) allow them to remotely and in real-time measure water temperature, provide contextual images and hydro-meteorological parameters (water surface height, water surface velocity, rain rates). Water parameters are transferred in real-time through GSM or SpaceIOT networks.  The technology has been entirely designed and patented by vorteX-io.

The combination of these in-situ data with satellite measurements is thus optimal for downstream services related to water resources management and assessment of flood/drought risks: calibration, validation and accuracy assessment of EO projects in space hydrology. The vorteX-io technology is selected by ESA for Sentinel-3 Altimetry CalVal for inland waters: installation of stations under the track and synchronization of in-situ acquisition with the passage of the satellite to operationally provide Fiducial Reference Measurements (FRMs). In addition, vorteX-io is involved in the definition of future inland water FRMs for the upcoming CRISTAL mission and also on the ESA DTE for hydrology. 

In June 2023, the European Innovation Council awarded the company to deploy 1000 micro-stations in France and Croatia

In June 2024, vorteX-io has completed a funding round, which notably includes Caisse des Dépôts (CdC represent a major public financial institution in France) and CNES as shareholders. This funding will specifically facilitate the continued deployment of the constellation across Europe.

The long-term vision is to cover river basins in Europe with an in-situ network, to be used at large scale as earth-observation in situ component either for monitoring water quality parameters or for extreme hazards monitoring such as floods and drought

How to cite: Gachelin, J. P. and Poisson, J. C.: Development of a new remote sensing device to be used at large scale as earth-observation in situ component, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12071, https://doi.org/10.5194/egusphere-egu25-12071, 2025.

EGU25-12476 | ECS | PICO | HS6.9

Small agricultural reservoirs detection with satellite data and OpenStreetMap integration for sustainable water management: a contribution to the CASTLE project. 

Noemi Mannucci, Gabriele Bertoli, Marco Lompi, Tommaso Pacetti, Mehdi Sheikh Goodarzi, Patrick Ebel, Davide Danilo Chiarelli, Margherita Azzari, and Enrica Caporali

Meteorological unpredictability, exacerbated by severe events caused by climate change, poses significant problems for water resource management (IPCC, 2023). Climate change has increased the frequency and severity of droughts, especially in mid-latitude regions, where reduced precipitation coupled with rising temperatures is expected to exacerbate water scarcity (https://doi.org/10.1007/s40641-018-0093-2). In this regard, Small Agricultural Reservoirs (SmARs) offer a strategic response, as they are designed to collect and store water for use in irrigation and other agricultural applications. This is the context in which the research activity described here is developed, contributing to the research project CASTLE - Creating Agricultural reSilience Through smaLl rEservoirs.

Despite their importance, the lack of comprehensive national databases for SmARs remains a major obstacle to their efficient management. Prior to this study, for example only eight of Italy's twenty regions had SmARs inventories, often based on non-standardised and incomparable approaches (https://indicatoriambientali.isprambiente.it/it/pericolosita-sismica/invasi-artificiali). This fragmentation of information makes the analysis and management of SmARs challenging. A possible option to overcome this problem is represented by satellite data, which provides accurate and continuous information over large geographical areas. Sentinel-2 satellite imagery - part of the European Space Agency's Copernicus programme - was particularly well suited to this study.

The objective of this research was to develop a methodology for detecting Small Agricultural Reservoirs from satellite imagery with integration of OpenStreetMap (OSM) and the ESA World Cover 2021 dataset and creating a comprehensive inventory of the existing reservoirs in Italy. The system was validated in Tuscany with the use of the ground truth database of LaMMA - CNR IBIMET (https://geoportale.lamma.rete.toscana.it/difesa_suolo/#/viewer/372).

Integration with OSM helped eliminate false positives such as ponds, glaciers, large dams, rivers, and canals, which spectral indices alone cannot distinguish from SmARs due to their similar reflectance characteristics, as they are also water surfaces. The ESA World Cover data were used to exclude urbanized areas, which were irrelevant to this study. 

The combined use of open-source data has enabled the development of a replicable methodology adaptable to various spatial scales, considerably enhancing the identification and mapping of SmARs. This strategy will help to manage agricultural water resources more efficiently and increase resilience to climate change.

ACKNOWLEDGMENTS

This study was carried out within the CASTLE project and received funding from the European Union Next-GenerationEU (National Recovery and Resilience Plan – NRRP, Mission 4, Component 2, Investment 1.1 – D.D. n. 104 02/02/2022 PRIN 2022 project code MUR 2022XSERL4 - CUP  B53D23007590006).

.

How to cite: Mannucci, N., Bertoli, G., Lompi, M., Pacetti, T., Goodarzi, M. S., Ebel, P., Chiarelli, D. D., Azzari, M., and Caporali, E.: Small agricultural reservoirs detection with satellite data and OpenStreetMap integration for sustainable water management: a contribution to the CASTLE project., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12476, https://doi.org/10.5194/egusphere-egu25-12476, 2025.

EGU25-14387 | PICO | HS6.9

Regression-based surface water fraction mapping using a synthetic spectral library for monitoring small water bodies 

Yalan Wang, Giles Foody, Xiaodong Li, Yihang Zhang, Pu Zhou, and Yun Du

Small water bodies (SWBs), such as ponds and on-farm reservoirs, play a crucial role in agriculture irrigation, carbon storage, and biogeochemical cycle. Medium-spatial-resolution satellite imagery such as Sentinel-2 imagery has been widely promoted to monitor SWBs, due to its relatively fine spatial and temporal resolution. However, the small size and diverse spectral characteristics of SWBs present significant challenges, particularly the mixed-pixel problem, where both water and land classes contribute to the observed spectral response of the image pixel. To address this issue, we propose a novel regression-based surface water fraction mapping method (RSWFM) that leverages a random forest regression model and a synthetic spectral library to generate 10 m spatial resolution surface water fraction maps from Sentinel-2 imagery. RSWFM incorporates a compact set of endmembers, representing water, vegetation, impervious surfaces, and soil, to simulate a spectral library using both linear and nonlinear mixture models, while accounting for spectral variability across diverse SWBs. Additionally, to enlarge the number of pure spectra and enhance their representativeness for training, RSWFM applies data augmentation based on Gaussian noise. The performance of RSWFM was assessed across ten study sites with hundreds to thousands of SWBs smaller than 1 ha and was compared with fully constrained least squares (FCLS) linear spectral mixture analysis, multiple endmember spectral mixture analysis (MESMA), and random forest (RF) regression without data augmentation. Results indicated that RSWFM generates a low root mean square error (RMSE) of less than 0.09, reducing by approximately 30%, 15%, and 11% compared to FCLS, MESMA, and nonlinear RF regression without data augmentation, respectively. Furthermore, RSWFM achieves an R² of approximately 0.85 in estimating the area of SWBs smaller than 1 ha. This study demonstrates the potential of RSWFM for addressing the mixed pixel problem in SWB monitoring across large areas.

How to cite: Wang, Y., Foody, G., Li, X., Zhang, Y., Zhou, P., and Du, Y.: Regression-based surface water fraction mapping using a synthetic spectral library for monitoring small water bodies, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14387, https://doi.org/10.5194/egusphere-egu25-14387, 2025.

EGU25-15594 | ECS | PICO | HS6.9

Delineating small water bodies in Pune City India using Machine Leaning in Google Earth Engine 

Shobhit Choubey, Saidutta Mohanty, and Chandranath Chatterjee

Freshwater is a valuable and scarce resource under constant threat due to global climate uncertainties, population growth, and economic expansion. Mapping water bodies can be useful in effective water resources management. The present study was an effort towards mapping inland water body and identifying areas suitable for water conservation in the Pune city of the Upper Bhima River Basin. In this study, land cover classification was performed using machine learning in Google Earth Engine (GEE) to identify water and non-water pixels to delineate small water bodies. Three machine learning models, namely Support Vector Machine (SVM), Random Forest (RF) and Gradient Tree Boost (GTB), were compared for their efficacy in mapping the water bodies. An open-source high-resolution multi-spectral image (MSI) information from Copernicus Sentinel-2 Level 2A harmonised data was used to generate a water body map. The classification models were further compared with the Modified Normalized Difference Water Index (MNDWI) thresholding method, which distinguishes water regions based on the reflectance difference between the Short Wave Infra-Red (SWIR) band and the Green band. As the study area covered a diversified spectral signature of land use and land cover, the analysis was performed under three scenarios. In scenario 1, the ML model was trained and validated using hilly and built-up region data, in scenario 2 agricultural and built-up areas were considered and in scenario 3 all three regions were covered. Results showed that the SVM model performed more accurately and detected the maximum area of water bodies followed by RF, GTB and MNDWI threshold methods. Moreover, scenario 3 which considers the entire dataset ranging from hilly, built-up and agricultural regions is the most robust analysis to perform water body mapping. Finally, the SVM model considering scenario 3, was used to detect the small water bodies for the entire catchment. In total, 20,479 water bodies were identified by the SVM model covering 279.42 sq.km area. Furthermore, river networks were removed from the classification, which resulted in a total of 17,616 small water bodies with an area of 243.97 sq.km. As this analysis was performed using Sentinel-2A data which has spatial resolution of 10 meters, ML models and MNDWI method cannot estimate water bodies smaller than 100 sq. meters. The water body map can be useful for water resources planning in the study area.

Keywords: Google Earth Engine, Random Forest, Support Vector Machine, Gradient Tree Boost and Modified Normalized Difference Water Index.

How to cite: Choubey, S., Mohanty, S., and Chatterjee, C.: Delineating small water bodies in Pune City India using Machine Leaning in Google Earth Engine, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15594, https://doi.org/10.5194/egusphere-egu25-15594, 2025.

EGU25-15643 | ECS | PICO | HS6.9

Remote sensing of ferric iron in inland surface waters 

Marit van Oostende and Ype van der Velde

Iron concentrations in inland waters play an important role in nutrient cycling and water quality, particularly through their interaction with phosphorus, a key driver of eutrophication. Both excessive and insufficient ferric iron (Fe³⁺) levels can disrupt aquatic ecosystems. Insufficient Fe³⁺ availability hampers primary productivity, nutrient cycling, and ecosystem structure. Conversely, elevated levels of Fe³⁺ in water can pose risks to human and ecosystem health.

In Dutch agricultural-dominated lowland catchments, ferric iron-bound phosphorus is the main form of phosphorus in suspended particulate matter, potentially driving rapid transformation of dissolved phosphorus in groundwater to phosphorus in surface water. Groundwater seepage, rich in ferrous iron (Fe²⁺), further contributes to these dynamics, with Fe²⁺ oxidizing to Fe³⁺ upon exposure to oxygen, forming hydroxides that bind phosphorus. Seasonal hydrological changes also influence these interactions, with distinct red colouring observed in Dutch waters during winter attributed to iron oxidation under reduced biological activity.

A novel method using Sentinel-2 MSI data and machine learning has been developed to estimate and monitor the optically active Fe³⁺ concentration levels across Dutch surface waters in autumn and winter with a high spatial resolution (10 m). The model incorporates predictors including spectral band ratios, spectral band slopes, spectral band derivatives, and environmental variables such as air temperature and cumulative rainfall, derived from in situ data. It was trained on ~2,000 in situ iron measurements collected between 2015 and 2023.

Until now, research on Fe³⁺ in water using satellite data has been limited. This study provides a detailed spatiotemporal perspective on iron dynamics in the Netherlands, advancing the monitoring and management possibilities of water quality and ecosystem health.

How to cite: van Oostende, M. and van der Velde, Y.: Remote sensing of ferric iron in inland surface waters, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15643, https://doi.org/10.5194/egusphere-egu25-15643, 2025.

EGU25-18138 | ECS | PICO | HS6.9

Agricultural Drought Assessment Using Remote Sensing Technologies: A Case Study in Souss Massa Region 

Sara Merzoug, Zine El Abidine El morjani, Youssef Es-saady, and Mohamed El hajji

Agricultural Drought presents a significant risk to food security, particularly in arid and semi-arid regions where crop production is highly dependent on irrigation and annual rainfall. Therefore, this research was conducted in Souss Massa region, a semi-arid region relying on agricultural production for its economy to develop drought early warning studies in this region. This study aims to assess the agricultural drought and its associated impacts, as accurate identification is crucial for effectively minimizing its negative impacts. In this work, we evaluate various Remote Sensing-based indices to create a composite drought index with distinct severity classes (No Drought, Moderate Drought, Severe Drought, Extreme Drought). This approach enables the identification and mapping of drought-affected areas. These findings provide valuable insights into the potential of remote sensing for drought monitoring and contribute to development of effective drought management strategies in Souss Massa region.

How to cite: Merzoug, S., El morjani, Z. E. A., Es-saady, Y., and El hajji, M.: Agricultural Drought Assessment Using Remote Sensing Technologies: A Case Study in Souss Massa Region, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18138, https://doi.org/10.5194/egusphere-egu25-18138, 2025.

Lake thermal dynamics provide critical insights into regional and global climate change, and play a regulatory role in lake biogeochemical cycles. In situ measurements, remote sensing, and hydrodynamic modelling are key sources for monitoring lake temperature. In situ data are essential for calibration and validation (cal/val) of satellite products and numerical models, but are often scarce or irregular for many lakes. Additionally, data assimilation of lake surface water temperature (LSWT) products can improve numerical models. Satellite thermal imagery has been widely used for LSWT monitoring at regional and global scales. However, current operational LSWT services are limited to 1 km resolution, thereby excluding small lakes. High-resolution, high-revisit Earth observation missions, such as ECOSTRESS, LSTM, TRISHNA, and SBG, extend LSWT services to smaller lakes, but require dedicated cal/val efforts due to their unique radiometric and geometric properties. Collecting reliable skin temperature and ancillary datasets across diverse lakes and optimizing LSWT retrieval algorithms is thus urgently needed.

Our research, within the ESA-funded TRISHNA – Science and Electronics Contribution (T-SEC) project, focuses on validating and improving high-resolution LSWT products, and openly publishing final products for lakes in the Alpine region. We operate automated reference stations in four Swiss lakes: Lake Geneva, Lake Aegeri, Lago Bianco, and Greifensee. These lakes comprise a variety of morphological, bio-physical, and meteorological features, and are located along an elevation gradient in pre-, sub-, and high-alpine environments. Skin, sky, and bulk temperatures, as well as meteorological data are available for all sites. We evaluated Landsat 7/8/9 LSWT products from USGS Collection-2 Level-2 data and the single-band Acolite-TACT algorithm. Our matchup comparisons yielded a Mean Absolute Error (MAE) of < 1.2 °C, a Mean Bias Error (MBE) < 0.1 °C and a correlation coefficient (R2) of > 0.94. However, official Level-2 ECOSTRESS data showed weaker performance (MAE > 2.4 °C, MBE < -2 °C, and R2 < 0.85), highlighting the need for further cal/val and algorithm refinements, particularly for emissivity corrections.

Landsat validated algorithms are used for operational monitoring via AlpLakes web platform (www.alplakes.eawag.ch), which integrates satellite data, in situ measurements, and hydrodynamic models. AlpLakes’ scalable design enables rapid integration of new lakes and products. For example, we aim to disseminate our tools across lakes in the Alpine region under the EU Interreg AlpineSpace project DiMark (https://www.alpine-space.eu/project/dimark/). This pipeline will also facilitate the adoption of upcoming missions and timely dissemination of validated products. Ultimately, our research and datasets will support lake monitoring and modelling activities in Switzerland and beyond. Moreover, integrating satellite data, hydrodynamic models, and in situ measurements (e.g., assimilating LSWT products into existing models) will enhance understanding of short-term events and long-term trends in lakes, fostering interdisciplinary research and providing deeper insights into underlying bio-physical processes.

How to cite: Irani Rahaghi, A., Odermatt, D., and Naegeli, K.: Advancing Alpine lake monitoring and modelling through calibration, validation, and dissemination of high-resolution thermal remote sensing products, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18287, https://doi.org/10.5194/egusphere-egu25-18287, 2025.

EGU25-19245 | ECS | PICO | HS6.9 | Highlight

Integrating satellite observations to enhance reservoir monitoring: a case study facing the emergency shortage of fresh water in Bogotá, Colombia 

Camilo Sanabria-Morera, David Zamora, and Sebastian Palomino-Ángel

Colombia, with a water surface of approximately 831,163.7 hectares distributed in swamps, reservoirs, lagoons and marshes, faces significant challenges in monitoring the status of its lentic water bodies. The lack of information places the country in a disadvantageous position for managing its water-related ecosystems. For instance, quantifying water resources and reporting on international initiatives such as SDGs requires implementing robust monitoring systems to track progress on several objectives and indicators (e.g., SDG indicator 6.6.1, "Change in the extent of water-related ecosystems over time"). Moreover, monitoring is crucial for decision-making under extreme hydrometeorological phenomena and climate change.

Bogotá, Colombia’s capital city, is the sixth most populated capital in Latin America, where domestic water demand, flow regulation, and energy generation are supplied by a set of reservoirs located inside and outside the basin where the capital is located. The in-situ monitoring of these bodies of water faces technical, logistical, and economic difficulties, such as high installation costs, low availability of measuring stations, vandalism, and restricted access to data captured by some organizations. These difficulties hinder efficient management and informed decision-making.

Since mid-2024, Bogotá has been experiencing one of its most challenging water shortage emergencies in recent history due to the "El Niño" phenomenon that has brought reservoir levels to critical conditions. This situation has generated the need to explore new sources of hydrological data that complement on-site observations and enable the inclusion of other actors in water management decisions for this region. Satellite data emerges as a viable solution to complement gauge observations. However, consistent assessments of the accuracy of these measurements at the local level are required for this purpose. This study evaluates observations from different types of satellite data to obtain hydrological measurements in high mountain reservoirs and lakes of Bogotá's water supply system. The data assessed includes observations from the Sentinel-3 radar altimeter to get water level data, and amplitude observations from Sentinel-1 to estimate the surface area of the reservoirs.

The evaluation was initially carried out at the Neusa reservoir with daily water level measurements. Data from January 2019 to December 2020 were used for validation. Preliminary results show that Sentinel-3 observations provide water level measurements with good accuracy for the evaluated reservoir, achieving an R2 of 0.99 and RMSE of 0.084 meters (n = 24). Observations were obtained with almost monthly periodicity.

The results provide a first assessment of Sentinel-3's potential for monitoring the reservoirs of the city's water supply system, opening new opportunities for integrating different actors in water monitoring. Future research will focus on using Sentinel-1 to obtain reservoir surface area data and integrate these with level observations to calculate changes in reservoir volume. Finally, the analysis is expected to be expanded to other reservoirs and lakes in Bogotá's water supply system.

How to cite: Sanabria-Morera, C., Zamora, D., and Palomino-Ángel, S.: Integrating satellite observations to enhance reservoir monitoring: a case study facing the emergency shortage of fresh water in Bogotá, Colombia, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19245, https://doi.org/10.5194/egusphere-egu25-19245, 2025.

EGU25-20477 | ECS | PICO | HS6.9

Evaluating the Sensitivity of Hydrological Models to Remotely Sensed Precipitation in a Transboundary Basin 

paula lady pacheco mollinedo, Frédéric Satgé, Renaud Hostache, Marie-Paule Bonnet, Jorge Molina Carpio, Ramiro Pillco, Edson Ramirez, and Daniel Espinoza

Accurate precipitation data is vital for hydrological modelling, particularly in transboundary basins with scarce hydro-climatic stations. This study evaluates the performance of 20 gridded precipitation products (GPPs), derived from remotely sensed data and reanalyses, in the transboundary Lake Titicaca basin. The methodology integrates two approaches: first, a spatial and temporal accuracy assessment of the GPPs, and second, their application as input data in hydrological models.

For spatial accuracy, annual precipitation maps were generated for each GPP, preserving their native resolution, and compared with gauge-based maps. Temporal accuracy was assessed using Taylor diagrams. To evaluate the impact of GPPs on hydrological modelling, streamflow simulations were performed using the GR4J (lumped) and MGB-IPH (semi-distributed) models for three sub-basins, with model performance assessed through Kling-Gupta Efficiency (KGE).

Results indicate that CHIRPS, IMERG, and MSWEP excel in spatial and temporal accuracy, capturing the north-to-south precipitation gradient shaped by Andean topography. Streamflow simulations showed that GPPs often outperform gauge-based precipitation in basins with uneven station distribution. In GR4J, MSWEP and CHIRPS yielded the highest KGE values across all sub-basins, while in MGB-IPH, SM2Rain_CCI and IMERG-FR performed best. Notably, the higher KGE scores observed for the GR4J model can be attributed to its lumped structure, which compensates for GPP over/under estimations and spatial distribution inconsistencies.

This comprehensive evaluation demonstrates the potential of remotely sensed precipitation products to address data scarcity in transboundary basins. By improving streamflow simulations, these products support informed water resource management, climate adaptation, and transboundary collaboration.

How to cite: pacheco mollinedo, P. L., Satgé, F., Hostache, R., Bonnet, M.-P., Molina Carpio, J., Pillco, R., Ramirez, E., and Espinoza, D.: Evaluating the Sensitivity of Hydrological Models to Remotely Sensed Precipitation in a Transboundary Basin, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20477, https://doi.org/10.5194/egusphere-egu25-20477, 2025.

EGU25-3176 | Posters on site | HS6.5

Multi-Sensor SAR-Based Flood Mapping for High-Temporal Monitoring of the 2020 Flood Event in Thừa Thiên Huế, Vietnam 

Felix Bachofer, Patrick Sogno, Elly Schmid, Kerstin Büche, André Assmann, and Hoang Khanh Linh Nguyen

The 2020 flood season in Thừa Thiên Huế province, Central Vietnam, was among the most severe in recent history, driven by consecutive tropical storms and prolonged heavy rainfall. Between October and November 2020, a series of storms, including Tropical Storm Linfa, Typhoon Molave, and Typhoon Goni, brought intense precipitation, causing widespread inundation and significant damage to infrastructure and livelihoods. The hydrological complexity of the region, characterized by mountainous terrain, low-lying floodplains, and the extensive Tam Giang-Cau Hai lagoon system, further exacerbated the flood impacts, underscoring the need for advanced monitoring tools to capture the event's dynamics.

This study leverages multi-sensor Synthetic Aperture Radar (SAR) data, including Sentinel-1, Cosmo-Skymed, and TerraSAR-X, to create a high-temporal flood inventory for this hydrologically challenging region. Multi-temporal SAR intensity and coherence data were processed using threshold-based change detection algorithms and normalized difference indices to delineate flood extents. These SAR-based methods, immune to cloud cover, provided continuous observations despite the adverse weather conditions during the flood. Validation was performed using in-situ flood markers and drone imagery, ensuring accuracy in the derived flood maps. To complement SAR data, hydrodynamic modeling using HEC-RAS simulated water flow, inundation depths, and river system behavior, enabling cross-comparison with SAR-derived flood extents.

The 2020 flood event highlighted a challenge often associated with satellite-based flood mapping: image acquisitions seldom capture the peak of the flood. However, the high temporal resolution provided by the combined SAR datasets allowed researchers to track the pulse of the flood, revealing its evolution and alignment with storm events and precipitation patterns. This capability provided critical insights into the timing, extent, and dynamics of flooding, even in a region with complex topography and hydrology.

The high-temporal flood inventory produced in this study enhances understanding of flood dynamics across diverse land-cover types, enabling improved flood risk assessments and adaptive management. The outcomes not only advance flood monitoring methodologies for Vietnam but also demonstrate the value of integrating Earth Observation data with hydrological modeling to support disaster risk reduction efforts. This approach offers scalable solutions for other regions prone to extreme weather events, contributing to global efforts in informed decision-making and adaptive flood management strategies.

How to cite: Bachofer, F., Sogno, P., Schmid, E., Büche, K., Assmann, A., and Nguyen, H. K. L.: Multi-Sensor SAR-Based Flood Mapping for High-Temporal Monitoring of the 2020 Flood Event in Thừa Thiên Huế, Vietnam, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3176, https://doi.org/10.5194/egusphere-egu25-3176, 2025.

EGU25-6094 | ECS | Orals | HS6.5

Enhanced Flood Hazard Assessment and Mapping Using SAR Data: A Case Study of Afghanistan’s Flood Events (2018–2024) 

M. Sulaiman Fayez Hotaki, Mahdi Motagh, and Mahmud Haghshenas Haghighi

Afghanistan faces severe flood risks, but challenges such as limited flood data, cloud cover, and difficulties in on-ground data collection hinder traditional flood mapping methods. This study introduces an automated flood mapping approach using Synthetic Aperture Radar (SAR) data to overcome these limitations. Combining SAR intensity and interferometric coherence analyses, the method improves flood detection accuracy, particularly in complex terrains and rapid-onset events. The study spans the period from 2018 to 2024, covering 17 flood events across the country.

Processed on the Google Earth Engine (GEE), the method enables near-real-time monitoring by analyzing dense Sentinel-1 SAR time series data. SAR intensity identifies floodwaters, while coherence detects subtle changes in vegetated and urban areas, where intensity alone may fall short. Interferometric coherence was derived using the Hybrid Pluggable Processing Pipeline (HyP3), a cloud-based SAR processing platform accessed via the Alaska Satellite Facility (ASF) Data portal.

Validated against high-resolution PlanetScope imagery, the approach achieved F1 scores exceeding 82% in key provinces like Faryab and Baghlan. Land cover analysis revealed irrigated agriculture as the most affected type (709 hectares), while coherence mapping highlighted vulnerable urban areas, such as Baghlan-e-Markazi and Charkiar cities.

Compared to the Global Flood Monitoring (GFM) system, this method significantly improves detection accuracy, capturing up to 83% more flood extent in certain areas. For example, in Baghlan Province, it detected 709 hectares of flooding versus GFM’s 114 hectares.

By leveraging SAR data, HyP3, and GEE’s processing capabilities, this method provides a scalable, rapid-onset, and efficient solution for flood monitoring in data-scarce regions. Covering seven years of flood events, it offers a valuable tool for disaster management in Afghanistan and other regions vulnerable to climate change-induced flooding.

How to cite: Hotaki, M. S. F., Motagh, M., and Haghshenas Haghighi, M.: Enhanced Flood Hazard Assessment and Mapping Using SAR Data: A Case Study of Afghanistan’s Flood Events (2018–2024), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6094, https://doi.org/10.5194/egusphere-egu25-6094, 2025.

EGU25-6671 | Posters on site | HS6.5

Earth Observation data for Advancing Flood Forecasting: EO4FLOOD project 

Angelica Tarpanelli, Guy Schumann, and Cecile Kittel and the EO4FLOOD team

Floods are among the most destructive natural disasters, causing severe damage to human health, the environment, cultural heritage, and economies. Over the past 50 years, Europe alone has experienced approximately 4,000 fatalities and $274 billion in economic losses due to floods. The situation is even more severe in developing regions, where the lack of infrastructure and resources intensifies the impacts of such disasters. As climate change exacerbates the frequency and intensity of flood events, there is an urgent need for innovative approaches to improve flood forecasting and reduce societal impacts.

EO4FLOOD is a project funded by ESA demonstrating the potential of advanced satellite data in enhancing the accuracy and timeliness of flood forecasting systems. The project focuses on integrating state-of-the-art satellite technologies and hydrological and hydraulic models to deliver reliable flood predictions up to seven days in advance.

EO4FLOOD is structured around three main objectives:

  • Development of an Advanced EO Dataset: The EO4FLOOD dataset integrates high-resolution satellite products from ESA and non-ESA missions, providing global coverage of critical variables such as precipitation, soil moisture, snow, flood extent, water level and river discharge.
  • Integration into Flood Forecasting Models: By combining these datasets with machine learning-enhanced hydrological and hydraulic models, the project achieves more accurate flood predictions while quantifying uncertainty.
  • Demonstration for Science and Society: EO4FLOOD showcases the application of these tools in flood risk management and explores the influence of human activities, such as land-use changes and dam construction, on flood dynamics.

By leveraging cutting-edge algorithms and satellite products, EO4FLOOD provides a robust framework for advancing flood forecasting and supporting effective disaster preparedness and response, highlight its broader implications for global flood risk management.

How to cite: Tarpanelli, A., Schumann, G., and Kittel, C. and the EO4FLOOD team: Earth Observation data for Advancing Flood Forecasting: EO4FLOOD project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6671, https://doi.org/10.5194/egusphere-egu25-6671, 2025.

EGU25-7115 | ECS | Posters on site | HS6.5

A catchment-scale screening tool for the assessment of bridge overtopping using GIS and LiDAR-derived digital elevation models 

Michele Amaddii, Fabio Castelli, and Chiara Arrighi

Bridges are critical infrastructures of the transport network given their high construction costs and limited alternative routes. Flood events are the most frequent cause of damage to transport infrastructure compared to any other natural hazard. Bridge overtopping is a phenomenon with serious safety consequences for drivers and leads to cascading effects such as traffic disruption and reduced efficiency of evacuation and emergency plans. Whereby, proactive management is essential to enhance bridge resilience and ensure user safety.
This work introduces a catchment-scale screening method using GIS and remotely sensed data to assess the propensity of riverine bridges to overtopping. The application of the method is based on the use of elements such as road network (OSM), hydrographic network, and LiDAR-derived Digital Elevation Models of the bare terrain (DTM) and of the surface (DSM). The propensity of bridges to overtopping is evaluated considering the geometric and morphological characteristics of river-roads intersections, independent of hydrological forcing. The method assumes that bridges with intersection heights (Hi), i.e. the difference between the road level (DSM) and river thalweg (DTM), lower than the corresponding cross-section heights (Hs), are more prone to overtopping during floods.
Intersections between roads and the hydrographic network were identified, and Hi values were calculated by extracting elevation differences within a defined buffer. To minimize noise from vegetation and other elements in the DSM, the topographic ruggedness index was employed as a filter, assuming that roads have smooth surfaces compared to the high roughness of vegetation. Field measurements of Hi were performed to validate the remotely sensed Hi values. Riverbanks and their corresponding Hs values were identified using the Iso Cluster Unsupervised Classification approach, testing various morphometric derivatives of the DTM. A combination of profile curvature and maximum difference from mean elevation provided the clusters of landforms corresponding to riverbanks.
The method was applied to the Magra River basin in Italy (970 km²), an area frequently impacted by flood events.
Results indicate that for roads intersecting streams with Strahler order (S) <4 the median height error (∆he) between remotely sensed and measured Hi is significant (2 m, i.e. 40%). In contrast, the method proved effective for S>3 (∆he= 0.4 m, i.e. 12%). The mean cross-section width for such streams is 35 m (excluding the main river), which is two orders of magnitude larger than the planimetric accuracy of the DTM (0.3 m). A total of 231 bridges were identified, and approximately 30% exhibited Hi<Hs, indicating a high propensity for overtopping. This approach enables large-scale screening to identify road-river intersections with geometric and morphological predispositions to overtopping. It provides a valuable tool for prioritizing bridges for further hydrologic-hydraulic and traffic disruption modeling, supporting infrastructure resilience, and flood risk management.

Acknowledgments
This study was founded by the European Union - Next Generation EU through the PRIN 2022 call powered by MUR, within the project “FLOOD@ROAD” (Prot. 202257JJSJ).

How to cite: Amaddii, M., Castelli, F., and Arrighi, C.: A catchment-scale screening tool for the assessment of bridge overtopping using GIS and LiDAR-derived digital elevation models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7115, https://doi.org/10.5194/egusphere-egu25-7115, 2025.

EGU25-7445 | ECS | Posters on site | HS6.5

Global Soil Moisture Products for Flood Modeling in a Semi-Arid Area 

El Mahdi El Khalki, Tramblay Yves, Massari Christian, Brocca Luca, Simonneaux Vincent, Gascoin Simon, and Saidi Mohamed Elmehdi

Devastating floods in the Mediterranean region are caused by heavy rainfall. Flood forecasting systems are essential in Maghreb countries like Morocco to reduce the consequences and impacts of floods. Developing such a system for ungauged areas is challenging. Even though there is a shortage of observed data, remote sensing products offer a promising solution to fill these data gaps. Different soil moisture and precipitation products are evaluated against in situ data for flood modeling applications. Using an event-based hydrological model with an hourly time step, the results show that observed soil moisture is strongly related to the SMOS-IC satellite product and the ERA5 reanalysis. The comparison of soil moisture records allowed us to calculate the initial soil moisture state using the Soil Conservation Service Curve Number (SCS-CN). Daily in situ soil moisture data may not represent basin soil moisture conditions; however, ASCAT, SMOS-IC, and ERA5 products performed similarly in terms of validation for flood modeling. The daily time step may not accurately represent the saturation state just before a flood, as soil moisture in these semi-arid areas is depleted more quickly after rainfall. For the hourly time step, the initial soil moisture conditions of the SCS-CN model were found to be more accurately represented by ERA5 and in situ data. This work highlights the potential of remote sensing products to improve flood forecasting in semi-arid regions, providing valuable information for the development of robust hydrological models where traditional data are scarce.

How to cite: El Khalki, E. M., Yves, T., Christian, M., Luca, B., Vincent, S., Simon, G., and Mohamed Elmehdi, S.: Global Soil Moisture Products for Flood Modeling in a Semi-Arid Area, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7445, https://doi.org/10.5194/egusphere-egu25-7445, 2025.

EGU25-7538 | ECS | Orals | HS6.5

An enhanced global terrain map using a vision transformer machine learning model 

Peter Uhe, Laurence Hawker, Chris Lucas, Malcolm Brine, Hamish Wilkinson, Anthony Cooper, and James Savage

Digital Elevation Models (DEMs) describe the earth surface’s topography and are an important source of information for applications of physical modelling, engineering and many others. Flood inundation modelling, where water flows are determined by terrain slope, is also highly dependent on DEM quality. The most accurate DEMs currently available are sourced from airborne LiDAR, however these only cover a small fraction of the globe, leaving the majority of the globe sourced from satellite imagery. Satellite based DEMs have limitations and are considered Digital Surface Models (DSMs) which represent the surface of vegetation canopy, buildings and other objects, rather than the bare earth surface which is represented by a Digital Terrain Model (DTM). 

Due to this, we have developed FathomDEM, a DTM generated from the best global satellite based DSM, Copernicus DEM. FathomDEM uses a novel vision transformer technique to improve on previous attempts to generate a DTM from Copernicus DEM.  FathomDEM reduces the Mean Absolute Error and Root Mean Squared Error to half of our previous work, FABDEM, and quarter of Copernicus DEM, while also improving the spatial correlation. 

Flood simulations of inundation using a given DEM shows its use in a real world application and we present results showing flood inundation maps from different global DEMs and LiDAR. FathomDEM gives similar scores to LiDAR data when compared to benchmark flood extents, tested across multiple sites. FathomDEM therefore provides a significant advance when applied to flood inundation modelling in locations without LiDAR DEMs. 

How to cite: Uhe, P., Hawker, L., Lucas, C., Brine, M., Wilkinson, H., Cooper, A., and Savage, J.: An enhanced global terrain map using a vision transformer machine learning model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7538, https://doi.org/10.5194/egusphere-egu25-7538, 2025.

EGU25-9140 | ECS | Orals | HS6.5

Building a global archive of flood events for the last decade based on Sentinel-1 

Andrea Betterle, Bernhard Bauer-Marschallinger, Franziska Kraft, Sandro Martinis, Patrick Matgen, Florian Roth, Tobias Stachl, Wolfgang Wagner, Claudia D'Angelo, and Peter Salamon

The observation of floods from space using Synthetic Aperture Radars (SAR) is a powerful means to understand how inundations unfold across space and time, together with the ensuing impacts. The systematic quantification of the extension of flooded areas and its dynamics is crucial to inform mitigation strategies and organize rescue efforts. Spatiotemporal trends in flood impacts can also help interpret the joint dynamics of climate and exposure, the first for example being associated with climate change while the second with socio-economical evolution. Furthermore, a comprehensive and consistent knowledge of flood events can help to quantify the effectiveness of legislative frameworks attempting to reduce flood impacts, such as the European Flood Directive (2007/60/EC).

This contribution presents the effort in building a global archive of flood events — featuring not only flood extent but also water depth — based on the flood delineations provided by the Copernicus Global Flood Monitoring (GFM). The flood delineations provided by GFM based on Copernicus Sentinel-1 SAR are enhanced using terrain topography, and they are complemented with water depth estimates obtained via the recently released algorithm FLEXTH (Betterle and Salamon, NHESS, 2024). The flood archive will have a global coverage at 20 m spatial resolution, spanning from 2015 until present. The procedure behind the construction of the dataset will be presented, together with the forthcoming steps of combining flood depth maps with exposed asset to further complement the database with flood impacts.

How to cite: Betterle, A., Bauer-Marschallinger, B., Kraft, F., Martinis, S., Matgen, P., Roth, F., Stachl, T., Wagner, W., D'Angelo, C., and Salamon, P.: Building a global archive of flood events for the last decade based on Sentinel-1, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9140, https://doi.org/10.5194/egusphere-egu25-9140, 2025.

EGU25-10980 | Posters on site | HS6.5

Integrated coastal-river water surface elevation datasets derived from SWOT to improve compound flooding simulations over the Mekong Delta 

Monica Coppo Frias, Cecile Marie Margaretha Kittel, Karina Nielsen, Aske Folkmann Musaeus, Christian Toettrup, and Peter Bauer-Gottwein

River deltas are home to more than 400 million people worldwide, being fundamental centers for industry, and ecosystems of great ecological and economic importance. Some of the most densely populated rural and urban areas are in low-lying deltaic regions, such as the Mekong Delta. These areas are highly vulnerable to the impacts of climate change on coastal-river floods, which are driven by several factors, such as sea level rise, extreme river flows or storm surges. To mitigate these effects, accurate integrated coastal-river hydraulic models are essential for enhancing predictive capabilities for compound flooding events and developing effective contingency plans. However, the accuracy of hydraulic models is often limited by the quality of available observations. Developing reliable datasets for coastal-river domains involves addressing several challenges, including a) the high spatial and temporal variability of coastal-estuary dynamics, b) the complex morphology of delta regions characterized by extensive floodplains, braided river channels, and man-made structures, and c) the lack of continuous coastal-river datasets.

Traditional in-situ monitoring provides data only at widely spaced stations, which limits coverage. As a results, satellite Earth Observation (EO) has emerged as a solution to generate datasets with large spatial coverage and high spatial resolution. The Surface Water and Ocean Topography (SWOT) mission is the first dedicated mission to monitor surface water, while also providing ocean height measurements, making it ideal to overcome the monitoring challenges in coastal-river domains. The SWOT mission, with a 120 km wide swath, offers large spatial coverage that can deliver water surface elevation (WSE) and surface water extent observations for rivers as narrow as 50 meters. Additionally, the mission offers a revisit time of 21 days, delivering 2-6 observations in each cycle.

In this study we utilize SWOT observations over the Mekong Delta to generate continuous datasets that span from the river to the ocean. These datasets are used to inform and validate an integrated coastal-river hydraulic model of the Mekong Delta. The SWOT L2_HR_Raster product is exploited at a 100-meter resolution, to derive coastal and estuarine WSE time series and surface water extent. This dataset has the capability to map complex river morphological structures at a temporal resolution previously unattainable by satellite EO missions. It can also capture the effects of ocean tides and storm surges on river water levels, as well as the impact of high river flows on coastal domains. Moreover, the 2D nature of the L2_HR_Raster product can deliver not only river-ocean WSE profiles, but also coastal longitudinal ocean height, to better understand the effect of high river flows in near-coastal areas.

The results provide continuous coastal-river datasets mapping the interplay between near coastal and estuarine dynamics, as well as the complex morphology of the Mekong Delta region. The datasets are used to calibrate and validate a hydraulic model of the Mekong Delta that integrates river and coastal zones to accurately simulate WSE and surface water extent in deltaic regions. The integrated model supports better prediction capabilities for compound flooding simulations and the impacts of climate change on the coastal and estuarine environments.

How to cite: Coppo Frias, M., Kittel, C. M. M., Nielsen, K., Musaeus, A. F., Toettrup, C., and Bauer-Gottwein, P.: Integrated coastal-river water surface elevation datasets derived from SWOT to improve compound flooding simulations over the Mekong Delta, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10980, https://doi.org/10.5194/egusphere-egu25-10980, 2025.

EGU25-11828 | ECS | Posters on site | HS6.5

Satellite Mapping Analysis of the November 2023 Flood in Prato, Tuscany 

Beatrice Carlini, Luca Baldini, Elisa Adirosi, Giovanni Serafino, Giovanni Scognamiglio, and Roberta Paranunzio

Climate change has increased the frequency and intensity of extreme weather events, leading to greater risks for vulnerable urban areas. Inadequate infrastructure often exacerbates vulnerability of many areas, resulting in significant socioeconomic losses from climate-related hazards and in particular flooding. Satellite services, smart technologies such as GIS-based Digital Twin help to simulate flooding scenarios to support urban planning and decision-making and provide monitoring and short-term forecasting of floods thus contributing to enhance climate resilience and to strengthen financial risk strategies.

To ensure that these systems operate effectively, the validation of their predicted  outputs in terms of flooding maps is crucial. This task is usually possibly carried out using the satellite-based data available and particularly those from Synthetic Aperture Radar (SAR), which are effective in various meteorological conditions. In urban areas, the application of state-of-the-art SAR-based methods for flood detection is challenging due to the complexity of effects caused by the radar backscattering from built environments.

This study focuses on validating flood maps for urbanized environments based on a consolidated approach that reprocesses the clustering result with fuzzy logic approach (Pulvirenti et al. 2023, DOI: 10.3390/w15071353) and here improved to better estimate flooding in urban areas. The method was applied to a severe precipitation event in November 2023 in Tuscany, Italy, which caused multiple flood episodes. Our focus was on the Florence-Prato-Pistoia plain, the most densely populated area in Tuscany. On November 2, heavy rainfall began in the early afternoon, accumulating 130-170 mm within 5-6 hours. This led to the first flood episodes after 19:00 local time, resulting in several casualties.

Copernicus Rapid Mapper was activated on 03/11/2023, 04:21 (Local time = UTC+1). It produced an estimate of flooded area mainly using one COSMO-SkyMed image, collected on November 6, after a second storm occurred in the night between 4 and 5 November. In our analysis we used two images. For the common image, good spatial correspondence was obtained. However, due to the late availability of satellite images, critical early floods were missing.

This work takes this case study to address the opportunity and challenges of flood detection in urban areas using satellite data. While highlighting the importance of having a satellite flood mapping service, some drawbacks are also pointed out, such as the lack of revising time that can imply missing early stages of floods to early implement search and rescue operations. Projects to improve revisiting time are related to the emergence of next generation constellations, such the ASI/ESA IRIDE multisatellite and multiservice constellation. In case of fast evolving phenomena, such as the one considered in this study, a higher time resolution of flood mapping would increase the chance to obtain data even in the first floods. In practice, resorting to modelling and sensor data coupled in digital twins eventually integrated with obtained from citizens science will be still unavoidable. This is demonstrated within the SCORE project (https://score-eu-project.eu/), a four-year EU-funded project aiming to increase climate resilience in European coastal cities (Coastal City Living Labs - CCLLs).

How to cite: Carlini, B., Baldini, L., Adirosi, E., Serafino, G., Scognamiglio, G., and Paranunzio, R.: Satellite Mapping Analysis of the November 2023 Flood in Prato, Tuscany, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11828, https://doi.org/10.5194/egusphere-egu25-11828, 2025.

EGU25-13306 | ECS | Posters on site | HS6.5

Ground observations and UAV mapping to support a GIS-based implementation of the Flash Flood Impact Severity Scale (FFISS) for the 2009 and 2020 flash floods in Evia, Greece. 

Nafsika-Ioanna Spyrou, Michalis Diakakis, Spyridon Mavroulis, Georgios Deligiannakis, Emmaouil Andreadakis, Christos Filis, Evelina Kotsi, Zacharias Antoniadis, Maria Melaki, Marilia Gogou, Katerina-Navsika Katsetsiadou, Eirini-Spyridoula Stanota, Emmanuel Skourtsos, Emmanuel Vassilakis Vassilakis, and Efthymios Lekkas

Flash floods have been responsible for some of the most catastrophic events globally. The extensive range of effects and the varying severity of impacts present significant challenges in accurately understanding the damage caused by a flood event, thereby hindering our capacity to predict future consequences. When evaluating flood impacts and their severity, most existing approaches rely on qualitative descriptions (e.g., major, catastrophic, etc.) or examine the impacts from a single perspective or discipline, such as economic losses. In this study, the Flash Flood Impact Severity Scale (FFISS) is employed to evaluate, map, and categorize the impacts of two flash floods that occurred in the Lilas River in Greece in 2009 and 2020. The goal of this application is to analyze the varying severity levels and how one flood event can influence the impacts of a subsequent event. The methodology involved extensive fieldwork, including the collection of ground-based and aerial observations using UAV technology to document the impacts. These observations were subsequently georeferenced, followed by applying the Flash Flood Impact Severity Scale (FFISS) and creating detailed maps to assess and evaluate the severity of impacts associated with the two flood events. The results indicate that, despite the higher water levels during the second flood, areas previously affected show lower severity values. This reduction is attributed to the community’s gradual adaptation, improvements in infrastructure, and significant local widening of the river channel. Conversely, newly flooded areas during the second event exhibit high severity levels. Overall, applying the FFISS reveals spatial patterns of impact severity, offering insights into the local nature of floods while suggesting a potential reduction in overall risk during the post-flood period.

How to cite: Spyrou, N.-I., Diakakis, M., Mavroulis, S., Deligiannakis, G., Andreadakis, E., Filis, C., Kotsi, E., Antoniadis, Z., Melaki, M., Gogou, M., Katsetsiadou, K.-N., Stanota, E.-S., Skourtsos, E., Vassilakis, E. V., and Lekkas, E.: Ground observations and UAV mapping to support a GIS-based implementation of the Flash Flood Impact Severity Scale (FFISS) for the 2009 and 2020 flash floods in Evia, Greece., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13306, https://doi.org/10.5194/egusphere-egu25-13306, 2025.

EGU25-13757 | ECS | Posters on site | HS6.5

Semi-Automatic Extraction and Morphometric Characterization of Paleochannels using LiDAR Data: A Case Study in the fens of Lincolnshire, England 

Gianpietro Imbrogno, Giuseppe Cianflone, Rocco Dominici, Giuseppe Maruca, Paolo De Cesare, Mark Schuerch, and Luca Mao

Paleochannels are natural features in floodplains, and their identification and geometric characterization can guide river restoration and natural flood management interventions. This study focuses on identifying the network of dendritic drainage patterns in a portion of the Lincolnshire fens near Billinghay. A semi-automatic approach was developed for identifying paleochannels and performing a morphometric analysis of these features.

A high-resolution LiDAR data survey from 2022 was downloaded from the UK environment portal. The LiDAR digital terrain model has a resolution of 2 m and vertical accuracy of +/- 15 cm. The raw LiDAR point cloud was pre-processes using CloudCompare. An initial ground-level extraction was performed with automatic filters and further refined by identifying and removing additional anthropogenic features such as roads, buildings, and artificial levees along canals, using a vector data analysis. The dendritic drainage channels of the particular study site (6.78 km2) were isolated using a semi-automatic selection with specific elevation filters. The differences in elevation between the paleochannel surface and the surrounding flat areas were used to define distinct elevation ranges for different altimetric bands. Points within these ranges were selected and reclassified to create a preliminary morphological model of the paleochannels. Discontinuous segments were interpolated, and areas with missing values were resampled, resulting in a consistent and detailed representation of the paleochannels.

The dendritic drainage network was characterized in terms of Strahler order, sinuosity, length, and location of connection nodes. Additionally, several cross-sectional profiles were generated and a Python script was developed to quantify the width, depth, and area between the crest of the paleosurface and the ground level. Reaches of paleochannels of higher Strahler order were found to be deeper and wider. The sinuosity is lower for the reaches in the upper part of the dendritic network. Interestingly, the channels are located in areas that are highly convex compared to the surrounding flat areas. The total surface area occupied by the identified paleochannels in the study site is approximately 1.8 km2, which represents a significant portion of the floodplain.

The geometry of the identified enclosed basin and of the dendritic network are being used to test a morphodynamic model in order to identify the sea level and tidal ranges responsible for the formation of the paleochannels.

How to cite: Imbrogno, G., Cianflone, G., Dominici, R., Maruca, G., De Cesare, P., Schuerch, M., and Mao, L.: Semi-Automatic Extraction and Morphometric Characterization of Paleochannels using LiDAR Data: A Case Study in the fens of Lincolnshire, England, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13757, https://doi.org/10.5194/egusphere-egu25-13757, 2025.

EGU25-14254 | ECS | Orals | HS6.5

A Database of Flood Maps using high-resolution Airborne Imagery and Machine Learning Models 

Dinuke Munasinghe, Sagy Cohen, Dan Tian, and Hongxing Liu

Optical Satellite imagery commonly suffers from the presence of cloud cover during flood events; Radar Satellites are disadvantaged from water look-alike conditions where the ground surface interacts with the incoming radar signal as if it were water; Regardless of modality of satellite, more importantly, satellite overpasses during a flood are chance occurrences where the capture of the maximum extent is a fortuitous incident. Low-altitude aerial remote sensing, on the other hand, can be used to survey the extent of flooding at the peak or soon after it has occurred, with a good measure of reliability. Opportune scheduling of these reconnaissance flights not only capture floods at ultra-high resolution, but also allows for seamless geographical coverage unhindered by cloud cover.

The National Oceanic and Atmospheric Administration (NOAA) Emergency Response Imagery is very high resolution (50 cm Ground Sampling Distance between pixels) airborne imagery acquired by the Remote Sensing Division of the National Geodetic Survey (NGS) during major flood events in the United States to support NOAA’s homeland security and emergency response requirements.

In this work, we evaluated the performance of four different machine learning models (Gradient Boosting, Random Forest, Support Vetor Machine, Convolutional Neural Network) on the ability to classify floods from raw aerial imagery. The classifier with the highest classification accuracy metrics - depending on geographical and hydrological setting – was used to produce flood inundation extent maps for 30 major flood events.

We demonstrate the utility of these high-fidelity flood maps via two use-cases: both synergistic studies to this work. 1) As benchmarks for validation of hydrodynamic model results: Historic flooding occurred on the Neuse River in North Carolina in the United States triggered by Hurricane Matthew in 2016. Several hydrodynamic models were deployed to simulate flood dynamics with an end goal of understanding flood susceptibility in the Neuse basin under changing climate conditions. The aerial imagery-based flood maps were used as benchmarks for model validation. 2) Enhancing the versatility of FIMPEF: Flood Inundation Mapping Predictions Evaluation Framework (FIMPEF) is an open-source, cloud-based geospatial venture by the University of Alabama that calculates accuracy statistics between benchmark and modeled flood extents. Integration of aerial imagery, in addition to the satellite-based benchmarks that FIMPEF was ingesting so far, has vastly enhanced its robustness and user-demand. Free access (no account/login credentials needed) to these high-quality flood maps is granted through the United Sates Flood Inundation Map Repository (USFIMR), an online geospatial warehouse that provides high-resolution inundation extent maps of past U.S. flood events.

How to cite: Munasinghe, D., Cohen, S., Tian, D., and Liu, H.: A Database of Flood Maps using high-resolution Airborne Imagery and Machine Learning Models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14254, https://doi.org/10.5194/egusphere-egu25-14254, 2025.

EGU25-15889 | Orals | HS6.5

Local hydrological and hydrodynamic modeling for flood forecasting in Burkina Faso 

Laetitia Gal, Pauline Casas, Kévin Larnier, Romulo J. Oliveira, and Adrien Paris

Burkina Faso climate is characterized by a short rainy season and  high rainfall variability, characteristic of tropical-equatorial regions, resulting in extreme rainfall events and high flood risks in its watersheds and cities. In the capital Ouagadougou, rapid urban development associated with low-permeability soils and high precipitation intensity lead to major flooding events (e.g. in 2009, 2016, 2020) affecting households and economy. This vulnerability to flooding also affects other strategic points in Burkina Faso, such as crossroads between national roads and rivers, where overflows almost every year lead to limited road access and hinder economical transportation.

This study presents an innovative integrated framework to improve forecasting capacity and manage flood risks at the local scale, for both (i) pluvial flooding over Ouagadougou city and (ii) fluvial flooding at six points of interest (POIs) across Burkina Faso. The methodology is based on a 2D hydrodynamic modeling using the DassHydro [1] framework and only publicly available data (soil properties, land cover, etc.). For pluvial flooding (Ouagadougou case), this model is forced with operational precipitation products. For fluvial flooding,  daily real-time discharge data computed with the MGB hydrological model [2] are used as boundary conditions for the hydrodynamic model set at the POIs. Both approaches produce local flood maps for different warning levels, based on precipitations  and/or discharge thresholds. Flood maps produced for each POI were validated through comparisons to Sentinel-2 images of  historical floods, on-site flood marks analysis and spatial altimetry.  Additionally, comparisons with previous studies conducted in Ouagadougou as well as historical informations,  demonstrated the relevance and reliability of the results obtained through our methodology at both local scale.

This preliminary approach showed the efficiency of the methodology for a flood risk warning and forecasting system in a data-sparse context and highlighted the strong need for in-situ data and finer-grained topology data, among others, in those regions. Further consideration of new in situ data provided by local agencies should permit increasing the accuracy of forecasts and provide refined risk analysis.

[1] https://dasshydro.github.io/

[2] https://www.ufrgs.br/lsh/mgb/what-is-mgb-iph/ 

How to cite: Gal, L., Casas, P., Larnier, K., J. Oliveira, R., and Paris, A.: Local hydrological and hydrodynamic modeling for flood forecasting in Burkina Faso, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15889, https://doi.org/10.5194/egusphere-egu25-15889, 2025.

EGU25-15998 | ECS | Posters on site | HS6.5

Enhancing Water Level Estimates with DEM-derived Stream Geomorphometry 

Søren Kragh, Jun Liu, Lars Troldborg, Simon Stisen, Raphael Schneider, and Julian Koch

Accurate water level predictions are increasingly crucial for mitigating flood risks. Hydrological and hydrodynamic models provide water level predictions, but their accuracy depends on detailed information about stream cross-sections and floodplain topography, which are data that are difficult to obtain at larger scale, especially in regions with perennial river systems. Stream discharge is a variable that is more straightforward to predict by conventional hydrological models. However, the relationship between discharge and water level is complex, depending on cross-section geometry and channel roughness. Here machine learning models offer an alternative opportunity to predict water level by ingesting readily available topographic data derived from high-resolution digital elevation models in combination with simulated stream discharge, thereby skipping the need to explicitly define rating curves or to run complex hydrodynamic simulations. The idea is that stream discharge provides information about the temporal variability, whereas the topographic data provides static information in the cross-section geometry.  

First, we present a method for extracting stream geomorphometry from a high-resolution (40 cm) digital elevation model in Denmark. The methodology is based on analyzing elevation changes along cross-sections throughout the entire Danish river network. Stream widths are estimated by identifying the most probable bank positions through a probabilistic count of all possible configurations within 100-meter stream reaches. The resulting dataset has been validated against 2,000 measured cross-sections along Danish rivers, showing similar spatial patterns across reach to river scales. Moreover, the slope and elevation of the water level as well as channel area and depth are derived from the high-resolution DEM for 100-meter stream reaches.

Second, we present the development of a machine learning-based model that utilizes the derived stream geomorphometry in combination with stream discharge simulated by the National Hydrological Model of Denmark to predict daily stream water levels. Timeseries of daily stream water level of 40 gauging stations are used to train a Long Short Term Memory network. The results demonstrate that incorporating topography-derived information of mean water level and slope, stream channel width, area, and depth, enhance the accuracy of the water level estimates. Overall, our approach provides a versatile approach providing crucial information on flood risks that can easily be scaled up to national scale.

How to cite: Kragh, S., Liu, J., Troldborg, L., Stisen, S., Schneider, R., and Koch, J.: Enhancing Water Level Estimates with DEM-derived Stream Geomorphometry, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15998, https://doi.org/10.5194/egusphere-egu25-15998, 2025.

EGU25-16950 | ECS | Posters on site | HS6.5

Closing the Gap: Towards Consistent Flood Extent Retrieval with Multi-Sensor Data Fusion 

Chloe Campo, Paolo Tamagnone, Guy Schumann, Suelynn Choy, Trinh Duc Tran, and Yuriy Kuleshov

Despite the significant increase in Earth Observation (EO) satellites, the frequency of cloud-free imagery at sufficiently high spatial resolutions for timely inundation mapping remains a significant challenge. Obtaining more frequent flood extent estimations would contribute to our understanding of flood dynamics and increase the likelihood of capturing the flood peak, which often evades EO acquisitions. Integrating complementary data from multiple sensors is a potential solution to overcome limitations posed by temporal resolution, spatial resolution, cloud cover, adverse weather, or light conditions. Surface water fractions, indicating the proportion of a pixel covered by water, can be derived from a variety of sensors that passively sense across different spectral ranges daily. However, the fractional coverages are derived at various spatial resolutions, necessitating a methodology to harmonize and combine the information to obtain a comprehensive flood map at a meaningful resolution. The present study proposes a methodology to seamlessly combine data from Low-Earth Orbiting (LEO) multispectral, Geostationary-orbiting (GEO) multispectral, and Passive Microwave (PMW) sensors. The proposed approach is tested on the February 2022 flood event in Brisbane, Australia, and fuse data from Visible Infrared Imaging Radiometer Suite (VIIRS), the Himawari 8/9 Advanced Himawari Imager (AHI), and the Special Sensor Microwave Imager/Sounder (SSMIS). These sensors offer complementary strengths in flood detection, including sub-daily imagery from VIIRS and AHI, and fractional water estimates beneath cloud cover from SSMIS.

Surface water fractions, representing the fraction of a pixel covered by water, are derived from VIIRS, AHI, and SSMIS at spatial resolutions of 375 m, 1 km, and 25 km, respectively. These surface water fractions are subsequently homogenized via downscaling and fused to obtain an aggregated flood map. A Digital Terrain Model and its derivatives, including the Slope, Topographic Water Index, Height Above Nearest Drainage, and Flow Accumulation, and water frequency information are utilized to downscale and distribute the surface water fractions in physically plausible ways. This disaggregation process produces comparable flood maps from all sensors. These maps are thereafter combined to yield a single detailed flood map. This multi-sensor framework ensures the consistent generation of flood maps at a meaningful spatial and temporal resolution, compensating for the unavailability of moderate- to high-resolution imagery due to satellite revisit timing and cloud obstruction. The proposed approach enables more frequent generation of detailed flood maps, providing valuable insights into inundation dynamics to scientists and decision makers.

How to cite: Campo, C., Tamagnone, P., Schumann, G., Choy, S., Duc Tran, T., and Kuleshov, Y.: Closing the Gap: Towards Consistent Flood Extent Retrieval with Multi-Sensor Data Fusion, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16950, https://doi.org/10.5194/egusphere-egu25-16950, 2025.

Floods exacerbated by climate change significantly increase the risk of dam failure, posing a critical threat to downstream regions. A cost-effective way to analyze the consequences of dam break floods is by using unsteady hydrodynamic models that incorporate St. Venant’s or diffusion wave equations. These models require detailed topographic data, land cover information, and a dam break hydrograph. This study assesses the influence of various remote sensing topographic datasets on 2-dimensional (2D) hydrodynamic flood modeling using HEC-RAS v6. The methodology is applied to İmranlı town in Türkiye, located downstream of an irrigation dam. Under a 500-year return period flood scenario, a breach hydrograph is simulated in HEC-RAS, assuming overtopping when the reservoir is at full capacity. Manning's roughness values are derived from the ESA-WorldCover satellite land use map. Two types of topographic data are tested: Digital Surface Models (DSMs) and Digital Terrain Models (DTMs). Specifically, datasets include field-based Light Detection and Ranging (LiDAR) DSM (0.5 x 0.5 m resolution), Turkish General Directorate of Mapping (HGM)-based DSM (5 x 5 m resolution), Advanced Land Observing Satellite – Phased Array type L-band Synthetic Aperture Radar (ALOS-PALSAR)-sourced DTM (12.5 x 12.5 m resolution), and Shuttle Radar Topography Mission (SRTM)-sourced DTM (30 x 30 m resolution).

The study also explores the impact of combining high-resolution and low-resolution topographic data by mosaicking LiDAR data, limited to urbanized areas, with other datasets. Results are evaluated using performance metrics such as Mean Absolute Error (MAE), Root Mean Square Error (RMSE), F-index, and correlation coefficient (R²). Additionally, comparisons are drawn using flood-related maps, including flood inundation area, water depth, velocity, duration, and hazard. The study highlights that nearly the entire İmranlı district center and the Doğançal settlement would be inundated in the event of a dam failure, exposing approximately 7,028 individuals to flood risk. The findings suggest that while high-resolution HGM-based data serve as a reliable reference, integrating satellite datasets like ALOS-PALSAR with LiDAR enhances model performance, making them valuable alternatives when high-resolution data are unavailable.

How to cite: Uysal, G. and Tasci, E.: Two-Dimensional Hydrodynamic Modeling and Comparison of Flood Propagation from İmranlı Dam Break Using Different Remotely Sensed Topographic Data, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17954, https://doi.org/10.5194/egusphere-egu25-17954, 2025.

EGU25-18097 | Orals | HS6.5 | Highlight

Integration of Remote Sensing and Hydraulic Modeling for Dynamic Flood Monitoring: A Copernicus Emergency Management Service for retrospective flood temporal analysis in Saarland, Germany 

Alexandros Konis, Vasiliki Pagana, Stavroula Sigourou, Alexia Tsouni, Emmanouil Salas, Michail-Christos Tsoutsos, Nikolaos Stathopoulos, Nikolaos Stasinos, and Charalampos (Haris) Kontoes

Floods affect many regions of the world every year and are the most deadly natural hazard. The increasing pressures of a growing global population, widespread ecosystem degradation, and the compounding effects of climate variability and change are significantly increasing flood risks worldwide. Hydrodynamic models, combined with Earth Observations (EO), play an increasingly important role in the comprehensive analysis and characterization of floods, providing a deeper understanding of their dynamics in past, present, and future scenarios.

Under the “Copernicus Emergency Management Service (CEMS) Risk and Recovery Mapping (RRM)” framework, this on-call study (i.e., CEMS activation “ΕMSN: Retrospective flood temporal analysis of floods in Saarland, Germany”) focused on the mid-May 2024 (16/05/2024-22/05/2024) flood in Saarland, Germany, which resulted in extensive damage across the Saarland state capital Saarbrücken and several districts in Saarland. Leveraging advancements in Earth observation (EO), this study integrated multi-source remote sensing data into a 2D hydraulic modeling framework to enhance the understanding of flood dynamics in the region through a comprehensive temporal analysis.

Using the HEC-RAS hydraulic modeling open-source software of the United States Army Corps of Engineers (USACE), a rain-on-grid approach was employed to simulate direct rainfall runoff to supplement fluvial model simulation of flood propagation over a 7-day period. Model calibration was based on observed water depth data from Gauging stations’ recordings, with adjustments made to improve accuracy. Validation was conducted using EO-derived flood delineations from multitemporal post-event imagery, spanning multi-Platform Satellite products including SAR (Sentinel-1A, RadarSat-2, COSMO-SkyMed and TerraSAR-X) and Optical (Planet Scope) imagery. Therefore, the outputs of the study including the water depth and the flood persistence were derived from the combination of the hydraulic modeling and remote sensing methodologies.

Despite the relatively lower flood thematic accuracy of EO-derived flood outlines in urban and forested areas given the inherent limitations of the SAR analysis techniques, the availability of multitemporal EO imagery was decisive in validating the hydraulic modelling accuracy and robustness. The study findings emphasize the emerging potential of EO data for validating hydraulic models and therefore enhancing flood mapping and monitoring capabilities. In this context, the availability of multitemporal EO datasets further enhanced the flood modelling performance in providing a better insight into the flood propagation and dynamics over the whole period of impact.

Acknowledgment: The service took place under the Framework Service Contract 945236–IPR–2023 “Copernicus Emergency Management Service (CEMS) Risk and Recovery Mapping (RRM) Tailor-Made Products. 

We would like to acknowledge the great support of the JRC CEMS team memebrs, namely Guido Di Carlo, Cristina Rosales Sanchez, and Emanuele Sapino, for the completion of this service contract.

How to cite: Konis, A., Pagana, V., Sigourou, S., Tsouni, A., Salas, E., Tsoutsos, M.-C., Stathopoulos, N., Stasinos, N., and Kontoes, C. (.: Integration of Remote Sensing and Hydraulic Modeling for Dynamic Flood Monitoring: A Copernicus Emergency Management Service for retrospective flood temporal analysis in Saarland, Germany, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18097, https://doi.org/10.5194/egusphere-egu25-18097, 2025.

EGU25-20493 | Orals | HS6.5

Toward Robust Evaluations of Flood Inundation Predictions Using Remote Sensing Derived Benchmark Maps 

Sagy Cohen, Anupal Baruah, Parvaneh Nikrou, Dan Tian, Hongxing Liu, and Dinuke Munasinghe

Remote Sensing-derived Flood Inundation Maps (RS-FIM) are an attractive and commonly used source of evaluation benchmarks. Errors in model-predicted FIM (M-FIM) evaluation results due to biases in RS-FIM benchmarking are quantified by introducing a high-confidence benchmark FIM, which was manually delineated from ultra-resolution imagery, as Ground Truth. The evaluation results show considerable differences in M-FIM accuracy assessment when using lower-quality benchmarks. A RS-FIM enhancement (gap-filling) procedure is presented and its effect on FIM evaluation results is analyzed. The results show that the enhancement is insufficient for significantly improving the robustness of the evaluation. The impact of including/excluding Permanent Water Bodies (PWB) on FIM evaluation results is analyzed. The results show that including PWB in FIM evaluation can significantly inflate the model accuracy. A novel evaluation strategy is proposed and analyzed. The proposed evaluation strategy is based on excluding low-confidence grid cells and PWB from the M-FIM evaluation analysis. Low-confidence grid cells are those that were estimated to be flooded by the gap-filling procedure, but were not classified as such by the remote sensing analysis. The results show that the proposed evaluation strategy can dramatically improve the robustness of the evaluation, except when a considerable number of false positives exist in the RS-FIM. The analyses showcase the many challenges in FIM evaluation. We provide an in-depth discussion about the need for standards, user-centric evaluation, the use of secondary sources, and qualitative evaluation.

How to cite: Cohen, S., Baruah, A., Nikrou, P., Tian, D., Liu, H., and Munasinghe, D.: Toward Robust Evaluations of Flood Inundation Predictions Using Remote Sensing Derived Benchmark Maps, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20493, https://doi.org/10.5194/egusphere-egu25-20493, 2025.

EGU25-1028 | Posters on site | NH3.3

The characterization of landslide heterogeneity in urbanized area using geophysical and machine learning methods: a case study from Cieszyn, Poland 

Małgorzata Sokołowska, Iwona Stan-Kłeczek, Artur Marciniak, Krzysztof Śliwiński, and Marta Palarz

Landslides in urban areas pose special challenges for engineering geology. Because of the high risk they pose, they require special attention. In the presented work, the key novelty is an approach using geophysical imaging methods and unsupervised machine learning to identify a high-risk landslide in an urban area. It proved insufficient in the case presented here, and the proposed approach made it possible to identify the slip surface much more accurately. The results obtained were verified and supplemented with borehole data. Combining model generation based on machine learning can be applied as a new solution.

The research presented concerns the analysis of the stability of a slope located in the centre of the city of Cieszyn (Voivodeship, Silesia, Poland). The research used geophysical methods, including electrical resistivity tomography, refraction seismic and multichannel surface wave analysis. The essence of the study was to identify the geological structure and determine the slip surface of the rock masses, which are expected to answer whether further urbanization and development of the area is possible on the studied slope and whether the recognized landslide threatens lower-lying structures. As a result of the research, the object of study was recognized, and the effectiveness of the assumed cost-effective methodology was presented. The described example and used approach can broadly apply to similar research problems in the Carpathian region and for imaging similar geotechnical problems in other parts of the world.

How to cite: Sokołowska, M., Stan-Kłeczek, I., Marciniak, A., Śliwiński, K., and Palarz, M.: The characterization of landslide heterogeneity in urbanized area using geophysical and machine learning methods: a case study from Cieszyn, Poland, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1028, https://doi.org/10.5194/egusphere-egu25-1028, 2025.

EGU25-2864 | ECS | Posters on site | NH3.3

Cutting-edge applied geophysics and data science in the evaluation of hydrogeological risk in urban areas  

Luigi Martino, Giuseppe Calamita, Sebastian Uhlemann, Francesco Cavalcante, Filomena Canora, and Angela Perrone

The frequency of extreme rainfall events has significantly risen in recent years, compelling administrators of urban areas to develop and adopt innovative strategies to effectively manage precipitation overload. This climatic trend has heightened the complexity of addressing hydrogeological risks, requiring a deeper understanding of the mechanisms driving catastrophic events. Among these, landslides represent one of the most critical challenges, necessitating multidisciplinary approaches to improve prediction, prevention, and mitigation strategies. Several studies have demonstrated how the spatial-temporal variation of moisture content in soil is crucial in the triggering and reactivation of landslide phenomena. Hydrogeophysics plays a pivotal role in understanding these processes at multiscale spatial-temporal resolutions. Its effectiveness is significantly enhanced when combined with detailed hydrological and environmental analyses. Multiparametric strategy, leveraging continuous multisensory monitoring systems, has proven to be one of the most effective methods for modeling soil moisture behavior. Recent advancements have optimized the selection of components for such systems, incorporating time-lapse ERT systems alongside various hydrologic and environmental sensors. This synergy enables sophisticated 2D and 3D dynamic thermo-hydro-geomechanical modeling of the subsurface, offering unprecedented insights into soil moisture dynamics and landslide mechanisms. Our work focuses on a peri-urban landslide located a few hundred meters from the centre of a small town in the southern Apennines of Italy, characterized by slow-moving displacements. We are establishing an open-air laboratory equipped with a monitoring station that integrates time-lapse ERT system with an array of several hydrological (tensiometers, soil moisture sensors, piezometers) and meteorological (thermometers, hygrometers, anemometers, pyranometers) sensors. The large quantity of data generated by this monitoring station will be managed through the development of innovative data processing methods, also leveraging advanced machine learning techniques. These approaches will enable efficient analysis and integration of geophysical, hydrogeological, and environmental datasets across laboratory and site scales, enhancing our ability to model and understand landslide behaviour with greater accuracy and precision.  This work is one of the activities carried out within the WP7-7.4 task of the ITINERIS "Italian Integrated Environmental Research Infrastructures System" project (PNRR M4C2 Inv.3.1 IR), funded by the EU's Next Generation program, an integrated geophysical approach for the assessment of geohazards in urban areas.

How to cite: Martino, L., Calamita, G., Uhlemann, S., Cavalcante, F., Canora, F., and Perrone, A.: Cutting-edge applied geophysics and data science in the evaluation of hydrogeological risk in urban areas , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2864, https://doi.org/10.5194/egusphere-egu25-2864, 2025.

EGU25-4033 | ECS | Posters on site | NH3.3

Evolution of the strain localization and shear-zone internal structure in the granular material: Insights from ring-shear experiments 

Yangshuai Zheng, Wei Hu, Yan Li, Huaixiao Gou, and Yi Ge

Shear zones are commonly observed in natural faults, landslides, and laboratory experiments involving granular materials. Gaining insight into the evolution of shear zones in these materials is essential for understanding the mechanics of faults and landslides, yet this process remains insufficiently understood. To address this, we conducted a series of ring-shear experiments to study the development of strain localization and the internal structure of shear zones in both cohesive and non-cohesive granular materials. Using high-resolution X-ray computed tomography (CT), we quantitatively analyzed shear-zone structures, including particle shapes, orientations, and grain-size distributions, at various levels of shear strain. Our results reveal that with increasing shear displacement, larger particles within the shear zones become progressively rounded, though without a preferred orientation. Additionally, wear and attrition processes generate a significant number of nanoparticles within the shear zones. Fine-particle layers composed of these nanoparticles were observed to form along the edges of the shear zones as shear localization developed, suggesting a transition of the shear process from a distributed zone to a more defined interface. These findings provide insights into the evolution of shear zones in granular materials, offering a deeper understanding of the mechanics underlying fault and landslide dynamics.

How to cite: Zheng, Y., Hu, W., Li, Y., Gou, H., and Ge, Y.: Evolution of the strain localization and shear-zone internal structure in the granular material: Insights from ring-shear experiments, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4033, https://doi.org/10.5194/egusphere-egu25-4033, 2025.

EGU25-6219 | Orals | NH3.3

Insights from Snow Avalanche Detection in Norway: A Distributed Acoustic Sensing (DAS) Study 

Antoine Turquet, Guro K. Svendsen, Andreas Wuestefeld, Finn K. Nyhammer, Espen Nilsen, Andreas Persson, and Vetle Refsum

Snow avalanches pose a significant hazard in mountainous areas, especially when snowpacks block roads, either burying vehicles directly or exposing traffic to subsequent avalanches during active cycles.

We have been monitoring avalanche activity along road stretches in Northern Norway since 2022 using Distributed Acoustic Sensing (DAS),  a technology capable of theoretically covering spans of up to 170 km. Traditional detection methods often focus on only a limited section of a road stretch, making effective risk management challenging. DAS powered alert system can work unaffected by visual barriers and in adverse weather conditions. The developed algorithm identifies avalanches affecting the road and estimates accumulated snow. Moreover, the system can also detect vehicles on the road, offering invaluable support to search and rescue operations.

Over 3 winters the system successfully identified 10 road-impacting avalanches (100% detection rate). Our results via DAS align with the previous works and indicate that low frequency part of the signal (<20 Hz) is crucial for detection and size estimation of avalanche events. We have identified subsets of snow avalanches based on the paths they followed and discuss the snow accumulation and deposition signatures on signals. Various fiber installation methods are explored to optimize sensitivity in detecting avalanches. The findings highlight the system’s robustness and low maintenance demands, offering a clear advantage over conventional systems, which are costly to install, have restricted coverage, or are vulnerable to environmental factors such as weather and lighting.

How to cite: Turquet, A., Svendsen, G. K., Wuestefeld, A., Nyhammer, F. K., Nilsen, E., Persson, A., and Refsum, V.: Insights from Snow Avalanche Detection in Norway: A Distributed Acoustic Sensing (DAS) Study, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6219, https://doi.org/10.5194/egusphere-egu25-6219, 2025.

EGU25-6320 | Orals | NH3.3

The June 2024 Mattertal slope destabilizations: zoom into the Gugla rock glacier 

Eric Larose, Maélys Strapazzon, Antoine Guillemot, Agnès Helmsetter, and Guillaume Favre-Bulle

The last weeks of June 2024 were a very active period in the Alps with various floods, landslides, rockfalls and debris flows. In particular, the Mattertal valley (Switzerland) was hit by intense rainfall on June 20-22, following a very snowy winter and rainy spring. This led to various floods and debris flows, including the cutting off of the road and railway to the famous town of Zermatt. Also, some exceptional slope destabilization were also observed before the late June storm activity. Forecasting such natural hazards and anticipating the effects of rapid erosion processes is key for public managers, especially for energy and communications infrastructures and tourist resorts in mountainous valleys.

Using passive seismic sensors placed on the Gugla rock glacier (2700 m a.s.l) above Herbriggen village, Mattertal, we have detected landslides and quakes around the rock glacier almost continuously from 2016 to 2024 [1]. Using the same seismic instrument, we were also able to measure relative seismic velocity changes on a daily basis, which are indicative for the variations in stiffness at depth undergone by the rock glacier [1]. We observe seasonal variations of relative velocity changes and rockfall activity, mainly controlled by the freeze-thawing cycles. Melting seasons and wet summer episodes (storms) generally lead to seismic velocity drops of 2-3% in May-June. In June 2024, however, we observed a significant decrease in seismic velocity (-6.5%), which corresponds to a significant decrease in stiffness (ice melting) and a high liquid water content (snow melting infiltration), both lowering ground stability. This reduction in ground stability is likely to be responsible for the observed faster kinematics of the frontal part of the rock glacier, as well as rockfall and debris flow activity increase downstream.

Since this reduction in ground stability is likely to have occurred further in the Mattertal catchment at the same elevation and orientation, our work emphasizes that this reduction in seismic velocity at the catchment scale may be a good proxy for the higher sensitivity of the catchment to environmental triggers such as rainfall, eventually leading to a higher probability of slope destabilization.

[1] A. Guillemot, et al: Seismic monitoring in the Gugla rock glacier (Switzerland) : ambient noise correlation, microseismicity and modelling, Geophys. J. Int. 221, 1719-1735 (2020).

This work was partially funded by the Wallis canton, and by the European Research Council (ERC) under grant No. 101142154 - Crack The Rock project.

How to cite: Larose, E., Strapazzon, M., Guillemot, A., Helmsetter, A., and Favre-Bulle, G.: The June 2024 Mattertal slope destabilizations: zoom into the Gugla rock glacier, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6320, https://doi.org/10.5194/egusphere-egu25-6320, 2025.

EGU25-6485 | Orals | NH3.3

Integrating Remote Sensing and Geophysics to Assess Landslide Risk in the Italian Apennines: A Case Study in Gorgoglione, Italy 

Giuseppe Calamita, Angela Perrone, Francesco Falabella, Antonio Pepe, Tony Alfredo Stabile, Maria Rosaria Gallipoli, Vincenzo Serlenga, Erwan Gueguen, Jessica Bellanova, Mario Bentivenga, and Sabatino Piscitelli

This study proposes an integrated methodology to investigate hydrogeological instability, combining remote sensing with in-situ geophysical surveys in Gorgoglione, a small town in Basilicata, southern Italy, located in a low mountain area (~800 m a.s.l.). The Italian Apennines, where Gorgoglione is situated, are highly susceptible to geomorphological instability due to the interplay of lithology, relief morphology, active tectonics, seismicity, climate, and vegetation. In recent decades, land abandonment around small towns and villages has exacerbated soil erosion and increased landslide occurrence. These challenges are further compounded by inadequate urban planning, poor construction practices, and ineffective water and wastewater management, along with a lack of sufficient landslide mitigation measures. Unlike regions experiencing rapid urbanization, these areas face issues tied to unregulated urban decline, making them critical test beds for developing innovative methods to study and mitigate natural processes that heighten urban risks.

The research aims to provide insights into residual landslide risks to support the development of effective mitigation and management strategies. The activity of instability processes was analyzed using SAR interferometry from both satellite and ground-based platforms. Subsurface geological and lithostratigraphic characteristics were reconstructed by integrating geological and geomorphological information with geophysical techniques, including Electrical Resistivity Tomography (ERT) and Horizontal-to-Vertical Spectral Ratio (HVSR) analysis of ambient seismic noise. The pronounced directional patterns observed in HVSR analysis are being investigated to determine their potential correlation with the landslide movement direction identified through SAR interferometric data.

How to cite: Calamita, G., Perrone, A., Falabella, F., Pepe, A., Stabile, T. A., Gallipoli, M. R., Serlenga, V., Gueguen, E., Bellanova, J., Bentivenga, M., and Piscitelli, S.: Integrating Remote Sensing and Geophysics to Assess Landslide Risk in the Italian Apennines: A Case Study in Gorgoglione, Italy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6485, https://doi.org/10.5194/egusphere-egu25-6485, 2025.

EGU25-8772 | ECS | Orals | NH3.3

Distributed Seismic Sensing of Debris Flow Dynamics at Illgraben, Switzerland 

Christoph Wetter, Fabian Walter, Brian McArdell, Zhen Zhang, Johannes Aichele, and Andreas Fichtner

Recent years have shown the destructive nature associated with debris flows in alpine regions, including densely inhabited regions in Central Europe. Surge fronts within debris flows increase peak discharge and the dynamical complexity, which contributes much to the hazard potential. In recent years, numerical models helped to gain insights into the surging behavior of debris flows, in particular into the formation of surging waves including roll waves and erosion-deposition waves (Edwards & Gray, 2014). In order to capture the dynamic processes involved in the formation and propagation of flow surges, it is necessary to obtain distributed observations in the spatio-temporal domain. However, demanding field installations have confined studies to theoretical or laboratory settings, and results have yet to be validated under large-scale, real-world conditions. In this study, we close this gap by utilizing distributed, near-torrent seismic measurements at the Illgraben debris flow observatory maintained by the Swiss Federal Institute of Forest, Snow and Landscape Research WSL.

In 2024, a chain of 33 seismic nodes was deployed along a 2-kilometer section of the Illgraben torrent, with a spacing of 70 m between each node. In total, 10 debris flows with front velocities varying between 0.2 and 6 m/s and maximum flow heights varying between 1 - 3 m were recorded. The nodal array detected debris flow signals up to 2 km away, at a stage when the flows were still mobilizing in Illgraben’s upper catchment. The seismic record is characterized by high-frequency signals commonly attributed to particle-ground impacts within the debris flow. Additionally, it is found that steps in torrent geometry (check dams) produce a strong, low-frequency (1 – 10 Hz) background signal that is detectable kilometers away from the torrent.

Our measurements provide novel data of the spatio-temporal evolution of debris flows: Bifurcations of surge fronts and spawning of erosion-deposition waves can be observed and traced along the torrent. The data furthermore reveal the interaction between surging waves and the debris flow front. Our dense seismic recordings thus show how and where surging waves develop and how they modify maximum discharge and thus allow inferring the debris flows destructive potential.

The distributed seismic measurements at Illgraben offer new perspectives on measuring flow instabilities such as surge fronts and roll waves, allowing us to track them along extended torrent sections. They furthermore enabled us to refine our understanding of the seismogenesis of torrential processes, which is often only investigated with single stations or sparse networks. In a next step we plan to use these findings to better represent pulsing behavior in numerical debris flow models.  

Edwards & Gray, 2014, J. Fluid Mech., doi:10.1017/jfm.2014.643

How to cite: Wetter, C., Walter, F., McArdell, B., Zhang, Z., Aichele, J., and Fichtner, A.: Distributed Seismic Sensing of Debris Flow Dynamics at Illgraben, Switzerland, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8772, https://doi.org/10.5194/egusphere-egu25-8772, 2025.

EGU25-8864 | Orals | NH3.3

Advantages of applying the Multichannel Analysis of Surface Waves (MASW) in ice-rich rock glacier environments: A case study 

Mirko Pavoni, Ilaria Barone, Jacopo Boaga, Steven Javier Gaona Torres, and Alexander Bast

Rock glaciers are typical landforms of mountain permafrost, composed of a surface of boulders and debris insulating an ice-bearing sediment layer, overlying a glacial till deposit and/or bedrock. It is well-known and documented that the degradation of mountain permafrost is influencing the triggering of slope mass movements (e.g. rock falls, debris flows, and floods), and the stability of infrastructures (e.g. ski resorts). Consequently, a reliable characterization of rock glacier’s structure is a key aspect for evaluating the risk related to their presence.

Since boreholes are challenging and expensive to realize in high mountain environments, geophysical methods are widely used to characterize the internal structure of rock glaciers. Electrical Resistivity Tomography (ERT) and Seismic Refraction Tomography (SRT) are among the most applied techniques to retrieve the electrical properties and the compressive wave velocities (Vp) in the subsurface.

In this work, we propose the application of the Multichannel Analysis of Surface Waves (MASW) to complement the information brought by ERT and SRT, and to overcome some limitations of the SRT method. For this purpose, the seismic data should be collected with low-frequency geophones (i.e., with 4.5 Hz natural frequency). The main advantage of the MASW approach is the possibility of obtaining shear wave velocity (Vs) profiles and to reveal velocity inversions in the subsurface, i.e., a lower velocity layer between two higher velocity layers (e.g., the unfrozen till deposit between the ice-bearing layer and bedrock). Furthermore, Vs are insensitive to the liquid phase in the medium, therefore MASW approach could be used to detect the ice-rich layer when it is surmounted by a water-saturated sediment layer (supra-permafrost flow), that could prevent P-waves from penetrating deeper.

In this work, we successfully tested the MASW method at the Flüela rock glacier (Engadine, Switzerland). ERT results clearly suggest the presence of an ice-rich layer, but the SRT analysis surprisingly does not show P-wave velocities consistent with this interpretation. The Vp model reveals in fact the typical values of liquid water. On the other hand, the Vs profiles retrieved from the MASW approach are in very good agreement with the ERT outcomes. Therefore, we hypothesise the presence of a thin water-saturated sediment layer on the top of the ice-rich layer, that would prevent P-waves penetration. In order to support our hypothesis, we performed a seismic full-wave forward modelling: the synthetic shot gathers are consistent with the real ones, both in terms of surface wave dispersion and P-wave first-arrival times.

How to cite: Pavoni, M., Barone, I., Boaga, J., Gaona Torres, S. J., and Bast, A.: Advantages of applying the Multichannel Analysis of Surface Waves (MASW) in ice-rich rock glacier environments: A case study, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-8864, https://doi.org/10.5194/egusphere-egu25-8864, 2025.

EGU25-9189 | ECS | Orals | NH3.3

Using a seismic network for automatic detection, localization and characterization of mass movements in the Mont-Blanc massif  

Jakub Kokowski, Agnès Helmstetter, Eric Larose, Ludovic Ravanel, and Xavier Cailhol

In the Mont-Blanc massif (western European Alps), seismic stations record numerous signals originating from surface mass movements, such as rockslides, rockfalls, and serac avalanches. The large number of recorded signals makes the automation of the processing workflow essential for practical application. These seismic waveforms differ significantly from those generated by earthquakes, making standard algorithms unsuitable for their analysis. The signals typically exhibit an emergent onset, making it challenging to precisely determine their start time. Moreover, the arrival times of P and S waves, routinely used for earthquake localization, cannot be easily identified. The seismic records also vary in length, reflecting the differing durations of the associated phenomena.

To analyze such data using a seismic network, we adapted selected algorithms to address these challenges. For detection, we chose the STA/LTA algorithm, and for localization, we used amplitude decay algorithm and BackTrackBB software, which exploits wave field coherence. To test these algorithms, we created a reference dataset consisting of large, well-documented mass movements. The dataset was developed using regional mass movement databases, webcam image analysis, direct observations made by a network of observers, and seismic data from the Sismalp network. This reference dataset enabled us to fine-tune the algorithms and automate the processing of waveforms related to mass movements.

How to cite: Kokowski, J., Helmstetter, A., Larose, E., Ravanel, L., and Cailhol, X.: Using a seismic network for automatic detection, localization and characterization of mass movements in the Mont-Blanc massif , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9189, https://doi.org/10.5194/egusphere-egu25-9189, 2025.

EGU25-11136 | Orals | NH3.3 | Highlight

Hydrogeophysical investigation of clay-rich landslides through combined electrical and electromagnetic methods 

Adrian Flores Orozco, Anna Hettegger, and Clemens Moser

Landslides are complex systems, which are commonly investigated using data from punctual sensors, e.g., installed in boreholes. Assuming lateral variations in the subsurface fabric by interpolating the data from sparse boreholes may provide biased insight into the processes and architecture of landslides. Geophysical methods can be used to overcome this issue, gaining information about the physical properties (e.g., electrical conductivity, seismic velocity) of the subsurface covering large areas with high resolution. In landslides, geophysical methods have been used to investigate the geometry of the geological units and the depth to the bedrock, of the position of sliding planes and to compute the volume of mobilized material. Moreover, recent studies have demonstrated the ability of geophysical methods to quantify variations in the hydrogeological properties in an imaging framework. While the use of borehole data helps to reduce ambiguities in the interpretation of the geophysical images, the combination of more of different geophysical methods allows to enhance the coverage and resolution of the investigation as well as to reduce modeling uncertainties.

In this contribution, we present the combination of electrical and electromagnetic methods for the hydrogeological characterization of a clay-rich landslide located in Upper Austria (Austria). The investigation considers a two-step approach: (1) mapping at the large scale using electromagnetic methods at low induction number, and (2) selection of particular areas for the conduction of spectral induced polarization (SIP) transects. The first step aims resolving the main variations of clay content as well as to identify preferential flow paths for near-surface run-off; while in the second step SIP measurements are used to quantify hydraulic conductivity and water content. EMI mapping was conducted using vertical and horizontal configurations with two different instruments, each one consisting of three receivers, resulting in mapping information along 12 different geometries reaching a maximal nominal depth of investigation of 7 m. SIP measurements were collected at 12 different frequencies in the range between 0.25 and 225 Hz using 64 electrodes in each transect, with a spacing of 2.5 m to reach a depth of investigation of ca. 50 m.

Maps of the electrical conductivity gained by EMI measurements reveal strong lateral variations in clay content across the entire site. The inversion of the SIP data permits to quantify vertical and lateral changes in the hydraulic conductivity and water content along the transects. Our results demonstrate that an adequate processing of the data and the use of cascade inversion of multi-frequency SIP data permit to resolve for consistent hydraulic properties using different petrophysical approaches. Inversion of the EMI data along the SIP profiles reveals consistent results in the variations of electrical conductivity, permitting to validate the SIP results in shallow areas. Additionally, we investigate the relationship between electrical and hydraulic conductivity along the SIP transects and use it for a quantitative interpretation of the EMI maps; thus, permitting a hydrogeological investigation of the entire study area.  Our results reveal the potential of combining EMI and SIP for quantitative investigations of landslides.

How to cite: Flores Orozco, A., Hettegger, A., and Moser, C.: Hydrogeophysical investigation of clay-rich landslides through combined electrical and electromagnetic methods, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11136, https://doi.org/10.5194/egusphere-egu25-11136, 2025.

EGU25-12267 | ECS | Posters on site | NH3.3

Similarity and Diversity of Debris Flow Footprints in Seismic Records 

Qi Zhou, Hui Tang, Michael Dietze, Fabian Walter, Dongri Song, Yan Yan, Shuai Li, and Jens Turowski

The ability of seismic instruments to monitor catastrophic channelized flows (e.g., bedload transport, debris flows, glacial lake outburst floods, and lahars) is becoming of interest to scientists and practitioners. However, using debris flows as an example, the variability in catchment geology, event properties, and seismic instrument configurations complicates the development of event detectors that can be transferred between sites without major adjustments of parameters and thresholds.

In this work, we built a global debris flow seismic data catalog comprising more than seventy events from three regions (Europe, China, and the USA). The collected events from nine catchments represent rainfall-triggered debris flows originating from diverse environmental contexts, such as post-fire catchments, post-earthquake catchments, and high-erosion catchments. We analyzed the similarities and differences among these events using dimensionless amplitude damping fitting. Furthermore, we evaluated the performance of a pre-trained machine learning detector applied to our event catalog to assess the feasibility of a generalized early warning approach. Our results will reveal the key signatures of debris flow footprints in seismic records within complex areas, which will guide the design of next-generation event detectors and warning systems. At the same time, the differences will guide us to customize the warning thresholds based on local site conditions and stakeholder interests. This study thus provides a foundation for affordable, seismic-data-driven early warning systems for debris flows and other channelized flows.

How to cite: Zhou, Q., Tang, H., Dietze, M., Walter, F., Song, D., Yan, Y., Li, S., and Turowski, J.: Similarity and Diversity of Debris Flow Footprints in Seismic Records, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12267, https://doi.org/10.5194/egusphere-egu25-12267, 2025.

EGU25-13168 | Orals | NH3.3

Multi-instrumental insights into the dynamics of an active rockslide near Spitze Stei, Switzerland 

Małgorzata Chmiel, Fabian Walter, Lena Husmann, Giacomo Belli, Clément Hibert, Nils Hählen, and Christian Kienholz

In recent years, the rockslide near Spitze Stei (Kandersteg, Switzerland) has shown elevated displacement rates exceeding 10 cm per day, indicating a growing instability of 20 million m3. This increased activity triggers frequent mass movements, including rockfalls, gravel flows, and debris avalanches, which elevates the potential for major events with secondary consequences such as debris flows and flooding.

To mitigate the risks associated with the Spitze Stei rockslide, extensive monitoring has been in place since 2018, including borehole measurements of temperature and water pressure and surface displacement observations. These measurements underline the presence of degrading permafrost and planes of enhanced gliding and shear deformation. However, the limited spatial coverage of these methods makes it challenging to understand slope-wide subsurface processes, which are crucial for characterizing instability and identifying mass movement triggers, especially in complex, highly active rockslides with multiple rock compartments.

Our study addresses these challenges through a passive seismic experiment to quantify mass movement activity and investigate subsurface processes at Spitze Stei. In this talk, I will discuss the two main research questions that motivate our study:

  • Are there correlations between meteorological factors and rock slope stability that reflect climate-induced changes? How can they be quantified?
  • Can seismology constrain subsurface processes, such as freeze-thaw cycles, water pressure variations, and progressive damage that affect rock slop stablitity? How these processes impact the dynamics of the rock slope?

To address these questions at the Spitze Stei rockslide, we develop a machine learning approach combining seismic and infrasound data to monitor rock falls, avalanches, and possibly stick-slip tremors reflecting frictional sliding within the slope. Furthermore, we use interferometric seismic noise analysis to detect small changes in elastic properties within the rock slope, which may be related to stability changes and permafrost degradation.

The rich ancillary data acquired at Spitze Stei offers a unique opportunity to validate our seismic methods against independent measurements and refine the interpretation of our results. Such analysis enhances warning efforts, deepens our understanding of triggering factors and their thresholds, and establishes a foundation for continuous seismic monitoring of rockslide dynamics in the context of climate change.

How to cite: Chmiel, M., Walter, F., Husmann, L., Belli, G., Hibert, C., Hählen, N., and Kienholz, C.: Multi-instrumental insights into the dynamics of an active rockslide near Spitze Stei, Switzerland, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13168, https://doi.org/10.5194/egusphere-egu25-13168, 2025.

EGU25-13632 | ECS | Posters on site | NH3.3

Automated seismic detection of surficial mass movements for volcano monitoring: the Stromboli case study 

Gaia Zanella, Sergio Gammaldi, Massimo Orazi, Walter De Cesare, Antonietta Esposito, Rosario Peluso, and Dario Delle Donne

Gravitational instabilities on active volcanic islands present a significant tsunami hazard, with waves capable of travelling vast distances and impacting far-off coastlines. A notable example is the tsunami triggered by Anak Krakatau's activity in 2018, along with earlier events that affected Montserrat in 1997 and 2003 and Rabaul in 1994. However, monitoring gravitational mass movements in volcanic settings remains challenging due to limited data and the complex dynamics of volcano-landslide interactions. This hampers accurately identifying some landslide key source parameters such as the path location, run-out distances, flow velocity, and mobilized volumes.

Stromboli, an active volcano in the Tyrrhenian Sea, frequently experiences various types of surficial mass movements—such as rockfalls, debris avalanches, and pyroclastic flows—along its Northwest flank, known as the Sciara del Fuoco. These events are closely monitored due to their tsunami-generating potential, as demonstrated during the 2002 eruption when two landslides produced ~2m high waves along the coast. Landslide activity at Stromboli is often linked to volcanic phenomena, such as effusive eruptions and paroxysmal explosions.

Here we used seismic data from the Stromboli monitoring network to investigate patterns of landslide activity along the Sciara del Fuoco and their relationship with the persistent Strombolian activity. The primary objective is to develop near-real-time automatic algorithms aimed at retrieving some landslide key parameters, such as duration, run-out distances, path location, flow velocity, and rate of occurrence. Monitoring these parameters provides valuable insights into ongoing volcanic processes and can help identify early warning signals for potential tsunami triggering.

The study focused on the year 2020, a period marked by varying volcanic activity levels. Automatic landslide detections were validated by manual inspection of seismic record. A total of 457 landslide events, with an average duration of ~200 seconds, were automatically detected and analyzed during the study period. The daily landslide event rate was ranging from 1 to 17 events per day. These findings are vital for improving volcano monitoring at Stromboli volcano as the developed automatic algorithm can be incorporated into the real-time monitoring systems, improving early warnings of volcanic eruptions and tsunamis.

How to cite: Zanella, G., Gammaldi, S., Orazi, M., De Cesare, W., Esposito, A., Peluso, R., and Delle Donne, D.: Automated seismic detection of surficial mass movements for volcano monitoring: the Stromboli case study, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13632, https://doi.org/10.5194/egusphere-egu25-13632, 2025.

EGU25-14583 | Posters on site | NH3.3

Distribution Characteristics of Submarine Landslides in the Ulleung Basin 

Gwang-Soo Lee, Roger Urgeles, Dong-Geun Yoo, and Seung-Won Jeong

The Ulleung Basin, located in the East Sea, exhibits numerous submarine landslides along its southern and western slopes, as previously documented. Recent seismic events in the southeastern region of the Korean Peninsula have raised concerns about the potential for additional submarine landslides in the slope of Ulleung Basin. To assess this risk, this study analyzes high-resolution bathymetric data and seismic profiles to establish classification criteria for submarine landslides and identify their distribution patterns. A GIS-based database was developed to catalog the identified features. According to the database, a total of 82 scarps, which represent distinct morphological displacements caused by submarine landslides, were identified. Additionally, 74 deposits, formed by the accumulation of displaced sediment at the base of slopes and within the basin, were mapped. Deposits often overlap in some areas of the basin, making it challenging to delineate their boundaries compared to scarps. The most prominent headscarps, corresponding to steep slope areas, are concentrated at depths of approximately 900 m, with slope angles ranging from 5° to 8°. The average area of identified deposits is approximately 500 km². The study also detected potential scarps that indicate a risk of future submarine landslides. Two of these scarps, which are continuous features, span widths of approximately 13 km and 25 km, respectively. Comparative analysis of seismic profiles and physical property data from deep drill cores obtained in 2019 revealed significant contrasts in sediment properties along the glide planes of existing submarine landslides. These findings suggest that changes in physical properties at the glide plane may play a crucial role in the initiation of submarine landslides in the Ulleung Basin.

How to cite: Lee, G.-S., Urgeles, R., Yoo, D.-G., and Jeong, S.-W.: Distribution Characteristics of Submarine Landslides in the Ulleung Basin, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14583, https://doi.org/10.5194/egusphere-egu25-14583, 2025.

EGU25-16509 | Posters on site | NH3.3

Geophysical prospections using 2D-ERT coupled with EMI survey for the spatial variability assessment of landslide related pedological discontinuities in the Turbolo basin (Calabria Region, Italy).  

Massimo Conforti, Luigi Borrelli, Elena Ceravolo, Gino Cofone, Fabio Ietto, Francesco Perri, Pasquale Ruocco, Fabio Scarciglia, Fabio Terribile, and Simona Vingiani

In the framework of the ongoing project “SOIL SHADES – SOIL features and pedogenic processes as predisposing factors of SHAllow landsliDES”, funded by Next Generation EU, National Recovery and Resilience Plan (PNRR) of Italy, M4.C2.1.1., National Research Programme (PNR)–Research Projects of Significant National Interest (PRIN), an integrated multi-scale and multi-analytical approach was applied in the Turbolo Stream catchment, in northwestern Calabria region (southern Italy). Due to its peculiar geological-geomorphological and pedological characteristics, this basin has been selected as pilot study area representative of several Mediterranean environments. It is about 30 km2 wide, elevation ranges between 75 and 1015 m asl and displays a dendritic pattern in mountainous sub-basins along with a trellis-like pattern in hilly reaches. Paleozoic metamorphic rocks (gneiss, phyllites, schists, metabasites interbedded with metapelites and metalimestone) outcrop in the western sector, while Miocene to Pleistocene deposits (clay, sand and conglomerate) in the eastern part, and Holocene sediments in the valley floor. The western sector is dominated by high relief and steep slopes dissected by deep V-shaped valleys, whereas the eastern hilly reaches are characterised by gentler slopes, fluvial terraces and broad valleys. The study area is recurrently affected by rainfall-triggered landslides damaging agricultural land, infrastructure and settlements. Geophysical prospections using 2D-electrical resistivity tomography (ERT) have been combined with electro-magnetic induction (EMI) surveys for the identification of possible shallow sliding surfaces, due to the effectiveness of both techniques in the detection of geological and pedological discontinuities in terms of particle size distribution, mineralogy, porosity, water content, solute concentration, etc. To support the geophysical data, several field observations were conducted along the landslide area. The most representative soil profile was selected at about 130 m asl on the southern slope of a Pleistocene fluvial terrace, in the eastern hilly reaches of the basin. A very deep soil profile (approximately 3 m of depth) was described on the scarp of a rotational slide that developed for some tens of meters downslope. Soils appear moderately to deeply weathered and have a matrix color ranging from reddish to yellowish brown with red and grey mottles indicating the persistence of stagnic conditions during the rainy season. Evidence of clay illuviation processes (i.e., clay coatings) is found in both the topsoil and the bottom soil, very likely due to alternating phases of slope stability and surface soil erosion. The soil texture varies from sandy loam to clay loam with relevant changes in the amount of subrounded to subangular coarse fragments. The soil reaction is from slightly acid to neutral, consistently with the absence of carbonates and the illuviation process evidence. Results of the geophysical surveys displayed some changes in the measured parameters in the surface layers, which are consistent with the depth of the landslide scarp and of the soil profile, as well as of the potential depth of the failure surface.

How to cite: Conforti, M., Borrelli, L., Ceravolo, E., Cofone, G., Ietto, F., Perri, F., Ruocco, P., Scarciglia, F., Terribile, F., and Vingiani, S.: Geophysical prospections using 2D-ERT coupled with EMI survey for the spatial variability assessment of landslide related pedological discontinuities in the Turbolo basin (Calabria Region, Italy). , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16509, https://doi.org/10.5194/egusphere-egu25-16509, 2025.

EGU25-16650 | ECS | Posters on site | NH3.3

Experimental Investigation of Particle Impacts Using Distributed Acoustic Sensing 

Zheng Chen and Siming He

Geophysical granular flows, such as rock avalanches, debris flows, and bedload transport, generate intense impact forces on the channel bed during downslope movement. These forces produce high-frequency seismic and acoustic waves, which can be detected by seismometers and acoustic sensors. The resulting vibration signals provide valuable insights into flow characteristics; however, quantitatively measuring granular flow processes remains challenging due to the complex mechanisms of particle impacts and the variability in particle locations, motion modes, and impact velocities. Distributed Acoustic Sensing (DAS) offers a promising approach for monitoring such granular flows, leveraging its ability to provide high-resolution, real-time spatial and temporal data across extensive areas. In this study, as a pre-experimental test, particle drop experiments were conducted using spherical objects (5 kg) with varying impact locations and drop heights to investigate the dynamic signal response of a DAS system deployed laterally over 50 m. The DAS system operated with a sampling frequency of 1000 Hz and a spatial resolution of 0.4 m. For each particle impact, key parameters including the number of signal impulses, amplitude, centroid frequency, and power spectral density (PSD) were extracted from the raw DAS data. Virtual shot gathers were analyzed and utilized for wave speed analysis, while beamforming techniques were applied to locate particle impact events spatially. The experimental results demonstrated how signal impulses, amplitudes, and PSDs vary with changes in particle size and impact location. These findings highlight the potential of DAS for monitoring granular flow processes, such as bedload transport, in natural settings.

How to cite: Chen, Z. and He, S.: Experimental Investigation of Particle Impacts Using Distributed Acoustic Sensing, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16650, https://doi.org/10.5194/egusphere-egu25-16650, 2025.

EGU25-17256 | Orals | NH3.3

Electrical Resistivity Tomography (ERT) Moisture Monitoring on the Müsch Landslide (Ahr valley, Germany) 

Rainer Bell, Anna Schoch-Baumann, Michael Dietze, and Lothar Schrott

The extreme Ahr flood 2021 caused 135 fatalities (and one still missing), severe damage and enormous geomorphological changes of the riverbanks, floodplains and adjacent slopes. Many slopes were undercut and several landslides have been reactivated. The Müsch landslide is located in a narrow section of the upper Ahr valley. The instability is 100 m wide, 200 m long, and of unknown age. Approximately 7000 m³ of the landslide toe were eroded by the 2021 flood. After the flood, the landslide was reactivated, resulting in minor changes on the surface (e.g. opening of cracks). A major reactivation of the entire landslide body, however, might potentially lead to a landslide dammed lake inundating buildings upstream. Thus, there is the need to better understand the landslide structure and behavior.

Since water saturation plays a crucial role in landslide activities, an electrical resistivity tomography (ERT) moisture monitoring system has been set up in January 2024 along one longitudinal and one cross profile (both 200m). We use permanently installed steel electrodes with a spacing of 2.5 m for both profiles. Monthly repeated manual ERT measurements (array: gradient) are analyzed with time-lapse inversions.

ERT results show an increasing reduction in resistivity values until June 2024 down to about 10-15 m along both ERT profiles correlating with increasing water saturation in the landslide body. The opening and widening of cracks indicate accelerating landslide activity from April onwards and continuing until July 2024 when the topsoil had started to dry out while the deeper layers were still sufficiently wet. Subsequently, landslide activity slowed down. This is in line with precipitation records and modelled soil moisture distribution over 2 m soil profiles by the German Weather Forecast (DWD) and observations made in 2023, in which similar dynamics occurred.

Continued measurements and analyses will enable us to better assess water saturation of the landslide and its spatial heterogeneity. Results will be correlated to rainfall data, on site measured soil moisture data (10 and 40 cm depth) as well as data of a passive seismic monitoring of the landslide, which is in place since October 2021. Deep drillings are scheduled for early 2025, with inclinometers and piezometers subsequently installed on behalf of the State Geological Survey. A combination of those measurements will help to better understand landslide behavior and assess potential hazards and risks.

How to cite: Bell, R., Schoch-Baumann, A., Dietze, M., and Schrott, L.: Electrical Resistivity Tomography (ERT) Moisture Monitoring on the Müsch Landslide (Ahr valley, Germany), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17256, https://doi.org/10.5194/egusphere-egu25-17256, 2025.

EGU25-17452 | Orals | NH3.3

Seismic noise measurements for the characterisation of Pays de Herve landslides 

Veronica Pazzi, Agnese Innocenti, Anne-Sophie Mreyen, Lèna Cauchie, David Caterina, Yawar Hussain, Valmy Dorival, and Hans-Balder Havenith

The Pays de Herve, located in the eastern part of Belgium, can be characterized as a multiple sections tableland with gentle slopes of less than 15°. It is located in the vicinity of the northern section of the Hockai Fault Zone, a 42 km-long seismogenic fault zone, that is characterized by the presence of fault scarps, multiple dissection elements and the presence of more than 20 paleo-landslides. Among these latter, the Manaihan landslide is the most studied and monitored landslide in the area. From a geological point of view, it developed in a Upper Cretaceous sedimentary setting, i.e., Vaals Clays overlaying Aachen sands. Even today, the slope is affected by instability and subsidence phenomena, likely linked to anthropogenic loading combined with prolonged periods of rainfall and possibly historic seismic events leading to liquefaction in the Aachen sands.

Recently, new geophysical surveys have been carried out using an integrated approach, combining electrical resistivity measurements, active seismic methods (interpreted as P-wave tomography and MASW), and passive seismic techniques (single-station H/V). The key question to address is: How deep is the sliding surface, and is it possible to identify it?

The combination of these surveys allowed the identification of two main layers. The first layer has a variable thickness ranging between 5 m and 20 m. On the electrical resistivity tomography, it corresponds to a more conductive layer with values between 5 and 20 Ωm, and on the seismic tomography, it shows velocities between 500 and 1300 m/s. From the electrical and seismic tomographies, the second layer appears to be more resistive, with values ranging between 30 and 50 Ωm, and P-wave velocities exceed 1500 m/s. Based on the geological map and their physical properties, the identified layers have been attributed to the Vaals clay formation and the Aachen sands, respectively.

The H/V measurements were processed to produce sections showing the variation of H/V amplitude (or the log of H/V) with depth. If multiple H/V measurements can be aligned in a linear array and the surface layer can be assumed to be homogeneous, i.e., shear wave velocity increasing with depth, the H/V curves along the alignment can be modeled together to create a 2D section. Several H/V sections could be developed for the Manaihan landslide, revealing a similar pattern of contrasts between the before mentioned layers. The main contrast is located at a depth ranging from 15 m to 40 m and corresponds to the interfaces identified by ERT and SRT. This interface is present beyond the landslide and even outside of it, suggesting that it may be associated to the geological contact between the Vaals clays and the Aachen sands. The conductive layer identified in the ERTs can furthermore be associated to very low log H/V amplitudes in the upper range of the H/V sections until 10-20 m depth. The H/V amplitude analysis of all the identified sections suggests that the sliding surface of the Manaihan landslide is located at the contact between the clay and the sand layers.

How to cite: Pazzi, V., Innocenti, A., Mreyen, A.-S., Cauchie, L., Caterina, D., Hussain, Y., Dorival, V., and Havenith, H.-B.: Seismic noise measurements for the characterisation of Pays de Herve landslides, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17452, https://doi.org/10.5194/egusphere-egu25-17452, 2025.

EGU25-19568 | Posters on site | NH3.3

A near-real-time public mass movement catalogue for Switzerland 

Philipp Kastli, John Clinton, Toni Kraft, Tobias Diehl, and Florian Haslinger

Since mid 2023, the Swiss Seismological Service (SED) maintain a public list of mass movements occurring in and around the Swiss Alpine region.We include all events that can be detected and characterised by the national monitoring service, limiting the list to only larger events. The SED operate a seismic network that includes over 400 seismic stations in and around the Swiss territory. Sites in the alps include about 40 broadband sensors in low noise hard rock vault conditions, as well as strong motions stations in Alpine valleys.

The network is optimised to detect earthquakes, but due to the station density, we also detect mass movements. Within minutes to hours of their occurrence, seismologists review all automatic events and if a landslide source is suspected, events are indicated as such and immediately made available, with approximate location and magnitude, but precise information on origin time. This information is also shared with a wide community of scientists and civil authorities in the Swiss domain. If a mass movement is confirmed either via this expert group or through the media, the event is labelled as confirmed and the location is fixed. In this presentation we will present the catalogue and how it has evolved over time; describe how we detect and characterise events; and demonstrate the growing importance and profile of this valuable new information resource.

How to cite: Kastli, P., Clinton, J., Kraft, T., Diehl, T., and Haslinger, F.: A near-real-time public mass movement catalogue for Switzerland, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19568, https://doi.org/10.5194/egusphere-egu25-19568, 2025.

EGU25-19906 | Posters on site | NH3.3

Shear strength and slip rate dependence of weathered volcanic ash soil controlled by water adsorption ability 

Ryousei Omori, Miki Takahashi, and Shinichi Uehara

Our research aims to clarify the process of slip acceleration in landslides in order to mitigate landslide disasters. Here we especially focus on what factors control the shear strength of rocks and soils that compose the landslide slip zone and what factors generate variety in sliding features. This is because knowing those factors could provide hints for predicting the onset of runaway slip. Although a method to predict the onset of slope failure has been proposed (Fukuzono, 1985), which is based on the inverse of surface slip rate converging to zero as the failure time approaches, it doesn’t always work. There have been reports that the slip rate turned to decrease, and the slide did not induce to the failure, even after obtaining enough slip acceleration (Matsuura et al., 2015; Doi et al., 2020).

We here bring the concept of rate-dependent shear strength, which has been developed in seismology and is related to fault slip stability (e.g., Dieterich, 1979). Whether the slip exhibits further acceleration or deceleration depends on whether the shear strength of the shear zone material shows negative rate-dependence or positive rate-dependence. The former is called velocity-weakening, and the latter is called velocity-strengthening, respectively. Thus, such materials could cause the sliding feature that turns to deceleration during slip acceleration, meaning the slip velocity will have an upper limit value. In this study, the concept of rate-dependent shear strength was applied to describe the sliding properties of clay-rich soils as simulants the landslide. Moreover, the clay-rich soils are naturally thought to be one of the causes of slope failure because of their low frictional property (Bromhead, 2013; Schulz and Wang 2014). We conducted the shear experiments on clay-rich soils to measure the shear strength and rate-dependence. Additionally, we measured various properties of the soils, such as mineral composition and content, liquid limit (WL), plasticity index (PI) and specific surface area (SSA), at the viewpoint what determine the lowness of strength and the variety of rate-dependence.

The samples we used were collected from eight locations in the landslide-prone area in western part of the Aso Caldera, Kumamoto Prefecture. The rotary shear experimental apparatus we used was set at Geological Survey of Japan, AIST (Togo and Shimamoto 2012). We varied the slip velocity from 10⁻⁴ to 10 mm/min (10⁻³ - 10² μm/s) that provided the rate-dependent shear strength functioned by the velocity. The samples were saturated in water (drainage) at room-temperature, and the normal stress was set at approximately 1 MPa

Samples with larger SSA showed the trend of negative rate-dependence at lower velocities (< 1 mm/min) but positive rate-dependence at higher velocities (> 1 mm/min), indicating they have the potential to suppress the acceleration at around 1 mm/min. On the other hand, the rate-dependence was always negative for the samples with small SSA, meaning they have the potential of runaway slip generation. Thus, it can be said that how much SSA is in the slip zone material might constrain the variety in slip at landslide.

How to cite: Omori, R., Takahashi, M., and Uehara, S.: Shear strength and slip rate dependence of weathered volcanic ash soil controlled by water adsorption ability, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19906, https://doi.org/10.5194/egusphere-egu25-19906, 2025.

EGU25-400 | PICO | HS1.2.1

From innovative sensors to steady data streams: The TEMBO Africa project 

Nick van de Giesen, Frank Annor, Sylvester Nsobire Ayambila, Kwame Duah, Tomáš Fico, Andrea Gatti, Olivier Hoes, Gordana Kranjac-Berisavljevic, Salvador Peña-Haro, Eugenio Realini, Hubert Samboko, and Hessel Winsemius

TEMBO Africa is a Horizon Europe project that seeks to improve in situ sensing of weather and water in sub-Saharan Africa. To ensure beyond-the-project sustainability, we are using innovative sensors to measure variables such as rainfall, bathymetry, river flow, and large-scale soil moisture. TEMBO also develops services for hydropower, agriculture, and disaster management. These services will produce societal and economical value, for which governments and companies are willing to pay. These payments, in turn, serve to maintain the observation networks. One guiding principle is that the new data gathering method should cost less than 10% of existing methods in term of total costs of ownership. This principle implicitly pays special attention to the local availability of human resources. Many monitoring projects in Africa consist of installation by experts from the Global North, followed by a short training of local technicians. This works nicely until something breaks down. In TEMBO, African universities and spin-off companies are co-developing the technologies such that any operational problems can be solved without flying in expensive foreign experts.  

In this presentation, we will go through the sensor innovations and how these feed into different products and services.

 

How to cite: van de Giesen, N., Annor, F., Ayambila, S. N., Duah, K., Fico, T., Gatti, A., Hoes, O., Kranjac-Berisavljevic, G., Peña-Haro, S., Realini, E., Samboko, H., and Winsemius, H.: From innovative sensors to steady data streams: The TEMBO Africa project, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-400, https://doi.org/10.5194/egusphere-egu25-400, 2025.

EGU25-1828 | ECS | PICO | HS1.2.1

Hydraulic conductivity estimation in Porous Media: Insights from Neural computing 

Abhishish Chandel and Vijay Shankar

Precise estimation of hydraulic conductivity (K) in porous media is vital for advancing hydrological and subsurface flow investigations. Groundwater experts have increasingly adopted neural computing approaches to indirectly determine K in porous media, offering a more efficient alternative to conventional methods. The research focuses on developing the Feed-Forward neural network (FFNN) and Kohonen Self-organizing maps (KSOM) models to compute the K using easily measurable porous media parameters i.e., grain-size, uniformity coefficient, and porosity. The observed data were split into 70% and 30% for the development and validation phase, respectively. The developed model's performance was examined via statistical indicators, including root mean square error (RMSE), determination coefficient (R²), and mean bias error (MBE). The findings suggest that the FFNN model significantly outperforms the KSOM model in estimating the K value, with the KSOM model achieving only moderate accuracy. During the validation phase, the FFNN model shows a stronger correlation with the measured values, yielding RMSE, R², and MBE values of 0.016, 0.94, and 0.006, while the KSOM model returns values of 0.024, 0.91, and -0.004 respectively. The FFNN model's superior predictive ability makes it a reliable tool for accurate K estimation in aquifers.

How to cite: Chandel, A. and Shankar, V.: Hydraulic conductivity estimation in Porous Media: Insights from Neural computing, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1828, https://doi.org/10.5194/egusphere-egu25-1828, 2025.

EGU25-1910 | ECS | PICO | HS1.2.1

Large-scale Soil Moisture Monitoring: A New Approach 

Ilektra Tsimpidi, Konstantinos Soulis, and George Nikolakopoulos

Soil moisture is a critical factor for understanding the interactions and feedback between the atmosphere and Earth's surface, particularly through energy and water cycles. It also plays a key role in land climate and hydrological processes. Recent advancements in autonomous robotic applications for precision agriculture have introduced significant solutions, particularly in remote sensing. Currently, these platforms enable autonomous soil parameter measurement and on-site data collection, which is essential for resource optimization and data-driven agricultural decision-making. However, challenges persist, especially in real-time soil moisture monitoring—a key focus for improving irrigation efficiency, water use, and crop yields. Soil moisture measurement in-situ techniques include the accurate oven-drying method and soil moisture sensors, while satellite remote sensing uses optical, thermal, and microwave imaging to estimate surface soil moisture from a broader perspective. However, fully autonomous robotised sampling procedures for optimising the process, increasing repeatability and overall accuracy, as well as increasing the reachability of the sampling of remote areas, are still not utilized.

Soil moisture measurement in-situ techniques include the accurate oven-drying method and soil moisture sensors, while satellite remote sensing uses optical, thermal, and microwave imaging to estimate surface soil moisture from a broader perspective. However, fully autonomous robotised sampling procedures for optimising the process, increasing repeatability and overall accuracy, as well as increasing the reachability of the sampling of remote areas, are still not utilized.

Measuring soil moisture presents a significant challenge due to its reliance on human labour, which is required to cover extensive areas for sensor measurements manually. Additionally, soil moisture measurements at a specific point vary with time and environmental conditions, making these values unstable. While satellites offer a potential solution to some of these issues, their accuracy is affected by environmental factors such as cloud cover and dense vegetation, while they only describe the upper soil layer. Moreover, ground measurements of surface soil moisture are still necessary for calibrating and training the satellite systems. To address these challenges, we propose an adaptable in situ method for automating soil moisture measurements.

Our approach introduces AgriOne, an autonomous soil moisture measurement robot leveraging a surface-aware data collection framework to achieve precise and efficient soil moisture assessments, thereby minimizing reliance on permanent sensors and reducing associated costs and labour. The hardware of AgriOne consists of a UGV Husky A200 from Clearpath Robotics loaded with the soil moisture sensor TEROS12 from Meter Group. The sensor is mounted on a linear actuator probe, as shown in the figure.  

To evaluate the proposed approach, we conducted two field experiments in different locations under different weather and soil conditions. The experiments were successful in both cases, and we collected 70 and 66 measurements, respectively, of surface soil moisture. For the first experiment, we covered an area of 380m2 in 57 minutes, and for the second experiment, we covered an area of 800m2 in 2,5 hours. The results showed proof of concept because of the workability of the AgriOne robot and the reliability of the data collection framework. 

 

How to cite: Tsimpidi, I., Soulis, K., and Nikolakopoulos, G.: Large-scale Soil Moisture Monitoring: A New Approach, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1910, https://doi.org/10.5194/egusphere-egu25-1910, 2025.

EGU25-2786 | PICO | HS1.2.1 | Highlight

Bathymetric Survey and Underwater Structure Inspection for Hydraulic Engineering in Shallow Waters Using Unmanned Surface Vehicles 

XiaoQing Gan, Peng Wan, Jianzhou Li, and Bangning Ding

Bathymetric surveys and underwater structure inspections are critical for ensuring the safe operation of hydraulic engineering projects. Accurate data on topographical changes and structural conditions help mitigate operational risks caused by erosion, scouring, or structural deficiencies. However, traditional manned vessels face significant limitations in shallow and complex areas, such as downstream spillways, due to accessibility and maneuverability challenges.

The development of unmanned surface vehicles (USVs) offers an efficient and precise alternative for surveying and inspection in shallow water environments. This study utilized the Huawei-3 USV to conduct a bathymetric survey and underwater structure inspection in the shallow downstream area of the spillway at the Wangfuzhou Hydropower Station, Hubei Province, China.

The survey employed the Huawei-3 USV, equipped with high-precision echo sounders and RTK systems, to collect bathymetric and structural data. Water surface elevation data were acquired using RTK measurements, with water levels observed five times before and after the survey to establish a reference elevation. In areas less than 2 meters deep, RTK was also used to directly measure the bottom elevation. The USV combined its draft depth and transducer depth with RTK-derived water surface elevations to calculate the bottom elevation. Satellite imagery was used for pre-planning survey lines, which were aligned parallel to the downstream protective apron, spaced 5 meters apart, ensuring a point spacing of approximately 2 meters. In complex or nearshore areas, manual control was applied to densify survey lines. Data processing involved converting depth to elevation, noise filtering, and generating CAD and 3D models.

The results revealed significant scouring near the downstream protective apron, forming a scour pit with an area of 2,897.2 m², a minimum elevation of 70.26 m, and a proximity of 6.87 m to the reinforced apron edge. The overall underwater topography of the reinforced apron section closely matched the design, with a minimum measured elevation of 70.937 m, differing by only 6.3 cm from the designed elevation of 71 m, indicating stability. However, a portion of the 73 m design elevation zone showed scouring depths up to 25 cm, with an average depth of 12.5 cm. No significant deepening of scour was observed between 2022 and 2024.

The findings demonstrate that USV-based bathymetric systems are highly applicable in shallow water environments, achieving data accuracy that meets regulatory standards. These systems effectively identify scour pits and structural changes, providing reliable data support for ensuring the safe operation of hydraulic engineering projects. Moreover, the method shows significant potential for application in other shallow, complex water environments in the future.

How to cite: Gan, X., Wan, P., Li, J., and Ding, B.: Bathymetric Survey and Underwater Structure Inspection for Hydraulic Engineering in Shallow Waters Using Unmanned Surface Vehicles, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2786, https://doi.org/10.5194/egusphere-egu25-2786, 2025.

Evaporation from the water surface is among the main water losses from natural and artificial lakes and ponds. Air temperature (Ta), wind speed (va), relative humidity (RH), atmospheric pressure (pa), surface water temperature (Tw) and radiation (R) are among the physical controls of this process. In recent years, water temperature data have increasingly become available so that the question arises, if the measurement of radiation (which, in turn, affects water temperature) may still be required.

The method employed in this study is modelling of daily evaporation by means of artificial neural networks (ANNs) of the multilayer perceptron type (backpropagation, one hidden layer), using varying sets of input variables. Evaporation data from a white Class A pan (Qiu et al., 2022) served as target (50 patterns of daily averages). A logistic activation function was used. Data records were divided 2:1 into training and testing sets, resp.

Data were scaled to the interval between 0.1 and 0.9, and for each run (105 epochs) the root mean square error (RMSE) of the scaled output was computed.

Learning rate (η), momentum (α) and number of hidden nodes were subject to optimization for three different sets of input variables. ANN runs of series S1 comprised Ta, va, RH, pa, Tw and incoming solar radiation (R) as inputs (6 in total). Series S2 and S3 were subsets of S1, with S2 using Ta, va, RH, pa and Tw as inputs. For the input data of Series S3, water temperature Tw  was replaced by radiation R.

The neural networks achieved a fair representation of the evaporation data. Optimization yielded a minimum RMSE for Series S1 of 0.0514 and 0.0669 for training and testing, resp. (6 hidden nodes, η=0.009 and α=0.0). 

Using the same input variables with the exception of the incoming radiation (in total, therefore, 5 inputs) S2 reached a minimum training RMSE of 0.0557 and a minimum testing RMSE of 0.0887 (5 hidden nodes, η=0.012 and α=0.0).

Series S3 with the 5 inputs Ta, va, RH, pa and R (with water temperature left out), finally achieved an RMSE of 0.0545 for training and 0.0775 for testing, resp. (6 hidden nodes, η=0.006 and α=0.2).

Comparison of Series S2 and S3 shows that, in the case of the data set studied here, the ANNs including incoming radiation among their input variables (but excluding water temperature) outperformed those explicitly accounting for water temperature in lieu of radiation. Using both radiation and water temperature as inputs (S1) resulted in a notable improvement of the ANN output as compared to the runs with either of these variables not accounted for explicitly.

References

Qiu, G. Y., Gao, H., Yan, C., Wang, B., Luo, J., & Chen, Z. (2022): An improved approach for estimating pan evaporation using a new aerodynamic mechanism model. Water Resources Research, 58, e2020WR027870. https://doi.org/10.1029/2020WR027870.

How to cite: Schmid, B.: On the relative importance of water temperature versus radiation for ANN-based pan evaporation modelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2944, https://doi.org/10.5194/egusphere-egu25-2944, 2025.

The primary objective of groundwater analysis is to determine the direction and velocity of water flow, which are essential for effective groundwater resource management and contaminant investigation. Conventional methods of evaluating groundwater flow direction, such as using solute or thermal tracers, require the installation of multiple observation wells and are typically laborious, expensive, and time-consuming. Furthermore, the uneven distribution of observation wells and the heterogeneity of aquifers often lead to inaccurate estimations of groundwater flow velocity and direction. Accordingly, this study proposes a novel approach: the thermal vector distributed temperature sensor (TV-DTS) method, combined with a heated line source, to overcome these challenges. The TV-DTS apparatus consists of a single heated fiber and four sensing fibers. The heated fiber functions as the heat source, while the sensing fibers are used to measure temperature changes. These measurements are then used to determine the direction and velocity of water flow by the analytical solution derived from the heat transfer with a heated line source. This method employs only a single-well heating test to estimate both the direction and velocity of groundwater flow, eliminating the need for multiple wells and significantly reducing the time and financial resources. Besides, the TV-DTS has several advantages, such as the ability to provide continuous spatial-temporal temperature data, ensuring reliable and high-resolution monitoring. Two groundwater contamination sites in northern and southern Taiwan have be selected to demonstrate the effectiveness of TV-DTS. The preliminary results showed that at the northern site, the flow direction was predominantly northeast to southwest, with velocities ranging from 0.25 - 0.34 m/day at different depths. In contrast, at the southern site, the flow direction was mainly toward west with higher velocities of 0.1 – 8.0 m/day. The estimated directions and velocities from both sites aligned with previous studies; however, uncertainties were higher at the southern site due to greater velocities observed. This method provides a high-resolution, cost-effective approach for hydrogeological investigations and contaminated sites assessment, serving as a valuable reference for the future investigation and evaluation of hydrogeological characterization.

Keywords: groundwater flow direction, groundwater flow velocity, heat transfer, distributed temperature sensors, borehole, uncertainty, contamination

How to cite: Liu, C. H. and Chiu, Y. C.: Utilizing distributed temperature sensors in a single well with a heating line source to simultaneously estimate the direction and velocity of groundwater flow, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3004, https://doi.org/10.5194/egusphere-egu25-3004, 2025.

EGU25-3849 | PICO | HS1.2.1

Performance of the Image Wave Velocimetry Estimation for physics-based non-contact discharge measurement in rivers 

Salvador Peña-Haro, Giulio Dolcetti, and Hessel Winsemius

Herein we present an analysis of the performance of the Image Wave Velocimetry Estimation (IWaVE), a python library for image-based river discharge calculations.  IWaVE simultaneously performs a 2D velocimetry analysis and calculates the stream depth through 2D Fourier transform, exploiting the sensitivity of water wave dynamics to flow conditions. Unlike existing velocimetry approaches such as Particle Image Velocimetry (PIV), Particle Tracking Velocimetry (PTV) or Space-Time Image Velocimetry (STIV), the uniqueness of this approach lies in: 1) velocities that are advective of nature can be distinguished from other wave forms such as wind waves. This makes the approach particularly useful in estuaries or river stretches affected strongly by wind, or in shallow streams in the presence of standing waves. 2) The velocity is estimated based on the physical behavior of the water surface, accounting for the speed of propagation of waves and ripples relative to the main flow. This makes the approach more robust than traditional methods when there are no visible tracers. 3)  If the depth is not known, it can be estimated along with the optimization of x and y-directional velocity. Depth estimations are reliable only in fast and shallow flows, where wave dynamics are significantly affected by the finite depth.

We analyzed 2 videos recorded from a drone on a site in the Netherlands over a tidal channel in Zeeland at Waterdunen - Breskens. One of the videos has strong winds, which creates waves moving upstream. ADCP measurements for both videos are available. The videos were taken at different moments during different tidal conditions, they were processed using IWaVE, a LSPIV and a STIV methods. The results show that the LSPIV, STIV and IWaVE are in good agreement with the ADCP measurements for the case where there is no wind. However when there is wind the LSPIV and STIV methods fail to obtain the correct surface velocity, while the velocity calculated with IWaVE is in good accordance with the ADCP.

How to cite: Peña-Haro, S., Dolcetti, G., and Winsemius, H.: Performance of the Image Wave Velocimetry Estimation for physics-based non-contact discharge measurement in rivers, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-3849, https://doi.org/10.5194/egusphere-egu25-3849, 2025.

EGU25-4525 | ECS | PICO | HS1.2.1

Enhancing Urban Resilience to Surface Water Flooding: A Novel Approach Using UAS-Derived Topographic Indices 

Rakhee Ramachandran, Monica Rivas Casado, Yadira Bajon Fernandez, and Ian Truckell

With the increase in urbanisation and climate change around the globe, there is an increased risk of surface water flooding. Although extreme flood events are commonly discussed, smaller, more frequent flood events also cause significant disruptions that impact human life and put financial stress on authorities. The majority of urban flooding is due to drainage failure. For effective surface water management, it is important to assess the effectiveness of existing surface drainage assets and accordingly plan asset maintenance or retrofitting of new drainage assets (both traditional and nature-based solutions). The surface drains usually fail because they are either not positioned where the surface water accumulates or are blocked and not maintained to meet the standards. Microtopography significantly influences the surface water flow movement, flow path, flow direction, and velocity, and consequently, dictates the areas of water accumulation.  Thefore, this study explores a novel approach to evaluate storm drain inlet positions using high-resolution topographic indices maps derived from Unmanned Aerial System (UAS) imagery. The Topographic Wetness Index (TWI) and Topographic Control Index (TCI) were employed to identify drains misaligned with surface water pathways and pinpoint critical drains in the sink points of the topography, respectively. 


Storm drain inlets were classified as functional or non-functional based on their intersection with the flow path defined by the optimal TWI threshold. The optimal threshold was determined to be the 90th percentile at a value of 6.19 based on the spatial similarity of the delineated runoff-contributing flow path with the 1 in 100 year surface water flood map produced by the Environment Agency. The validation of the classification of storm drains effectiveness based on TWI using field data yielded an overall accuracy of 53 %, 75% precision, and an F1 score of 62%, indicating a moderate success of TWI in identifying functional drains. Although validation with LIDAR data showed a slight improvement in accuracy and precision, the results generally demonstrated that TWI has a strong capability to correctly identify functional drains; however, it is slightly more challenging to identify nonfunctional drains. 


A comparison of the UAS-derived TCI map with the LIDAR-derived TCI map demonstrated a 90% match in the identified sink areas and a high accuracy of 93% in identifying critical drains in the sink areas. The results suggest that the combined use of TWI and TCI offers a promising approach for assessing storm drain effectiveness, based on its position and guiding authorities in identifying areas with drainage deficits and preparing targeted drainage maintenance strategies. The findings of this research provide valuable insights for urban planners and decision-makers to not only optimise the placement and maintenance of storm drain inlets but also highlight the potential for alternative nature-based low-impact development (LID) solutions in locations where traditional drainage is found to be inefficient. This would ultimately enhance the resilience of urban areas to surface-water flooding.

How to cite: Ramachandran, R., Rivas Casado, M., Bajon Fernandez, Y., and Truckell, I.: Enhancing Urban Resilience to Surface Water Flooding: A Novel Approach Using UAS-Derived Topographic Indices, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-4525, https://doi.org/10.5194/egusphere-egu25-4525, 2025.

The ability to record high-resolution data for extended periods using affordable systems can improve the study of hydrological and environmental processes. Unlike commercial alternatives, publicly available open-source sensors can be implemented at a significantly lower cost, allowing higher spatiotemporal resolution and continuous, real-time monitoring. In this presentation, I will outline the fundamental principles, advantages, and challenges of using open-source, self-assembly hardware for hydrological related monitoring using two novel systems. The first system is an incubation chamber system composed of O₂, CO₂, CH₄ low-cost sensors for monitoring gas fluxes from sludge samples, specifically tested on wetland samples under different temperature, oxygen, and light conditions. The second system consists of a portable photoreactor/spectrophotometer driven by Raspberry Pi and Arduino UNO microcontrollers. Validation tests of the photoreactor system were performed in one preliminary design for Rhodamine B dye photodegradation, in which the spectral module was constituted by seven arrays of high-power LED of different wavelengths (UVC and VIS), bismuth ferrite (BiFeO₃) catalyst, and hydrogen peroxide. Results showed significant dye degradation (39.7%) at high chamber temperature (45 °C). The performance of this system is improved in a new design, which includes an exchangeable light module, sampling system, and a spectrophotometer for real-time monitoring of the photocatalytic process in water. Complete technical guides on design, assembly, and installation are provided for both systems, aiming to promote their reproducibility and application for new microbial activity studies and laboratory water treatment applications.

How to cite: Orozco, D.: New open-source, self-assembly tools to study microbial activity and water treatment applications, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5229, https://doi.org/10.5194/egusphere-egu25-5229, 2025.

Soil Aquifer Treatment (SAT) is a widely adopted managed aquifer recharge technique that employs natural soil filtration processes to improve the quality of secondary treated wastewater. As treated wastewater percolates through the unsaturated zone, complex interactions occur between dissolved organic matter (DOM) and the soil matrix, leading to the transformation or retention of organic contaminants. Understanding the fate of DOM within SAT systems is essential for optimizing water quality outcomes and ensuring the sustainability of water reuse practices.

Fluorescent dissolved organic matter (fDOM) has emerged as an effective tracer for characterizing DOM dynamics in water systems. By utilizing excitation-emission matrices (EEMs) in conjunction with parallel factor analysis (PARAFAC), fDOM allows for the identification of distinct molecular fractions, their origins (such as microbial or terrestrial), and their reactivity within SAT environments. However, the mechanisms that govern the retention and transformation of specific fDOM fractions during soil passage remain inadequately understood.

In this study, we employed advanced fluorescence spectroscopy to monitor the behaviour of fDOM molecules in a full-scale SAT basin recharging treated wastewater. By integrating EEM-PARAFAC analysis with in-situ water sampling along the vertical profile of the soil, we uncovered complex and varied transformations in DOM as treated wastewater permeated through the soil. Shifts in fluorescence signals indicated a dynamic interplay of processes affecting DOM fractions, including changes in composition and reactivity throughout the infiltration pathway. These patterns illuminate the evolving interactions between organic matter and the soil environment, influenced by biotic and abiotic factors.

This research underscores the potential of fluorescence-based monitoring tools to provide high-resolution, molecular-level insights into DOM dynamics in SAT systems. Such advancements can enhance the design and operation of SAT basins for improved water quality management and resource sustainability.

How to cite: Adler‬‏, O., Nakonechny, F., and Arye, G.: The fate of fluorescent dissolved organic matter molecules in recharged secondary treated wastewater within soil aquifer treatment (SAT) basin, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5916, https://doi.org/10.5194/egusphere-egu25-5916, 2025.

EGU25-6748 | ECS | PICO | HS1.2.1

Development of a Low-Cost Soil Moisture Sensor Station for Hydrological Monitoring 

Veethahavya Kootanoor Sheshadrivasan and Jakub Langhammer

The growing demand for high-resolution hydrological data necessitates innovative, scalable, and cost-effective monitoring solutions. This study presents the development of a low-cost soil moisture sensor station designed to address these challenges by leveraging advancements in open-source hardware and software.

The sensor station employs modified versions of commercially available capacitive soil moisture sensors, selected after a thorough review of existing technologies and preliminary evaluations to balance affordability and robustness. Built around the Raspberry Pi Pico microcontroller, the station features modular MicroPython programming, combined with a real-time clock (RTC) and an SD card module for robust data logging. Reconfiguration is streamlined through a JSON-based setup, avoiding the need for firmware modifications.

A custom-designed power supply unit, powered by a Li-Poly battery recharged using a 5W solar panel, ensures long-term operation. The station employs power-saving sleep modes during dormant periods, enabling continuous logging at intervals as low as 15 minutes even under suboptimal sunlight conditions in continental Europe, as per conservative estimates. Housed in a 3D-printed enclosure, the main control unit integrates ports for connecting up to three capacitive soil moisture sensors (3.3/5 V Analogue Out) at various depths, a (DHT 11) temperature and relative humidity sensor, and a UART interface for real-time access to runtime logs.

The affordability of the proposed design potentially allows for the deployment of multiple stations for the cost of a single commercially available system. This scalability is particularly critical for applications requiring dense sensor networks, such as watershed-scale studies, hydrological forecasting, or localized climate impact assessments. While acknowledging that the precision and robustness of such systems may not fully match commercial counterparts, this trade-off is expected to be offset by their adaptability and wide applicability in aforementioned cases.

Advancements in monitoring and communication technologies brought about by the "Industry 4.0" phenomena have been instrumental in enabling the design and development of this sensor station. By harnessing these innovations, the study demonstrates how innovative, cost-efficient technologies can be adapted for hydrological monitoring applications. This work wishes to not only showcase the potential of such advancements to bridge the technological and economic barriers in environmental monitoring but also wishes to highlight their role in addressing the growing gap between the demand for hydrological data and its availability.

This study aspires to facilitate and encourage further translation of advancements in monitoring and communication technologies from the "Industry 4.0" era into hydrological monitoring systems in the hope that such advancements could help democratize access to hydrological monitoring technologies, potentially addressing critical data gaps, and in-turn enabling better-informed water management and research practices.

How to cite: Kootanoor Sheshadrivasan, V. and Langhammer, J.: Development of a Low-Cost Soil Moisture Sensor Station for Hydrological Monitoring, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6748, https://doi.org/10.5194/egusphere-egu25-6748, 2025.

EGU25-6948 | PICO | HS1.2.1

Hydrological Monitoring of Rivers and Reservoirs Using Innovative Image Processing and Satellite-based Approaches 

Issa Hansen, Salvador Peña-Haro, Beat Luethi, Kerstin Stelzer, and Marcel König

The combination of in-situ measuring systems and non-intrusive optical technologies can highly improve the monitoring of water quantity and water quality in rivers and reservoirs. This paper presents two applications about innovative camera-based and satellite-based approaches to estimate flow velocity, water level, discharge, turbidity and chlorophyll concentration. The river site of the first case study presented is equipped with a DischargeKeeper, a camera-based discharge measuring system for a continuous measurement of water level, velocity and discharge in real time, and with a Multi-Parameter System MPS for water quality measurement. The MPS measures water temperature, turbidity, oxygen concentration, oxygen saturation, electric conductivity and total suspended solids TSS. The MPS probe is connected to a data logger with data transmission module to deliver measured data in real time. The DK offers. The DK consists of a video camera, an infrared beamer for illumination, a central unit for data processing, a modem for data transfer and a power supply. In operational use the camera takes video sequences of around 5s in predefined intervals, usually ranging from a few minutes to several hours. To determine the surface flow velocity of the river a processing technique called Surface Structure Image Velocimetry (SSIV) is applied. The transmitted proof images with time stamp are very helpful for the optical verification of the measurement especially during flood events.  Furthermore, the camera used can be installed at almost any position with respect to the flow, regardless of the presence of a bridge, as far as the flow is in the view of the camera with a good resolution.

Optical satellite sensors, which is the second case study of this paper, provide the opportunity to determine water constituents for whole water bodies. It is possible to derive optically active substances, which leads to good assessment of chlorophyll concentration as a proxy for algal blooms, of the water turbidity, coloured dissolved organic matter and suspended sediment. If the concentration of algae is high enough (appr. > 10 µg/l), also the occurrence of cyanobacteria can be detected. For deriving these parameters, atmospheric correction and in-water retrieval are most important processing steps. The products derived from satellite data can be aligned with the in-situ measurements acquired within DIWA which provides a complementary view on a water body. In our case we aim in combining high temporal, but punctual in-situ data with the spatial information derived from satellite data. They both contribute to the warning system for exceptional high algal blooms or occurrence of cyanobacteria. In case of river systems, the detection of a bloom that occurs upstream can already help to prepare for measures further downstream. Besides the added value that satellite data provide, limits come with reduced data availability due to cloud coverage and limits in spatial resolution for very small water bodies or very narrow river systems.

Both case studies presented showed a very good applicability of image processing technologies for measuring various hydrological and water quality parameters.

How to cite: Hansen, I., Peña-Haro, S., Luethi, B., Stelzer, K., and König, M.: Hydrological Monitoring of Rivers and Reservoirs Using Innovative Image Processing and Satellite-based Approaches, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-6948, https://doi.org/10.5194/egusphere-egu25-6948, 2025.

EGU25-9375 | ECS | PICO | HS1.2.1

Stereo photogrammetry for river water level and cross-section update: classical and deep learning approaches 

Pedro Alberto Pereira Zamboni, Robert Krüger, and Anette Eltner

Water level information is essential for monitoring and modelling river systems. Traditional, water level monitoring is done using intrusive gauging methods, such as pressure gauges; however, these sensors might be lost during an intense flood. Furthermore, in extreme flood or droughts events, measurements may become insufficient. Camera gauges have gained attention over recent years. These techniques emerge to be a low-cost and remote sensing approach for river monitoring. Camera gauges provide a more flexible and convenient setup, with cameras installed in a safe location. Moreover, they can efficiently monitor a wide range of water level values. Additionally, image sequences can be used to estimate water surface velocity and to eventually measure river discharge. Common camera gauge setups use one camera requiring additional information, e.g., a 3D model of river reach and ground control points (GCPs). On this setup, the water area is extracted from the images, and the water surface contour is reprojected into the 3D model, with the reprojection process being supported by GCPs. However, capturing 3D models can be challenging and is sometimes not possible. Further, due to cross-section change over time, there is a need to update the 3D model to ensure precise measurement. Here, we propose to change the camera gauge paradigm by using two cameras and applying stereo-photogrammetry. Using a traditional stereo-photogrammetry approach, points from two images can be projected into a 3D space, without the need for a 3D model. In this setup, the only required additional information besides the interior camera geometry is the distance between the two cameras, e.g., the baseline. After retrieving the relative camera positions, images can be densely matched to produce high resolution point clouds of the river cross-section.

For stereo-reconstruction, one of the first steps is the matching of key points between the images. Matched points are used to retrieve the relative camera poses (position and orientation). The matching can be done using standard matching algorithms (e.g., SIFT, and SURF). However, these can fail in cases where images have low texture or when they are captured in challenging light conditions. Deep learning has gained attention as an alternative to improving stereo processing. Neural networks for the feature matching achieved state-of-the-art results, being more robust in challenging conditions. Attempts to fully replace the traditional stereo reconstruction have been made (e.g., DUSt3R and MASt3R). These approaches can be used in stereo reconstruction without any prior information; however, they were not yet evaluated for camera gauge applications.

The overall goal of our research is to produce an easy and robust stereo camera gauge setup that to flexibly estimate a 3D model of river cross-sections. Thereby, we can deliver a more robust long-term camera gauge, lowering system deployment costs and maintenance efforts, allowing for a flexible densification of the hydrological monitoring network.

How to cite: Zamboni, P. A. P., Krüger, R., and Eltner, A.: Stereo photogrammetry for river water level and cross-section update: classical and deep learning approaches, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9375, https://doi.org/10.5194/egusphere-egu25-9375, 2025.

EGU25-9460 | ECS | PICO | HS1.2.1

Prediction of Temporal Dissolved Oxygen Concentrations in a Lake Using Remote Sensing and Machine Learning 

Utku Berkalp Ünalan, Onur Yüzügüllü, and Ayşegül Aksoy

Dissolved oxygen (DO) levels are crucial for aquatic life, especially under climate change, making continuous monitoring essential for effective lake management. However, local measurements are often costly and time-intensive, whether collected through field campaigns or permanent gauges. This study investigates the feasibility of using remote sensing techniques, coupled with machine learning; to track and estimate DO in a shallow eutrophic lake. Because DO cannot be directly measured with optical sensors, we first identify optically sensitive parameters—chlorophyll-a (Chl-a), temperature, and water depth—that correlate statistically with ground-measured DO. A two-step pipeline is then introduced: (1) estimating water level changes, Chl-a, and surface temperature from satellite data, and (2) predicting DO based on these derived parameters.

 

Model development starts with developing three separate models to estimate Chl-a (Sentinel-2), water level changes (Sentinel-1), and lake surface temperature (MODIS), using the Google Earth Engine Python API for data processing and analysis. Subsequently, both remotely sensed parameters and local measurements are used to train a DO prediction model. The training procedure explores 16 machine learning frameworks with hyperparameter tuning, using a 70%–15%–15% time-series split for training, validation, and testing, implemented in scikit-learn and Optuna. Search stopped with the model with R² values of 0.89 and 0.64 and mean absolute errors of 0.81 mg/L and 1.29 mg/L for locally measured and predicted test datasets, respectively. These results highlight the potential of combining remote sensing-derived parameters with machine learning to estimate DO, an otherwise non-optically measurable parameter.

 

This approach offers a cost-effective alternative for modeling continuous temporal variations in DO and supports comprehensive temporal assessments of DO concentrations in shallow eutrophic lakes. Ultimately, this framework shows promise for broader applications and generalizations, thereby contributing to the effective monitoring of non-optical water quality parameters and advancing sustainable aquatic ecosystem management.

How to cite: Ünalan, U. B., Yüzügüllü, O., and Aksoy, A.: Prediction of Temporal Dissolved Oxygen Concentrations in a Lake Using Remote Sensing and Machine Learning, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9460, https://doi.org/10.5194/egusphere-egu25-9460, 2025.

EGU25-9477 | PICO | HS1.2.1

Using citizens recorded videos to estimate water surface velocity and dischargefor urban flash flood monitoring 

Raffaele Albano, Muhammad Asif, Silvano Dal Sasso, and Aurelia Sole

Flash floods in Mediterranean regions pose significant threats to lives, infrastructures, and economies. Recent episodes of extreme rainfall in one such region led to devastating flash floods, resulting in loss of life, destruction of homes, and widespread disruption of transportation networks. Therefore, there is a critical need for advanced methods to monitor and analyze the flood dynamics, especially in urban areas. This study investigates the use of two advanced image-based techniques, Fudaa-LSPIV (Coz et al., 2014) and SSISM-Flow (Ljubičić et al., 2024) for surface velocity and discharge estimation of urban flash floods. The research used videos or images of historical urban flood events and estimated the surface velocity. To analyze the urban floods, Matera, a city of southern Italy, was selected as case study. Matera was chosen because its historical city center, the “Sassi”, was affected by extreme rainfall events in the last few years, e.g. 2014, 2018, 2019, and 2023. Five extreme past flood events occurred on 3 Aug 2018, 12 Nov 2019, 2 Jun 2023, and 2 & 21 July 2024 were recorded for estimation of surface velocity. Fudaa-LSPIV works according to the Particle Image Velocimetry (PIV) principles, while SSISM-Flow is a user-friendly and Python-based innovative tool with OpenCV integration for precise surface velocity filed extraction. These methods involve steps such as image stabilization, camera calibration, orthorectifications, and velocity calculation. Both techniques were evaluated based on their accuracy, performance, and application to overcome the limitations of analyzing the surface flow of urban floods. This study is innovative in comparing methods to estimate surface velocity of real-time flash floods in urban areas. Using these techniques, the surface velocities were estimated along key transects, and results were cross-validated using the Float Time method as benchmark. The outcomes of both approaches turned out to be consistent with benchmark data, confirming their reliability in monitoring urban floods. This comprehensive flow analysis provided insights for calibrating flood models and enhanced risk management. This study introduced a novel application of these techniques in real-time urban flood monitoring. Furthermore, it contributes to the development of an early warning system, enhances management strategies, and mitigates flood risks in vulnerable areas.

Reference

Ljubičić, R., et al., 2024.  SSIMS-flow: image velocimetry workbench for open-channel flow rate estimation. Environ. Model. Softw. 173, 105938.

Coz, Jérôme Le, wt al., 2014. Image-Based Velocity and Discharge Measurements in Field and Laboratory River Engineering Studies Using the Free Fudaa-LSPIV Software. In Proc.of the Inter.  Conf. on Fluvial Hydraulics, River Flow, 1961–67.

Acknowledgments

This work was funded by the Next Generation EU - Italian NRRP, Mission 4, Component 2, Investment 1.5, call for the creation and strengthening of 'Innovation Ecosystems', building 'Territorial R&D Leaders' (Directorial Decree n. 2021/3277) - project Tech4You - Technologies for climate change adaptation and quality of life improvement, n. ECS0000009. This work reflects only the authors’ views and opinions, neither the Ministry for University and Research nor the European Commission can be considered responsible for them.

How to cite: Albano, R., Asif, M., Dal Sasso, S., and Sole, A.: Using citizens recorded videos to estimate water surface velocity and dischargefor urban flash flood monitoring, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9477, https://doi.org/10.5194/egusphere-egu25-9477, 2025.

The growing prevalence of urban floods necessitates the development of cost-effective and scalable monitoring solutions. Traditional water-level sensors are often prohibitively expensive for widespread deployment. Moreover, existing image-based methods frequently encounter limitations in generalizability, particularly the difficulty of harmonizing selected reference features in large-scale quantitative measurements. To address this research gap, we present a novel method that utilizes traffic camera imagery to provide a lightweight solution for quantitatively monitoring urban flood inundation depths with high spatial and temporal resolution. Specifically, the waterline in flood images is recognized by a neural network and localized using world coordinates calibrated by common road markings, allowing for accurate inundation depth measurement based solely on the imagery. This method eliminates the need for costly point cloud data collection or pre-calibrated measurement objectives in urban settings. Additionally, this method enables the simultaneous collection of waterlogging depths from multiple reference objectives within the same image, yielding more robust measurements. This innovative approach paves the way for cost-effective, high-resolution, and reliable quantitative monitoring of urban flood inundation depths, ultimately providing crucial data support for emergency responses and long-term flood mitigation strategies.

How to cite: Qin, J. and Ping, S.: A Scalable and Lightweight Urban Flood Monitoring Solution Utilizing Only Traffic Camera Images, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10710, https://doi.org/10.5194/egusphere-egu25-10710, 2025.

EGU25-11952 | PICO | HS1.2.1

Development of a low-cost IoT system for monitoring storm drain overflow during urban flooding  

Antonino Cancelliere, Gaetano Buonacera, Nunziarita Palazzolo, Alberto Campisano, Aurora Gullotta, and David J. Peres

Urban flooding, intensified by climate change, presents significant risks to public safety and infrastructure, necessitating effective early warning systems. These events are categorized into fluvial and pluvial flooding, with the latter becoming increasingly challenging to predict due to its localized nature and short lead times.

In this work we develop a novel low-cost device based on Internet of Things (IoT) useful for urban flooding monitoring. The proposed sensor leverages advances in open-source technology, using ESP32 development boards, to create an accessible and cost-effective solution based on ultrasonic and reed-switch mechanisms. The system features innovative design principles, including a 3D-printed structure, low power consumption, and reliable connectivity through LoRaWAN and MQTT protocols used as potential early warning system.

The system’s primary objective is to detect storm drain overflow caused by intense rainfall, triggering timely alerts to mitigate flood impacts. Functional requirements emphasize ease of installation, durability, and cost-effectiveness, enabling widespread adoption in diverse urban contexts. The sensor design incorporates a float mechanism, reed switch, and microcontroller housed in a compact, water-resistant case.

Preliminary testing demonstrated the system's ability to detect water level changes and transmit alerts efficiently. Further work includes refining the design to minimize false positives and enhance system reliability under various environmental conditions. This development represents a significant step toward scalable, low-cost flood monitoring systems, contributing to global efforts in urban flood risk management.

 

How to cite: Cancelliere, A., Buonacera, G., Palazzolo, N., Campisano, A., Gullotta, A., and Peres, D. J.: Development of a low-cost IoT system for monitoring storm drain overflow during urban flooding , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11952, https://doi.org/10.5194/egusphere-egu25-11952, 2025.

EGU25-12047 | ECS | PICO | HS1.2.1

Global Framework for Chlorophyll-a Monitoring in Inland Lakes: Integrating Remote Sensing, Machine Learning, and Databases - Achievements and Challenges 

Aung Chit Moe, Khim Cathleen Saddi, Ruodan Zhuang, Domenico Miglino, Jorge Saavedra Navarro, and Salvatore Manfreda

Eutrophication is a significant environmental concern, which is often monitored through Chlorophyll-a (Chla) concentrations in inland and coastal waters. While traditional in-situ measurement methods are accurate, these are time-intensive, labor-demanding, and limited in spatial and temporal resolution. In recent years, remote sensing and machine learning approaches have emerged as promising alternatives for environmental monitoring, although their effectiveness is limited by challenges such as constrained in-situ data availability, the variability of water characteristics, and difficulties in transferring models across regions. Existing global models prioritize data quantity over quality, often lacking in comprehensive analysis of relationships between water quality parameters and remote sensing bands and indices. This study aimed to enhance global Chla prediction accuracy by improving data quality and identifying key predictive features using Earth Observation (EO) data. Two feature groups were examined: Group 1 (reflectance values from single bands and band ratio indices) and Group 2 (reflectance values from single bands combined with mathematical transformations of multiple bands). Machine learning models, including Random Forest (RF), Least Squares Boosting (LSBoost), Support Vector Regression (SVR), and Gaussian Process Regression (GPR), were assessed for overall performance, cross-validation accuracy, and transferability to external datasets. Among tested models with their own dataset, GPR achieved the highest overall accuracy (R² = 0.95, RMSE = 2.82 µg/L with Group 2 features), while SVR exhibited the weakest performance. For transfer validation using data from external lakes, RF (R² = 0.73, RMSE = 12.39 µg/L) and LSBoost demonstrated the greatest transferability. Spatial-temporal predictions of Chla over 2023–2024 successfully captured seasonal trends by revealing reliable and consistent patterns of Chla distribution. The present study highlights the potential of the proposed framework for global Chla monitoring in inland waters, also, emphasizing the potential in areas outside the training dataset.

Keywords: global chla monitoring, transferability, remote sensing, machine learning

How to cite: Moe, A. C., Saddi, K. C., Zhuang, R., Miglino, D., Saavedra Navarro, J., and Manfreda, S.: Global Framework for Chlorophyll-a Monitoring in Inland Lakes: Integrating Remote Sensing, Machine Learning, and Databases - Achievements and Challenges, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12047, https://doi.org/10.5194/egusphere-egu25-12047, 2025.

EGU25-14585 | ECS | PICO | HS1.2.1

Bridging Data Gaps in Soil Matric Potential for Enhanced Water Management 

Mohammad Zeynoddin, Silvio José Gumiere, and Hossein Bonakdari

Handling unstructured and missing data (UMD) remains a significant challenge in environmental monitoring and precision agriculture. This study focuses on the imputation of UMD in soil matric potential (SMP) datasets, a critical parameter in assessing soil water availability and managing irrigation systems. Missing data can distort trends, complicate analysis, and hinder decision-making in critical areas such as water management and precision irrigation. Using Extreme Learning Machine (ELM) and Time Series Models with Exogenous Inputs (TSMX), the research reconstructs missing SMP records by integrating adjacent sensor datasets and explanatory environmental variables. This approach demonstrates the potential of advanced data-driven techniques to enhance the reliability of agricultural and hydrological datasets. The dataset encompasses hourly SMP measurements and explanatory variables, including meteorological inputs such as relative humidity, air temperature, and soil properties, collected across multiple sensors in a precision agriculture setup. Exploratory analysis revealed variations in data structure, including non-stationary trends and significant statistical differences between training and testing datasets. These insights guided the selection of inputs and model configurations, emphasizing the importance of autocorrelation analysis in determining the most significant predictors. The ELM model exhibited superior performance in imputing missing SMP values, achieving an R-value of 0.992, RMSE of 0.164 cm, and NSE of 0.983 using five key inputs. This robustness highlights ELM's capability to generalize across diverse input combinations effectively. Additionally, TSMX has also been explored for its potential to leverage temporal dependencies and explanatory variables for consistent imputation. The incorporation of adjacent sensor data in modeling efforts underscores the importance of spatial and temporal relationships in enhancing accuracy, particularly in heterogeneous environmental conditions. This research underscores the critical role of input selection and model tuning in addressing UMD in SMP datasets. The findings demonstrate the complementary strengths of ELM and TSMX, offering practical insights for improving data reliability in precision irrigation and environmental monitoring. Future studies could explore integrating additional explanatory variables and employing advanced machine learning architectures to optimize imputation performance under varying environmental conditions further.

Keywords: Missing Data Imputation; Soil Matric Potential; Extreme Learning Machine; Time Series Models; Exogenous Inputs; Precision Agriculture; Environmental Monitoring.

How to cite: Zeynoddin, M., Gumiere, S. J., and Bonakdari, H.: Bridging Data Gaps in Soil Matric Potential for Enhanced Water Management, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14585, https://doi.org/10.5194/egusphere-egu25-14585, 2025.

EGU25-16334 | ECS | PICO | HS1.2.1

Hydrological monitoring in small catchments: the MagicHydroBox 

Simone Noto, Nicola Durighetto, Flavia Tauro, Ciro Apollonio, Andrea Petroselli, and Salvatore Grimaldi

The monitoring of small headwater catchment represents a major issue in hydrology, especially in remote areas, where gathering real-time hydrological data is often prohibitive due to the limited availability of power and connectivity. However, recent advances in non-contact computer vision and informatic technology offer an opportunity to fill such technical gap. In this regard, we designed and developed the MagicHydroBox prototype (MHB), and all-in-one camera and processing unit system aimed at monitoring the water level in small headwater rivers. The tasks performed by MHB include image collection, image processing, storage and transmission of the processed data. Since the MHB is equipped with NIR (NearInfrared) leds and camera, the image collection can be carried out both during the day and the night period. The image processing takes place directly in the MHB, to guarantee the onsite analysis, it is based on the Otsu’s segmentation method to identify a properly placed target within the images, and results in the direct estimation of water depth. Finally, we built in the MHB the possibility to transmit the processed data both through Gprs (mobile data) and LoRaWan (a long-range, low-power system). The MHB is also equipped with a GUI that allows the user to set and calibrate the instrument. We carried out preliminary field tests to evaluate the effectiveness of the MHB in providing an accurate measure of the target and transmitting the processed data. The preliminary results highlight the potential of the MHB to estimate the water level, especially in NIR images, and to provide a real-time hydrological monitoring where Internet signal is available. The main innovation of the MHB is represented by the fact that it automated a series of tasks that were instead manually performed in previous works. The concentration of all the necessary tasks within the MHB simplify the data acquisition, the processing and the management providing an useful tool where frequent maintenance or monitoring surveys are not possible. Moreover, the MHB is promising for future implementation of algorithms to measure surface velocimetry and discharge.

How to cite: Noto, S., Durighetto, N., Tauro, F., Apollonio, C., Petroselli, A., and Grimaldi, S.: Hydrological monitoring in small catchments: the MagicHydroBox, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16334, https://doi.org/10.5194/egusphere-egu25-16334, 2025.

Flow velocity and water surface elevation (WSE) are fundamental for understanding hydraulic phenomena in river engineering. Although underwater flow properties are not directly observable, these two parameters encapsulate the hydraulic properties governing river flow, as described by the conservation of mass and momentum equations. This information enables the understanding of actual hydraulics and facilitates the creation of digital twins, even during large scale flood events.

To measure flow velocity from UAV imagery, we developed a novel, reference-free image analysis method based on image conversion. This method eliminates the need for physical reference points, addressing practical challenges in field deployments. It leverages readily available camera information, including position (x, y, z) and orientation (pitch, roll, yaw). Complementary WSE data, obtainable from various sources, completes the required input. This allows accurate conversion of video pixel data to surface coordinates, enabling velocity measurements at any point within the river flow. Particle image velocimetry (PIV) is then applied to the converted images to derive the velocity field.

For WSE determination, we explored three approaches: Light Detection and Ranging (LiDAR), Structure from Motion (SfM), and edge-based downscaling of SfM. LiDAR data, while valuable and easy to observing, exhibits lower point density on the water surface compared to the surrounding non-water areas, depending on water surface conditions. However, even sparse LiDAR data in the mid-channel provides crucial hydraulic information. For SfM, we employed multiple UAVs capturing images at appropriate timing to resolve temporal WSE changes. As a downscaling approach using a single UAV, WSE data extracted solely from the riverbank can also be utilized.

We have begun accumulating observations of large-scale flow phenomena. Our results reveal cellular secondary currents and flow patterns over bedforms. Observations of cellular secondary currents show boiling-type phenomena occurring on the order of seconds, and more persistent cellular structures when averaging the flow field over one minute. From an engineering perspective, although these events are infrequent, they can significantly impact float-based discharge measurements when they occur. Observations of flow over bedforms show spatial variations in velocity and WSE along the flow direction, exhibiting wave-like patterns. The out-of-phase relationship between these wave patterns suggests they are associated with micro-bedforms, indicating active sediment transport. Furthermore, this understanding of sediment hydraulics can be used to estimate water depth.

How to cite: yorozuya, A. and kudo, S.: Flow structures in actual rivers obtained by areal measurement of flow velocity and water surface elevation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-16904, https://doi.org/10.5194/egusphere-egu25-16904, 2025.

EGU25-17076 | ECS | PICO | HS1.2.1

Transforming water resources management at river basin scale with digital twin technology 

Amir Rouhani, Ainhoa Mate Marin, Antonio Moya Diez, J. Jaime Gómez-Hernández, Michael Rode, and Seifeddine Jomaa

Digital twin, as a virtual representation of physical systems, is increasingly recognised as a core component of timely and accurate water management, particularly for interconnected and rapidly changing systems. Digital twin supports the simultaneous monitoring, simulation, and optimisation of real-world operations by integrating multiple data sources, including in-situ measurements, remote sensing and modelling data. By enabling a detailed characterisation of catchment functioning and its ecological boundary conditions, a digital twin facilitates equitable water allocation across sectors and supports timely and evidence-based decision-making.

Developing a digital twin requires extensive datasets, robust scientific evidence, and a clear grasp of ecological boundaries, reflecting the interconnected nature of multi-sectoral decision-making. The Bode River Basin, one of the best-monitored catchments in central Europe, serves as a showcase for designing and implementing a digital twin system for multi-sectoral and sustainable water management at catchment scale. The recent prolonged droughts (2017–2021) and their impacts on various water bodies offer a real-world “experiment” of extreme climate scenarios, highlighting the vulnerabilities and risks within the catchment and illustrating the complex trade-offs inherent in water resource management.

This study integrates long-term, high-resolution monitoring strategies with coupled surface water, groundwater, and water quality models into a unified framework that addresses both quantitative and qualitative aspects of water systems. Such a comprehensive approach enables forecasting climate change impacts and optimising water resource allocation across sectors. Overall, this work demonstrates the potential of digital twins to advance sustainable water resource management under changing climatic conditions.

Acknowledgment

This work was supported by the OurMED PRIMA Program project funded by the European Union’s Horizon 2020 research and innovation under grant agreement No. 2222.

How to cite: Rouhani, A., Mate Marin, A., Moya Diez, A., Gómez-Hernández, J. J., Rode, M., and Jomaa, S.: Transforming water resources management at river basin scale with digital twin technology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17076, https://doi.org/10.5194/egusphere-egu25-17076, 2025.

EGU25-17492 | ECS | PICO | HS1.2.1

The Role of Riverbed Background Reflectance in Long Term Turbidity Monitoring Using Camera Systems 

Domenico Miglino, Seifeddine Jomaa, Khim Cathleen Saddi, Aung Chit Moe, Michael Rode, and Salvatore Manfreda

The use of digital cameras in river monitoring activities can increase our knowledge of water quality status, solving the cost and spatial and temporal data resolution limitations of the existing techniques. The challenge of image-based procedures using camera systems is the proper red, green, and blue (RGB) bands signal interpretation and processing. The actual water upwelling light that reaches the camera lens is the sum of various reflectance components of the suspended particles, the riverbed background and the water itself. One component could prevail over the others, depending on the variability of hydrological (water level, flow velocity, etc.) and environmental (suspended solids concentration, floating pollutants, etc.) characteristics of the river. The effect of water level and turbidity concentration on the riverbed component of the total water upwelling light can be substantial, especially for shallow water. As a result, the riverbed reflectance component, if neglected, can significantly affect the evaluation of the water reflectance, and hence, water turbidity.

In our field campaign, a synthetic turbidity event was recreated by adding a natural clay tracer into the river, and we monitored it using a camera system. Two turbidimeters were installed within the river section to validate the results. Moreover, a submerged panel was fixed directly on the riverbed. This choice was prompted by the shallow water conditions during the experiment, where the riverbed reflectance significantly contributed to the total upwelling light captured by the camera, particularly under low turbidity levels. We defined a clear water condition in which the panel was fully visible, where turbidity level was considered equal to zero. As turbidity increased and the panel visibility decreased, we applied an image-based procedure to assess the actual river turbidity level. In addition, we applied a pixel-by-pixel mean of the camera frames every 2 minutes, for minimizing the signal distortions due to the effect of ripples, sun glare and shadows within the analyzed region of interest of the river surface.  These methodological steps allowed us to properly decompose the image into different reflectance components, and to enhance long-term monitoring practices that are subject to a wide range of environmental and hydrological variability.

This study focuses on implementing camera systems in real-world settings, supporting existing river monitoring techniques with early warning networks, and developing innovative solutions for water resource management.


Keywords: camera system, river monitoring, turbidity, image processing, remote sensing, water quality

How to cite: Miglino, D., Jomaa, S., Saddi, K. C., Moe, A. C., Rode, M., and Manfreda, S.: The Role of Riverbed Background Reflectance in Long Term Turbidity Monitoring Using Camera Systems, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17492, https://doi.org/10.5194/egusphere-egu25-17492, 2025.

EGU25-18462 | ECS | PICO | HS1.2.1

Optimizing Image-Based Techniques for River Monitoring: Insights into Graphical Enhancement and Parameter Sensitivity 

Francesco Alongi, Robert Robert Ljubičić, Dario Pumo, Silvano Fortunato Dal Sasso, and Leonardo Valerio Noto

Image-based techniques have gained significant attention for monitoring natural and artificial rivers, thanks to their many advantages over traditional methods. These non-intrusive and highly-versatile optical approaches provide accurate flow discharge measurements, even in challenging conditions like flood events, while ensuring the safety of both operators and equipment. However, the accuracy of optical measurements is affected by several factors, including environmental conditions, river flow characteristics, field acquisition protocols, and the parameterization of the processing software.

Image-based techniques follow a three-phase workflow: (i) seeding, (ii) recording, and (iii) processing. Seeding introduces natural or artificial tracers onto the water surface to detect motion. Recording captures video sequences from stationary or mobile platforms (e.g., UASs – Unmanned Aerial Systems). Processing extracts the surface velocity field and flow metrics. The latter phase is divided into three sub-steps: pre-processing, surface velocity evaluation, and post-processing. Pre-processing includes stabilization, orthorectification, and graphical enhancement; surface velocity evaluation uses correlation-based or similar algorithms to track tracers across frames; finally, post-processing refines velocity data by filtering noise, interpolating missing data, and extracting relevant metrics.

Among the steps of optical techniques, graphical enhancement is particularly critical. By increasing the contrast between tracers and the background, it enhances the ability of software algorithms to accurately track motion, thereby reducing errors. However, an inadequate parametric setup of the processing software can also result in the estimation of biased velocities. To investigate these interdependencies, this study conducted a comprehensive sensitivity analysis, evaluating the combined effects of graphical enhancement techniques and processing parameters on the performance of image-based analyses. The analysis compares traditional algorithms with more innovative approaches, including colorspace transformation, and assesses the impact of varying processing parameters under different operational conditions. A dataset of videos acquired from UAS platforms and fixed stations during discharge measurement campaigns on Sicilian rivers, in Italy, was used. The videos were analyzed using PIVlab and SSIMS-Flow software, and the results were benchmarked against ADCP measurements.

The findings reveal that both the choice of graphical enhancement methods and the optimization of key software parameters significantly affect the accuracy of velocity and discharge estimates. The study also provides valuable insights into selecting the most appropriate enhancement techniques and configuring processing parameters, tailored to specific field conditions and operational requirements, further demonstrating the potential of image-based methods for hydraulic monitoring.

How to cite: Alongi, F., Robert Ljubičić, R., Pumo, D., Dal Sasso, S. F., and Noto, L. V.: Optimizing Image-Based Techniques for River Monitoring: Insights into Graphical Enhancement and Parameter Sensitivity, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18462, https://doi.org/10.5194/egusphere-egu25-18462, 2025.

EGU25-18603 | ECS | PICO | HS1.2.1

Evaluating the Potential of Personal Weather Stations (PWS) for Semi-distributed Hydrological Modelling  

Ranka Kovačević, Andrijana Todorović, Carlo De Michele, Roberto Nebuloni, and Alessandro Ceppi
 

Personal Weather Stations (PWS) have gained attention in recent years as a potential complement to operational meteorological networks, which are often sparse and may not adequately capture localized rain events, especially in areas with complex orography. PWS, on the other hand, can improve the spatial resolution of rainfall data due to their affordability and, thus, widespread distribution. However, their effectiveness and reliability depend on overcoming certain challenges. PWS often lack adherence to World Meteorological Organization standards, as they may not be properly placed nor regularly maintained, and there are no standardised approaches for data quality check. Frequent gaps in the series (mainly due to data transmission issues), and a constantly changing network layout further limit reliability and consistency of PWS data for hydrological modelling. Therefore, the application of PWS rain data for hydrological modelling is still in its infancy.  

This research focuses on evaluating PWS rainfall data for hydrological modelling in the peri-urban Lambro catchment in northern Italy, by comparing characteristics of hourly rainfall data obtained from the MeteoNetwork (Giazzi et al., 2022; https://doi.org/10.3390/atmos13060928, 2022) to those of the rain gauge data obtained from the  Regional Agency for the Protection of the Environment of Lombardy (ARPA). This study focuses on the characteristics of the subcatchment-averaged rainfall series are compared. The rain depths in each of the 15 subcatchments are calculated by using the inverse-distance weighting method with the power of 2, and with increasing maximum distance between the station and the centroid of a subcatchment (10km, 25km and 50km). The two subcatchment-averaged rainfall series are compared in terms of (1) accumulated rain depth, (2) maximum rainfall intensity, and (3) timing of the peak rainfall intensity during a rain event. 

Our results indicate that, compared to ARPA rainfall data, PWS data can both underestimate and overestimate rainfall values with similar frequency. Specifically, the magnitude of error in rain depths ranges from -44% to +56% across the subcatchments, and this range does not change significantly with increasing maximum distance. With the maximum distance of 10 km, in eight out of 15 subcatchments the absolute value of the error is smaller than 15%, while the median value amounts to 1.9%, and decreases to -17% and -19% with increasing maximum distance. The errors in maximum rainfall intensity are slightly larger, ranging from -67% to 76%, when compared to the official ARPA gauges with the maximum distance of 10 km. The median error amounts to 15.5%, -26% and -30% for the three maximum distance values. Concerning the timing of peak intensity, there are no discrepancies between the two datasets, and PWS data can be considered accurate in this regard. However, large errors in rain depths and intensities suggest that PWS rain data alone cannot be expected to yield accurate outputs in hydrological simulations. This conclusion will be tested by running a hydrological model with these datasets.  

 

Acknowledgement  

This research is part of the work within the COST Action “Opportunistic Precipitation Sensing Network” (OpenSense, CA20136)

How to cite: Kovačević, R., Todorović, A., De Michele, C., Nebuloni, R., and Ceppi, A.: Evaluating the Potential of Personal Weather Stations (PWS) for Semi-distributed Hydrological Modelling , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18603, https://doi.org/10.5194/egusphere-egu25-18603, 2025.

EGU25-19738 | PICO | HS1.2.1

Impact of intense rainfall event on the physicochemical and microbiological characteristics of an urban stream 

Sándor Kun, Imre Boczonádi, Péter Tamás Nagy, Andrea Szabó, Florence Alexandra Tóth, Zsolt Zoltán Fehér, Tamás Magyar, Lili Adrienn Madar, István Szűcs, János Tamás, and Attila Nagy

Extreme weather events, including sudden and intense rainfall, have become increasingly frequent due to the growing impact of climate change. This rapid influx of water often carries a variety of pollutants, including nutrients, heavy metals, and microbial contaminants, significantly modifying the physicochemical and microbiological characteristics of urban streams. This study aims to evaluate the effects of rainfall events on the physicochemical and microbiological properties of the Tócó Stream, focusing on changes in key water quality parameters and microbial dynamics. Two sampling points were selected to represent different environmental areas: one site was located in a near-natural area, and the other was situated in an industrial zone, surrounded by facilities and a highway connecting road. Measurements were conducted both before and after the rainfall event. On-site measurements were performed included precipitation (mm), water level, dissolved oxygen content, and water temperature, while water samples were collected for laboratory analysis. The collected samples were tested for pH and electrical conductivity (EC) as well as for nutrient-concentrations of NH₄⁺, NO₂⁻, NO₃⁻, PO₄³⁻, K⁺, SO₄²⁻, chemical oxygen demand (COD) and biological oxygen demand (BOD5) were also determined from the samples. In case of microbiological parameters, total coliforms, yeasts, and total plate count were determined. Our results revealed differences between the two sampling sites and the pre- and post-rainfall conditions. At the industrial site the nutrient contents have decreased due to the rainfall, while at the near natural site we did not determine such change in connection with these elements. The same trend were detected in the case of EC as well. The microbiological analysis of the water samples clearly showed that while both total bacterial count and total coliform count showed an increasing trend after the rainfall at the first site, this trend was much less pronounced at the site reflecting the natural state. Our objecitve was to study the influence of sudden rainfall events, for the reason that these effects remain understudied, particularly in terms of their short- and long-term impacts on water quality and microbial properties.

The research presented in the article was carried out within the framework of the Széchenyi Plan Plus program with the support of the RRF 2.3.1 21 2022 00008 project.

 

How to cite: Kun, S., Boczonádi, I., Nagy, P. T., Szabó, A., Tóth, F. A., Fehér, Z. Z., Magyar, T., Madar, L. A., Szűcs, I., Tamás, J., and Nagy, A.: Impact of intense rainfall event on the physicochemical and microbiological characteristics of an urban stream, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19738, https://doi.org/10.5194/egusphere-egu25-19738, 2025.

Discharge estimation at a river site depends on local hydraulic conditions identified by recording water levels. In fact, stage monitoring is straightforward and relatively inexpensive compared with the cost necessary to carry out flow velocity measurements which are, however, limited to low flows and constrained by the accessibility of the site. In this context, the mean flow velocity is hard to estimate for high flow, affecting de-facto the reliability of discharge assessment for extreme events. On the other hand, the surface flow velocity can be easily monitored by using radar sensors allowing to achieve a good estimate of discharge by exploiting the entropy theory applied to rivers hydraulic (Chiu,1987). The growing interest towards the use of no-contact methods to estimate discharge (Tauro et al., 2018) in field applications has shown that the cross-track velocity distribution can be inferred with sufficient accuracy using the surface velocities, usurf, sampled using Surface Velocity Radars (SVR) (Fulton and Ostrowski, 2008; Moramarco et al., 2017, Alimenti et al. 2020), the quantitative imaging techniques as LSPIV (Fujita et al., 1998) or PTV (Tauro et al., 2019). In this context, overall the velocity-area method is applied to estimate the mean flow velocity starting from the depth-averaged velocity, uvert, which is inferred through the velocity index, k=uvert/usurf.. For many river gage sites configurations, k has been set to 0.85. However, considering k refers to a monotonous velocity profile, not taking account of dip phenomena, the application may fail in estimating the depth-averaged velocity (Moramarco et al., 2017; Koussis et al., 2022, Pumo et al., 2025). Based on that, this work proposes a new entropy-based approach to estimate the depth-averaged velocity starting from the measured surface velocity retrieved by conventional and/or no-contact measurements. The approach exploits the dependence of the entropy parameter M with the hydraulic and geometric characteristics of channel (Moramarco and Dingmann, 2017), allowing to derive formulations on Manning’s roughness, shear velocity and water surface slope. Based on these features, the entropy-based method by using the measured surface velocity and the geometry of the river site is able to turn usurf  into uvert considering for each  usurf  an index which depends on the local water surface slope. The application to river sites along the Tiber River, Po River and Amazon River has shown the effectiveness of the approach in estimating the depth-averaged velocities with a fair accuracy along all verticals. Therefore, the method well lends itself to be integrated in the field of no-contact streamflow measurements.

 

 

How to cite: Moramarco, T.: Entropy-based depth-averaged velocity assessment from surface flow velocity, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19742, https://doi.org/10.5194/egusphere-egu25-19742, 2025.

EGU25-20109 | PICO | HS1.2.1

Long-Term Evolution and Challenges of Hydrological Observations at the Fiumarella Basin in Southern Italy 

Htay Htay Aung, Beniamino Onorati, Mauro Fiorentino, Silvano Fortunato Dal Sasso, Biagio Sileo, Teresa Pizzolla, Salvatore Manfreda, and Maria Rosaria Margiotta

Hydrological observations are essential for understanding the complex interactions between the land surface and the atmosphere, improving water resource management, strengthening flood defense, and advancing hydrological modeling. However, the long-term maintenance of experimental basins like Fiumarella di Corleto presents significant challenges, requiring continuous updates to address environmental changes and technological advancements. This study reviews over 20 years of observations at the Fiumarella basin in Southern Italy, focusing on its evolution, challenges, and future directions. The Fiumarella basin, covering an area of 32.5 km², includes a sub-basin of 0.65 km². Since 2002, a hydrometeorological network has been monitoring key variables such as rainfall, temperature, wind, and streamflow, capturing hydrological variability across spatial and temporal scales. In 2006, 22 soil moisture probes were installed along a 60-meter transect at depths of 30 and 60 cm. Additionally, high-resolution LiDAR data and pedological studies have enhanced the understanding of the basin’s morphology and soil characteristics. The maintenance of this experimental basin has posed substantial challenges. Frequent extreme flood events have resulted in significant damage to hydrometric stations, requiring reconstruction and recalibration. Moreover, the sediment and debris accumulation in the retention basin of the sub-basin necessitated periodic clearing to maintain functionality and ensure continuous data collection. These challenges underscore the effort and adaptability required to sustain long-term monitoring in dynamic environments. Data collected from the basin have significantly contributed to hydrological science. Analyses of peak flow events and antecedent soil moisture conditions have provided insights into flood response mechanisms. Spatial and temporal variability in hydrological processes has informed the calibration and validation of semi-distributed hydrological models, enhancing their accuracy and reliability. These findings highlight the importance of integrating diverse datasets such as soil moisture, precipitation, topography, and land use—for comprehensive hydrological research. Looking ahead, planned upgrades aim to further enhance the basin’s capabilities. The installation of a meteorological radar would improve rainfall measurement precision and expand spatial coverage, thereby addressing existing data gaps. Additional hydrometric sensors and automated systems would increase the granularity and reliability of observations, supporting high-resolution analyses. These advancements will ensure that the Fiumarella basin remains a state-of-the-art research facility capable of addressing emerging challenges in hydrology and climate science.

This abstract is part of the project NODES which has received funding from the MUR-M4C2 1.5 of PNRR funded by the European Union - NextGenerationEU (Grant agreement no. ECS00000036).

The present research has been carried out within the RETURN Extended Partnership and received funding from the European Union Next-Generation EU (National Recovery and Resilience Plan -NRRP, Mission 4, Component 2, Investment 1.3 - D.D. 1243 2/8/2022, PE0000005).

 

How to cite: Aung, H. H., Onorati, B., Fiorentino, M., Dal Sasso, S. F., Sileo, B., Pizzolla, T., Manfreda, S., and Margiotta, M. R.: Long-Term Evolution and Challenges of Hydrological Observations at the Fiumarella Basin in Southern Italy, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20109, https://doi.org/10.5194/egusphere-egu25-20109, 2025.

Lateritic landscapes are structurally complex systems formed through intense chemical weathering under tropical paleoclimates. These profiles are found in stable, low-relief landscapes across tropical, subtropical, and Mediterranean climates, particularly between 35°N to 35°S. Their vertical structure reflects long-term shifts in climatic, hydrological, and tectonic conditions, offering a valuable "memory" of past environmental changes. Despite their environmental and economic significance, lateritic landscapes remain underrepresented in CZ research, a bias compounded by the concentration of Critical Zone Observatories in the Northern Hemisphere, where shallow, truncated profiles prevail due to glacial erosion. This underrepresentation limits our understanding of long-term CZ processes and how they have shaped subsurface architecture.

This study investigates the subsurface architecture of a lateritic hillslope at the Avon River Critical Zone Observatory (AR-CZO) in Western Australia. Prolonged subaerial weathering since the Cretaceous, followed by mid-Miocene aridification, has created a stratigraphically complex regolith hillslope shaped by weathering, erosion, and colluvial deposition. To resolve the structural complexity of this hillslope, we applied a multi-method geophysical approach, combining electrical resistivity tomography (ERT), horizontal-to-vertical spectral ratio (HVSR) passive seismic methods, and borehole observations. ERT captured fine-scale stratigraphy, delineating the pallid zone, saprolite, and duricrust, while HVSR resolved broader interfaces, such as the duricrust-bedrock boundary and the base of the colluvial deposit.

The results reveal how landscape position influences CZ structure. The hilltop is capped by a duricrust that transitions downslope into an erosional surface, where the pallid zone of the lateritic weathering profile is exposed at the surface. At the foot slope, approximately 11 m of colluvial sediment has accumulated from the erosion of the hillslope material. Bedrock depth estimates differed between methods, with ERT indicating depths of 23 m on the slope and 32 m at the foot slope, while HVSR revealed deeper depths of 31 m and 39 m, respectively. The discrepancy highlights the limitations of ERT in saline environments, where conductivity masks key interfaces, while HVSR’s broader resolution provides more reliable bedrock detection in such conditions. Together, these methods reveal a laterally variable weathering profile that responds to shifts in landscape position, erosion, and deposition.

The complementarity of ERT and HVSR underscores the value of a multi-method geophysical approach for resolving the structural complexity of lateritic CZs. Our conceptual model demonstrates how weathering, erosion, and colluvial processes shape the structure of a deeply weathered hillslope, while also providing a transferable framework for characterizing saline, regolith-dominated systems. Given their depth, age, and capacity to preserve past climatic and tectonic conditions, lateritic CZs offer a vital opportunity to enhance global understanding of long-term CZ evolution. This research addresses the Northern Hemisphere bias in CZ science, highlights the underexplored role of stable, deeply weathered landscapes, and underscores the need for future comparative studies to understand the drivers of heterogeneity in subsurface architecture across CZs worldwide.

How to cite: Weller, J., Jakica, S., Thompson, S., and Leopold, M.: Combining electrical resistivity tomography and passive seismic to characterise the subsurface architecture of a deeply weathered lateritic hill within the Avon River Critical Zone Observatory, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1401, https://doi.org/10.5194/egusphere-egu25-1401, 2025.

The ground substrate is a new concept in the field of natural resources proposed by Chinese scientists in 2020 (Ministry of Natural Resources, 2020). It is the basic material that supports and nurtures various natural resources such as soil, forests, grasslands, wetlands, and water. The layer of ground substrate is the most active geological space for the exchange of substances and energy such as water, heat, salt, gas, carbon, etc. It is also serving as a bridge link between the land cover layer and the underground resource layer. The proposal on concept of ground substrate has clarified new directions and goals for geological survey to support ecological civilization construction and natural resource management, has great significance.

In different climate zones such as humid, semi humid, semi-arid, and arid in China, there are significant differences in the material composition, genetic types, and characteristic physicochemical properties of ground substrates, which call ground substrate heterogeneity by us. In recent years, based on multiple ground substrate surveys and research projects, some important conclusions has been gained. The first is we revealed the constraint mechanisms of the physical structure, mineral element composition, and chemical properties of ground substrates on the types, NDVI, NPP of vegetation ecology in the key ecological functional areas in northern China and hilly mountainous areas in southern China. The second is the determination of the bottom boundary of the ground substrate layer requires comprehensive consideration of five factors: they are depth of the underground variable temperate zone, the roots depth of crop and vegetation, the depth of the surface karst development zone, the thickness of the weathering crust, and the burial depth of the groundwater level. It is generally believed that the depth of the ground substrate layer is less than 20 meters. The third is the key constraint layer of ground substrate (rock and soil layers that have important control and influence on vegetation and crop growth, water and salt storage and transport, etc.) is a special layer that should be given special attention in ground substrate filed survey.

More detailed about the scientific connotation and theoretical framework of ground substrate, please see the published paper(Hao Aibing, Yin Zhiqiang*, Li Hongyu, Lu Qinyuan, Peng Ling, Shao Hai, Jiang Qida, Zhao Xiaofeng, Liu Jiufeng, Pang Jumei, Yang Ke, Chen Peng, Kong Fanpeng, Hou Hongxin, Lu Min. 2024. The scientific connotation and theoretical framework of ground substrate. Acta Geologica Sinica. 98(11):3225-3237)

How to cite: yin, Z., peng, L., and hao, A.: The concept of ground substrate and its physical structure & mineral element composition constrain mechanisms on vegetation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1441, https://doi.org/10.5194/egusphere-egu25-1441, 2025.

EGU25-1952 | ECS | Orals | HS8.2.8

Clogging model of hyporheic exchange based on coupled lattice Boltzmann discrete element simulations 

Xudong Zhang, Atsushi Takai, Tomohiro Kato, and Takeshi Katsumi

The hyporheic exchange between the surface water and the underground water is considered a significant process in the natural water cycle system. Some sediment particles in the riverbed can be carried to the exchange channel under the stream effect. Over time, these particles accumulate on the channel can decrease the exchange efficiency of water resources, and induce clogs. The clogging problem of the exchange channel may further induce various geological and environmental disasters such as the shrinkage of lakes and desertification.

To detail the clogging mechanism in the exchange channel, we simulated the exchange clogging process on the exchange channel based on a coupled lattice Boltzmann method (LBM) and discrete element method (DEM). The results indicated particles could form an arch structure clogging the channel orifice. The formation of the clogging arch prevented the discharge of soil particles and greatly decreased the fluid velocity. Notably, the fluid velocity distribution around the orifice is in a certain shape according to the velocity of the LBM cells—the size of the shape regularly changes with the distance to the channel orifice. The variation of the average fluid velocity in the orifice first increases to a peak (about 0.497 cm/s) in the initial time and then decreases to an approximate value after clogging (about 0.037 cm/s). The maximum velocity is almost thirteen times the minimum, indicating that the clogging effect can reduce the water velocity of hyporheic exchange by more than one order of magnitude. In addition, it was found that the soil skeleton was necessary for forming clogs in polydisperse particle systems by analyzing the clogging arch-forming process. The sediment particles in different scales have different effects on the clogging arch. The large particles in the sediments are closely related to the formation of the soil skeleton. The fine particles were involved in the filling and enhancing of the soil skeleton.

Based on our simulation analysis, an explanation for the clogging formation under microscopic conditions was proposed, leading to a detailed description of the exchange clogging in the hyporheic exchange channel. In addition, some mechanism statements to better understand the exchange phenomenon in the water cycling ecosystem are also provided.

How to cite: Zhang, X., Takai, A., Kato, T., and Katsumi, T.: Clogging model of hyporheic exchange based on coupled lattice Boltzmann discrete element simulations, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1952, https://doi.org/10.5194/egusphere-egu25-1952, 2025.

EGU25-2490 | Orals | HS8.2.8

Spring and stream intermittency in an instrumented steep Himalayan Mountain catchment 

John Armitage, Kapiolani Teagai, Niels Hovius, Luc Illien, and Christoff Andermann

The pathway for rainfall into stream flow in mountain catchments can be fast via surface run-off or short-lived storage in the weathered zone, or slow via the deep fractured bedrock groundwater system. In mountainous topography, springs can be found at almost all elevations, suggesting that groundwater storage occurs at all elevations. There is however uncertainty as if this storage is short lived, confined to the weathered zone, or longer lived and is part of the groundwater system. Intermittent streams and springs might reflect the storage of water within the subsurface. To measure stream intermittency and the migration of the associated headwater springs we installed intermittency loggers based on repurposed HOBO luminosity loggers along five gulleys within the Kahule Khola catchment in central Nepal.

The intermittency loggers measure an electric current when the circuit is closed by surface moisture and flowing water. The loggers were installed in spring 2023 before the pre-monsoon and were removed in November 2024. At low elevation, three series of loggers were installed in gullies below the village of Listi. These below Listi loggers had perennial springs at their lowest elevation. Furthermore, one series of loggers ended at an ERT repeat survey that showed evidence of year-round shallow subsurface saturation. At high elevation, two series of loggers were installed near the village of Bagham, below an open meadow where ephemeral springs were mapped (we call these the meadow loggers). A coincident ERT repeat survey showed evidence of lateral flow of groundwater within this region.

The loggers recorded three distinct phases: (1) The pre-monsoon, where individual storm events can be registered along each gulley as separate wetting events. (2) Monsoon, where there is a continuous and high conductivity measurement for all loggers, representing continuous flow of surface water. (3) The dry season, which starts with a recession in the electric current observed, followed by sparce wet events. The below Listi systems dried completely within the dry season, while the meadow gulleys recorded low but non-zero electric currents even throughout the dry season. The loggers did not record any evidence of spring migration down the gulleys, rather a uniform drying after rainfall events at all locations, with prolonged wetness post monsoon only seen for loggers that were situated just above known perennial springs. The observations would therefore suggest that intermittent run-off comes from the temporary storage in the weathered zone that dries out at the same rate across the catchment, while persistent flow is from points where the topography intersects with the deeper groundwater reservoir. Run-off within the steep catchment therefore operates through two coexisting systems, (1) an intermittent system that is fed from temporary storage of water in the weathered zone, where there is no distinct headwater spring, and (2) perennial streams fed by groundwater springs.

How to cite: Armitage, J., Teagai, K., Hovius, N., Illien, L., and Andermann, C.: Spring and stream intermittency in an instrumented steep Himalayan Mountain catchment, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2490, https://doi.org/10.5194/egusphere-egu25-2490, 2025.

EGU25-2917 | Posters on site | HS8.2.8

Accelerating Critical Zone Science with an International Network of Networks 

Jeffrey Munroe, Bhavna Arora, Kevin Bishop, Theresa Blume, Heye Bogena, Elizabeth Boyer, Isabelle Braud, Jérôme Gaillardet, Ralf Kiese, and Steffen Zacharias

The international Critical Zone Network of Networks (CZ-NoN) project, launched in January 2025 and funded by the US National Science Foundation, promotes the study of the Earth’s Critical Zone (CZ), the vital near-surface environment where essential life-supporting processes converge.  Building on previous investments in CZ research, CZ-NoN fosters collaboration and communication between existing and emerging environmental observatories and monitoring networks worldwide.  By establishing a unified framework for collaboration and discussion, CZ-NoN addresses long-standing challenges such as fragmented methodologies, redundancies, poor communication, and barriers to data discoverability and accessibility.  Key project components include planning meetings, workshops, and an online webinar series aimed at building community, showcasing new efforts, and increasing awareness of ongoing CZ research.  In parallel, a global polling effort will compile a crowdsourced list of grand research questions to guide future CZ studies.  By bringing together researchers from different countries and disciplines, and prioritizing cooperation over competition, CZ-NoN will accelerate scientific research and position the international research community for future funding opportunities to support complex, integrated study of the global CZ across diverse socio-environmental conditions.

How to cite: Munroe, J., Arora, B., Bishop, K., Blume, T., Bogena, H., Boyer, E., Braud, I., Gaillardet, J., Kiese, R., and Zacharias, S.: Accelerating Critical Zone Science with an International Network of Networks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-2917, https://doi.org/10.5194/egusphere-egu25-2917, 2025.

Tropical vegetation plays a vital role in global ecosystem services, with one critical aspect lying in its hydrological functions of water cycle regulation. Climate change and accelerated human interventions threaten the stability of tropical vegetation, associated with profoundly hydrological changes particularly in recent decades. Despite various studies on land-atmosphere feedback using earth system models, the regulation of terrestrial hydrological components remains unclear over tropical regions, due primarily to inherent limitations of models in accurately simulating terrestrial water storage (TWS) and runoff. Here, we combine multisource observations to reveal a disparity pattern in storage-runoff interactions over tropical regions for the past two decades. Using satellite-based Landsat optical archives, Global Ecosystem Dynamics Investigation, GRACE gravimetry, and gauge-based runoff database, we show that large-scale forest degradation and cropland expansion have weakened moisture recycling over the eastern tropical South America and eastern tropical Africa (Region I), indicated by a significant decrease in net precipitation input (precipitation minus evapotranspiration). This further causes declines in both TWS and streamflow, shown as a pattern of “less storage and less runoff” due to vegetation degradation. In contrast, over the western tropical South America, western tropical Africa, and tropical Asia (Region II), we did not find marked changes in land cover but a significant increasing trend in vegetation greenness and leaf area index. This is associated with a significant increase in net precipitation input and an enhanced moisture recycling. The increased water input over Region II causes an increase in TWS but a decline in streamflow, shown as a pattern of “more storage but less runoff” due to the decrease in rainfall-runoff generation induced by vegetation growth. The disparity patterns between Region I and Region II highlight different responses of tropical terrestrial water system to a changing environment. Unlike most past studies relying on land surface or earth system models, this study leverages strengths in advanced observation techniques to explore different mechanisms underlying changes in the tropical terrestrial water system. Findings from this study provide valuable supplements to the current model-based analysis, and inform adaptive strategies for changes over tropical regions.

How to cite: Li, X. and Peng, J.: Multisource observations reveal different roles of tropical vegetation in terrestrial water regulation, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-7653, https://doi.org/10.5194/egusphere-egu25-7653, 2025.

EGU25-9089 | Posters on site | HS8.2.8

Monitoring the triple oxygen isotope composition of water and biogenic silica at the soil-plant-atmosphere interface: benefits for investigating West African present and past water cycles 

Anne Alexandre, Clément Outrequin, Christine Vallet-Coulomb, Christophe peugeot, Manuela Grippa, Julie Aleman, Claudia Voigt, Amaelle landais, Eric mougin, Ousmane Ndiaye, Corinne Sonzogni, David Au Yang, Jean-Charles Mazur, Martine Couapel, Jérome Ogée, Theodore Ouani, Simon Afouda, Nogmana Soumaguel, Torbern Tagesson, and Rasmus Fensholt

Quantitative data are needed to constrain vegetation-hydroclimate in water cycle modelling. Here, we use the triple oxygen isotope composition (δ'18O and 17O-excess) of water compartments to track water transfers and mixing within the soil-plant-atmosphere continuum. At three AMMA-CATCH sites in Benin and Senegal we monitored the δ'18O and 17O-excess of precipitation, groundwater, soil water and plant water, as well as the 17O-excess of phytoliths, an indicator of atmospheric relative humidity. We found that : 1) the 17O-excess in precipitation is very stable over several years; 2) groundwater has δ'18O and 17O-excess values consistent with a multi-year recharge by modern precipitation; 3) the 17O-excess in soil water shows a limited contribution of evaporated water, despite high evaporation conditions, which has important implications for our knowledge of water transfers within soils; 4) extrapolating linear relationships between δ'18O and excess 17O-excess of leaf and stem water allows us to determine the origin of the water absorbed by the roots. At the savanna and dry forest sites, during the rainy season, grasses absorb soil water supplied by precipitation. In contrast, during the dry season, trees reach the perennial groundwater recharge. 5) the 17O-excess of grass and tree leaf water follow the dynamics of relative humidity; 6) the 17O-excess of grass phytoliths records daily relative humidity during the growing season. These results provide a solid basis for using the triple oxygen isotope composition of water and phytoliths to trace present and past water cycles at the soil-plant-atmosphere interface.

This study was conducted in the framework of the HUMI-17 and PAST-17 projects supported by the ANR (ANR-17-CE01-0002-01 and ANR-22-CE01-0027-01), JA and CV have benefited from a Marie Sklodowska-Curie grant from the European Union (n°101063961 for JA and 101063961 for CV). TT acknowledge funds from FORMAS (Dnr 2021-00644), and the European Union under the Development Smart Innovation through Research in Agriculture (DeSIRA) Initiative (FOOD/2019/410-169).

How to cite: Alexandre, A., Outrequin, C., Vallet-Coulomb, C., peugeot, C., Grippa, M., Aleman, J., Voigt, C., landais, A., mougin, E., Ndiaye, O., Sonzogni, C., Au Yang, D., Mazur, J.-C., Couapel, M., Ogée, J., Ouani, T., Afouda, S., Soumaguel, N., Tagesson, T., and Fensholt, R.: Monitoring the triple oxygen isotope composition of water and biogenic silica at the soil-plant-atmosphere interface: benefits for investigating West African present and past water cycles, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9089, https://doi.org/10.5194/egusphere-egu25-9089, 2025.

The Bassée Observatory, located in the heart of the Seine catchment and part of the Zone Atelier Seine network, is an essential research platform for understanding the hydrological processes associated with the strategic challenges of sustainable water resource management. It focuses on the behaviour of the alluvial plain as a complex and anthropised hydrosystem, considering its long-term geohistorical evolution. Through an extensive network of surface water and groundwater monitoring stations, the observatory highlights the central role of groundwater and its interactions with surface water in the current dynamics of this region. We introduce the new groundwater model of the Bassée, developed as a tool combining the CaWaQS hydrogeological platform with the groundwater utilities of the PEST parameter estimation approach. This integration improves the representation of the heterogeneity of the alluvial plain and provides a solid basis for quantitative decision making. The model is designed to assist stakeholders in addressing the challenges of operating and conserving the alluvial plain in the context of a changing environment.

How to cite: Jost, A., Saias, C., and Renaud, A.: Groundwater modelling in the Bassée alluvial plain: A tool for understanding the dynamics of a complex socio-hydrosystem, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9355, https://doi.org/10.5194/egusphere-egu25-9355, 2025.

EGU25-10061 | ECS | Posters on site | HS8.2.8

Quantifying hydrogeological drivers influencing daily fluctuations in shallow groundwater levels within an altiplanic pristine catchment in Chile. 

Amanda Peña-Echeverría, Cristina Contreras, Jorge Renaud, Sarah Leray, and Francisco Suárez

Daily fluctuations in shallow groundwater levels provide valuable insights into hydro-ecological dynamics and aquifer hydraulic properties. These fluctuations usually depend on hydrological/hydrogeological processes, such as precipitation, evaporation, snow/ice melting/thawing, as well as soil characteristics that influence aquifer response times. The Salar del Huasco basin (20.2°S, 68.8°W; 4,164 m a.s.l.; 1,470 km2) is an endorheic system located in the arid Chilean Altiplano, hosting wetlands and a saline lagoon that sustain part of the region essential biodiversity such as chilean, andean, parina and chica’s flamingos, and it serves as a refuge for migratory birds (e.g., peregrine falcon, golden plover and baird's sandpiper). The area experiences extreme thermal oscillations (4–14°C daily averages; winter lows of -20°C), high potential evaporation (1,200 mm/year), and variable summer precipitation (11–400 mm/year). To explore shallow groundwater dynamics, we monitored for ~1 year two sites near the basin’s salt flat: the north and the south sites. Meteorological, soil, and groundwater levels data were collected at 30-min intervals. At the northern site, daily groundwater level fluctuations ranged from 6 to 45 mm, with a sharp and abrupt 300 mm rise in austral spring. In contrast, the southern site showed daily groundwater level fluctuations between 7 and 58 mm, with multiple rises during winter, ranging from 100 to 300 mm. Distinct patterns emerged at these sites: in the northern site, the maximum diurnal fluctuations correlated with solar radiation, while the southern site showed a more stable behavior, with no clear daily peaks. We applied a water balance to determine how the amplitude of possible input and output fluxes in the system altered the daily level fluctuations, and whether, despite the proximity of both sites (~9 km), soil texture, vegetation cover, and local meteorological-hydrogeological conditions explain the differences in groundwater level behavior.

How to cite: Peña-Echeverría, A., Contreras, C., Renaud, J., Leray, S., and Suárez, F.: Quantifying hydrogeological drivers influencing daily fluctuations in shallow groundwater levels within an altiplanic pristine catchment in Chile., EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10061, https://doi.org/10.5194/egusphere-egu25-10061, 2025.

In times of climatic unpredictability driven by a quickly changing climate, it is critical to investigate hydrological processes and water availability in different climatic and geomorphological contexts. Mountains have long been acknowledged as fundamental sources of abundant high-quality water for the densely populated downstream areas. The large volumes of water stored in mountain lakes, reservoirs, and snow caps are extremely important to buffer precipitation variability and sustain ecological and anthropic water uses during droughts. So far, the flow and storage of water in the deeply fractured rock formations constituting the core of mountain massifs have mostly been neglected, even for the long-term water balance. However, recent experimental evidence has shown that poorly porous and conductive fractured bedrock can host aquifers whose contribution to streamflow can be substantial, particularly during droughts.

This study systematically assesses, under a wide range of geomorphoclimatic conditions, how deep subsurface storage and flows affect critical hydrological and hydrogeological variables such as the age of streamflow (as opposed to the age of baseflow), surface seepage, and permanent drainage density. These critical hydrological processes are investigated via a large set of steady-state numerical experiments by modulating surface topography, groundwater recharge, and hydrogeological properties of the subsurface (e.g., formation depth, hydraulic conductivity, and its heterogeneity).

The results quantitatively show, for example, how different morphological and hydrogeological conditions may respond to climate change and can be useful in identifying vulnerable areas where mitigation strategies should be prioritized to cope with water shortages. The study can also help understand where ecological alterations driven by the lack of water can have a more profound impact on riverine habitats and where to expect the shift of species in the future.  

How to cite: Bellin, A. and Betterle, A.: Assessing the Impact of Deep Subsurface Storage and Flows on Hydrological Processes and Water Availability in Mountainous Regions, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11743, https://doi.org/10.5194/egusphere-egu25-11743, 2025.

EGU25-12719 | Orals | HS8.2.8

 Infiltration depth, rooting depth, and regolith flushing—A global perspective 

Gonzalo Miguez-Macho and Ying Fan

How deep does the rain regularly infiltrate into the ground? Do plant roots follow? How much infiltration is pumped back to the atmosphere (short-circuiting)  and how much passes below plant roots reaching the water table, flushing the regolith, recharging aquifers and rivers, and eventually reaching the ocean (long-circuiting) thus regulating global biogeochemical cycles and long-term climate? What is the depth that supplies evapotranspiration, and what is the regolith flush rate? What are the implications to global material and energy cycles? The answers depend on local climate–terrain–vegetation combinations. We use observations and high resolution numerical modeling at the global scale to shed light on multiscale causes–feedbacks among climate, drainage, substrate, and plant biomass that interactively create a global structure in the depths and rates of hydrologic plumbing of the Earth's critical zone, informing global models on critical depths and processes to include in Earth-system predictions.

How to cite: Miguez-Macho, G. and Fan, Y.:  Infiltration depth, rooting depth, and regolith flushing—A global perspective, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12719, https://doi.org/10.5194/egusphere-egu25-12719, 2025.

EGU25-12772 | Orals | HS8.2.8

Groundwater controls on headwater stream dynamics 

Clément Roques, Ronan Abhervé, Etienne Marti, Ronny Figueroa, Nicolas Cornette, Alexandre Gauvain, Jean-Raynald de Dreuzy, Sarah Leray, Camille Bouchez, Alexandre Boisson, Luc Aquilina, and Philip Brunner

Headwater catchments, defined as the uppermost segments of drainage networks with intermittent and/or perennial third-order streams, are vital sources of freshwater and nutrients for downstream river basins. Despite their critical role in sustaining natural ecosystems and supporting human services, these systems remain poorly understood and are often referred to as 'aqua incognita1.' A key challenge lies in unraveling the hidden groundwater processes that contribute to storage-discharge dynamics. Recent advances in both in-situ and remote monitoring, combined with innovative modeling techniques, now offer opportunities to capture the complex interactions between surface and subsurface processes across diverse climatic, topographic, and geological contexts.

In this presentation, we will present recent findings from field investigations conducted in headwater observatories, complemented by numerical modeling experiments designed to evaluate the controls of key geomorphic factors on groundwater-surface water interactions. The presentation will explore how landforms, lithologies, subsurface stress, and faults shape hydrological behaviors, including stream baseflow recession, groundwater seepage distribution, flow intermittency, and water residence times. Additionally, we will highlight advances in numerical modeling techniques, particularly through the HydroModPy community modelling platform2, which enhance the representation and calibration of groundwater processes in catchment-scale hydrological models. Through the application of these models on pilot sites, we will illustrate how subsurface heterogeneity influences the predictions of water availability under future climate change scenarios, emphasizing the importance of integrating hydrogeological insights for supporting resilient water resource management.

1 Bishop, K., Buffam, I., Erlandsson, M., Fölster, J., Laudon, H., Seibert, J., Temnerud, J., 2008. Aqua Incognita: the unknown headwaters. Hydrological Processes 22, 1239–1242. https://doi.org/10.1002/hyp.7049

2 Gauvain, A., Abhervé, R., Coche, A., Le Mesnil, M., Roques, C., Bouchez, C., Marçais, J., Leray, S., Marti, E., Figueroa, R., Bresciani, E., Vautier, C., Boivin, B., Sallou, J., Bourcier, J., Combemale, B., Brunner, P., Longuevergne, L., Aquilina, L., and de Dreuzy, J.-R.: HydroModPy: A Python toolbox for deploying catchment-scale shallow groundwater models , EGUsphere [preprint], https://doi.org/10.5194/egusphere-2024-3962, 2025.

How to cite: Roques, C., Abhervé, R., Marti, E., Figueroa, R., Cornette, N., Gauvain, A., de Dreuzy, J.-R., Leray, S., Bouchez, C., Boisson, A., Aquilina, L., and Brunner, P.: Groundwater controls on headwater stream dynamics, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12772, https://doi.org/10.5194/egusphere-egu25-12772, 2025.

EGU25-12792 | ECS | Orals | HS8.2.8

Deep flow behavior and the critical zone in a deep well: A hydrogeological study in Mexico City 

Zaida Martínez Casas, Eric Morales Casique, Selene Olea Olea, and Jose Luis Lezama Campos

In Mexico City, where population growth has significantly increased water demand, a well was drilled to a vertical depth of 1992 meters. 
To understand the groundwater dynamic in the critical zone- an area extending from the surface to the base of the groundwater system, where complex interactions occur between the atmosphere, lithosphere, hydrosphere, and biosphere- various tools were employed, including geophysical log analysis, pumping tests, and groundwater sampling for hydrochemical and isotopic (stable and radioactive) analyses.

The results revealed consistent ion concentrations during hydrogeochemical monitoring, classifying the water as sodium-chloride type with minor nitrate contamination attributed to the use of drilling mud.

Isotopic analysis indicated that the water likely originated from precipitation infiltrating at approximately 3000 meters above sea level, possibly from nearby mountain ranges. Radiocarbon dating estimated a residence time of 2840 years, although additional testing is necessary for confirmation.

Hydraulic tests determined a transmissivity of 768 m²/day and a specific storage of 3.11 × 10⁻⁶ m⁻¹, corresponding to an average hydraulic conductivity of 0.885 m/day. This is a complex hydrogeological system characterized by deep, highly fractured saturated zones. Groundwater in this well originates from the deep infiltration of rainfall in the surrounding sierras, circulating through fractures in volcanic rocks. Initially, the water quality showed temporary mixing with surface water due to the interaction between formation water and drilling mud; however, it later exhibited a distinct chemical composition.

The residence time of the water indicates a dynamic system with varying water ages. The results suggest hydraulic connectivity between different hydrogeological units and an endorheic behavior of groundwater flow in the area. In summary, this study enhances the understanding of groundwater flows in Mexico City, emphasizing the critical zone's role in shaping subsurface processes and highlighting the importance of considering the complexity of these systems for sustainable management.

How to cite: Martínez Casas, Z., Morales Casique, E., Olea Olea, S., and Lezama Campos, J. L.: Deep flow behavior and the critical zone in a deep well: A hydrogeological study in Mexico City, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12792, https://doi.org/10.5194/egusphere-egu25-12792, 2025.

EGU25-13922 | Orals | HS8.2.8

A Scale-Adaptive Framework for Modeling Critical Zone Processes and River Water Quality in the East River Watershed 

Dipankar Dwivedi, Ilhan Özgen Xian, Bhavna Arora, Boris Faybishenko, Michelle Newcomer, Patricia Fox, Carl Steefel, Kenneth Williams, Peter Nico, Susan Hubbard, and Eoin Brodie

Critical Zone processes encompass interactions among rock, soil, water, air, and living organisms, essential for quantifying water and nutrient fluxes and predicting downstream river water quality. High-fidelity reactive transport models (RTMs) are important for understanding Critical Zone processes but are typically computationally expensive, which limits their applicability across large catchments. To address these challenges, we developed a scale-adaptive reactive transport simulation framework that balances process fidelity with computational efficiency. We developed the RiverFlotran Module, which employs fully dynamic 1D shallow-water equations for river hydrodynamics, and integrated it into PFLOTRAN, a subsurface reactive transport model. This integration enables us to simulate bidirectional exchanges at the land-water interface. Subsequently, we developed a machine learning-based exchange function, trained on the simulated data, and tailored for the East River. This function allows us to predict river water quality along the river continuum. This framework was applied to the East River Mountainous Watershed in Colorado, a study site of Berkeley Lab's Watershed Function Scientific Focus Area, to demonstrate its effectiveness in capturing intricate Critical Zone interactions and predicting downstream river water quality. Our study of the East River Floodplain's alluvial aquifer revealed that prevailing anoxic conditions generate pronounced redox gradients, resulting in the downstream export of dissolved iron and nitrogen near meander bends. These bends consistently serve as nitrogen hotspots, irrespective of water levels, driven by variations in river stage, bathymetry, and meander geometry, such as sinuosity. This modeling framework provides a foundation for quantifying river water quality at the catchment scale.

How to cite: Dwivedi, D., Özgen Xian, I., Arora, B., Faybishenko, B., Newcomer, M., Fox, P., Steefel, C., Williams, K., Nico, P., Hubbard, S., and Brodie, E.: A Scale-Adaptive Framework for Modeling Critical Zone Processes and River Water Quality in the East River Watershed, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-13922, https://doi.org/10.5194/egusphere-egu25-13922, 2025.

EGU25-14421 | Posters on site | HS8.2.8

Groundwater dynamics in a steep Himalayan catchment: the role of a widespread weathering layer in water storage and transfer 

Kapiolani Teagai, John-Joseph Armitage, Niels Hovius, Léo Agélas, Nobuaki Fuji, Luc Illien, Basanta Raj Adhikari, and Christoff Andermann

The Himalayan region is crucial for providing water resources to millions of people in downstream regions across Asia. However, the processes governing groundwater storage and flow in steep mountain catchments remain poorly understood, particularly regarding the interplay between monsoonal rainfall, infiltration, and groundwater recharge in these highly dynamic landscapes. This study investigates the Kahule Khola watershed in central Nepal, combining field-based approaches encompassing Electrical Resistivity Tomography (ERT), infiltration measurements, and hydrogeochemical analyses, to investigate the pathways and storage mechanisms of groundwater across pre-, during, and post-monsoon seasons. Our findings highlight the critical role of a laterally extensive weathering layer, 10–25 m thick, in regulating hydrological processes. The weathering layer exhibits high infiltration capacities (<24.1 cm/h) that exceed even intense monsoonal rainfall rates (<162.8 cm/h), allowing surface water to rapidly penetrate the subsurface and replenish groundwater stores. The 2D ERT profiles reveal seasonal variations in the saturation of this layer, with significant vertical and lateral flow dynamics linking it to deeper fractured bedrock aquifers. Hydrogeochemical analyses of spring water further demonstrate a bi-compartmentalized flow regimes, where fast and shallow pathways dominate during the monsoon, while slower and long-term storage within the fractured bedrock sustains perennial spring discharge and stream baseflow throughout the dry season. This study enhances our understanding of the hydrological functioning of steep mountain landscapes, emphasizing the dual role of the weathering layer as both a temporary water reservoir and a conduit for deeper aquifer recharge, demonstrating heightened efficiency during monsoon season. By proposing a conceptual model of water transfer and storage in Himalayan catchments, this research provides critical insights into groundwater processes that are fundamental for sustainable water resource management under increasing pressures from climate variability and tectonic activity.

How to cite: Teagai, K., Armitage, J.-J., Hovius, N., Agélas, L., Fuji, N., Illien, L., Adhikari, B. R., and Andermann, C.: Groundwater dynamics in a steep Himalayan catchment: the role of a widespread weathering layer in water storage and transfer, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14421, https://doi.org/10.5194/egusphere-egu25-14421, 2025.

EGU25-14670 | Orals | HS8.2.8

Extreme Winter Precipitation Drives Recharge of Deep Mountain Groundwater 

W. Payton Gardner, Matthew Swarr, Donald Argus, Hilary Martens, Zachary Young, and Zachary Hoylman

Extreme winter precipitation events, associated with frequent and intense atmospheric rivers, deposit significant quantities of water in mountain regions over short periods of time. Precipitation is forecast to become more variable as climate change intensifies; however, it is unclear how that will affect mountain aquifer recharge. Here we use high-precision Global Navigation Satellite Systems (GNSS) surface displacements and elastic deformation models to surface loading to estimate total water storage changes.  Using independent estimates of water stored within shallow subsurface and surface reservoirs, we isolate changes in mountain groundwater storage in two important mountain regions of the western US at high spatial (~30km) and temporal (~ 1 week) resolution. We find that groundwater storage is the dominant component of long-term total water loss within the Sierra Nevada and Cascades, composing up to 95% of the total water lost over the past two decades. However, extremely wet winters, such as that of 2023, can recharge groundwater storage by more than twice the average annual amount, driving the state of groundwater storage from historical lows to above or near-normal conditions over relatively short periods. Further, we find gains in groundwater storage associated with these events are relatively durable, persisting over several proceeding years following the extreme recharge event. Mountain aquifers have been increasingly recognized as a dynamic and critical source of water storage and release to adjacent low-elevation communities; however, persistent declines in mountain aquifer storage have been observed across the western US over the past two decades. In a future with increasingly variable precipitation, the strong influence of extreme events may act to maintain mountain groundwater, sustaining ecosystem health and buffering adjacent areas against drought conditions in between events.

How to cite: Gardner, W. P., Swarr, M., Argus, D., Martens, H., Young, Z., and Hoylman, Z.: Extreme Winter Precipitation Drives Recharge of Deep Mountain Groundwater, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-14670, https://doi.org/10.5194/egusphere-egu25-14670, 2025.

EGU25-15141 | ECS | Orals | HS8.2.8

Down Under(ground) – Introducing the Australian Critical Zone Observatory Network 

Simone Gelsinari, Konrad Miotliński, Matthias Leopold, Jessie Weller, and Sally Thompson

The growing global network of Critical Zone Observatories provides exciting insights into how terrestrial and subsurface environments are interconnected, emphasising the value of understanding the Critical Zone as a vertically integrated system.  Yet this network is situated overwhelmingly in the frequently young and post-glacial or glacially-influenced landscapes of the Northern Hemisphere.  The Southern Hemisphere offers diverse landscapes with geologic parent materials spanning the Archaean to the Cenozoic, which have experienced little glaciation relative to the Northern Hemisphere.  The Australian Critical Zone Observatory Network was established in 2020 to provide insights into the structure and functioning of such landscapes on the ancient, chemically depleted, dry and diverse Australian continent. Five sites have been established with a common suite of instrumentation and operating principles, and are working collaboratively to develop Critical Zone datasets in landscapes ranging from rainforest to eucalyptus woodlands, dryland mallee, tropical savannah and rain-dependent agricultural lands.

This talk will introduce the OzCZO – the Australian Critical Zone Observatory Network, the five sites, their instrumentation and opportunities for scientific research within and by making comparisons among the sites.  It will then share some of the initial observations being collected at one of the observatories – the ancient lateritic landscape of the Avon Critical Zone Observatory.  We will illustrate how CZ structure, illuminated by bore logs and geophysics, organises soil physical and chemical properties across the landscape, and reveal how these properties then feed into land management decisions, hydrological functioning, and large-scale ecological health.  The Avon CZO is located within a biodiversity hotspot in the South-West of Australia, where the health of land and waters, and the ecosystems and agricultural production that depend on them, is threatened by both dryland salinity and a drying climate – with outcomes all mediated by the Critical Zone.

All data from OzCZO will be publicly available for use, and the sites are intended to act as an open platform where researchers can develop and test their ideas.  Given the scope for valuable cooperation and comparisons across these sites, we invite researchers at EGU to engage with OzCZO and keep progressing towards a global Critical Zone science.

How to cite: Gelsinari, S., Miotliński, K., Leopold, M., Weller, J., and Thompson, S.: Down Under(ground) – Introducing the Australian Critical Zone Observatory Network, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15141, https://doi.org/10.5194/egusphere-egu25-15141, 2025.

EGU25-17724 | ECS | Posters on site | HS8.2.8

Understanding surface - groundwater interactions in central European upland catchments: the Ahr valley, Germany 

Benoit Abadie, Laura Fracica, Christoff Andermann, Niels Hovius, Michael Dietze, and John Armitage

With a changing climate, major flood events are an increasing risk in many parts of the world, including temperate zones in Western Europe. Recent examples of destructive flooding in central European upland catchments, such as the 2021 Eifel floods in western Germany, highlight the importance of improving our understanding of the mechanisms behind stream response and sediment transport to precipitation events in upland catchments in temperate Western-Europe. The HIdden water and LANDscape ERosion (HILANDER) project that started in spring 2024 has two major goals: 1. To put in place an observatory in the Ahr catchment to characterize how water travels through the critical zone. 2. To incorporate surface/groundwater interactions in models of landscape evolution and river erosion.

The Ahr valley, ranging from 50m to 737m of elevation, is characterized by gently sloped hilltops and a steep, incised river valley. Preliminary recession analyses of the Ahr catchment, performed on data from four existing hydrographs, show a faster flowing aquifer in the upper parts of the catchment and a slow flowing aquifer in the lower regions. This implies that the upper parts of the catchment may be dominated by sub-surface flow through a more permeable shallow layer whereas the streamflow in lower reaches of the catchment is dominated by the deeper underlying aquifer. Two sub-catchments of the upper Ahr river, the Michelsbach, mainly forested and the Huhnenbach, largely agricultural with engineered drainage systems were chosen as study sites. The catchments are instrumented with pressure sensors, turbidimeters and seismometers, to continuously measure streamflow, suspended sediment concentrations, bedload transport and groundwater saturation. Furthermore, springs have been mapped and sampled for stable isotopes, dating and major elements.

Springs are found at both high and low elevations within both sub-catchments, and the locations of these springs do not vary from summer to winter. Observations from the summer spring mapping campaign of June 2024 found that the age of spring-water at high elevation is a mix of young water (ages of 2 to 3 years) and old water (age of 16 years). The presence of both young and old components in the spring water implies multiple pathways for groundwater within the catchment. In January 2025 we found that the ridge tops were saturated with substantial ponding of surface water. Down slope there was either diffuse release of this water or point release at the same locations of springs that were mapped and sampled in the summer. This, along with higher winter oxygen saturation in the springs, points to the potential for interflow during high rainfall events, where water flows laterally through the shallow soil and rock moisture layers (weathering zone) mixing with the groundwater supply. The future continuous monitoring in this critical zone observatory will give insight to the interplay between lateral water pathways in the weathering zone, and deep groundwater reservoirs allowing for a better understanding of how water flow through the catchments can impact erosion and landscape evolution.

How to cite: Abadie, B., Fracica, L., Andermann, C., Hovius, N., Dietze, M., and Armitage, J.: Understanding surface - groundwater interactions in central European upland catchments: the Ahr valley, Germany, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17724, https://doi.org/10.5194/egusphere-egu25-17724, 2025.

EGU25-18247 | Posters on site | HS8.2.8

The International Soil Moisture Network (ISMN): A global hub for in situ observations serving earth system science 

Matthias Zink, Tunde Olarinoye, Fay Boehmer, Kasjen Kramer, Stephan Dietrich, and Wolfgang Korres

Soil moisture is a critical component of the Earth’s hydrological cycle, influencing weather, climate, agriculture, and ecosystems. In situ soil moisture measurements are indispensable for validating satellite observations, calibrating hydrological and land surface models, and advancing our understanding of regional and global water cycles. Unlike remote sensing, in situ measurements provide direct observations of soil moisture variability across temporal and spatial scales, offering a benchmark for numerous environmental applications.

The International Soil Moisture Network (ISMN) serves as a vital repository of harmonized in situ soil moisture data collected from diverse networks worldwide. Since its inception, the ISMN has integrated measurements from over 80 networks with more than 3000 stations at various depths, standardizing and curating them to ensure accessibility and comparability. Beyond offering comprehensive in situ soil moisture data, ISMN disseminates additional environmental variables, including soil temperature, snow depth, snow water equivalent, precipitation, air temperature, surface temperature and soil water potential if they are available from our data providers. ISMN’s quality control framework addresses inconsistencies and errors, enabling researchers and practitioners to confidently utilize its datasets for applications ranging from hydrological modeling to climate change studies. ISMN’s free data access (https://ismn.earth) has fostered global collaboration and supported hundreds of studies in Earth system science.

Ongoing efforts are concentrated on expanding the database by incorporating additional stations and networks from institutional or governmental sources. Further resources are directed towards fortifying the operational system and improve usability to better serve our users. ISMN further contributes to the data-to-value chain on international initiatives like WMO, FAO and GCOS. One example is the contribution to WMO’s yearly Global State of the Water Resources report.  To enhance data quality, ISMN is researching AI-based methods for detecting anomalies such as spikes, dips, and plateaus, showing promising initial results.

How to cite: Zink, M., Olarinoye, T., Boehmer, F., Kramer, K., Dietrich, S., and Korres, W.: The International Soil Moisture Network (ISMN): A global hub for in situ observations serving earth system science, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18247, https://doi.org/10.5194/egusphere-egu25-18247, 2025.

EGU25-18699 | ECS | Orals | HS8.2.8

Elaboration of a geological and hydraulic mapping project of infiltrability potential on the Aix-Marseille Provence Metropole (SE, France) 

Lilas Ruttyn, François Fournier, Philippe Leonide, Borgomano Jean, Bruno Arfib, Sophie Viseur, Laurent Goulet, Olivier Vignoulle, and Narimane Zaabar

The Aix Marseille Provence metropolitan area experiences rapid urbanization that reinforces the need for infrastructure and implies considerable sealing of the substratum This region is a typical arid and Mediterranean environment where rain precipitation can be exceptionally catastrophic. This two factors creates runoff, overflow and flooding in the urban area. One solution to manage the flooding and overflow is to allow more water to penetrate into the soil, by removing the impermeable and anthropic materials where the geological substratum is naturally able to infiltrate the water.

Usually, standard parameters such as: topography, drainage density and hydrological balances, are used to estimate runoff and indirectly find the infiltrability values and ultimately tackle infiltration problematics. These approaches are informatic and mathematics-based that work in a small, delimited and homogeneous area. To integrate this problematics to large scale and heterogenous systems, reservoir geology concepts such as geomorphology, uncertainties of scale change processes or structural geology can be addressed. Therefore, this project aims to understand the geological processes that controls the infiltration potential in the geological substratum and its spatial distribution for the purpose of creating an infiltrability map of the Aix Marseille metropolis.

The goal of this study is to develop a method for predicting the infiltration capacity on a large scale and heterogenous area including urban zone. This involves acquiring local observational data points which classify rock outcrops in 4 “hydraulic types” (HT) defined as follows: HT-1 represents impermeable rocks or soils, where no infiltration is possible; HT-2 represents thin soils with variable porosity and permeability; HT-3 describes rocks with low to very high matrix porosity influenced by clay matrix presence and variable permeability; HT-4 describes rocks with fractures and/or karst networks with low to very high permeability depending on fracture/cavity density, with variable porosity. With the geolocated data points, a map is created on QGIS (a Geographic Information System free software) in order to up-scale the hydraulic types over a larger scale grid by spatial interpolation.

For an even acquisition area, geological heterogeneity and accessibility of outcrops determines the data number needed to upscale hydraulic types. This approach is well-known in reservoir geology and this large-scale project is the opportunity to apply the methodology to  hydrogeology field.

Additionally, to address the lack of visibility of outcrops, subsurface data (shallow well data from the BRGM, Bureau of Geological and Mining Research) will be combined with field observations. Furthermore, a calibration of this method will be required to quantify and to establish thresholds within the Hydraulic Types classification. This project will ultimately provide specific values for infiltration capacity and facilitate flood risk management without having to use complex and costly technologies.

 

Keywords : SIG mapping, infiltration, runoff, geological substratum, stratigraphy, structural geology, heterogeneity, precipitation, de-sealing, available water

 

How to cite: Ruttyn, L., Fournier, F., Leonide, P., Jean, B., Arfib, B., Viseur, S., Goulet, L., Vignoulle, O., and Zaabar, N.: Elaboration of a geological and hydraulic mapping project of infiltrability potential on the Aix-Marseille Provence Metropole (SE, France), EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18699, https://doi.org/10.5194/egusphere-egu25-18699, 2025.

EGU25-20210 | Posters on site | HS8.2.8

Integrating Data into the Hydrogeophysical Model: A Case Study of the Orgeval Critical Zone Observatory 

Agnès Rivière, Ludovic Bodet, Maxime Gautier, Alexandrine Gesret, Roland Martin, Sylvain Pasquet, Nicolas Radic, Jose Cunha Teixeira, Marine Dangeard, and Didier Renard

Quantifying the water and heat fluxes at the interface between surface water (SW) and groundwater (GW) is a key issue for hydrogeologists to consider for safe yield and good water quality. However, such quantification with field measurements is not straightforward because the SW-GW changes depend on the boundary conditions and the spatial description of the hydrofacies, which aren't well known and are usually guessed by calibrating models using standard data like hydraulic heads and river discharge. We provide a methodology to build stronger constraints to the numerical simulation and the hydrodynamic and thermal parameter calibration, both in space and time, by using a multi-method approach. Our method, applied to the Orgeval Critical Zone Observatory (France), estimates both water flow and heat fluxes through the SW-GW interface using long-term hydrological data, time-lapse seismic data, and modeling tools. We show how a thorough interpretation of high-resolution geophysical images, combined with geotechnical data, provides a detailed distribution of hydrofacies, valuable prior information about the associated hydrodynamic property distribution. The temporal dynamic of the WT table can be captured with high-resolution time-lapse seismic acquisitions. Each seismic snapshot is then thoroughly inverted to image spatial WT variations. The long-term hydrogeological data (such as hydraulic head and temperature) and this prior geophysical information are then used to set the parameters for the hydrogeological modeling domain. The use of the WT geometry and temperature data improves the estimation of transient stream-aquifer exchanges. Future developments to achieve the fully coupling of the hydrogeophysical model will be presented.

How to cite: Rivière, A., Bodet, L., Gautier, M., Gesret, A., Martin, R., Pasquet, S., Radic, N., Cunha Teixeira, J., Dangeard, M., and Renard, D.: Integrating Data into the Hydrogeophysical Model: A Case Study of the Orgeval Critical Zone Observatory, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20210, https://doi.org/10.5194/egusphere-egu25-20210, 2025.

Statistical models are a frequently used tool in hydrology, especially when it comes to estimating design floods, i.e. flood events that used to design flood protection systems or reservoirs. The often complex hydrological data, which are affected by e.g. missing values, extremes or time-varying processes, require sophisticated statistical models that take these challenges into account. As a scientist, developing such models can be a lot of fun and provide interesting insights. After months of thinking about the best model under certain statistical assumptions, proving asymptotic theorems and testing the model with synthetic data, you are happy and proud to have developed a new model. This model will hopefully be widely used in future research. The next step is to apply the model to a large real data set. The results look good on average. The results will be shared with practitioners, because of course you want the model to be useful for science and practice. And then: the phone call. You are told that your results are not plausible for a certain catchment area. And in general, the new model is not needed in practice because there is an established model. This example describes such a case and discusses ways of dealing with it. It is intended to illustrate the importance of communication between science and practice and a general understanding between both sides.

How to cite: Fischer, S.: When practical considerations impact your scientific model, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1620, https://doi.org/10.5194/egusphere-egu25-1620, 2025.

EGU25-1660 | Orals | EOS4.8

The Minkowski–Bouligand dimension of a clay brick 

Nick van de Giesen and John Selker

In the early 1990's, fractals and chaos were hot. In 1987, James Gleick had published "Chaos: Making a New Science", popularizing non-linear dynamics. Hydrologists played an important role in the development of fractal theory. Hurst had discovered that sequences of dry and wet years for the Nile showed very long memory effects. Instead of the chance of a dry year following a dry year being 50%, Hurst found that there were surprisingly many long series of dry or wet years. Seven fat years, seven lean years, as it is noted in Genesis. Scott Tyler found fractals in soils ("Fractal processes in soil water retention"). At Cornell, where we were at the time, David Turcotte described "Fractals in geology and geophysics". A few years later, Ignacio Rodríguez-Iturbe and Andrea Rinaldo would publish "Fractal River Basins: Chance and Self-Organization". In short, fractals were exciting scientific gold.

A fractal is not just an obscure mathematical object but something that can actually be found everywhere in nature. Early on, a paper was published in Nature with the title "Fractal viscous fingering in clay slurries" by Van Damme, Obrecht, Levitz, Gatineau, and Laroche. They "only" did an experiment on a fractal embedded in 2D; we should be able to do one better and find the fractal dimension of the surface of cracking clay embedded in 3D. So out we went, collected some clay, mixed it with water in a cement mixer, siliconed together a shallow "aquarium", and poured in the slurry. To observe the cracking of the drying slurry, a video camera was mounted above the experiment, looking down and taking time-lapse images. To access the views from the sides, mirrors were installed at 45 degrees at each of the four sides. Lights made sure the camera captured high quality images. The whole set-up was enclosed in a frame with dark cloth to ensure that lighting was always the same.  We already had some box-counting code ready to calculate the fractal dimension of the surface, called the Minkowski–Bouligand dimension. One variable needed some extra attention, namely the boundary between the clay slurry and the glass sides. If the clay would cling to the sides, it would be difficult to understand the effects that this boundary condition had on the outcome of the experiment. Moreover, the cracks may not have become visible in the mirrors when the sides were covered with mud. So, instead, it was decided to make the sides hydrophobic with some mineral oil. This ensured that when the clay would start to shrink, it would come loose from the sides. Now, all we had to do was wait. It took only a week or so before the consolidated slurry started to shrink and to come loose from the sides. After that, the clay continued shrink for many weeks. This is how we learned that the fractal dimension of a shrinking brick of clay is (very close) to 3.0. 

How to cite: van de Giesen, N. and Selker, J.: The Minkowski–Bouligand dimension of a clay brick, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-1660, https://doi.org/10.5194/egusphere-egu25-1660, 2025.

EGU25-5035 | Orals | EOS4.8

Landslides and hillslope erosion increase relief 

Stefan Hergarten and Jörg Robl

In 2018, we found exciting new results in landform evolution modeling by coupling the two simplest models of fluvial erosion and hillslope processes. While the stream-power incision model is the simplest model for detachment-limited fluvial erosion, the diffusion equation is the simplest description of hillslope processes at long timescales. Both processes were added at each grid cell without an explicit separation between channels and hillslopes because fluvial erosion automatically becomes dominant at large catchment sizes and negligible at small catchment sizes.

We found that increasing diffusion reduces the relief at small scales (individual hillslopes), but even increases the large-scale relief (entire catchments). As an immediate effect, the hillslopes become less steep. In turn, however, we observed that the network of the clearly incised valleys, which indicates dominance of fluvial erosion over diffusion, became smaller. So a smaller set of fluvially dominated grid cells had to erode the material entering from the hillslopes. To maintain a morphological equilibrium with a given uplift rate, the rivers had to steepen over long time. This steepening even overcompensated the immediate decrease in relief of the hillslopes.

This result was counterintuitive at first, but we were happy to find a reasonable explanation. So we even prepared a short manuscript for a prestigious  journal. We just did not submit it because we wanted to explain the effect quantitatively from the physical parameters of the model. From these theoretical considerations, we found that our numerical results did not only depend on the model parameters, but also on the spatial resolution of the model and noticed that this scaling problem was already discussed in a few published studies. Beyond the scaling problem, we also realized that applying the concept of detachment-limited fluvial erosion to the sediment brought from the hillslopes into the rivers is quite unrealistic. A later study including fluvial sediment transport and a model for hillslope processes that avoids scaling problems did not predict any increase in large-scale relief. So we finally realized that our original findings were mainly the result of a specific combination of models that should not be coupled this way and are not  as relevant for landform evolution as we thought.

This example illustrates many of the pitfalls of numerical modeling beyond purely technical issues. In particular, combining models that are widely used and make sense individually may still cause unexpected problems.

 

How to cite: Hergarten, S. and Robl, J.: Landslides and hillslope erosion increase relief, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5035, https://doi.org/10.5194/egusphere-egu25-5035, 2025.

EGU25-5091 | ECS | Orals | EOS4.8

(Re)(De)bugging tragedies with Hector 

Guillemette Legrand

In this presentation, I will discuss my research into the simple climate model Hector, which calculates temperature change based on the impact of various climate scenarios. More specifically, I will discuss how an artistic-led approach through (un)voluntary-caused computational bugs can help document the model's logic and socio-political implications. I will describe methods for collective 'debugging' to produce transdisciplinary knowledge (beyond solely scientific inquiry) to foster conversation about the potential and limits of current climate infrastructure to foster concrete climate actions. This research investigates the field of climate science through artistic practice, software and infrastructure studies, and participatory methods. To expand on the role of bugs in my investigation, I will elaborate on concrete examples of differences in perception of 'error' in the fields of arts and science, looking at case studies where mistakes or glitches have been valorised and mobilised through artistic practice to grapple with, appropriate, and/or repurpose scientific instruments.

How to cite: Legrand, G.: (Re)(De)bugging tragedies with Hector, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5091, https://doi.org/10.5194/egusphere-egu25-5091, 2025.

EGU25-5951 * | Orals | EOS4.8 | Highlight

Improving extreme temperature definitions until they are wrong 

Lukas Brunner, Maximilian Meindl, and Aiko Voigt

"Doesn't this look a bit strange?" 

It began with an innocent question during one of our Master's colloquia. And it could have ended there. "We were just following an approach from the literature". And who could argue against following the literature?

But it bugged me. During a long train ride, I began to think about the issue again. 10 hours and many papers later, I was only more confused: was it really that obvious, and why had no one picked up on it before? But sometimes the most obvious things are the most wicked, and after a few conversations with knowledgeable colleagues, I was sure we were in for an unexpected surprise. 

A commonly used approach to defining heat extremes is as exceedances of percentile-based thresholds that follow the seasonal cycle. Such relative extremes are then expected to be evenly distributed throughout the year. For example, over the 30-year period 1961-1990, we expect three (or 10%) of January 1s to exceed a 90th percentile threshold defined for the same period - and the same for all other days of the year. In a recent study, we show that there are many cases where this does not hold, not even close (Brunner and Voigt 2024).

Here, we tell the story of how this blunder spread in the literature out of the desire to improve extreme thresholds. We show that seemingly innocent changes can sometimes have unintended consequences and that taking the time to check the obvious can help avoid mistakes in science. 

 

Brunner L. and Voigt A. (2024): Pitfalls in diagnosing temperature extremes, Nature Communications, https://doi.org/10.1038/s41467-024-46349-x

How to cite: Brunner, L., Meindl, M., and Voigt, A.: Improving extreme temperature definitions until they are wrong, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-5951, https://doi.org/10.5194/egusphere-egu25-5951, 2025.

When economists estimate the expected economic damages from current-day CO2 emissions, they usually calculate the social cost of carbon – that is, the aggregated damage caused by the emission of an additional ton of CO2. Several cost-benefit integrated assessment models (IAMs) are built to assess this quantity, and among them is the META model. This model is built specifically to assess the effects of tipping points on the social cost of carbon, and it usually operates stochastically. When integrating a deterministic, but small carbon cycle tipping point into the model, however, the social cost of carbon seems to explode: a few gigatons of additional emissions almost double the impact estimates of CO2 emissions! Well, maybe. In fact, these results are a pure artifact of two things: 1) the way in which social cost of carbon estimates are calculated with IAMs; and 2) the way that tipping points are implemented in the META model. And, of course, 3): a lack of initial thoughtfulness on behalf of myself. A thorough look into this issue shows that, as expected, a marginal change in emissions leads to a marginal change in damage estimates. While that result is rather boring, the previous blunder can actually be instructive about the scarcely-known methods used to obtain economic impact estimates of climate change.

How to cite: Schaumann, F.: Drastic increase in economic damages caused by a marginal increase in CO2 emissions?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-9145, https://doi.org/10.5194/egusphere-egu25-9145, 2025.

EGU25-10285 | ECS | Orals | EOS4.8

How robust are modeled non-local temperature effects of historical land use changes really? 

Felix Jäger, Petra Sieber, Isla Simpson, David Lawrence, Peter Lawrence, and Sonia I. Seneviratne

Historically, large areas across the globe have been affected by deforestation or irrigation expansion. The replacement of forests with agricultural land and increased water availability in irrigated croplands altered the land’s surface properties, leading to influences of biogeophysical changes on near-surface temperature. From limited observations and mostly idealized simulations, we know that sufficiently large alterations of land surface properties can theoretically lead to systematic temperature and precipitation changes outside and even far from the altered areas. Not only the advection of temperature anomalies, but also changes in circulation and ocean feedbacks have been shown to be potential drivers of such non-local responses in single and multi-model studies.

We tested the robustness of non-local temperature signals to internal variability in the fully coupled Community Earth System Model 2 (CESM2) simulations of the historical period (1850 – 2014) with all forcings vs. all-but-land-use-change forcings. Doing so, we first found seemingly robust non-local temperature effects of land use change on the global and regional scale. But when accounting for the sampling of internal variability in the model using a large initial condition ensemble, the global scale signal was found to be indistinguishable from noise. Only regionally in some hotspots, we found robust and historically important non-local temperature signals. Through increasingly rigorous analysis, we reached a partly negative and unexpected but important finding, which may have implications for future assessments of comparably weak or spatially heterogeneous forcings to the Earth system.

How to cite: Jäger, F., Sieber, P., Simpson, I., Lawrence, D., Lawrence, P., and Seneviratne, S. I.: How robust are modeled non-local temperature effects of historical land use changes really?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10285, https://doi.org/10.5194/egusphere-egu25-10285, 2025.

EGU25-10615 | Orals | EOS4.8

Think twice – pitfalls in hydrological modelling 

Jan Seibert, Franziska Clerc-Schwarzenbach, Ilja van Meerveld, and Marc Vis

Failures are only common in science, and hydrological modelling is no exception. However, we modellers usually do not like to talk about our mistakes or our overly optimistic expectations and, thus, “negative” results usually do not get published. While there are examples where model failures indicated issues with the observational data, in this presentation the focus is on modelling studies, where some more (realistic) thinking could have helped to avoid disappointments. Examples include the unnecessary comparison of numerically identical model variants, naively optimistic expectations about increasing the physical basis of bucket-type models and excessively hopeful assumptions about the value of data.

How to cite: Seibert, J., Clerc-Schwarzenbach, F., van Meerveld, I., and Vis, M.: Think twice – pitfalls in hydrological modelling, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-10615, https://doi.org/10.5194/egusphere-egu25-10615, 2025.

EGU25-11357 | Orals | EOS4.8

Two steps forward, one step back: four years of progress and setbacks on invisible ship tracks 

Peter Manshausen, Anna Tippett, Edward Gryspeerdt, and Philip Stier

The idea of invisible ship tracks for the study of aerosol-cloud interactions sounds promising: We have been studying the effects of aerosols on clouds for many years, among others by investigating the bright lines of clouds left in low marine clouds by ships. However, only a small fraction of ships leaves behind visible tracks. This means we can only study aerosol-cloud interactions under certain meteorological conditions, biasing our understanding. Instead, by studying all clouds polluted by ships ('invisible ship tracks') with a methodology we developed, we should be able to get a full picture of aerosol-cloud interactions. A number of interesting and impactful results have come out of this research, along with several setbacks and corrections to initial results. Here, we examine them in order, showing how correcting for one identified bias can introduce two new ones. Unexpected glitches arise from sources as varied as: choices regarding ship track definition, retrieval geometry, specific weather systems biasing results, and mathematical subtleties. What can we conclude after four years of progress on this methodology? While some results still stand, others had to be significantly corrected. This makes us see invisible ship tracks as an example of research that is closer to a method of 'tinkering' than to a 'magnificent discovery'.

How to cite: Manshausen, P., Tippett, A., Gryspeerdt, E., and Stier, P.: Two steps forward, one step back: four years of progress and setbacks on invisible ship tracks, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-11357, https://doi.org/10.5194/egusphere-egu25-11357, 2025.

EGU25-12720 | ECS | Posters on site | EOS4.8

Physical understanding of bugs to improve the representation of the climate system   

Hans Segura, Cathy Hohenegger, Reiner Schnur, and Bjorn Stevens

Earth system models are important tools used to understand our climate system and project possible changes in our climate due to anthropogenic and natural forcings. Human errors can occur in the development of Earth System models, i.e., bugs, giving an unphysical representation of our climate. A way to identify and solve bugs is to apply physical concepts. Here, we present an experience that occurred in the development of the ICOsahedral Non-hydrostatic model (ICON) as a kilometer-scale Earth System model, in which physically understanding a bug in the surface energy budget fixed land precipitation. 

In a simulation of ICON, referred to as ICON-bug, precipitation over tropical land continuously decreased across the simulation. This led to a ratio of land-ocean precipitation in the tropics of less than 0.7, which, otherwise, should be more than 0.86. As part of the possible explanations, the surface energy budget over land was targeted as a culprit. This idea relies on the influence of the interaction between soil moisture, surface heat fluxes, and winds to generate circulation favoring precipitation over dry land surfaces (Hohenegger and Stevens 2018). Indeed, the surface energy budget over dry surfaces in the ICON-bug showed an error in sensible heat flux. The sensible heat flux transmitted to the atmosphere was 70% of what was calculated for the surface module. Fixing this error closed the surface energy budget and increased land precipitation over the tropics, leading to a ratio of land-ocean precipitation of 0.94, close to observations. 

How to cite: Segura, H., Hohenegger, C., Schnur, R., and Stevens, B.: Physical understanding of bugs to improve the representation of the climate system  , EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-12720, https://doi.org/10.5194/egusphere-egu25-12720, 2025.

Whenever you study a phenomenon of mm to a few cm-scale in the laboratory which involves an interface, the question of surface tension arises. Surface tension is due to the fact that molecules prefer to stay with their own kind. Therefore, the creation of an interface between two fluids requires energy, and this influences the dynamics around the interface.

Surface tension can be a blessing: it produces the round shape of rain drops or the nice bubble shapes of colorful liquid in a lava lamp. It allows objects with a higher density to float on a liquid (such as an insect on water, or a silicone plate on sugar syrup). It can generate flow up a capillary.

However, it can also be a curse in the case of thermal convection. Purely thermal convection  develops when a plane layer of fluid is heated from below and cooled from above. The engine of motion is the thermal buoyancy of the fluid. This is what is happening in a planetary mantle on scales of hundreds to thousands kilometers. This is also what is happening in a closed box in the laboratory. But as soon as an interface exists, either between an upper and a lower experimental mantle, or in the case of a free surface at the top of the fluid layer, surface tension effects can become important. For exemple, the variation of surface tension with temperature was responsible for the beautiful honey-comb patterns imaged by Benard (1901) in the first systematic study of thermal convection with a free-surface. Surface tension is also going to act against the initiation of subduction (which acts to break the surface). 

We shall review in this presentation the signatures of surface tension in a convective context, and the different ways to minimize and/or remove the effects of surface tension in convection experiments, such as using miscible liquids, or a layer of experimental « sticky air ».

How to cite: Davaille, A.: Analog studies of mantle convection: the curse of surface tension (or not) ?, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15059, https://doi.org/10.5194/egusphere-egu25-15059, 2025.

EGU25-15457 | Orals | EOS4.8

The crux with variability: too much or too little 

Markus Weiler

In hydrology we measure and follow the water. What if there is too much or too little? It happens a lot. As a field hydrologist, I frequently have to determine the location of a measurement, the time to take the measurement, the location to set up a field experiment, or the amount of a tracer to inject to study a hydrological system. However, this is a very bumpy road, as variability is often not in favor of my decisions because the distribution is wider than expected, bimodal instead of unimodal, or the probability of an event is theoretically small, but still an extreme event occurs during our experiment. I will showcase some examples to demonstrate what I mean and what I experienced, as well as how frequently the PhD students or Postdocs have suffered as a result of my decisions or of the unexpected variability: Climatic variability resulted in a winter without snow, just as new sensors were already deployed. Or the winter snowpack was extremely high, preventing any work at high altitudes in the Alps until mid of July, thereby reducing our field season by half. An ecohydological study to observe the effects of drought in a forest with a rainout shelter was ineffective because it occurred during an extremely dry year, making the control just as dry as our drought treatment. The automatic water sampler was set-up to collect stream water samples, but it was washed away four weeks later by the 50-year flood. The calculated amount of artificial tracer was either way too low, because the transit times of the system were much longer than expected, or it was far too high, resulting in colored streams or samples that had to be diluted by a factor of 100 due to much faster transit times Finally, and most expensively, we installed many trenches along forest roads to measure subsurface stormflow but after three years, we abandoned the measurements because we never measured a drop of water coming out of the trenches, as the bedrock permeability was much higher due to many high permeable fissures that prevented the formation of subsurface stormflow.  These experiments or observations failed because of unexpected variability in input, system properties or a lack of technical variability in the equipment. I will reflect on residual risk of failure in fieldwork related to that crux and discus approaches to reduce this risk.

How to cite: Weiler, M.: The crux with variability: too much or too little, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15457, https://doi.org/10.5194/egusphere-egu25-15457, 2025.

EGU25-15826 | ECS | Posters on site | EOS4.8

Output regridding can lead to Moiré pattern in km-scale global climate model data from ICON 

Benjamin Poschlod, Lukas Brunner, Benjamin Blanz, and Lukas Kluft

The emergence of global km-scale climate models allows us to study Earth's climate and its changes with unprecedented local detail. However, this step change in spatial resolution to grid spacings of 10 km or less also brings new challenges to the numerical methods used in the models, the storage of model output, and the processing of the output data into actionable climate information. The latest versions of the ICON-Sapphire model developed in the frame of the NextGEMS project address these challenges by running on an icosahedral grid while outputting data on the so-called HEALPix grid. Both grids are unstructured grids, which avoids, for example, the issue of longitude convergence. In addition, HEALPix allows data to be stored in a hierarchy of resolutions at different discrete zoom levels, making it easier for users to handle the data.  

The transition from the native 10 km grid to the output grid is made by a simple but very fast nearest-neighbour remapping. An advantage of this simple remapping approach is that the output fields are not distorted, i.e. the atmospheric states in the output remain self-consistent. As HEALPix only provides discrete zoom levels in the setup of the run, it was decided to remap to the closest available resolution of 12 km rather than to the next finer resolution of 6 km. This decision was made to avoid artificially increasing the number of grid points and to avoid creating duplicates through the nearest neighbour remapping.

As a consequence of this approach, wave-like patterns can emerge due to the Moiré effect that can result from the interaction of two grids. We find these patterns when looking at certain derived precipitation extremes, such as the annual maximum daily precipitation, the 10-year return level of hourly precipitation, or the frequency of dry days. At first, we interpreted these patterns as a plotting issue, as the figures might have too low resolution to cope with the high-resolution global plot (aliasing) leading to a Moiré pattern.

However, zooming in on the affected regions and closer examination of the data revealed that the pattern is in fact in the data. Further investigation with synthetic data confirmed the suspicion that the Moiré pattern was indeed caused by the remapping of the native 10 km icosahedral grid to the slightly coarser 12 km HEALPix grid. We hypothesise that precipitation is particularly affected by this issue, as it typically contains many grid cells with zero precipitation, with local clusters of non-zero values at the 15-minutely output interval. Yet, we cannot exclude the possibility that other variables are also affected.

As a consequence, if remapping is required, it is recommended to first remap from the native resolution to a finer resolution grid. As a next step, the conservative nature of the HEALPix hierarchy can be used to compute the coarser level. In this way it is likely to be possible to avoid aliasing and still keep the amount of output data the same.

How to cite: Poschlod, B., Brunner, L., Blanz, B., and Kluft, L.: Output regridding can lead to Moiré pattern in km-scale global climate model data from ICON, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-15826, https://doi.org/10.5194/egusphere-egu25-15826, 2025.

EGU25-17676 | ECS | Posters on site | EOS4.8

What river plastic hotspots do not have in common 

Rahel Hauk, Adriaan J. Teuling, Tim H.M. van Emmerik, and Martine van der Ploeg

Plastic pollution is a global issue, across all environmental compartments. Rivers connect the terrestrial with the marine environment, and they transport various materials, among these plastic pollution. Rivers not only transport plastic, but also accumulate and store it, especially on riverbanks. In fact, plastic deposition and accumulation on riverbanks is a common occurrence. However, our understanding of why plastic is deposited on a certain riverbank is rather limited. Riverbanks along all major Dutch rivers have been monitored for plastic and other litter twice a year by citizen scientists, in some locations since 2018. This provides an extensive dataset on plastic accumulation, and we used these data with the aim of understanding the factors determining plastic concentration/accumulation variability over time and space. We tested multiple riverbank characteristics, such as vegetation, riverbank slope, population density, etc., hypothesized to be related to plastic litter. After having exhausted a long list of auxiliary data and analysis strategies, we found no significant results. Ultimately, we had a close look at ten consistent hotspots of macroplastic litter, along the Meuse, and Waal river. And once again, they seem to have nothing in common. But, there is a pattern, because some riverbanks have consistently very high densities of plastic litter so it does not seem completely random. We have been looking to explain spatial variability, whereas we might have to look at temporal consistency, and we shall not give up our efforts to bring order to this chaos.

How to cite: Hauk, R., Teuling, A. J., van Emmerik, T. H. M., and van der Ploeg, M.: What river plastic hotspots do not have in common, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17676, https://doi.org/10.5194/egusphere-egu25-17676, 2025.

EGU25-17811 | Posters on site | EOS4.8

Temporal variation of ambient noise at the Grande Dixence reservoir recorded by a nodal deployment 

Mita Uthaman, Laura Ermert, Angel Ling, Jonas Junker, Cinzia Ghisleni, and Anne Obermann

Grande Dixence, the tallest gravity dam in the world, is located in the Swiss Alps on the Dixence River with a catchment area of 4 km2 at a towering elevation of 2000m. The lake serves as a collecting point of melt water from 35 glaciers and reaches full capacity by late September, subsequently draining during winter and dropping to lowest levels in April. For a reservoir as large as the Grande Dixence, the variation in hydrological load can be expected to induce changes in crustal stress. The goal of this study was to harness the loading effect of the time-varying level of reservoir load as a source of known stress to investigate the variation in seismic velocity of the bedrock due to changes induced in crustal stress and strain rates. 22 seismic nodes were thus deployed along the banks of the reservoir which were operational from mid-August to mid-September, corresponding to the time period when the lake level reaches its maximum. Of the 22 nodes, 18 were deployed in closely spaced patches of six in order to carry out coherent stacking and to increase the signal-to-noise ratio, besides one group of three nodes and one single node. Measurement quality appears satisfactory: small local earthquakes are recorded well, and the probabilistic power spectral densities (PPSDs) computed for data quality validation evidence the ambient noise levels to be well within the global noise limits. However, the recorded noise is unexpectedly complex and, at periods shorter than 1 second, varies strongly by location. The 0.5--5s (0.2--2 Hz) period band at lakes generally records a diurnally varying noise level, often associated with lake generated microseism. Diurnal variations around 1 second of period are observed in our study as well. The amplitude of ambient noise level around 1 second of period is observed to be highest when the lake level changes, along with the prominent diurnal variation. A similar variation is observed in the seismic velocity variation (dv/v) computed from cross-correlated and auto-correlated ambient noise filtered between 0.5--1 Hz, with dv/v exhibiting a drop with rising lake level. These results provide preliminary evidence for possible change in crustal stress state with changing hydrological load. Future direction of this study consists of analytically modeling the results to quantify the influence of thermobarometric parameters on PPSDs and dv/v, and deconvolve it from the lake induced variations.

How to cite: Uthaman, M., Ermert, L., Ling, A., Junker, J., Ghisleni, C., and Obermann, A.: Temporal variation of ambient noise at the Grande Dixence reservoir recorded by a nodal deployment, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-17811, https://doi.org/10.5194/egusphere-egu25-17811, 2025.

EGU25-18185 | Orals | EOS4.8

Advancing river plastic research through serendipity and stupidity 

Tim van Emmerik and the WUR-HWM River Plastic Team

Rivers play an important role in the global distribution of plastic pollution throughout the geosphere. Quantifying and understanding river plastic pollution is still an emerging field, which has advanced considerably thanks to broad efforts from science, practice, and society. Much progress in this field has been achieved through learning from failures, negative results, and unexpected outcomes. In this presentation we will provide several examples of serendipity and stupidity that has led to new insights, theories, methods, and completely new research lines. We will share what we learned from rivers flowing in the wrong direction, sensors that disappear, equipment blocked by invasive plants, and dealing with suspicious local authorities. Pushing the science sometimes requires an opportunistic approach, embracing surprises and chaos you may face along the way.

How to cite: van Emmerik, T. and the WUR-HWM River Plastic Team: Advancing river plastic research through serendipity and stupidity, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18185, https://doi.org/10.5194/egusphere-egu25-18185, 2025.

With the advent of parallel programming in the late 1990s. A port of the than available Max Planck Institutes for Meteorology spectral atmospheric model echam5 to MPI and OpenMP was done. For testing and validation of the hybrid parallelization a coherence algorithm was developed. The implementation has been incorporated into todays NWP and climate model ICON as well. The coherence algoritm consists of several stages: first one MPI rank is running the serial model against an n-task MPI parallelized model. During runtime the state vector is checked for binary-identity. If successfull a m-task MPI version can be compared to an m-task MPI version for high processor counts. The same schema can be used OpenMP parallelization. ONe MPI task runs the model serial using one OpenMP thread and a second MPI task runs k OpenMP threads. Again, the results are compared for binary-identity. As the testing needs to be done automatically, bit-identity is important for testing not necessarily for production.

The tesing revealed plenty of problems during the initial parallelization work of echam5 and showed constant appearing problems in the ICON development phase.

However, far in a couple of century long simulation the bit-identity was just by accident found to be broken: the search of the cause started!

How to cite: Kornblueh, L.: MPI and OpenMP coherence testing and vaildation: the hybris of testing non-deterministic model code, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18400, https://doi.org/10.5194/egusphere-egu25-18400, 2025.

EGU25-18981 | ECS | Posters on site | EOS4.8

Publishing BUGS: Insights from the Journal of Trial and Error 

Stefan Gaillard

Addressing positive publication bias and clearing out the file drawer has been at the core of the Journal of Trial and Error since its conception. Publishing the trial-and-error components of science is advantageous in numerous ways, as already pointed out in the description of this panel: errors can lead to unexpected insights and warning others about dead ends can prevent wasted time and other resources. Besides those advantages, publishing negative and null results facilitates conducting robust meta-analyses. In addition, predictive machine learning models benefit from training on data from all types of research rather than just data from studies with positive, exciting results; already researchers are reporting that models trained on published data are overly optimistic.

Besides publishing negative and null results as well as methodological failures, the Journal of Trial and Error couples each published study with a reflection article. The purpose of these reflection articles is to have a philosopher, sociologist or domain expert reflect on what exactly went wrong. This helps contextualize the failure, helping to pinpoint the systematic factors at play as well as helping the authors and other scientists to draw lessons from the reported research struggles which can be applied to improve future research.

Publishing failure brings with it some practical challenges: convincing authors to submit manuscripts detailing their trial-and-error; instructing peer reviewers on how to conduct peer review for the types of articles; differentiating between interesting … and uninformative, sloppy science; and determining the best formats to publish various failure-related outcomes in. Authors are still hesitant to publish their research struggles due to reputational concerns and time constraints. In addition, authors often fear that peer reviewers will be more critical of articles describing research failures compared to articles reporting positive results. To counteract this (perceived) tendency of peer reviewers to be more critical of research without positive results, we provide specific instructions to peer reviewers to only assess the quality of the study without taking into account the outcome. This then also ensures that we only publish research that adheres to the standards of the field rather than sloppy science. Whether submitted research provides informative insights is assed by the editor-in-chief and the handling editor.

Finally, we are constantly evaluating and innovating the types of articles we publish. Various types of errors and failures benefit from differing ways of reporting. For example, recently we introduced serendipity anecdotes, a format where scientists can anecdotally describe instances serendipity which occurred during their research. This format allows researchers to focus on the conditions which allowed for the serendipitous discovery rather than the research itself.    

How to cite: Gaillard, S.: Publishing BUGS: Insights from the Journal of Trial and Error, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-18981, https://doi.org/10.5194/egusphere-egu25-18981, 2025.

It is common to perform two-dimensional simulations of mantle convection in spherical geometry. These have commonly been performed in axisymmetric geometry, i.e. (r, theta) coordinates, but subsequently we (Hernlund and Tackley, PEPI 2008) proposed using (r, phi) spherical annulus geometry and demonstrated its usefulness for low-viscosity-contrast calculations. 

When performing scaling studies in this geometry, however, strange results that did not match what is expected from Cartesian-geometry calculations were obtained when high-viscosity features (such as slabs) were present. It turns out that this is because the geometrical restriction forces deformation that is not present in 3 dimensions. Specifically, in a 2-D spherical approximation, a downwelling is forced to contract in the plane-perpendicular direction, requiring it to extend in the two in-plane directions. In other words, it is "squeezed" in the plane-perpendicular direction.  If the downwelling has a high viscosity, as a cold slab does, then it resists this forced deformation, sinking much more slowly than in three dimensions, in which it could sink with no deformation. This can cause unrealistic behaviour and scaling relationships for high viscosity contrasts. 

This problem can be solved by subtracting the geometrically-forced deformation ("squeezing") from the strain-rate tensor when calculating the stress tensor. Specifically, components of in-plane and plane-normal strain rate that are required by and proportional to the vertical (radial) velocity are subtracted, a procedure that is here termed "anti-squeeze". It is demonstrated here that this "anti-squeeze" correction results in sinking rates and scaling relationships that are similar to those in 3-D geometry whereas without it, abnormal and physically unrealistic results can be obtained for high viscosity contrasts. This correction has been used for 2-D geometries in the code StagYY (Tackley, PEPI 2008; Hernlund and Tackley, PEPI 2008) since 2010.

How to cite: Tackley, P.:  Adventures in Modelling Mantle Convection in a Two-Dimensional Spherical Annulus and Discovering the Need for "Anti-Squeeze”, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-19890, https://doi.org/10.5194/egusphere-egu25-19890, 2025.

EGU25-20057 | Posters on site | EOS4.8

Some Perfectly Reasonable Ideas that Didn’t Work: Snow Hydrology 

Ross Woods

The science question: how can we use hydrological process knowledge to understand the timing and magnitude of seasonal streamflow in snow-influenced catchments.

What was known: in general, catchments with colder climates have later and larger seasonal streamflow peaks, because more snow tends to accumulate in colder catchments, and it melts later because the time when melt can occur is later in the year in colder climates. Numerical models with fine space and time resolution were able to resolve these phenomena, but there was no theory which directly linked long term climate to seasonal streamflow.

In 2009 I published a very simple deterministic theory of snow pack evolution. I tested it against snow observations at 6 locations in the western USA and it apparently worked well (although I later discovered that I'd been lucky).

In 2015 I used the snowmelt derived from this deterministic theory to predict timing and magnitude of seasonal streamflow. It did poorly, and revealed untested assumptions in my theory. I tried making the theory slightly more complicated by considering within-catchment variation in climate. This did not help.

In 2016 I created a stochastic version of the theory (a weakness identified in 2015), and then also considered the within-catchment variation in climate. It did better at reproducing measured snow storage, but did not help in understanding seasonal streamflow.

My next step will be to consider all forms of liquid water input, i.e. not just snowmelt but also rainfall.

What survived: I will continue to use the stochastic version of the theory as it is clearly an improvement. I will continue to examine whether within-catchment climate variability is important, but it seems unlikely after two negative results. But whether introducing liquid water input will be sufficient, who can say? I will also try to examine in more detail how it is that the finely-resolved numerical models can do an adequate job, but the theory cannot - it is in this gap that the answer probably lies.  However the models are very complicated, and it is not easy to get a good understanding of exactly what they are doing, even though we know which equations the are implementing.

 

How to cite: Woods, R.: Some Perfectly Reasonable Ideas that Didn’t Work: Snow Hydrology, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20057, https://doi.org/10.5194/egusphere-egu25-20057, 2025.

EGU25-20866 | ECS | Posters on site | EOS4.8

A case for open communication of bugs in climate models 

Jan Gärtner, Ulrike Proske, Nils Brüggemann, Oliver Gutjahr, Helmuth Haak, Dian Putrasahan, and Karl-Hermann Wieners

Climate models are not only numerical representations of scientific understanding but also human-written software, inherently subject to coding errors. While these errors may appear minor, they can have significant and unforeseen effects on the outcomes of complex, coupled models. Despite existing robust testing and documentation practices in many modeling centers, bugs broader implications are underexplored in the climate science literature.

We investigate a sea ice bug in the coupled atmosphere-ocean-sea ice model ICON, tracing its origin, effects, and implications. The bug stemmed from an incorrectly set logical flag, which caused the ocean to bypass friction from sea ice, leading to unrealistic surface velocities, especially in the presence of ocean eddies. We introduce a concise and visual approach to communicating bugs and conceptualize this case as part of a novel class of resolution-dependent bugs - long-standing bugs that emerge during the transition to high-resolution models, where kilometer-scale features are resolved.

By documenting this case, we highlight the broader relevance of addressing bugs and advocate for universal adoption of transparent bug documentation practices. This documentation complements the robust workflows already employed by many modeling centers and ensures lessons from individual cases benefit the wider climate modeling community.

How to cite: Gärtner, J., Proske, U., Brüggemann, N., Gutjahr, O., Haak, H., Putrasahan, D., and Wieners, K.-H.: A case for open communication of bugs in climate models, EGU General Assembly 2025, Vienna, Austria, 27 Apr–2 May 2025, EGU25-20866, https://doi.org/10.5194/egusphere-egu25-20866, 2025.

CC BY 4.0