Abstract
The Copernicus Regional Reanalysis for Europe, CERRA, is a high-resolution regional reanalysis dataset for the European domain. In recent years, it has shown significant utility across various climate-related tasks, ranging from forecasting and climate change research to renewable energy prediction, resource management, air quality risk assessment, and the forecasting of rare events, among others. Unfortunately, the availability of CERRA is lagging 2 years behind the current date, due to constraints in acquiring the requisite external data and the intensive computational demands inherent in its generation. As a solution, this paper introduces a novel method using diffusion models to approximate CERRA downscaling in a data-driven manner, without additional informations. By leveraging the lower resolution ERA5 dataset, which provides boundary conditions for CERRA, we approach this as a super-resolution task. Focusing on wind speed around Italy, our model, trained on existing CERRA data, shows promising results, closely mirroring the original CERRA. Validation with in-situ observations further confirms the model’s accuracy in approximating ground measurements.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
In this study, we propose the utilization of super-resolution neural network architectures for the downscaling of reanalysis datasets in the field of meteorology and climate science. Recent advancements in super-resolution techniques have significantly enhanced their performance across a variety of applications, rivalling traditional physics-based models. This promising development suggests that these neural architectures could offer a more effective and practical approach in the field of climate data analysis, potentially transforming our ability to obtain high-resolution insights from coarse-scale meteorological data.
The downscaling models typically used for reanalysis in meteorology share a problem formulation similar to super-resolution, where a low-resolution image is transformed into a high-resolution counterpart. This similarity opens up the possibility of utilizing data-driven neural methods to approximate the performance of physics-inspired downscaling models. Current downscaling models are often hindered by high computational costs and typically require a vast amount of additional information for effective downscaling. This additional information frequently depends on various sources, which can delay the availability of the data. Coupled with the high computational demands, these factors tend to postpone the availability of downscaling results, often lagging years behind the current date and usually falling behind more available lower-resolution reanalyses. This delay can impair the ability to access fine-resolution datasets of recent events, thus delaying the response of the scientific community. This was particularly evident in the case of the Italian peninsula, especially with respect to the May 2023 Emilia-Romagna floods and the November 2023 Ciarán Storm.
With our research, we aim to demonstrate the feasibility of neural super-resolution architectures in the data-driven downscaling of reanalysis datasets. Our objective is to harness the capabilities of our super-resolution, diffusion-based approach to provide researchers with rapid and effective access to downscaled data, facilitating a swift response to current events and eliminating lengthy delays in data availability. We also seek to demonstrate the feasibility of neural methods for downscaling altogether, with possible future integration with physics inspired model to create more efficient and better all-around performing models. We argue that the application of super-resolution models in downscaling could significantly aid in the understanding and prediction of weather and climate patterns at a local scale, proving particularly beneficial for specialized areas such as agriculture, urban planning, and disaster management.
Our methodology involves creating a framework to evaluate the efficacy of state of the art neural super-resolution models in a recognized downscaling task. We selected ERA5 [1] and CERRA [2] as our low-resolution and high-resolution datasets, respectively (Fig. 1). This choice is particularly advantageous because CERRA is derived using ERA5 as a baseline, establishing a robust connection between the two datasets and minimizing their differences. At the time of writing, data availability issues made CERRA reanalysis lag more than 2 years behind ERA5, impairing the research on current events, and showcasing the need for alternative downscaling methods. For our experimental setup, we focused on a region encompassing the Italian peninsula, bounded by the following geographical coordinates: North at 47.75\(^\circ\), South at 35\(^\circ\), East at 18.75\(^\circ\), and West at 6\(^\circ\). We chose a training period of 10 years, from 2010 to 2019, with a temporal resolution of 3 h, and selected two years, 2009 and 2020, as our test set. Our primary strategy employs a conditioned diffusion model for downscaling, a model that has recently demonstrated top-tier performance in various applications and can effectively benefit from ensemble approaches. In addition to the diffusion model, we evaluated a range of other well-known architectures for comparative analysis. To validate our results, we incorporated in-situ observations from ground stations, which aided in assessing the real-world accuracy and effectiveness of our downscaling approach.
In this research, we present several significant contributions to the field of meteorological data analysis and climate modeling. Firstly, we demonstrate the feasibility of using neural network methodologies for the downscaling of wind speed data from ERA5 to CERRA, providing a novel approach to refining climate data resolution. Furthermore, we conduct a comprehensive comparison of multiple state-of-the-art models to establish a benchmark in this domain. Our findings indicate that ensemble diffusion models outperform other techniques, marking a pivotal advancement in the field. Additionally, we validate our proposed model along with the CERRA dataset against in-situ ground station measurements, ensuring the accuracy and reliability of our approach. Lastly, we have computed and made freely available an approximated CERRA dataset for recent years, which are not yet available from the original CERRA sources. This contribution not only enhances the accessibility of high-quality climate data but also supports ongoing climate research and modeling efforts.
This document starts by examining the correlation between the downscaling problem and the super-resolution problem. Following this, we provide an overview of the super-resolution research field, with an emphasis on diffusion models. We then detail the datasets utilized in our experiments, specifically ERA5, CERRA, and IGRA V2 [3]. Subsequently, we introduce our methodology and delve into the specifics of our diffusion model, including an analysis of training and evaluation processes, as well as comparisons with competing models. Our validation framework is then described. Finally, we outline our experimental procedures and discuss the results obtained, concluding with reflections on potential future work and applications in this field.
2 Super-resolution and downscaling
Super-resolution (SR) and statistical downscaling, while distinct in their methodologies and objectives, share a common goal of refining data resolution, particularly in the context of weather datasets. Super-resolution, primarily used in image processing, involves increasing the resolution of an image by adding more pixels and thereby enhancing the detail and clarity of the image. In recent years, super-resolution has played a pivotal role in the evolution of deep learning models, demonstrating significant improvements in performance over traditional non-neural methods.
In contrast, statistical downscaling in meteorology is concerned with translating large-scale climate or weather data (often from global climate models such as ERA5) into more localized, high-resolution data. This is crucial for understanding local climate impacts, with the global models often unable to capture small scale phenomenons in detail.
With both statistical downscaling and super-resolution aiming to provide higher resolution outputs, we argue that a super-resolution deep learning approach can be applied to the downscaling of meteorological variables.
2.1 Problem definition
Super-resolution is a process in digital image manipulation that enhances image resolution. This problem can be defined as an ill posed inverse problem, where the task is to invert an unknown degradation function applied to a high resolution image. Given a low resolution image denoted by y and the corresponding high resolution denoted by x, we define the degradation process as:
where \(\Theta\) is the degradation function, and \(\rho _n\) denotes the degradation parameters (such as the scaling factor, noise, etc.). In a real-world scenario, only y is available while no information about the degradation process or the degradation parameters \(\rho\) exists. Super-resolution seeks to nullify the degradation effect and recovers an approximation \(\hat{x}\) of the ground-truth image x as:
where \(\rho _r\) are the parameters for the function \(\Theta ^{-1}\). The degradation process is unknown and can be quite complex. It can be affected by several factors, such as noise, compression, blur and other artefacts. Therefore, most research works [4,5,6] prefer the following degradation model over that of Eq. 1:
where k is the blurring kernel and \((x \otimes k)\) is the convolution operation between the HR image and the blur kernel, \(\downarrow _s\) is a downsampling operation with a scaling factor s.
Let us consider now the task of statistical downscaling in weather sciences. Statistical downscaling is a technique employed to convert broad-scale meteorological data, typically obtained from global climate models or reanalysis products, into finer, local-scale weather information. This approach relies on establishing statistical relationships between large-scale atmospheric variables (like temperature, pressure, and wind) and local-scale weather phenomena. This process is usually implemented via a sophisticated physical model, such as the HARMONIE-ALADIN [7, 8] model. These models have as input the low resolution ground truth plus the additional information necessary to compute the downscaling result. Thus, if we represent, coherently with the Super-resolution problem, the high resolution image as x, the low resolution images as y, the statistical downscaling model as HA and the additional information as \(\rho _\textrm{HA}\), we can define the downscaling process as:
We observe that this equation is closely related to Eq. 2, which establishes a relationship between high-resolution and low-resolution images. This similarity effectively equates the task of super-resolution in image processing to the problem of statistical downscaling in meteorological studies.
The primary distinction between the two problems lies in the inverse nature of their low-resolution/high-resolution relationship. Contrary to the super-resolution process described in Eq. 3, where the low-resolution image is derived from an unknown high-resolution counterpart, in statistical downscaling, the situation is reversed. Here, the high-resolution image is an enhanced version of the low-resolution image.
We can thus define the new super-resolution task as the problem of approximating the behaviour of the HA model. Considering that our goal is to proceed in a data driven manner and to avoid the utilization of additional information for the downscaling process, we define our approximate model without the use of \(\rho _\textrm{HA}\).
Where \(\Phi\) is our super-resolution model and \(\hat{x}\) is the approximation of the ground truth x generated by the statistical downscaling model.
2.2 Super-resolution methods
Traditional methods of Super-resolution typically involved interpolation techniques like bicubic or bilinear interpolation, which enlarged images by inserting new pixels based on the color values of neighboring pixels. These methods, while straightforward and computationally inexpensive, often led to underwhelming results, especially when dealing with significant upsampling.
In recent years, the research has instead focused on deep learning approaches [9,10,11,12,13,14,15,16]. Deep learning has revolutionized the field of super-resolution by substantially enhancing the quality and effectiveness of image upsampling. This advancement includes not only improved image clarity but also the capability to learn from data and perform context-aware upsampling. Central to this advancement are Convolutional Neural Networks (CNNs), which form the core of many state-of-the-art super-resolution methods, with architectures like SRCNN [10] and VDSR [17] leading the way.
Subsequent improvements to the CNN architecture include ESPCN [15], which is designed to operate in real-time both for images and videos, implementing a sub-pixel convolution layer at the very end of the network to aggregate LR feature maps and simultaneously perform projection to high-dimensional space to reconstruct the HR image. Subsequent advancements have been significantly enhanced by the introduction of residual networks. These networks incorporate skip connections in their design, effectively addressing the issue of vanishing gradients and enabling the construction of deeper networks, which in turn boosts overall performance. Example of residual networks for SR include Enhanced Deep Super-Resolution (EDSR) [18], which modifies the ResNet architecture [19] proposed originally for image classification to work with the SR task, and CARN [20], which utilize residual blocks in intermediate layers which are then cascaded and converged onto a 1 × 1 convolutional layer.
The research in super-resolution methods based on CNN architectures has seen a remarkable diversification in the following years, novel architectural innovations include the use of attention mechanism [21, 22], multi-branch designs [23, 24] and dense connections [25, 26].
In more recent years, the focus of research in super-resolution has undergone a significant paradigm shift, moving away from traditional CNN-based predictive models towards more advanced generative approaches. The first steps in this direction were based around the Generative Adversarial Network (GAN) architecture [27], which employs a game-like approach, with the model being comprised of two competing elements, the generator, and the discriminator. The generator is trained to produce an HR image in such a way that the discriminator cannot distinguish it from the original HR image. Training in this adversarial manner, HR images with better perceptual quality are generated. The most notable applications of GAN to the super-resolution task include SRGAN [28] and ESRGAN [29].
The state of the art in super-resolution has since shifted from GAN-based approaches to diffusion and transformer based approaches. Transformers [30] are the most relevant architecture for natural language processing. Recently, their adaptation to the visual domain, known as visual transformers, has also begun to set new benchmarks, establishing themselves as state-of-the-art for various visual tasks. Transformers excel in handling sequential data. They utilize self-attention to process input data in a way that each element of the sequence can directly attend to every other element, enabling the model to capture complex relationships. In the case of images, the process involves treating an image as a sequence of patches or pixels. Thanks to the ability to learn contextual relationships, transformer-based models such as SWINIR [31] and SWIN2SR [32] are at the forefront of the research in super-resolution.
Similarly to transformers, diffusion models [33] have recently demonstrated the capability to achieve state-of-the-art performance in super resolution tasks [34]. Diffusion models are a type of generative model that simulate a process similar to heat diffusion. They start by gradually adding noise to an image (or any data) until it turns into random noise. The model then learn to reverse this process, reconstructing the original data from the noisy version. This distinctive capability positions diffusion models at the forefront of generative tasks, avoiding the training instability that can affects GAN and Transformers. In this work, we focus on the Denoising Diffusion Implicit Model (DDIM) as our primary architecture.
2.3 Metrics
In our research, we will evaluate the quality of our super-resolution methodologies using three established metrics: Mean Squared Error (MSE), Peak Signal-to-Noise Ratio (PSNR), and Structural Similarity Index Measure (SSIM). Subsequently, a validation of the results obtained will be carried out, using Mean Absolute Error (MAE) and Mean Squared Error. Mean Squared Error (MSE) is a widely used metric in a wide range of applications, including super-resolution. It quantifies the difference between the predicted values and the actual values by averaging the squares of the errors. The error in this context is the difference between each predicted value and its corresponding true value.
Mean Absolute Error (MAE) is another commonly used metric to assess prediction accuracy. It calculates the average magnitude of the errors between predicted and actual values, without considering their direction. Unlike MSE, which squares the errors, MAE uses the absolute values of the errors. This makes MAE less sensitive to outliers compared to MSE, and it can give a clearer understanding of the entity of the error.
Peak Signal-to-Noise Ratio (PSNR) [35] is a prominent metric used in the field of image processing to evaluate the quality of reconstructed, compressed, or denoised images in comparison to the original image. It is defined as the ratio between the maximum possible power of a signal (represented by pixel values in an image) and the power of corrupting noise that affects its fidelity.
Finally, Structural Similarity Index (SSIM) [35] is an advanced metric used for measuring the similarity between two images. Developed to provide a more accurate and perceptually relevant assessment of image quality, SSIM differs from traditional metrics like MSE and PSNR by considering changes in structural information, luminance, and contrast. Unlike MSE and PSNR, which focus on pixel-to-pixel differences, SSIM evaluates changes in structural patterns, making it more aligned with human visual perception. SSIM can be written as:
where the metric is computed between two corresponding patches represented by x and y, with \(\mu _x\) and \(\mu _y\) being the respective pixel average, \(\sigma _x^2\) and \(\sigma _y^2\) the variance and \(\sigma _{xy}\) the covariance. \(C_1\) and \(C_2\) are two custom parameters to stabilize the division with weak denominator. The aforementioned metrics have been employed in a multitude of recent studies with objectives analogous to those of the present investigation. [36,37,38,39]
2.4 Diffusion models
Diffusion models are a class of probabilistic generative models that are particularly effective in modeling complex, high-dimensional data distributions. At the core of diffusion models lies the mathematical concept of a diffusion process, that is, a stochastic process that describes the continuous random movement of particles over time, modeling the spread or diffusion of some quantity in space or time, where the particles tend to move from regions of high concentration to regions of low concentration, resulting in a gradual blending or mixing of the quantity. In the context of machine learning, diffusion models leverage the principles of diffusion processes to model the generation of data. Instead of directly sampling data points from a fixed distribution, these models iteratively transform a simple initial distribution, typically a known distribution like a Gaussian or uniform distribution, into the desired complex data distribution. The main idea is to perform a series of diffusion steps, where each step updates the probability distribution of the data. This is achieved by adding Gaussian noise to the current data samples and iteratively refining them.
From a mathematical perspective, considering a distribution \(q(x_0)\) which generates the data, generative models aim to find a parameter vector \(\theta\) such that the distribution \(p_\theta (x_0)\) parameterized by a neural network approximates \(q(x_0)\).
Denoising Diffusion Probabilistic Models (DDPM) [40] assume the generative distribution \(p_\theta (x_0)\) to have the form
given a time range horizon \(T>0\). where the Markov Chain formulation is :
Training is traditionally based on a variational lower bound of the negative log likelihood. Considering that the Kullback–Leibler Divergence \(D_\text {KL} = D_\text {KL}(q(x_{1:T}\vert x_0) \Vert p_\theta (x_{1:T}\vert x_0) )\) is positive, we obtain:
We can thus expand the second term to derive the training loss \(L_\theta\):
In the case of diffusion models, the latent space has typically the same dimension of the visible space.
The inference procedure \(q(x_{1:T} \vert x_0)\) is fixed, in contrast with different latent variables models like Variational Autoencoders (VAEs) [41,42,43] where it is learned along with the generative distribution.
In the particular case of Denoising Diffusion Implicit Models (DDIMs) [44], used in this work, the authors considered a non-Markovian diffusion process
where \(q_\sigma (x_T \vert x_0) = \mathcal {N}(x_T \vert \sqrt{\alpha _T} x_0, (1 - \alpha _T) \cdot I)\), and
with
The definition of \(q(x_{t-1} \vert x_t,x_0)\) is cleverly chosen in order to ensure two important aspects of the diffusion process of DDPM: the Gaussian nature of \(q(x_{t-1} \vert x_t,x_0)\) and the fact that the marginal distribution \(q_\sigma (x_t\vert x_0) = \mathcal {N}(x_t \vert \sqrt{\alpha _t} x_0; (1 - \alpha _t) \cdot I)\) is the same as in DDPM. Thanks to the latter property, \(x_t\) can be expressed as a linear combination of \(x_0\) and a noise variable \(\epsilon _t \sim \mathcal {N}(\epsilon _t \vert 0; I)\):
The next step consists in defining a trainable generative process \(p_\theta (x_{0:T})\) where \(p_\theta (x_{t-1}\vert x_t)\) leverages the structure of \(q_\sigma (x_{t-1} \vert x_t, x_0)\). The idea is that given a noisy observation \(x_t\), one starts making a prediction of \(x_0\), and then use it to obtain \(x_{t-1}\) according to Eq. 13.
In practice, we train a neural network \(\epsilon _\theta ^{(t)}(x_t, \alpha _t)\) to map a given \(x_t\) and a noise rate \(\alpha _t\) to an estimate of the noise \(\epsilon _t\) added to \(x_0\) to construct \(x_{t}\). As a consequence, \(p_\theta (x_{t-1} \vert x_t)\) becomes a \(\delta _{f_\theta ^{(t)}}\), where
Using \(f_\theta ^{(t)}(x_t, \alpha _t)\) as an approximation of \(x_0\) at time step t, \(x_{t-1}\) is then obtained as follows:
As for the loss function, the term in Eq. 11 can be further refined expressing \(L_\theta\) as the sum of the following terms [45]:
where
All previous distributions are Gaussian and their KL divergences can be calculated in closed form using, obtaining the following formulation:
This can be simply interpreted as the weighted Mean Squared Error between the predicted and the actual noise a time t.
Due to experimental evidence, the weighting parameters are usually ignored in practice, since the training process works better without them.
The pseudocode for training and sampling is given in the following Algorithms.
Sampling is an iterative process, starting from a purely noisy image \(x_T \sim \mathcal {N}(0,I)\). The denoised version of the image at time step t is obtained using Eq. 16.
2.5 Related works
The body of literature surrounding statistical downscaling is extensive and well established. Initially, this technique utilized foundational statistical methods such as linear regression and ridge regression [46]. Prominent examples of downscaling methods that integrate multiple statistical techniques include SDSM [47] and LARS-WG [48]. Subsequently, researchers have incorporated machine learning elements into statistical downscaling, adopting methods such as k-nearest neighbors [49], random forest [50, 51], and support vector machines (SVM) [52]. Ensemble methods have also emerged, enhancing predictions by synthesizing outputs from various models [53, 54]. Moreover, the inclusion of neural networks has been pivotal, with successful applications including Dense Networks [46], Convolutional Neural Networks (CNN) [55, 56], Long Short-Term Memory networks (LSTM) [57], U-Net [58], and Generative Adversarial Networks (GAN) [59].
In recent years, the analogy between SR and statistical downscaling gained momentum, with multiple relevant examples present in literature. In [60], the authors effectively utilize super-resolution neural techniques to perform downscaling of precipitation data. The study introduces and evaluates three super-resolution deep learning frameworks: the Super-Resolution Convolutional Neural Network (SRCNN), Super-Resolution Generative Adversarial Networks (SRGAN), and Enhanced Deep Residual Networks for Super-Resolution (EDSR). These frameworks are applied to refine the spatial resolution of daily precipitation forecasts in Southeast China. Specifically, they enhance the resolution from approximately 50 kms to finer scales of about 25 kms and 12.5 kms. For comparison, Bias Correction Spatial Disaggregation (BCSD) as a traditional SD method is also performed under the same framework. The precipitation forecast data utilized in this research are sourced from various Ensemble Prediction Systems (EPSs), including the European Centre for Medium-Range Weather Forecasts (ECMWF), the National Centers for Environmental Prediction (NCEP), and the Japan Meteorological Agency (JMA). The results demonstrate that all three SR models effectively capture intricate local details in precipitation patterns, with EDSR achieving the best overall performance. In [61], the authors propose a climate research library presenting a series of tasks, including statistical downscaling. They choose two different settings, the first is downscaling 5.625\(^{\circ }\) ERA5 to 2.8125\(^{\circ }\) ERA5 with hourly intervals and at global scale, using as target variables Z500, T850 and T2m. In the second setting, the authors consider downscaling from 2.8125 \(^{\circ }\) ERA5 to 0.75\(^{\circ }\) PRISM over the continental United States at hourly intervals for the variable daily max T2m. With PRISM being a dataset of interpolated in-situ observations, this setting is particularly interesting as it performs downscaling from a reanalysis to an observational dataset. For both settings, the training period is from 1981 to 2016, and the testing period is the year 2017. The authors test a series of model, including ResNet, U-Net, and ViT, nearest and bilinear interpolation. In [39], the authors perform statistical downscaling with super-resolution neural models, while guaranteeing physical constraints via the use of hard constraining layers enforcing conservation of mass. In the experimental setting, the ERA5 dataset was used, focusing on the total column water (tcw) variable. The super-resolution setting was achieved by downsampling ERA5 by factors ranging from 2 to 16. The obtained results reveal increased performance with respect to non-constrained models.
In this work, our focus is also on the validation of results. The task of validating downscaled data is frequently encountered in literature. In [62] the authors analyze the performance of ERA5, ERA5-Land and CERRA on the territory of Sicily in estimating air temperature, air relative humidity, solar radiation, and wind speed. The estimated weather variables are then compared to in-situ measurement, with the outcomes revealing that CERRA provides the best estimations overall. In [63] the authors perform an in-situ validation for various modeling systems of global and regional reanalyses including ERA5, ERA5-Land, ERA5-Crocus, CERRA-Land, UERRA MESCAN-SURFEX and MTMSI. The validation data included observational references (in situ, gridded observational datasets and satellite observations) across the European Alps from 1950 to 2020 for a set of weather variables including snow cover and its first-order drivers, temperature, and precipitation. Their results reveal that no modeling strategy outperforms all others within their experimental sample. In [64], the authors assess the accuracy of CERRA for human-bio meteorological and heat-health applications on the territory of Greece, taking into consideration temperature and wind speed. The findings indicate that CERRA outperformed in the majority of statistical assessments and demonstrated performance on par with ERA5-Land in the remaining ones. The study carried out by [65] examines the performance of various reanalysis datasets, including ERA5, ERA5-Land, CERRA, MERIDA, VHR-REA_IT, in accurately reproducing the temperature patterns characteristic of the Italian peninsula. Although the research is not yet completed, initial results reveal that the utilized reanalyses effectively replicate the temperature climatology across Italy for a range of spatial resolutions, from 31 to 2.2 km, exhibiting only minimal degradation in performance when modeling smaller-scale processes.
3 Datasets
In our study, we utilized three distinct datasets to perform super-resolution and validation. We employed the ERA5 [1] dataset as our low-resolution baseline, characterized by its spatial precision of around 30 km, and compared it with the high-resolution CERRA [2] dataset, having a finer spatial resolution of around 5.5 km over the European region. For real-world validation of our results, we used the IGRA V2 [66] in-situ dataset.
Both ERA5 and CERRA are cutting-edge in the field of numerical weather reanalysis, widely utilized in a spectrum of applications including forecasting, predicting extreme weather events, and conducting climate studies. While ERA5 offers a global perspective, CERRA provides a more region-specific focus, covering only Europe. Our selection of datasets was strategic; CERRA uses ERA5 as a boundary condition, aligning well with our super-resolution objectives where a robust correlation exists between low-resolution and high-resolution images.
Notably, CERRA adopts a conical projection optimal for high latitudes, in contrast to ERA5’s cylindrical projection. This necessitated a conversion for direct comparison. Focusing on the Italian peninsula, an area with relatively lower latitudes, we conducted our comparative analysis using the cylindrical projection. In this converted format, the reference area for CERRA amounted to 256 × 256 pixels, while for ERA5, it was 52 × 52 pixels.
For this comparative study, we chose wind speed as our primary variable, offering a valid example through which examine the downscaling performance.
3.1 ERA5
ERA5 [1], short for the Fifth Generation of the European Centre for Medium-Range Weather Forecasts (ECMWF) Reanalysis, is a well known and vastly utilized dataset in the realm of climate and atmospheric research. ERA5, encompassing a period from 1940 to the present, provide a detailed portrait of the Earth’s atmospheric system. It provides hourly estimates of a large number of atmospheric, land and oceanic climate variables with a horizontal resolution of 0.25 degrees (around 30 km) and 137 levels in the vertical from the surface to 0.01 hPa (about 80 km). An innovative aspect of The ERA5 reanalyses is its timely availability. Collecting the necessary data and their computation results in a production latency of approximately 5 days. This timely availability makes ERA5 a valuable tool for studying events in the recent past and forecasting future weather.
The assimilation system used in ERA5, based on the ECMWF Integrated Forecasting System (IFS), incorporates advancements in numerical weather prediction models, radiative transfer schemes, and satellite data utilization, contributing to the dataset’s reliability and precision [67].
ERA5 is a state-of-the-art global atmospheric reanalysis, and it serves as a robust tool for understanding and analyzing the Earth’s climate system, offering a global representation of atmospheric, land, and oceanic conditions spanning several decades. ERA5 over the years it has been a crucial asset for scientists from a variety of fields, with the dataset appearing in a vast amount of studies. Its applications span several key areas:
-
Analysis of precipitation trends: ERA5 data is extensively used for tracking precipitation changes, as evidenced in multiple recent works [68,69,70,71]. These works highlight its utility in observing and understanding changes in precipitation patterns relative to climate change.
-
Investigation of temperature trends: The dataset is instrumental in analyzing long-term temperature trends, with studies ranging from global temperatures [72, 73] to on urban heat [74].
-
Wind sanalysis: Recent works demonstrates its application in studying wind patterns for renewable energy monitoring [75] and addressing climate changes [76].
-
Extreme events study: ERA5’s role is pivotal in advancing our understanding of extreme climate events, exploring storm surge [77] and heatwaves [78] and extreme precipitations [79, 80].
The high temporal resolution of ERA5 is particularly valuable in detecting shifts in climate patterns, which is crucial for understanding global warming and climate extremes. This information enables researchers to refine climate models, project future scenarios, and assess the potential impacts of climate change on both regional and global scales.
Additionally, ERA5 data finds important applications in other fields such as agriculture [81], water resources management [82, 83], and urban planning and infrastructure design [84]. In summary, ERA5 represents an invaluable asset in weather and climate studies, playing a pivotal role in ongoing scientific research.
3.2 CERRA
The Copernicus Regional Reanalysis for Europe (CERRA) is a sophisticated high-resolution regional reanalysis (RRA) dataset specifically designed for the European region. It is a product of the European Copernicus program, executed through a contract with the Swedish Meteorological and Hydrological Institute (SMHI), in collaboration with subcontractors Meteo-France and the Norwegian Meteorological Institute. CERRA offers data at a fine 5.5 km horizontal resolution.
This dataset achieves its detailed resolution by utilizing the global ERA5 reanalysis dataset, which provides both initial and boundary conditions. In addition to inputs from ERA5, CERRA’s regional reanalysis incorporates higher-resolution observational data and physiographic datasets that describe surface characteristics. This comprehensive approach is illustrated in Fig. 2.
CERRA’s output includes both forecasts and reanalyses. Weather forecasting in this context involves analyzing the current state of the atmosphere and terrestrial and marine surfaces. These forecasts are generated using mathematical and physical numerical models that start from this analysis. During the reanalysis phase, a weather forecasting model is initially used to estimate the atmospheric state at a specific time. This ’first guess’ is then refined through a process known as ’data assimilation’, which involves correcting the model based on observational data. This process ensures that the reanalysis provides an accurate historical record of weather conditions.
Three different stages of reanalysis: (left) the Global Reanalysis ERA5 will be used as boundary condition, (middle) CERRA Regional Reanalysis, and (right) a CERRA 2D Reanalysis for the near surface. The arrows indicate that the amount of observational information used for the reanalysis progressively increases from the global to the regional reanalysis. Source: Copernicus-ECMWF CERRA Presentation, 2024 [85]
CERRA contains a wide range of meteorological variables, including temperature, humidity, wind speed and direction, precipitation, and cloud cover. The data is available on a regular grid, while the availability in terms of temporal resolution follows a different structure with respect to the hourly one present in ERA5, differentiating between reanalysis and forecast time. For each day, 8 reanalyses are available, starting from 00 UTC every 3-h until 21 UTC. The prevailing assumption is that analysis data for these time periods exhibits superior quality compared to forecasts for the same time frames, as it is inherently more aligned with observations. Starting from each reanalysis data available, the next 6 h are processed through forecasting. And for reanalyses starting at 00 UTC and 12 UTC, the next 30 h are processed, with a 3-h interval after the first 6 h. In this way users can select, even for the same hour, the forecast they prefer, or if available, the reanalysis result. However, this means that even if CERRA provides data for each hour, only the 8 present in the reanalysis schedules guarantee a corrective step using observations, while the others are forecast results.
The CERRA dataset is a valuable resource for a wide range of applications, including the study of climate trends, renewable energy forecasting, water resource management, and risk assessment. For climate research, CERRA provides a historical reconstruction of meteorological variables, which is essential for understanding long-term climate trends and the impacts of climate change in Europe. A study [86] demonstrated that CERRA outperforms the ERA5-Land reanalysis in replicating temperature and other bio-meteorological variables, making it particularly useful for assessing heat-related health risks. In the field of renewable energy, CERRA’s accurate weather observations are crucial for developing forecasting tools for wind and solar power, aiding the integration of these renewable sources into the power grid [87]. Water resource management also benefits from CERRA’s ability to represent precipitation [88] and snow [89], helping to manage resources and assess the risks of drought, flooding and economic risks. Lastly, CERRA has been demonstrated as a viable substitute for ground-based agro-meteorological measurements, effectively capturing data on air temperature, actual vapor pressure, wind speed, and solar radiation, thus making CERRA a valuable tool for regional studies in water resource management [62]. These applications underscore the importance of the CERRA dataset in supporting adaptation actions, policy development, and climate monitoring and research across Europe.
Given the importance of CERRA in a variety of applications, it would be ideal to have a real-time production of this data, to effectively study recent events. Unfortunately, delays in the production and data gathering slowed the release of CERRA, with a delay of more than 2 years at the time of writing.
3.3 IGRA V2
For the validation task, data from the Integrated Global Radio-sounding Archive (IGRA) V2 [3] dataset has been employed. IGRA is a comprehensive collection of global radio-sounding profiles, encompassing temperature, humidity, and wind measurements, and it is freely accessible.
The in-situ data used refer to meteorological stations located in the same range of longitude and latitude represented by the other two datasets. The result of this selection is 14 stations based spread on the selected area.
IGRA V2 undergoes validation through various processes to ensure data quality and reliability. The validation includes quality control of radiosonde observations of temperature, humidity, and wind at stations. The data is required to have adequate documentation of its digital format, and only observed values are included, while estimated values are not, so this dataset represents a good source of information to validate results for the field of reanalysis and downscaling.
4 Methodology
In this section, we will provide a detailed exposition of our diffusion technique applied to super-resolution in statistical downscaling, which includes a comprehensive validation component. Our discussion begins with an overview of the preliminary steps required to prepare our case study, including the harmonization of features from various datasets. We then offer a detailed exploration of our primary and most effective model, the diffusion model. Following this, we compare the diffusion model with other super-resolution approaches to highlight relative strengths and weaknesses. Next, we outline the validation methods employed to ensure the accuracy of our model’s output against real-world measurements. Lastly, we evaluate our results and provide an in-depth analysis of our findings.
4.1 Data selection and preprocessing
In this section, we analyze the essential preprocessing steps required for defining the downscaling process from ERA5 to CERRA. Both ERA5 [1] and CERRA [2] are freely available from the Copernicus Program [90] website.
In our study, we have chosen to concentrate on the meteorological variable of wind speed. Several factors influenced this decision. Firstly, we aimed to simplify our experimental framework by focusing on a single variable. We selected wind speed because, compared to variables like rainfall, it is more physically self-contained and thus easier to analyze in isolation. Additionally, wind speed holds significant societal relevance, particularly in the context of extreme weather events and its implications for energy generation in wind farms.
We utilized wind speed at 10 ms above the surface. In the case of CERRA, wind speed is already included as a main variable, while for ERA5 we computed it from both the zonal u and the meridional v wind components by \(\sqrt{u^2 + v^2}\). The utilized values are instantaneous.
For our analysis, we extracted a specific region from the ERA5 dataset, delimited by the following geographical coordinates: North at 47.75\(^\circ\), South at 35\(^\circ\), East at 18.75\(^\circ\), and West at 6\(^\circ\). This targeted extraction ensures a focused study area for our research. This area coincides with a square containing Italy, Switzerland, Slovenia and partially Tunisia, Algeria, France, Austria, Croatia, Hungary, Bosnia & Erzegovina and Montenegro, as shown in Fig. 3.
As mentioned in Sect. 3, the two datasets have a different projection mechanism, with ERA5 using a cylindrical projection and CERRA a Lambert Conformal Conic projection. Given that our study area is located at mid-latitudes, where distortions in cylindrical projections are relatively minimal, we chose to equalise the two projections by re-projecting CERRA using a cylindrical projection consistent with ERA5. The projection was realized utilizing the CDO [91] library.
CERRA provides data samples with a 3-h temporal resolution, whereas ERA5 offers an hourly resolution. To align these datasets, we extracted ERA5 measurements corresponding to the times 00:00, 03:00, 06:00, 09:00, 12:00, 15:00, 18:00, and 21:00, matching CERRA’s schedule. Our training set encompasses data from 2010 to 2019. For testing purposes, we selected the years 2020 and 2009. The choice of non-contiguous test years is motivated by aiming to widen our evaluation range and the availability of in-situ observations for validation.
The training sets for both ERA5 and CERRA have been normalized to a range between 0 and 1, with each entry being divided by the maximum value within its respective dataset. Similarly, the test sets for each dataset have been normalized using the same maximum value as determined from their corresponding training sets.
4.2 Diffusion model architecture
In this section, we describe our super-resolution diffusion model. This includes an examination of the super-resolution (SR) framework, the specifics of the conditioning process, ensemble methods, and various implementation details.
4.2.1 SR framework
Super-resolution for neural networks implies the use of some form of upsampling, which may be implemented in different ways and in different positions within the network. Generally speaking, we can define two frameworks for SR: pre-upsampling, when the upsampling is done in the first part of the network, and post-upsampling, when the upsampling part is implemented at the very end of the network. If the upsampling is performed at the start, the spatial dimension of the features is increased, usually granting the network better overall performance at the expense of computational complexity. The first framework is generally used in setting where the main target is performance, such as SRGAN [29], the latter is more common in performance-oriented model such as ESPCN [15]. For our diffusion model, we opted for a pre-upsampling framework.
The process of upsampling can be executed using various methods. Some methods employ learnable parameters, such as transposed convolution, while others, like 2D upsampling or bilinear and bicubic upsampling, do not use learnable parameters. In our model, we opted for pre-emptive bilinear upsampling, due to its practicality. Additionally, our experiments with methods involving upsampling with learnable parameters yielded similar results.
For what concerns the super resolution task setting, we decide to utilize a sequence-to-one approach, as represented in Fig. 4. In this approach, the super resolution model is trained to produce the high resolution images while having access to multiple timestamps of the low resolution counterpart. Specifically, experimental results showed that the optimal sets of conditioning information for the prediction of the high resolution image at time \(t_0\) contains the low resolution images at hours \(t_{-6}\), \(t_{-3}\), \(t_0\), \(t_{+3}\). In our implementation, the main bulk of our data is kept on disk while a data generator feeds the model the sequences for both training the testing operations, with the first being a random set of sequences and the second a contiguous set.
4.2.2 Denoising and conditioning
Diffusion models essentially operate as iterative denoising algorithms. Their main trainable component is the denoising network, denoted as \(\epsilon _\theta (x_t, \alpha _t)\). This network receives as input the noisy input, \(x_t\), and a corresponding noise variance, \(\alpha _t\), with the objective of estimating the amount of noise infiltrating the image. The training of this underlying denoising network is done conventionally. An initial sample, \(x_0\), is extracted from the dataset and subjected to a predefined amount of random noise. The network is then tasked with estimating the noise within these corrupted images.
Our model of choice for the denoising network is the U-net. The U-net is one of the most common architectures for denoising [92,93,94,95] and it is often implemented in diffusion models [96]. Originally introduced for semantic segmentation [97], the U-Net architecture has gained widespread popularity and found applications in diverse image manipulation tasks. The network comprises a downsampling sequence of layers, followed by an upsampling sequence while incorporating skip connections between layers of the same size, as reported in Fig. 5. Our implementation employs bilinear interpolation for the upsampling process and utilizes average pooling for the downsampling procedure.
The U-Net configuration can be determined by defining the number of downsampling blocks and the number of channels for each block. The upsampling structure follows a symmetric pattern, and the spatial dimension is dependent on the number and resolution of the input and output images. Our U-net utilize 4 downsampling blocks with increasing number of channels, with the number of channels being 64, 128, 256, 384. With the number of channels increasing as the spatial dimension decrease. This dimension of U-net proved to be the most effective experimentally, and amounts to around 20 million parameters. To improve the sensibility of the U-net to the noise variance, \(\alpha _t\) is taken as input, which is then embedded using an ad-hoc sinusoidal transformation by splitting the value in a set of frequencies, in a way similar to positional encoding in Transformers [98]. The embedded noise variance is then vectorized and concatenated to the noisy images along the channel axes before being passed to the U-Net. This helps the network to be highly sensitive to the noise level, which is crucial for good performance.
Conditioning of the model is necessary to guide the diffusion towards a forecast defined by the known previous weather conditions. Our conditioning is applied in a classifier-free manner, by concatenating the conditioning frames to the noisy images alongside the channel axis. Practically, the model \(\epsilon _\theta (x_t, t, y)\) takes as input the noisy images \(x_t\), which represents the high resolution wind speed at time \(t_0\). The conditioning information \(y = \{w_{-6}, w_{-3}, w_{0}, w_{+3}\}\) contains the low resolution wind timestamps at -6 h, -3 h, current time and +3 h ahead.
The input fed into our denoising U-net consists of four pre-emptively upsampled ERA5 images, along with a single image that represents the noisy input. This configuration outputs a singular, denoised image. Given that the image size is set at 256 × 256 pixels, the input dimension is structured as (batch_size, 256, 256, 5), while the output dimension is formulated as (batch_size, 256, 256). Our implementation directly provides the U-net with the conditioning information, specifically, each temporal slice in the input data is treated analogously to a color channel in an RGB image.
4.3 Ensemble diffusion
Similarly to [99], in our work, we leverage the ensemble properties of diffusion models. With the goals of diffusion models being the approximation of the distribution of the training data, it is expected that our generative outcome is the outcome of a probabilistic point in the data distribution. Thus, it is common that for diffusion processes with the same conditioning that the results may vary considerable with different executions. This behaviour of diffusion models may impair the prediction effort, because a single generation may diverge considerably from the mean of the distribution, producing a convincing although highly unlikely outcome inside the distribution. A possible solution to this problem lies in computing the mean of a set of diffusion generations, thus moving our generated image close to the mean value of the distribution and therefore generating a more probable outcome.
In our experiments, we settled for an ensemble numbering 15 executions of the single diffusion model, with each diffusion being comprised of 5 steps. This combination proved to be the optimal configuration with respect to our testing years. Unlike [99], our approach simply calculates the mean between the predicted instances instead of utilizing a post-processing model; this simpler approach showed comparable results with a lower computational overhead.
In Fig. 6, a general overview of the ensemble architecture for evaluation is reported. Starting from the left, we showcase that the inputs are the noise and 4 low-resolution conditioning images. Pre-emptive bilinear upsampling is performed on the low-resolution images before the insertion into the model. Next, the model performs 5 diffusion steps of sequential denoising, with a noise scheduler managing the re-addition of the noise and the computation of noise rates. The noise scheduling process requires the generation of a linear schedule, with denoising operation alternated to re-addition of noise at a lower rate, as reported in Algorithm 3. This operation is performed 15 times as part of the ensemble effort, using the same conditioning information but different input noise. The cumulated outputs are finally combined via a mean operation to produce the downscaled output. This operation is effectively performed in a parallelized manner with a batch size of 32.
The super-resolution diffusion framework employs a sequence-to-one strategy, where pre-upsampling is achieved via bilinear interpolation. The diffusion process comprises five diffusion steps, using a denoising U-net to progressively denoise the input. A noise scheduler orchestrates this process as reported in Algorithm 3. This procedure is replicated 15 times, generating an ensemble of potential super-resolution outcomes. These outcomes are then amalgamated using a mean operation to produce the final output
4.3.1 Training and evaluation
The training procedure of diffusion models, as reported in Algorithm 1, is implemented by randomly selecting an amount of noise, which is subsequently added to the input images, with the denoising U-net tasked to denoise the image. For the evaluation phase the process is reversed, applying the so called reverse-diffusion, as reported in Algorithm 2. The reverse diffusion process involves a sequence of steps, beginning with the introduction of input noise. This procedure alternates between denoising steps and re-injection of noise, as depicted in Fig. 7. These steps are managed via a noise scheduler, which dictates a sequence of noise rates following a linear progression. For instance, in a 5-step process, the noise rates might be set as 1, 0.8, 0.6, 0.4, and 0.2, respectively. The process starts with a denoising operation at a noise rate of 1, then noise is re-injected at a rate of 0.8, followed by another denoising step at the same rate, and the pattern continues through the sequence.
Our model is developed using the TensorFlow/Keras framework. We trained the model on data from 2010 to 2019, and tested it on data from 2009 and 2020, which allowed us to verify its performance across time periods before and after our training years. The training employs a data generator for efficient handling of training batches, creating custom sequences on the fly, and uses memory mapping to load only the necessary data into RAM. The selection of data batches is randomized over the 10-year training period. The training was conducted in batches of 8 over 200 epochs, using the AdamW optimization algorithm. To optimize training, we implemented a callback function that adjusts both the learning rate and weight decay. The learning rate starts at 1e\(-04\) and is gradually reduced to 1e\(-05\), while weight decay starts at 1e\(-05\) and decreases to 1e−06. The training loss, coherently with most diffusion model implementation utilized MAE computed on the noise differences. In the evaluation phase, we employed a generator to produce continuous sequences for each testing year. We configured the batch size to be 32. Taking into account the 3-h time resolution, this approach resulted in batches representing 96-h (4-day) periods.
4.4 Comparative models
In our work, we selected a set of well-known neural models to compare our diffusion architecture with. Our selection include, Efficient Sub-Pixel Convoluted Neural Network (ESPCNN) [15], Enhanced Deep Super-Resolution Network (EDSR) [18], the residual U-net, Super-Resolution Generative Adversarial Network (SRGAN) [28], Vision Transformer (ViT) [100] and Swin Transformer with Long Short-Term Memory (SwinLSTM) [101].
ESPCNN is a well known neural architecture for super-resolution, which utilize sub-pixel convolution as a form of efficient post-upsampling at the very end of the network. This design choice enables the convoluted portion of the network to function at a lower resolution feature dimension, significantly enhancing the network’s efficiency. As a result, ESPCNN maintains a low parameter count, making it not only resource-efficient but also faster in processing. This efficient architecture ensures that ESPCNN delivers high-quality super-resolution images while requiring fewer computational resources, a key advantage in practical applications.
The Enhanced Deep Super-Resolution Network (EDSR) is a well-known architecture in the realm of super-resolution, renowned for its versatility and effectiveness across a broad spectrum of tasks. EDSR architecture is distinguished by its assembly of multiple residual blocks. A key innovation in EDSR is its use of constant scaling layers, which replace the traditional batch normalization layers. EDSR’s design, focused on optimizing image detail and quality, has solidified its status as a top-tier solution in the ever-evolving field of super-resolution technology.
The residual U-net is a U-net implementation which is internally used by the diffusion model to perform the denoising part. The network utilizes residual blocks and implements downscaling via average pooling and upsampling via interpolative resizing. It also excels as a standalone super-resolution model. Its effectiveness in this setting is noteworthy, providing an effective benchmark against the diffusion model. This comparison is particularly significant due to the architectural parallels and similarities in parameter count between the two models. Such shared characteristics allow for a more accurate understanding of their respective strengths and capabilities in the field of super-resolution.
SRGAN [28] is a renowned method for super-resolution that employs the generative adversarial network (GAN) framework, featuring a generator-discriminator setup engaged in a min-max optimization game. GAN-based methods continue to maintain a significant presence in super-resolution tasks, as evidenced by numerous reviews and challenges, underscoring the necessity of including SRGAN in comparative analyses within our research context. The SRGAN study detailed a broad spectrum of architectural decisions, with the most prevalent implementation incorporating a perceptual loss component into the generator’s loss function. However, our experiments indicated that utilizing a mean squared error (MSE)-based non-perceptual loss, in conjunction with a Patch-GAN [102] discriminator, proved to be more efficacious than its perceptual counterpart. This approach enabled us to obtain optimal performance outcomes.
Vision Transformers (ViT) represent a significant advancement in the field of computer vision, primarily by adapting the transformer architecture, which has been highly successful in natural language processing, to image analysis tasks. Introduced by researchers in 2020 [100], ViTs approach image manipulation in a novel way by treating images as sequences of patches, similar to how transformers treat text as sequences of words. This method begins by splitting an image into fixed-size patches, then linearly embedding each of them, similar to token embedding in NLP. The sequence of embedded patches, supplemented by a learnable positional encoding, is then processed through a series of transformer layers. This process allows the model to focus on different parts of the image and learn contextual relationships between patches, leveraging the self-attention mechanism inherent to transformers. This process is then followed by a reconstruction structure which reassembles the patches into an output image. Our implementation utilizes a patch size of 8 × 8, which is projected into an embedding of 256 float values via a dense layer. We utilize 6 transformer layers in parallel, each with 12 heads and concatenated outputs. The reconstruction process is implemented via transposed convolution upsampling alternated with residual blocks.
SwinLSTM [101] is a state of the art transformer-based architecture that combines the strengths of Swin Transformer and Long Short-Term Memory (LSTM) networks. This architecture was successfully applied to a variety of spatio-temporal prediction tasks, where the model needs to capture both spatial and temporal dependencies in the data. This can be critical for super-resolution tasks where the model needs to understand the context and dependencies between pixels or frames. This makes this transformer based technique especially viable for setting, where using a single low-resolution image to predict the high-resolution one would not be enough to provide the information necessary to achieve optimal results.
In Table 1, we report a comparison between the tested models, reporting for each one the number of parameters and the execution time necessary to compute one year of downscaled data (Fig. 8).
Visual evaluation of super-resolution methods for ten selected samples from the years 2009 and 2020 showcasing different weather conditions, including a Mediterranean cyclone. The comparison contains all tested neural downscaling methods alongside the CERRA model. Visual evaluation reveals that U-net, ViT, GAN and Diffusion methods provide the visually best results, with the diffusion models achieving the higher amount of details. The numerical results reported in Table 2 provide a more effective comparison of the respective performances
4.5 Validation
The validation process utilizes in-situ observations from weather stations as the ground truth against which the model predictions are evaluated. This process also results in an assessment of the quality of the reanalysis proposed by ERA5 and CERRA.
The IGRA V2 dataset provides for each measurement the value of the wind speed at different pressure levels, where the maximum value coincides with the altitude of the weather station. To remain consistent with the ERA5 and CERRA data, only the observation with the highest air pressure value was selected for each measurement.
Since the data available for the other datasets refer to measurements relative to the previous hour and with 3-h intervals, the in-situ data considered must also follow a coherent structure. Therefore, given the irregular timestamp distribution of IGRA data, being mainly concentrated around midnight and midday, we decided to remove all detections that did not occur in these two time slots. For what concerns multiple measurements carried out by the same meteorological station during the hour preceding these two slots, these have been reduced to a single average value and the related timestamp rounded to the stroke of the following hour.
To allow a spatially coherent comparison we decided to map all the values to the spatial representation of the in-situ data, in order to keep track of only the values present in ERA5, CERRA and in the results of the models for which there is a correspondence in the validation dataset, thus greatly reducing the space required.
The procedure for assembling the validation dataset is described in Fig. 9. The obtained dataset can be used to calculate the difference between the results obtained from the various models used to approximate CERRA and the ground truth represented by the observations. Although this does not represent our main objective, this type of analysis also allows us to obtain valuable information on the performance of ERA5 and CERRA themselves as reanalysis datasets.
Construction process of the dataset used for the validation of ERA5, CERRA and the generated data. IGRA V2 observations in the reference period undergo filtering depending on the reference time and the air pressure value, after which the values selected for an observation at a given time are averaged and become part of the dataset. At this point, the corresponding values of ERA5 (after calculating the wind speed from the two u and v components) and CERRA are concatenated to each observation. The same process is implemented for the values resulting from the various generative models
5 Experiments and results
In this section, we detail the experiments conducted using the models and data referenced in the preceding sections and provide an analysis of the results obtained.
For conducting our experiments, we strategically chose two distinct years, 2009 and 2020, as our testing periods. For both these years, we utilized the lower resolution ERA5 dataset as our conditioning alongside the CERRA dataset, which acted as our high-resolution ground truth. Additionally, we leveraged the IGRA V2 in-situ dataset, containing ground-based observations from meteorological stations, for validating the data generated by our experiments.
The primary objective of our experiment is to evaluate whether a neural-network-based super-resolution model can effectively replace a traditional downscaling model, like the physics-inspired HARMONIE model currently employed for the generation of CERRA. A key advantage of our neural model is its independence from supplementary data requirements, alongside significantly reduced computational demands. Our comparative analysis studies eight competing neural models: ESPCNN, EDSR, SwinLSTM, SRGAN, U-net, ViT, Single Diffusion, and Ensemble Diffusion. This general comparison includes a variety of models from small, efficient networks all the way to state of the art, high computational demand approaches. This assessment aims to determine not only the feasibility but also the relative performance of these neural models in the context of super-resolution tasks for weather variables.
All the models are trained on a 10-year period with a sequence to one approach, utilizing 4 conditioning low resolution time frames, as described in Sect. 4.1. The models are then evaluated on the two respective years. A summary of the obtained results is reported in Table 2, where we can note that all neural models have a considerably better performance with respect to our traditional downscaling baseline of bilinear interpolation. The model that achieves the best result, coherently for all three metrics, is Ensemble Diffusion. Ensemble Diffusion in comparison to bilinear interpolation improves the MSE value from 2.50e−03 to 1.02e−03 and provides respectively a 15% improvement in PSNR and 20% improvement in SSIM. A set of results for visual inspection is reported in Fig. 8.
The data generated through our process underwent validation against in-situ measurements, providing a crucial comparison to determine if the downscaled version offered tangible improvements over real-world observations. As indicated in Table 3, the validation confirmed that our downscaled data align more closely with the actual wind speed measurements in the area, demonstrating enhanced fidelity.
For our evaluation dataset, we chose the Mean Absolute Error (MAE) as our preferred metric for assessing errors. We selected MAE due to its tangible physical interpretation; being measured in meters per second, it provides a clear indication of the magnitude of the error.
Our validation analysis revealed that CERRA outperforms ERA5 in terms of real-world data correlation, achieving a 12.54% improvement in Mean Squared Error (MSE). Among the downscaling models we tested, The Ensemble Diffusion, SRGAN, and ViT methods emerged as the most effective, demonstrating a performance enhancement of 8.82% over ERA5. Additionally, all three models achieved parity with CERRA in terms of Mean Absolute Error (MAE). These results showcase the capabilities of neural models to effectively approximate the inner mechanics of state-of-the-art physics inspired model, providing both lower computational costs and avoiding the need for additional information that are often difficult to obtain, with common delays in the publication of current data. Our validation was possible only for the year 2009, as validation data was not available for the year 2020.
The Structural Similarity Index (SSIM) is widely acknowledged as a crucial metric for evaluating super-resolution performance, and we are interested to explore the changes in its performance within the training year. Figure 10 presents a comparative analysis of SSIM metrics across the various downscaling methods for the year 2020. Notably, there is a significant fluctuation in performance throughout the year. These variations, consistently observed across all models including bilinear interpolation, suggest that temporal weather variations significantly influence the efficacy of the downscaling models. This consistent trend across different models implies that the performance changes are predominantly driven by weather-related factors rather than the intrinsic characteristics of the models themselves.
Furthermore, an in-depth geographical analysis of SSIM performance can provide insights into which areas are harder to effectively downscale. Given that SSIM is calculated over patches, it’s feasible to compile a comprehensive SSIM output. This output takes the form of a matrix, where each pixel value denotes the SSIM computed over a window centered on that corresponding pixel in the input images. In Fig. 11, we showcase mean images derived from averaging these full SSIM matrices for the entire testing years of 2009 and 2020, respectively. These images reveal that coastal regions typically exhibit higher error rates, with the most pronounced discrepancies observed in the high Adriatic Sea, the Strait of Messina, and the Ligurian Sea. The results also indicate that the geographical errors are remarkably similar across the two testing years, suggesting that there has been no significant change in the error distribution between these comparative periods.
As a final experiment, considering the unavailability of CERRA for the years ranging from 2021 to 2024, we also make publicly available for further studies our computed approximation of CERRA, computed via Ensemble Diffusion.
Mean image representing the spatial SSIM error for both testing years of 2009 and 2020, showcasing the downscaling performance relative to different geographical areas. The similarity in the results demonstrates that the errors made by the model are consistent with respect to the two years used for the test
6 Discussion and conclusion
In this research, we successfully employed state of the art super-resolution diffusion models to downscale wind speed data from ERA5 to CERRA. This approach approximates the existing model used for generating CERRA, but it innovatively does so in a data-driven manner without the need for supplementary data. Our experiment was focused on the Mediterranean area surrounding Italy and utilizing wind speed as our main weather variable.
Our experiments, conducted over two distinct testing years, demonstrated that the downscaling computed by our method aligns closely with the outcomes of traditional physics-based models. This alignment was evident both in the similarity of results and in their concordance with measurements from ground-based in-situ stations. Among all the models tested, the Ensemble Diffusion approach achieved the most accurate results. However, it’s noteworthy that even less complex super-resolution models consistently outperformed basic methods like bilinear interpolation. Our findings reveal the practicality of using super-resolution diffusion models in providing timely and informative downscaled data for meteorological studies. Given the significant delays often encountered in downscaled reanalyses—for example, CERRA’s data lagging over 2 years behind the current date at the time of this study—our findings are particularly relevant for the scientific community. These delays typically arise from the unavailability of additional information needed by physics-based models, which depend on various sources and are subject to time-consuming computational processes. Our approach, leveraging a neural model trained on previously generated data, offers a timely alternative for accessing current downscaled results. This method not only reduces computational costs but also maintains a high quality of output. Importantly, our model produces data that aligns more closely with actual measurements than the original low-resolution data, demonstrating its practical potential in enhancing real-time weather analysis.
In our study, we encountered differing time scales between the high-resolution dataset (CERRA) and the low-resolution dataset (ERA5). ERA5 offers hourly temporal resolution, whereas CERRA provides data at a 3-h resolution. Given that our model can generate sequences conditioned only on ERA5 data, we can utilize our model to augment the temporal resolution of our downscaled data to match the temporal resolution of ERA, providing a higher temporal resolution when compared to the physics-inspired reanalyses model.
Regarding the main limitations of our approach, it’s important to note that our data-driven method requires pre-existing downscaled data for training, thus relying on already established models. Additionally, our experiments represent a preliminary phase, focusing solely on a single weather variable and a relatively limited geographical area. Future developments could encompass a broader range of weather variables and potentially cover the entire region addressed by CERRA. This expanded scope could lead to improved results, as the model would gain a more comprehensive understanding of the weather system with access to diverse meteorological data. Moreover, hardware constraints limited our training dataset to only 10 years, which is noteworthy given the high quality of our results. In future research, it would be beneficial to utilize a more extensive dataset, potentially spanning the full 40 years available from CERRA, to enhance the model’s accuracy and predictive capabilities.
In conclusion, our research presents a viable alternative for obtaining downscaled data of current events. We demonstrate the potential of diffusion models to effectively approximate the processes of traditional downscaling models, avoiding delays in data availability. Furthermore, our findings suggest a promising future for reanalysis models, potentially shifting towards neural network-based approaches that leverage the advantages identified in this study. This shift could mark a significant advancement in the field, offering more efficient and accurate ways to handle meteorological data.
Code availability
The code and downscaled data relative to the presented work is archived at the following GitHub https://github.com/fmerizzi/ERA5-to-CERRA-via-Diffusion-Models/. We also provide the full list of the environmental libraries and respective versions.
References
Hersbach H, Bell B, Berrisford P, Biavati G, Horányi A, Muñoz Sabater J, Nicolas J, Peubey C, Radu R, Rozum I, Schepers D, Simmons A, Soci C, Dee D, Thépaut J.-N (2018) ERA5 hourly data on single levels from 1940 to present. Copernicus climate change service (C3S) climate data store (CDS). Accessed from 22 Nov 2023
Schimanke S, Ridal M, Le Moigne P, Berggren L, Undén P, Randriamampianina R, Andrea U, Bazile E, Bertelsen A, Brousseau P, Dahlgren P, Edvinsson L, El Said A, Glinton M, Hopsch S, Isaksson L, Mladek R, Olsson E, Verrelle A, Wang Z.Q (2021) CERRA sub-daily regional reanalysis data for Europe on single levels from 1984 to present. Copernicus climate change service (C3S) climate data store (CDS). Accessed from 22 Nov 2023
Durre I, Yin X, Vose RS, Applequist S, Arnfield J, Korzeniewski B, Hundermark B (2016) Integrated global radiosonde archive (IGRA), Version 2. NOAA National Centers for Environmental Information
Anwar S, Khan S, Barnes N (2020) A deep journey into super-resolution: a survey. ACM Comput Surv 53(3):1–34
Lepcha DC, Goyal B, Dogra A, Goyal V (2023) Image super-resolution: a comprehensive review, recent trends, challenges and applications. Inf Fusion 91:230–260
Bashir SMA, Wang Y, Khan M, Niu Y (2022) A comprehensive review of deep learning-based single image super resolution
Bengtsson L, Andrae U, Aspelien T, Batrak Y, Calvo J, de Rooy W, Gleeson E, Hansen-Sass B, Homleid M, Hortal M et al (2017) The harmonie-arome model configuration in the aladin-hirlam nwp system. Mon Weather Rev 145(5):1919–1935
...Termonia P, Fischer C, Bazile E, Bouyssel F, Brožková R, Bénard P, Bochenek B, Degrauwe D, Derková M, El Khatib R, Hamdi R, Mašek J, Pottier P, Pristov N, Seity Y, Smolíková P, Španiel O, Tudor M, Wang Y, Wittmann C, Joly A (2018) The aladin system and its canonical model configurations arome cy41t1 and alaro cy40t1. Geosci Model Dev 11(1):257–281
Dong C, Loy CC, He K, Tang X (2016) Image super-resolution using deep convolutional networks. IEEE Trans Pattern Anal Mach Intell 38(2):295–307
Dong C, Loy C.C, He K, Tang X (2014) Learning a deep convolutional network for image super-resolution. In: Computer vision–ECCV 2014: 13th European conference, Zurich, Switzerland, September 6–12, 2014, Proceedings, Part IV 13. Springer, pp 184–199
Wang Z, Liu D, Yang J, Han W, Huang T (2015) Deep networks for image super-resolution with sparse prior. In: Proceedings of the IEEE international conference on computer vision, pp 370–378
Li Z, Yang J, Liu Z, Yang X, Jeon G, Wu W (2019) Feedback network for image super-resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 3867–3876
Dai T, Cai J, Zhang Y, Xia S.-T, Zhang L (2019) Second-order attention network for single image super-resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 11065–11074
Hu X, Mu H, Zhang X, Wang Z, Tan T, Sun J (2019) Meta-sr: a magnification-arbitrary network for super-resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 1575–1584
Shi W, Caballero J, Huszár F, Totz J, Aitken AP, Bishop R, Rueckert D, Wang Z (2016) Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1874–1883
Lu Z, Li J, Liu H, Huang C, Zhang L, Zeng T (2022) Transformer for single image super-resolution. In: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pp 457–466
Kim J, Lee JK, Lee KM (2016) Accurate image super-resolution using very deep convolutional networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1646–1654
Lim B, Son S, Kim H, Nah S, Mu Lee K (2017) Enhanced deep residual networks for single image super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 136–144
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Ahn N, Kang B, Sohn K-A (2018) Fast, accurate, and lightweight super-resolution with cascading residual network. In: Proceedings of the European conference on computer vision (ECCV), pp 252–268
Zhang Y, Li K, Li K, Wang L, Zhong B, Fu Y (2018) Image super-resolution using very deep residual channel attention networks. In: Proceedings of the European conference on computer vision (ECCV), pp 286–301
Choi J-S, Kim M (2017) A deep convolutional neural network with selection units for super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 154–160
Ren H, El-Khamy M, Lee J (2017) Image super resolution based on fusing multiple convolution neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 54–61
Hu Y, Gao X, Li J, Huang Y, Wang H (2018) Single image super-resolution via cascaded multi-scale cross network. arXiv preprint arXiv:1802.08808
Haris M, Shakhnarovich G, Ukita N (2018) Deep back-projection networks for super-resolution. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1664–1673
Tong T, Li G, Liu X, Gao Q (2017) Image super-resolution using dense skip connections. In: Proceedings of the IEEE international conference on computer vision, pp 4799–4807
Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2020) Generative adversarial networks. Commun ACM 63(11):139–144
Ledig C, Theis L, Huszár F, Caballero J, Cunningham A, Acosta A, Aitken A, Tejani A, Totz J, Wang Z et al. (2017) Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4681–4690
Wang X, Yu K, Wu S, Gu J, Liu Y, Dong C, Qiao Y, Change Loy C (2018) Esrgan: enhanced super-resolution generative adversarial networks. In: Proceedings of the European conference on computer vision (ECCV) workshops, pp 0–0
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser Ł, Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems 30
Liang J, Cao J, Sun G, Zhang K, Van Gool L, Timofte R (2021) Swinir: image restoration using swin transformer. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 1833–1844
Conde MV, Choi U-J, Burchi M, Timofte R (2022) Swin2sr: Swinv2 transformer for compressed image super-resolution and restoration. In: European conference on computer vision. Springer, pp 669–687
Song J, Meng C, Ermon S (2020) Denoising diffusion implicit models. arXiv preprint arXiv:2010.02502
Li H, Yang Y, Chang M, Chen S, Feng H, Xu Z, Li Q, Chen Y (2022) Srdiff: single image super-resolution with diffusion probabilistic models. Neurocomputing 479:47–59
Horé A, Ziou D (2010) Image quality metrics: Psnr vs. ssim. In: 2010 20th international conference on pattern recognition, pp 2366–2369
Chiang C-H, Huang Z-H, Liu L, Liang H-C, Wang Y-C, Tseng W-L, Wang C, Chen C-T, Wang K-C (2024) Climate downscaling: a deep-learning based super-resolution model of precipitation data with attention block and skip connections. arXiv preprint arXiv:2403.17847
Kurinchi-Vendhan R, Lütjens B, Gupta R, Werner L, Newman D (2021) Wisosuper: benchmarking super-resolution methods on wind and solar data. arXiv preprint arXiv:2109.08770
Tie R, Shi C, Wan G, Hu X, Kang L, Ge L (2022) Cldassd: reconstructing fine textures of the temperature field using super-resolution technology. Adv Atmos Sci 39(1):117–130
Harder P, Yang Q, Ramesh V, Sattigeri P, Hernández-García A, Watson C, Szwarcman D, Rolnick D (2022) Generating physically-consistent high-resolution climate data with hard-constrained neural networks. arXiv:2208.05424
Ho J, Jain A, Abbeel P (2020) Denoising diffusion probabilistic models. In: Larochelle H, Ranzato M, Hadsell R, Balcan M, Lin H (eds.) Advances in neural information processing systems 33: annual conference on neural information processing systems 2020, NeurIPS 2020, December 6–12, 2020, Virtual
Rezende D.J, Mohamed S, Wierstra D (2014) Stochastic backpropagation and approximate inference in deep generative models. In: Proceedings of the 31th international conference on machine learning, ICML 2014, Beijing, China, 21–26 June 2014. JMLR workshop and conference proceedings, JMLR.org. vol 32, pp 1278–1286
Kingma DP, Welling M (2019) An introduction to variational autoencoders. Found Trends Mach Learn 12(4):307–392
Asperti A, Evangelista D, Piccolomini EL (2021) A survey on variational autoencoders from a green AI perspective. SN Comput Sci 2(4):301
Song J, Meng C, Ermon S (2020) Denoising diffusion implicit models. arXiv: 2010.02502
Sohl-Dickstein J, Weiss EA, Maheswaranathan N, Ganguli S (2015) Deep unsupervised learning using nonequilibrium thermodynamics 37: 2256–2265
Khan MS, Coulibaly P, Dibike Y (2006) Uncertainty analysis of statistical downscaling methods. J Hydrol 319(1–4):357–382
Wilby RL, Dawson CW (2013) The statistical downscaling model: insights from one decade of application. Int J Climatol 33(7):1707–1719
Semenov MA, Barrow EM, Lars-Wg A (2002) A stochastic weather generator for use in climate impact studies. User Man Herts UK, pp 1–27
Gangopadhyay S, Clark M, Rajagopalan B (2005) Statistical downscaling using k-nearest neighbors. Water Resour Res 41(2)
Pham QB, Yang T-C, Kuo C-M, Tseng H-W, Yu P-S (2019) Combing random forest and least square support vector regression for improving extreme rainfall downscaling. Water 11(3):451
He X, Chaney NW, Schleiss M, Sheffield J (2016) Spatial downscaling of precipitation using adaptable random forests. Water Resour Res 52(10):8217–8237
Pour SH, Shahid S, Chung E-S (2016) A hybrid model for statistical downscaling of daily rainfall. Procedia Eng 154:1424–1430
Liu J, Yuan D, Zhang L, Zou X, Song X (2016) Comparison of three statistical downscaling methods and ensemble downscaling method based on bayesian model averaging in upper hanjiang river basin, china. Adv Meteorol 2016:1–12
San-Martín D, Manzanas R, Brands S, Herrera S, Gutiérrez JM (2017) Reassessing model uncertainty for regional projections of precipitation with an ensemble of statistical downscaling methods. J Clim 30(1):203–223
Vandal T, Kodra E, Ganguly S, Michaelis A, Nemani R, Ganguly AR (2017) DeepSD: generating high resolution climate change projections through single image super-resolution
Baño-Medina J, Manzanas R, Gutiérrez JM (2020) Configuration and intercomparison of deep learning neural models for statistical downscaling. Geosci Model Dev 13(4):2109–2124
Misra S, Sarkar S, Mitra P (2018) Statistical downscaling of precipitation using long short-term memory recurrent neural networks. Theor Appl Climatol 134:1179–1196
Sha Y, Gagne DJ II, West G, Stull R (2020) Deep-learning-based gridded downscaling of surface meteorological variables in complex terrain. Part II: daily precipitation. J Appl Meteorol Climatol 59(12):2075–2092
Chaudhuri C, Robertson C (2020) Cligan: a structurally sensitive convolutional neural network model for statistical downscaling of precipitation from multi-model ensembles. Water 12(12):3353
Ji Y, Zhi X, Tian Y, Peng T, Huo Z, Ji L-Y (2020) Downscaling of precipitation forecasts based on single image super-resolution
Nguyen T, Jewik J, Bansal H, Sharma P, Grover A (2023) Climatelearn: benchmarking machine learning for weather and climate modeling. arXiv preprint arXiv:2307.01909
Pelosi A (2023) Performance of the copernicus european regional reanalysis (cerra) dataset as proxy of ground-based agrometeorological data. Agric Water Manag 289:108556
Monteiro D, Morin S (2023) Multi-decadal analysis of past winter temperature, precipitation and snow cover data in the European alps from reanalyses, climate models and observational datasets. Cryosphere 17(8):3617–3660
Galanaki E, Giannaros C, Agathangelidis I, Cartalis C, Kotroni V, Lagouvardos K, Matzarakis A (2023) Validating the copernicus European regional reanalysis (cerra) dataset for human-biometeorological applications. Environ Sci Proc 26(1):111
Cavalleri F, Viterbo F, Brunetti M, Bonanno R, Manara V, Lussana C, Maugeri M et al (2023) Inter-comparison and validation of high-resolution surface air temperature reanalysis fields over italy
Durre I, Yin X, Vose RS, Applequist S, Arnfield J (2018) Enhancing the data coverage in the integrated global radiosonde archive. J Atmos Ocean Technol 35(9):1753–1770
Hersbach H, Bell B, Berrisford P, Hirahara S, Horányi A, Muñoz-Sabater J, Nicolas J, Peubey C, Radu R, Schepers D et al (2020) The era5 global reanalysis. Q J R Meteorol Soc 146(730):1999–2049
Yuan P, Hunegnaw A, Alshawaf F, Awange J, Klos A, Teferle FN, Kutterer H (2021) Feasibility of era5 integrated water vapor trends for climate change analysis in continental Europe: an evaluation with gps (1994–2019) by considering statistical significance. Remote Sens Environ 260:112416
Lavers DA, Simmons A, Vamborg F, Rodwell MJ (2022) An evaluation of era5 precipitation for climate monitoring. Q J R Meteorol Soc 148(748):3152–3165
Chiaravalloti F, Caloiero T (2022) The long-term era5 data series for trend analysis of rainfall in Italy. Hydrology 9(2):18
Shen L, Wen J, Zhang Y, Ullah S, Cheng J (2022) Changes in population exposure to extreme precipitation in the yangtze river delta, China. Clim Serv 27:100317
Yilmaz M (2023) Accuracy assessment of temperature trends from era5 and era5-land. Sci Total Environ 856:159182
Liu J, Hagan DFT, Liu Y (2020) Global land surface temperature change (2003–2017) and its relationship with climate drivers: airs, modis, and era5-land based analysis. Remote Sens 13(1):44
Lee J, Dessler AE (2024) Improved surface urban heat impact assessment using goes satellite data: a comparative study with era-5. Geophys Res Lett 51(1):2023–107364
Olauson J (2018) Era5: the new champion of wind power modelling? Renew Energy 126:322–331
Cai Y, Bréon F-M (2021) Wind power potential and intermittency issues in the context of climate change. Energy Convers Manage 240:114276
Dullaart JC, Muis S, Bloemendaal N, Aerts JC (2020) Advancing global storm surge modelling using the new era5 climate reanalysis. Clim Dyn 54:1007–1021
Al-Sakkaf AS, Jiahua Z, Fengmei Y, Hamed MM, Simbi CH, Ahmed A, Shahid S (2024) Assessing exposure to climate extremes over the Arabian peninsula using era5 reanalysis data: spatial distribution and temporal trends. Atmos Res 107224
Dong T, Dong W (2021) Evaluation of extreme precipitation over asia in cmip6 models. Clim Dyn 57(7–8):1751–1769
Wei W, Zou S, Duan W, Chen Y, Li S (2023) Spatiotemporal variability in extreme precipitation and associated large-scale climate mechanisms in central asia from 1950 to 2019. J Hydrol 620:129417
Rolle M, Tamea S, Claps P (2020) Improved large-scale crop water requirement estimation through new high-resolution reanalysis dataset, 19289
Zhang R, Li L, Zhang Y, Huang F, Li J, Liu W, Mao T, Xiong Z, Shangguan W (2021) Assessment of agricultural drought using soil water deficit index based on era5-land soil moisture data in four southern provinces of china. Agriculture 11:411
Tarek M, Brissette F (2020) Evaluation of the era5 reanalysis as a potential reference dataset for hydrological modelling over north America. Hydrol Earth Syst Sci 24:2527–2544
Jiao D, Xu N, Yang F, Xu K (2021) Evaluation of spatial-temporal variation performance of era5 precipitation data in china. Sci Rep 11
Copernicus Regional Reanalysis for Europe (CERRA). https://climate.copernicus.eu/copernicus-regional-reanalysis-europe-cerra
Galanaki E, Giannaros C, Agathangelidis I, Cartalis C, Kotroni V, Lagouvardos K, Matzarakis A (2023) Validating the copernicus European regional reanalysis (cerra) dataset for human-biometeorological applications. Environ Sci Proc 26(1):111
Nicodemo G, Pecoraro G, Rianna G, Reder A, Luongo D, Peduto D, Calvello M (2023) Remote sensing meteorological and dinsar historical data to analyse the kinematic behaviour of slow-moving landslides at municipal scale. In: National conference of the researchers of geotechnical engineering. Springer, pp 242–250
Verrelle A, Glinton M, Bazile E, Moigne PL (2021) Cerra-land : a new land surface reanalysis at 5.5 km resolution over Europe
Masloumidis I, Dafis S, Kyros G, Lagouvardos K (2023) Snow depth trends of european ski resorts. Environ Sci Proc 26(1)
Copernicus Climate Change Service (C3S) (2023) ERA5 hourly data on single levels from 1940 to present. Copernicus climate change service (C3S) climate data store (CDS). Accessed from 22 Nov 2023
Schulzweida U (2023) CDO user guide. Zenodo
Gurrola-Ramos J, Dalmau O, Alarcón TE (2021) A residual dense u-net neural network for image denoising. IEEE Access 9:31742–31754
Lee S, Negishi M, Urakubo H, Kasai H, Ishii S (2020) Mu-net: multi-scale u-net for two-photon microscopy image denoising and restoration. Neural Netw 125:92–103
Heinrich MP, Stille M, Buzug TM (2018) Residual u-net convolutional neural network architecture for low-dose ct denoising. Curr Dir Biomed Eng 4(1):297–300
Komatsu R, Gonsalves T (2020) Comparing u-net based models for denoising color images. AI 1(4):465–486
Dhariwal P, Nichol A (2021) Diffusion models beat gans on image synthesis. Adv Neural Inf Process Syst 34:8780–8794
Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention. Springer, pp 234–241
Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser L, Polosukhin I (2017) Attention is all you need. In: Advances in neural information processing systems 30: annual conference on neural information processing systems 2017, December 4–9, 2017, Long Beach, CA, USA, pp 5998–6008
Asperti A, Merizzi F, Paparella A, Pedrazzi G, Angelinelli M, Colamonaco S (2023) Precipitation nowcasting with generative diffusion models. arXiv preprint arXiv:2308.06733
Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S et al (2020) An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929
Tang S, Li C, Zhang P, Tang R (2023) Swinlstm: improving spatiotemporal prediction accuracy using swin transformer and lstm. In: Proceedings of the IEEE/CVF international conference on computer vision, pp 13470–13479
Isola P, Zhu J-Y, Zhou T, Efros AA (2017) Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1125–1134
Acknowledgements
ERA5 and CERRA were downloaded from the Copernicus Climate Change Service (C3S) (2023). This research was partially funded and supported by the following Projects: European Cordis Project “Optimal High Resolution Earth System Models for Exploring Future Climate Changes” (OptimESM), Grant agreement ID: 101081193 Future AI Research (FAIR) project of the National Recovery and Resilience Plan (NRRP), Mission 4 Component 2 Investment 1.3 funded from the European Union - NextGenerationEU. ISCRA Project “AI for weather analysis and forecast” (AIWAF)
Funding
Open access funding provided by Alma Mater Studiorum - Università di Bologna within the CRUI-CARE Agreement.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no conflict of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Model calibration
Model calibration
This appendix contains the hyper-parameters set used to calibrate each described model. See Tables 4, 5, 6, 7, 8, 9, 10
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Merizzi, F., Asperti, A. & Colamonaco, S. Wind speed super-resolution and validation: from ERA5 to CERRA via diffusion models. Neural Comput & Applic 36, 21899–21921 (2024). https://doi.org/10.1007/s00521-024-10139-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-024-10139-9