Open Access
Issue
J. Space Weather Space Clim.
Volume 15, 2025
Article Number 34
Number of page(s) 18
DOI https://doi.org/10.1051/swsc/2025029
Published online 21 August 2025

© M. Cobos-Maestre et al., Published by EDP Sciences 2025

Licence Creative CommonsThis is an Open Access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

1 Introduction

The solar wind is a stream of plasma emitted from the upper atmosphere of the Sun. Coronal holes (CHs) – areas of the solar corona characterized by open magnetic field lines – often result in high speed streams (HSS; Verbanac et al., 2011; Hofmeister et al., 2022) within the solar wind. These HSSs are typically associated with an increase in drag in the upper Earth atmosphere, potentially impacting satellite operation (Nwankwo, 2018). In particularly intense cases, they may lead to geomagnetic storms, potentially disrupting communications and the operation of sensitive services (e.g., Vybornov & Sheiner, 2021).

Studies centered on forecasting solar wind typically focus on its bulk speed, V, or its proton density, ρ. Techniques applied range from physical and empirical modeling methods, like the Wang-Sheeley-Arge (WSA) model (Arge et al., 2003) or EUHFORIA (Pomoell & Poedts, 2018), to deep neural networks like the ones proposed by Upendran et al. (2020), Raju & Das (2021), Brown et al. (2022), and Cobos-Maestre et al. (2024).

Upendran et al. (2020) use a deep neural network with Solar Dynamics Observatory (SDO; Pesnell et al., 2012) Atmospheric Imaging Assembly (AIA; Lemen et al., 2012) extreme ultraviolet (EUV) data as input to predict daily averages of V. The model proposed combines convolutional and long short-term memory (LSTM; Hochreiter & Schmidhuber, 1997) layers to produce a forecast with a lead time of up to four days, obtaining a root mean squared error (RMSE) of 82.3 ± 1.7 km s−1.

Raju & Das (2021) extend the results obtained by Upendran et al. (2020) by applying ballistic tracing to match solar wind speed measurements to specific AIA samples on the 19.3 nm wavelength. A convolutional neural network (CNN) is applied in this approach. When forcing a fixed lead time of four days, an RMSE of 76.3 ± 1.9 km s−1 is reported, improving upon previous results. The ballistic tracing itself is performed by assuming constant velocity matching the measurement taken at the L1 Lagrange point. The authors note that several different solar wind speed measurements can be mapped to a single image sample, and that some image samples may be left unmapped to solar wind measurements. Because arrival time is not accounted for, uncertainties are explicitly stated to surface in the forecasting process. As a result, the authors highlighted that quantifying predictive uncertainty improves reliability of the resulting model.

Brown et al. (2022) further build upon Raju & Das (2021)’s work by applying attention mechanisms to replace the convolutional layers present in the work of Upendran et al. (2020), and by focusing on a specific subset of the image sample around the center of the image. AIA samples on the 21.1 nm wavelength are used, and provided as a sequence. After these modifications, an RMSE of 71.9 ± 0.4 km s−1 is reported, offering a new improvement over previous results.

Most recently, we explored the possibility of applying a fully autoregressive scheme based on the application of self-attention (Cobos-Maestre et al., 2024). We named this model the Solar Wind Attention Network (SWAN), and we found that, by using the persistence of CHs over several Carrington rotations (Cranmer, 2009), we could reduce computational complexity. Notably, for aggregations of one day, training time was reduced to 45 s while using CPU. Without using solar disk information, we obtained an improvement over results previously reported in the literature, with an RMSE of 68.8 ± 1.1 km s−1 for a lead time of three days. One of the primary open questions identified in our study was whether incorporating solar disk information could yield a meaningful enough performance boost to offset the added computational cost.

Out of all Machine Learning (ML) works discussed, only Raju & Das (2021) quantify the uncertainty of their model. They do so by applying Monte Carlo dropout (Gal & Ghahramani, 2016) over the parameters of the model, and sampling repeatedly to obtain a measure of stability in the output. The uncertainty quantified in this method corresponds exclusively to that related to the model and the knowledge it has extracted, serving as a quantification of reliability.

The need to ascertain the reliability of any given prediction when applying ML models in Space Weather is discussed at length by Camporeale (2019). Indeed, strategies for uncertainty estimation on this field are later provided in Camporeale et al. (2019), with an emphasis on deferring the prediction of uncertainty to a second, auxiliary model tasked exclusively with predicting the standard deviation σ of a conditional probability distribution. Under this approach, the output of the primary model is treated as the mean of a normal conditional probability distribution function, and the combination of both models yields a measure of uncertainty.

More generally, uncertainty quantification of neural networks is usually performed using ensembles (Rahaman & Thiery, 2021) or Monte Carlo dropout (Gal & Ghahramani, 2016). Both approaches aim to treat the generated model as a conditional probability distribution that can be sampled repeatedly. From the range of sampled values, it becomes possible to extract a measure of the uncertainty associated with any given prediction. However, the adequacy of Monte Carlo dropout as a general method for uncertainty quantification has been brought into question on the grounds of it arguably not being properly Bayesian (Le Folgoc et al., 2021).

An alternative solution is the usage of probabilistic neural networks (PNN; Zaknich, 1998). These modified neural networks, when applied to regression, replace the traditional pointwise output of the neural network with a probability density function. The goal is, then, for this probability density function to estimate the conditional probability of the target output for any given input. To ensure applicability of common gradient descent methods, the so called reparameterization trick (Jankowiak & Obermeyer, 2018) is often used. The regression model predicts, then, the parameters defining a probability density function of the desired family, enabling training of the model by backpropagating the divergence from the empirical conditional probability distribution.

Bayesian Neural Networks (BNN; MacKay, 1995; Wilson & Izmailov, 2020; Charnock et al., 2022; Magris & Iosifidis, 2023) replace some or all the weights present in a neural network with Bayesian inference processes (Fortuin, 2022). By representing weights in this fashion, predictions made by the neural network become non-deterministic, instead stemming from the sampling of an underlying probability distribution. This provides a direct framework for estimating prediction uncertainty when applied in a PNN to form a probabilistic Bayesian neural network (PBNN). Indeed, the works of Caceres et al. (2021) and Le Folgoc et al. (2021) experimentally shows that PBNNs offer better reliability in uncertainty estimation than techniques like Monte Carlo dropout.

In this paper, we present BaLNet, a PBNN for solar wind speed forecasting that relies on an initial LSTM layer to extract time series features. We use a combination of solar wind and CH measurements as input for the network, and produce a normal distribution of values as the output. By leveraging the variational nature of the model, we show that it is possible to quantify uncertainty, and assign specific contributions to the input data and to the network’s own shortcomings. Furthermore, the architecture is lightweight enough to enable training in a CPU, with an average training time of 2.17 h on it.

The remainder of this paper is structured as follows: Section 2 describes the problem formulation taken and the dataset used. Section 3 describes the BNN proposed, the parameter optimization procedure applied, and an overview of relevant operational considerations. Section 4 describes and discusses the experimental results obtained. Finally, Section 5 provides a summary of the study conducted, as well as conclusions drawn from it.

2 Problem formulation and dataset

2.1 Problem formulation

Following the work of Valdenegro-Toro & Mori (2022), the problem of forecasting V at the L1 point is taken as a regression with uncertainty quantification. To this end, the regression target is an expectation distribution for a dependent variable (Y) given a set of independent variables (X) that can be estimated as

P(Y|X=x)N(μβ(x),σγ2(x)),$$ P(Y|X=x)\sim \mathcal{N}({\mu }_{\beta }(x),{\sigma }_{\gamma }^2(x)), $$(1)

where N$ \mathcal{N}$ denotes a normal probability distribution with mean defined by a parametric function μβ and variance defined by a second parametric function σγ2$ {\sigma }_{\gamma }^2$. β and γ refer to the parameters of each function. Because the model is a BNN, β and γ are stochastic functions of X generated through a Bayesian inference process. Thus, the predicted V will be presented as a normal distribution denoting the probability of each possible value.

2.2 Dataset description

The dataset used is comprised of solar wind observations and coronal hole measurements. Solar wind observations are taken from the OMNI 1 h resolution level 2 dataset (Mathews & Towheed, 1995). Solar wind observations in the dataset are comprised of tabular measurements made by the Advanced Composition Explorer (ACE; Stone et al., 1998), Solar wind electron proton alpha monitor (SWEPAM; McComas et al., 1998) and magnetic fields experiment (MAG; Smith et al., 1998) instruments. The measured bulk speed, V, plasma temperature, T, proton density, ρ, and the magnitude of the magnetic field vector, |B| are used as inputs.

To characterize CHs, we use their area, location and average magnetic flux. These measurements are computed from SDO AIA 19.3 nm level 1 image samples and Helioseismic and Magnetic Imager (HMI; Schou et al., 2012) level 1 magnetogram samples. These data have minimal processing and are available in almost real time. As a result, the model is forced to learn how to perform in the presence of corrupted samples that might lead to misidentified or missing CHs. Identification and feature extraction are performed algorithmically, as presented in the next subsection, with a daily sampling rate.

The combined dataset encompasses from 2012-09-26 6:00 to 2021-01-31 23:00, both included, as forecasted dates. A total of 81,210 samples are used for all experiments presented in this work.

2.3 Coronal hole feature extraction

The CH identification process is based on the ASSA algorithm.1 ASSA is chosen as a basis because it identifies visually dark structures that do not present the magnetic imbalance that defines CHs so that they can be discarded. Our algorithm expands upon this approach by leveraging Gaussian expansion over candidate CH regions to provide closer adherence to CH boundaries. Specifically, it consists of three steps (Cobos Maestre, 2020). First, potential CHs are identified using a thresholding technique based on intensity measurements. Next, a Gaussian expansion is applied to capture transitional regions and merge overlapping candidates into cohesive clusters. Finally, the skewness of the magnetic field distribution in each candidate is evaluated, and those with skewness values below 1 are discarded to remove filaments (Krista & Gallagher, 2009; Boucheron et al., 2016). Figure 1 presents an example of CH identification using this method compared to the original ASSA identification.

thumbnail Figure 1

Comparison of CH identification using the ASSA algorithm (left panel) and our algorithm (right panel). The sample shows a large, centered CH during 2019-09-25.

Once CHs are identified, three main properties are extracted: area, location, and magnetic flux. The area is measured in arcsec−2 using Helioprojective Cartesian coordinates (Thompson, 2006). The location is determined using a bounding box defined by its bottom-left and top-right corners in pixel coordinates. The average magnetic flux is calculated by summing the per-pixel magnetic flux values within the CH and dividing by the total area, expressed in G arcsec−2. Including the magnetic flux as a feature results in a reduction of the root mean square error (RMSE) by approximately 10 km s−1. Similarly, removing the area from the input results in an increase in RMSE of ∼10 km s−1. Furthermore, using the average magnetic flux rather than the total magnetic flux improves RMSE by an additional 3 km s−1.

To ensure that the entire dataset fits into memory during training, CH features are extracted on a daily basis. During validation, it was verified that increasing the measurement frequency up to once every three hours does not significantly impact numerical results or model behavior, confirming that daily measurements are sufficient.

2.4 Preprocessing

To enable model training, all data must be sampled at the same frequency. Additionally, data cleaning and normalization are essential to ensure model convergence and to prevent invalid features containing non-numerical values from corrupting the graph structure. Since the solar wind data is sampled hourly, the CH measurements – originally recorded at a daily frequency – are resampled to match this rate. This resampling introduces missing values that need to be filled. We initially tested two different methods to handle the missing entries: forward filling and linear interpolation.

In our tests, forward filling increased the required convergence time by approximately 12% without yielding any improvement in numerical results. Because of this, we selected linear interpolation for all subsequent experiments.

Two methods for linear interpolation are explored: one where interpolation is performed first, followed by computing the average absolute magnetic flux of each CH, and another where the average flux is computed first, followed by interpolation. The former is mathematically grounded, since interpolating first reduces the risk of distorting the data by not accounting for potential non-linearities. However, empirical results do not show significant differences in model performance between the two approaches in any of the metrics evaluated. Additionally, interpolating first requires six times more floating point operations, which significantly increases computational cost and scales poorly to large datasets. As results are equivalent, it is decided to perform interpolation last in favor of computational efficiency. To further enhance numerical stability, we standardize all features using the formula:

x'=x-μxσx,$$ {x\prime}=\frac{x-{\mu }_x}{{\sigma }_x}, $$(2)

where x is the original feature value, μx is the feature’s mean, σx is its standard deviation, and x′ is the standardized value. This ensures that all features have zero mean and unit variance, making the optimization process more stable.

Finally, we split the dataset into training, validation, and testing subsets with a 70%–15%–15% ratio. Table 1 details the exact distribution of samples between these three subsets. The dataset is split sequentially to prevent the risk of information leakage associated with the use of training examples appearing in the time series after the test data. Information leakage in this experimental setup could lead to unreasonably low error values by allowing the model to learn from future data, which would not be available in an operational setting (Zhou et al., 2024). Additionally, since the dataset spans different phases of the solar cycle, distributional shifts (Kuznetsov et al., 2015) may distort results obtained through cross-validation (Moreno-Torres et al., 2012).

Table 1

Division of the dataset between training, validation, and testing subsets.

To further assess the generalizability of the sequential results, five-fold cross-validation is performed, ensuring that the dataset is split sequentially to preserve the causal structure of the time series. Then, each fold is used for testing and for validation once, and the results are collated. The procedure is depicted in Figure 2. To verify the presence and potential impact of information leakage and distributional shift, aggregated and fold-wise results are then compared to those obtained in the sequential setting.This comparison enables testing on the complete solar cycle while controlling for information leakage and distributional shift. Therefore, random sampling is not considered, as it disrupts the sequential dependencies inherent to the time series.

thumbnail Figure 2

Schematic representation of the 5-fold cross-validation procedure. The time series is split into sequential folds to preserve causality. For each step of the experiment, the fold marked in orange is used as the test set, and the fold marked in blue is used as the validation set. The remainder of the data is used for training.

2.5 Input data construction

A fixed-size feature vector is used to represent CH properties for graph construction, limited by available RAM. The CHs are ordered from left to right, corresponding to the direction of solar rotation, and CH information is added following that order until the feature vector is full. This ordering enables the model to always consider CHs that will impact Earth later, allowing for forecasts with a lead time of up to five days. Moreover, when used as a time series, this approach allows BaLNet to account for CHs that may no longer fit into the feature vector but were present in previous timesteps. Figure 3 presents a schematic representation of the vectorization process.

thumbnail Figure 3

Schematic representation of the CH information vectorization process. First, CHs are identified on available AIA and HMI samples. Then, measurements are taken and stored in tabular form. Finally, CHs are ordered from left to right, and the first four are taken and linearized into a single vector.

Roughly 50% of samples in the dataset include between one and four CHs, while approximately 0.5% of samples contain more than four. An increase in performance is observed when using up to four CHs, with a 14% improvement on all metrics when using four instead of three. However, increasing the number of CHs beyond four leads to a decrease in performance. Using five CHs results in a 2% degradation in performance, along with a 38% increase in training time and a 41% increase in memory requirements. Similarly, using six CHs further degrades performance by 3%, while training time increases by 75% and memory consumption rises by 65%. This trend suggests that using more than four CHs introduces unnecessary complexity. Although the additional CHs may provide useful information, the scarcity of samples with more than four CHs means that BaLNet cannot effectively learn effects on V because of the associated feature sparsity (Petrini et al., 2022). As a result, the model is unable to capitalize on the potential value of this additional information, leading to a performance decline and increased computational costs.

A fixed window size W = 5 is used to represent the time series input, as it yielded the best numerical results during preliminary testing. The lead time, denoted as H, specifies the forecast horizon and is set as a parameter for the experiments. We conduct experiments for H ∈ {1, 3, 5} days. Because hyperparameter optimization is performed for a lead time of five days, and no projection correction is applied to CH information, we assess whether decoupling the offset of CH and solar wind data improves performance at shorter lead times. To this end, we introduce a hyperparameter HCH, which allows independent shifting of CH features relative to solar wind targets. For a lead time of one day, we test HCH ∈ {1, 3, 5}, and for a lead time of three days, we explore HCH ∈ {3, 5}. Since the effective lead time is given by min(H, HCH), this selection ensures that all CH features used in the input would have been observable at the time the forecast is made, preserving operational validity. To capture the persistence of CHs over multiple Carrington rotations, an additional segment of the time series with the same window size W is included as input, positioned C days before the target prediction time. The parameter C serves as an approximation of a Carrington rotation period and is treated as a hyperparameter in the model. For each experiment, we explore values of C ranging from 24.26 to 30.26 days, using a daily step size. Figure 4 provides a schematic illustration of this temporal alignment procedure.

thumbnail Figure 4

Schematic representation of the temporal alignment scheme. Two time windows of size W are selected and treated as parallel streams of features. The first time window is offset C days from the target forecast date. The second time window is offset some time H with respect to the forecasted time window. In this visual example, a single set of CH measurements is displayed in the interest of brevity, with W = 5, H = 5, and C = 27.26.

3 The BaLNet model

3.1 Model description

BaLNet is comprised of three main layers, not counting the input: an LSTM layer, a variational dense layer, and a traditional dense layer. This architecture compresses the input time series into a vector representation, which is then used to generate a variational encoding to extract probability parameters from. Dense layers operate by directly applying tensor operations. Given an input vector x, a weights matrix A, a bias vector b, and an activation function θ, the output of a dense layer d(x) is computed as

d(x)=θ(AxT+b).$$ d(\mathbf{x})=\theta \left(A\cdot {\mathbf{x}}^{\mathrm{T}}+\mathbf{b}\right). $$(3)

A variational dense layer builds on the general idea of traditional dense layers by introducing stochasticity. The weight matrix is replaced with a Bayesian inference process that generates a multivariate distribution representing the probability of each set of weights. This distribution is generated on a per-inference basis, and sampled to generate the realized weights applied as part of the classic forward propagation. The Bayesian inference process is conditioned on the minimization of the Kullback-Leibler divergence (Kullback & Leibler, 1951) with respect to the empirical conditional probability distribution estimated during backpropagation via Monte Carlo sampling.

To construct the prior distribution of the variational dense layer, we assume a mixture of multivariate normal probability distributions. For this purpose, the training dataset is split by the value of V observed at forecast time, t0, denoted Vt0$ {V}_{{t}_0}$. We generate five subsets conditioned to Vt0$ {V}_{{t}_0}$ from the training dataset: Vt0<400$ {V}_{{t}_0} < 400$, 400Vt0<600$ 400\le {V}_{{t}_0} < 600$, 600Vt0<700$ 600\le {V}_{{t}_0} < 700$, 700Vt0<800$ 700\le {V}_{{t}_0} < 800$, and 800Vt0$ 800\le {V}_{{t}_0}$; with all speed values expressed in km s−1. For each subset, the mean and standard deviation of each time aligned input feature are computed to configure the corresponding multivariate normal. The mixture weight of each multivariate normal in the final distribution is taken as the proportion of the training dataset represented in it. Values are normalized prior to computing the corresponding statistics. Table 2 presents the weight of each multivariate normal, as well as the corresponding mean, μ, and standard deviation, σ, for Vt0$ {V}_{{t}_0}$. Because the total number of variates under consideration is 56, and the prior serves only as an initial inductive state, the rest are omitted in the interest of brevity. The values of Vt0$ {V}_{{t}_0}$ are presented in a denormalized form to provide an intuition of the value ranges under consideration for each component.

Table 2

Weight, mean value (μ) and standard deviation (σ) of Vt0$ {V}_{{t}_0}$ for each multivariate normal used in the Gaussian mixture prior. The rest of the variates, totalling 56, are omitted in the interest of brevity.

The posterior distribution is a multivariate normal, with each covariate representing a weight of the dense layer. The activation function chosen is the hyperbolic tangent, tanh, because it reseults in consistent convergence during the initial exploration performed. The layer is comprised of 60 neurons. LSTM units are made up of several cells that work in tandem. Each cell has a hidden state path and an input path. The hidden state modulates the processing of inputs, and the result is passed through a forget gate to generate the hidden state for the next cell in the unit. This mechanism enables LSTM layers to effectively capture local temporal relationships and process time series data.

Because of the chosen prior structure, the number of LSTM cells is equal to the number of input variables. The LSTM layer’s role is to compress temporal information on a per-input basis, ensuring accurate input for the prior function while encapsulating relevant input features. Again, the tanh activation function is used because only logistic functions have demonstrated consistent convergence.

Finally, the output dense layer is comprised of two independent neurons with linear activation functions. The first neuron represents μ, and the second neuron represents σ. The output is encapsulated into an independent normal distribution to facilitate fitting and usage.

Figure 5 displays a summary of the complete model structure used. This architecture uses a total of 5,878,764 parameters because of the Bayesian inference process applied.

thumbnail Figure 5

Summary of the proposed architecture. The weights of the variational layer are structured as a Bayesian inference process that generates a normal distribution for each weight. The resulting distribution is then sampled each time the model is run, leading to stochastic outputs.

3.2 Optimization procedure

Hyperparameter optimization for the model architecture is focused on a five day lead time. This prioritization is motivated by the goal of extending the predictive range, as well as by the full hyperparameter optimization procedure requiring over 48 h of compute time. To further explore the impact of hyperparameter selection across different lead times, the value of C is systematically evaluated for one day, three day, and five day forecasts.

Because the model’s output is a probability density function, traditional metrics such as mean squared error (MSE) and mean absolute error (MAE) are ill-suited for the training task. Instead, the Kullback-Leibler divergence with respect to the empirical conditional probability distribution is minimized.

Direct optimization of the Kullback-Leibler divergence is computationally expensive, and renders the problem practically untractable when applied to the complete model (Blei et al., 2017). Instead, its Evidence Lower Bound (ELBO; Duan et al., 2017) is optimized. To this end the combined negative log-likelihood of each batch given the model’s current weights is minimized (Oleksiienko et al., 2023).

Optimization is performed using the RMSProp algorithm (Reddy et al., 2018), which is well-suited for non-stationary learning tasks due to its adaptive learning rate properties. To ensure stable training and prevent issues related to exploding gradients, gradient clipping is applied with a maximum norm of 1.0 (Ramaswamy, 2023). This technique helps control the magnitude of parameter updates, maintaining the stability and convergence of the training process. Early stopping and learning rate scheduling strategies are implemented to prevent overfitting.

3.3 Validation procedure

To validate model results, the model is trained for each combination of C and H seven times to limit the time dedicated to training. A total of 49 training runs are executed, with an average training time in CPU of 2.17 h. The CPU used is an Intel Xeon with a clock speed of 2.20 GHz. All processes are run single threaded and with a single worker. The resulting model’s predicted mean is characterized in terms of the following metrics. For the following formulas, Y denotes the observation, Ŷ$ \widehat{Y}$ denotes the model’s predicted mean, Y¯$ \bar{Y}$ denotes the mean value of the observation, N refers to the number of elements in Y and Ŷ$ \widehat{Y}$, and yi and ŷi$ {\widehat{y}}_i$ are elements of Y and Ŷ$ \widehat{Y}$, respectively:

  • Root mean squared error (RMSE), to validate the average error obtained by the model, weighted proportional to the magnitude of the error. It is defined as

RMSE ( Y , Y ̂ ) = i = 1 N ( y i - y ̂ i ) 2 N . $$ \mathrm{RMSE}(Y,\widehat{Y})=\sqrt{\frac{\sum_{i=1}^N ({y}_i-{\widehat{y}}_i{)}^2}{N}}. $$(4)

  • Mean absolute error (MAE), to validate the average error obtained by the model without weighting. It is defined as

MAE ( Y , Y ̂ ) = i = 1 N | y i - y ̂ i | N . $$ \mathrm{MAE}(Y,\widehat{Y})=\frac{\sum_{i=1}^N |{y}_i-{\widehat{y}}_i|}{N}. $$(5)

  • Pearson’s correlation coefficient between observation and prediction (CC), to verify the degree of trend similarity between model prediction and observation. It is defined as

CC ( Y , Y ̂ ) = cov ( Y , Y ̂ ) σ Y σ Y ̂ . $$ \mathrm{CC}(Y,\widehat{Y})=\frac{\mathrm{cov}(Y,\widehat{Y})}{{\sigma }_Y{\sigma }_{\widehat{Y}}}. $$(6)

  • Coefficient of determination (R2), to measure the degree of improvement obtained compared to simply applying a trivial estimation that always outputs the observed distribution mean. It is defined as

R 2 ( Y , Y ̂ ) = 1 - i = 1 N ( y i - y ̂ i ) 2 i = 1 N ( y i - Y ¯ ) 2 , $$ {R}^2(Y,\widehat{Y})=1-\frac{\sum_{i=1}^N ({y}_i-{\widehat{y}}_i{)}^2}{\sum_{i=1}^N ({y}_i-\bar{Y}{)}^2}, $$(7)

where Y¯$ \bar{Y}$ denotes the mean value found in Y.

  • Dynamic time warping (DTW) distance between observed and predicted time series, computed following the algorithm defined by Berndt & Clifford (1994). After computing the optimal warping path P between Y and Ŷ$ \widehat{Y}$, the distance is computed as

DTW ( Y , Y ̂ ) = ( i , j ) P | y i - y ̂ j | . $$ \mathrm{DTW}(Y,\widehat{Y})=\sum_{(i,j)\in P} |{y}_i-{\widehat{y}}_j|. $$(8)

To contextualize numerical results of the forecasted mean we benchmark with persistence models, a classic linear regression (Zou & Hastie (2005), and an ElasticNet linear regression. We also train and run a CNN following the design proposed by Raju & Das (2021), and compare with SWAN (Cobos-Maestre et al., 2024) to validate the degree of improvement obtained in comparison with state-of-the-art neural networks. Finally, we retrieve WSA-ENLIL (Arge et al., 2003) runs from the NASA Community Coordinated Modeling Center.

To ensure a fair comparison, all Machine Learning models are trained on the same training set. Evaluation is likewise constrained to the exact same test set to prevent deviations in the ground truth from affecting evaluation results. In the case of the CNN and SWAN, we apply the hyperparameter sets provided in their respective papers to guarantee that the estimator follows exactly the same architecture.

To account for WSA-ENLIL being a physical model that predicts multiple physical parameters, we constrain the comparison to the forecasts performed for V at the L1 Lagrange point with lead times of one, three and five days. Because V is the first physical parameter estimated by WSA, it is not affected by propagated errors in previous prediction steps.

To evaluate the quality of the predicted conditional probability distribution, we compute the Prediction Interval Coverage Probability (PICP; Huang et al., 2023). This is calculated using the number of samples within the prediction interval, denoted as Nin, and the total number of samples in the test dataset, N:

PICP=NinN.$$ \mathrm{PICP}=\frac{{N}_{\mathrm{in}}}{N}. $$(9)

We target a 95% prediction interval for calibration. A PICP value close to 95% indicates that the model is well-calibrated. A PICP below 95% suggests the model underestimates uncertainty, while a value nearing 100% indicates an overestimation of uncertainty. To account for the effects of a limited number of random initializations, we consider a model with PICP between 94% and 96% to be well calibrated.

3.4 Operational considerations

BaLNet is designed to provide operational forecasts. During operation, the variational design of the model needs to be considered. The probability of a set of extreme, outlier weights being selected for every selection tends to zero because of the normal distribution implemented. However, fluctuations should still be accounted for.

To obtain a stable prediction, we leverage the central limit theorem Dudley (1978). As per the work of Valdenegro-Toro & Mori (2022), a larger number of evaluations N yields more precise values. A minimum N = 100 is suggested to ensure stable results. We investigate whether a smaller N can be utilized by applying the Jensen-Shannon divergence, DJS, to measure stability similarity (Guzmán-Martínez & Alaiz-Rodríguez, 2011). Specifically, a criterion of DJS ≤ 0.05 is established to indicate similar stability between values of N. For N = 10 and N = 100, we find that DJS = 3.44 × 10−3, which fulfills the criterion. Therefore, N = 10 can be used during the operational implementation of BaLNet without compromising the results.

The main advantage of this architecture is the ability to estimate the influence of aleatoric and epistemic uncertainty in the inference process. Aleatoric uncertainty is attributed to the input data used, while epistemic uncertainty stems from the model. To unravel both uncertainties, we use the approach defined by Valdenegro-Toro & Mori (2022): given a predictive variance σ2, it can be decomposed as

σi2(x)=Ei[σi2(x)]+Vari[μi(x)],$$ {\sigma }_i^2(x)={\mathbb{E}}_i[{\sigma }_i^2(x)]+\mathrm{Va}{\mathrm{r}}_i[{\mu }_i(x)], $$

where Ei[σi2(x)]$ {\mathbb{E}}_i[{\sigma }_i^2(x)]$ denotes the mean variance across the predicted distributions from several model runs, and Vari[μi(x)]$ \mathrm{V}\mathrm{a}{\mathrm{r}}_i[{\mu }_i(x)]$ denotes the variance of their predicted means.

In this expression, the term Ei[σi2(x)]$ {\mathbb{E}}_i[{\sigma }_i^2(x)]$ captures the aleatoric uncertainty of the prediction. The term Vari[μi(x)]$ \mathrm{Va}{\mathrm{r}}_i[{\mu }_i(x)]$ captures the epistemic uncertainty. In an operational context, this approach provides a measure of reliability to facilitate interpretation of the forecasts obtained (e.g., low reliability when a high speed CME is expected to arrive).

4 Results and discussion

4.1 Model evaluation

Tables 35 present the numerical results obtained by each model for forecast lead times of one, three, and five days, respectively. For conciseness, only the best configuration of BaLNet (based on the parameter C) is shown in these tables. The optimal value of C is set to 25.26 days for a one day lead time, and to 27.26 days for three day and five day lead times. A detailed breakdown of the results for all BaLNet configurations is available in Appendix.

Table 3

Comparison of numerical results with a lead time of one day. Models denoted as “pers.” refer to persistence models. The best numerical results per metric are bolded.

Table 4

Same as Table 3 but for a three days lead time.

Table 5

Same as Table 3 but for a five days lead time.

Overall, BaLNet performs best for a five day lead time, which aligns with architectural hyperparameters having been optimized for that use case. For this lead time, the minimum and maximum PICP values for the 95% confidence interval are approximately 90% and 95%, respectively. In comparison, the one day lead time shows lower PICP values, ranging from 75% to 83%, while the three day lead time achieves PICP values between 85% and 89%. This behavior is consistent with all other evaluation metrics used.

Across all lead times, BaLNet outperforms all other models on the test dataset in terms of numerical metrics. Specifically, for the five day lead time, BaLNet achieves a 55% reduction in RMSE compared to the CNN, while for the one day lead time, the reduction is 15%. Furthermore, the negative R2 values found suggest a high relative error in most models’ predictions, higher than that obtained using the mean value of V as estimation (Kvålseth, 1985). This does not happen in BaLNet’s case, further highlighting its accuracy.

The nearly constant performance observed for both ElasticNet and CNN across all lead times reflects a conservative predictive behavior. This behavior exhibits the typical traits of non-convergence associated with excessive regularization: small fluctuations in output, difficulty in tracking the shape of the observations, and consistence of the models’ performance with respect to the training set (Ghojogh & Crowley, 2019; Obi & Jecinta, 2023). Furthermore, since regularization is the only common factor between these models, it further underscores its apparent influence on their performance.

Persistence models demonstrate superior performance in short-term DTW evaluations because they replicate the temporal structure of the time series through direct propagation of past values. As a result, persistence models inherently preserve the temporal alignment of features, making them unsuitable for direct comparisons against methods that generate forecasts based on distinct predictive mechanisms. Excluding persistence models from the evaluation, BaLNet achieves the highest agreement with the observed temporal patterns for three and five days lead times, as reflected by its DTW scores. Furthermore, the best initializations of BaLNet outperform Carrington persistence models for the five days lead time.

For the one day lead time, linear regression and the SWAN model show lower DTW values than BaLNet; however, when considering the complete set of evaluation metrics both linear regression and SWAN exhibit numerically inferior results compared to BaLNet. This discrepancy suggests that the high temporal autocorrelation of V, of 0.73 for a one day lead time, results in these models producing imitative behaviors rather than generalizable forecasts. BaLNet, in contrast, presents a clear alignment in all performance metrics, suggesting consistent, stable learning of predictive behaviors.

Table 6 displays the PICP value obtained for each lead time, expressed as a percentage. For the one and three day lead times, BaLNet underestimates the true uncertainty. In contrast, the five day model accurately captures the uncertainty present in the data. These results are consistent with the numerical evaluation of the predicted means, confirming that BaLNet is most reliable when forecasting with a five day lead time.

Table 6

PICP of BaLNet’s 95% CI with respect to lead time.

Figures 6, 7, and 8 illustrate example forecasts for one, three, and five days lead times, respectively, along with their corresponding uncertainty estimates. The most relevant statistics for each figure are summarized in Table 7. A CME arrival occurs between late 2021-11-03 and 2021-11-05, reaching a peak speed of 770 km s−1. For all three lead times, for target times affected by this arrival, BaLNet’s prediction increases in mean value until it reaches roughly 550 km s−1, and then stabilizes around that value. Furthermore, for a five days lead time, punctual decreases under this threshold are accurately captured. Using the CME database provided by Richardson & Cane (2010), BaLNet is observed to consistently produce this behavior for CMEs that exceed 550 km s−1 bulk speed. This is particularly notable because BaLNet is not provided any CME-specific data, such as coronograph images or extracted features.

thumbnail Figure 6

Top panel presents the comparison of forecasted values and observation with a one day lead time. The dashed line corresponds to the mean of the distribution forecasted by the model. The area shaded in orange corresponds to the 95% confidence interval. The bottom panel presents the corresponding uncertainty disentanglement. The area shaded in grey in both panels denotes a CME arrival with a peak speed of 770 km s−1.

thumbnail Figure 7

Same as Figure 6, but for a three days lead time.

thumbnail Figure 8

Same as Figure 6, but for a five days lead time.

Table 7

Maximum and minimum predicted μβ(x), and properties of the distributions of aleatoric and epistemic uncertinaty presented in Figures 6, 7 and 8. All measurements are provided in km s −1.

Across all lead times, BaLNet captures the general behaviors of solar wind speed. For the one and three days lead times, it primarily captures these general behaviors, while for the five days lead time, the model also predicts finer-grained detail. For example, in the high speed stream starting on 2021-09-15, it can be seen that BaLNet captures the presence of the initial peak for all three lead times. However, the second and third peaks are only captured for a five days lead time. Furthermore, for a five days lead time, BaLNet effectively distinguishes situations where the dominant source of uncertainty is epistemic, as opposed to aleatoric, suggesting its capability to capture the nature of uncertainty even with longer lead times.

4.2 Cross-validation results

The dataset used for this study spans from 2012-10-01 6:00 to 2022-01-01 00:00, covering most of Solar Cycle 24 and the rising phase of Cycle 25. The sequential evaluation, performed in accordance with best practices for time series regression (Clark et al., 2020), is therefore limited to the early stages of Cycle 25. However, this raises the question of whether the numerical results obtained from such a split are generalizable to other phases of the solar cycle.

To further assess the generalizability of the numerical results across different phases of the Solar Cycle, including data from both Cycles 24 and 25, we perform five-fold cross-validation. The dataset is split into five equally-sized, sequentially ordered subsets, each of which is used as the test set once. The results are then aggregated so that they can be compared to the sequential approach by treating each fold as analogous to an independent run. This approach allows us to evaluate BaLNet’s performance across different Solar Cycle phases, despite the limited data. Additionally, the sequential experiment helps identify whether known pitfalls of applying cross-validation to time series, such as information leakage and distributional shift, are present in this case. Table 8 presents a comparison between the numerical results obtained in the cross-validation experiment and those obtained in the sequential setting. While neither any individual folds nor the aggregated results of the cross-validation procedure exceed the sequential setting, this consistency does not indicate evidence of information leakage. However, interfold fluctuations in the relationship between evaluation metrics suggest the presence of distributional shift. This fluctuation is most evident when comparing RMSE and DTW to R2: the increase in numerical error for both metrics is accompanied by a larger improvement relative to a trivial estimator than in other folds, indicating a shift in the underlying conditional distribution function. Furthermore, using the Kolmogorov-Smirnov test (Massey, 1951), clear differences are observed between the distributions of metrics obtained using cross-validation and the sequential approach, further reinforcing the presence of distributional shift. The observed shift could be due to a more prominent presence of CMEs during periods of high solar activity. This would make periods around solar minimum ideal to compare models that do not aim to predict CMEs by ensuring minimal impact on the obtained performance metrics.

Table 8

Comparison of results obtained in the cross-validation experiment with results obtained in the sequential experiment for a lead time of five days. Each cross-validation fold is labeled using its start and end dates.

For all metrics, all folds and the aggregated cross-validation results display numerical improvements when compared to CNN, SWAN, the linear models and the persistence baselines. The highest RMSE obtained represents a reduction by 38% when compared to the CNN’s results in the sequential setting, despite the latter being evaluated on a more advantageous subset of the data. The consistent improvements found in the cross-validation procedure suggest that BaLNet is robust to fluctuations in the data, and underline its ability to generalize to different periods of the solar cycle.

Figure 9 displays the forecasts and uncertainty disentanglement produced by the model whose test set encompasses between 2012-10-01 and 2014-08-08, for the period between May 2014 and the end of the test set. Due to the higher mean speed value associated with the maximum of the Solar Cycle (Larrodera & Cid, 2020), the predictive mean stays elevated. However, the observed predictive behaviors align with those found in the sequential approach (Fig. 8) when accounting for the observed distributions in each phase, further corroborating BaLNet’s stability and its ability to generalize across both minimum and maximum solar conditions. Together with the numerical results, it demonstrates BaLNet’s ability to generalize across diverse solar conditions, offering robust predictions even in the presence of solar activity fluctuations.

thumbnail Figure 9

Top panel shows the comparison of forecasted value and observation with a five days lead time for the period between May and early August 2014. The area shaded in orange represents the 95% confidence interval. The bottom panel presents the corresponding uncertainty disentanglement. The area shaded in grey in both panels denotes a CME arrival with peak speed of 610 km s−1.

4.3 Critical analysis of temporal alignment between solar wind data and coronal hole information

Hyperparameter optimization for the previous experiments is performed primarily for a lead time of five days. Furthermore, no projection correction is applied to CH information. It is hypothesized that this should not negatively affect the presence of disk-centered CHs within the input vector for the corresponding lead times. If this hypothesis is wrong, it would lead to the observed reduction in performance for the shorter lead times.

To address this concern, we perform an additional experiment introducing a new hyperparameter, HCH. HCH enables independent offsetting of CH information relative to solar wind data, decoupling their offset to assess whether this improves performance by maximizing the duration for which relevant CHs remain centered on the solar disk within the input window. We evaluate HCH ∈ {1, 3, 5} for a one day lead time and HCH ∈ {3, 5} for a three days lead time. We note that the effective lead time when utilizing this new hyperparameter is min(H, HCH), so the list of values for a lead time of three days is adjusted accordingly. When HCH = H, this setting reproduces the previously reported same offset results. Table 9 summarizes the results obtained with each hyperparameter configuration. For a lead time of one day, the time-aligned setting (HCH = H) achieves the best performance in PICP, RMSE, MAE, CC and R2. HCH = 3 apparent improvement of DTW (∼1.6%) is within the uncertainty bounds of the value found for HCH = 1, and it underperforms in all other measurements. Similarly, for a three days lead time, HCH = H yields the best results for RMSE, MAE, CC, R2, with a statistically insignificant increase in PICP (∼1.1%) for HCH = 5.

Table 9

Comparison of performance results obtained using different values of HCH. The best numerical results for each combination of H and measurement are bolded.

Thus, temporally decoupling CH and solar wind information does not yield any improvements, as observed differences remain within the uncertainty bounds of the base configuration. This result reinforces that maintaining temporal alignment is preferable within the evaluated conditions, as it minimizes the number of preprocessing operations and ensures full consistency for every lead time.

4.4 Ablation study

To further understand the behavior of the variational dense layer and its impact on model performance, we remove specific components and quantify their impact on the obtained modeling results. Specifically, we reduce the number of components in the prior of the variational layer, we remove the LSTM layer, and we remove the variational layer. Removed layers are replaced with dense layers with the same number of units and activation function.

Table 10 presents a comparison of various performance metrics, including RMSE, CC, DTW, PICP, training time in CPU, and memory usage of the training graph between BaLNet and the ablated models.

Table 10

Summary of ablation study results. Models are labeled using the component of BaLNet that is ablated. MAE and R2 display the same tendencies as RMSE and CC, respectively. The best values for each measurement evaluated are bolded.

The results indicate that removing any components reduces the computational cost of training, at the expense of a reduction in performance. The combination of increased RMSE and reduced PICP suggests that, independently of the element removed, modeling results fail to properly capture the conditional probability distribution of V, with a consistent underestimation of uncertainty. Notably, none of the ablated models are well calibrated as per our criterion, as their PICP is strictly less than 94%. Furthermore, utilization of GPU reduces the average training time of the full model to an average of 15 min. Thus, the main disadavantage of applying the complete model is minimized.

5 Conclusions

In this paper, we have presented BaLNet, a variational Bayesian neural network architecture for probabilistic solar wind speed forecasting. We optimize model hyperparameters for a lead time of five days, and evaluate it for lead times of one, three and five days, and find consistent improvements when compared to state-of-the-art solar wind speed forecasting models. Notably, error is reduced by between 38% and 55%, depending on the test set, for a lead time of five days, suggesting a marked increase in forecasting ability. While cross-validation results show slightly higher RMSE and variability due to temporal distributional shifts, BaLNet consistently outperforms benchmark models across all valuation metrics, affirming its generalizability and reliability for operational use. Furthermore, BaLNet provides uncertainty estimation, and can be inspected to disentangle source of uncertainty. This enables users to gain a better understanding of the reliability of any given forecast.

We also evaluated the accuracy of the predicted conditional probability distribution by contrasting the derived 95% confidence interval with the prediction interval coverage probability (PICP) of the computed range. For a lead time of five days, a PICP of between 88% and 95% is obtained, suggesting a strong agreement with the empirical conditional probability distribution function underlying the generative process. The confidence interval provided is observed to lose accuracy when the lead time is reduced. This behavior suggests an excess of confidence by the model in such cases, leading to smaller variances than that observed in the empirical conditional probability distribution.

A key advantage of BaLNet is its lightweight architecture, which enables efficient training on a CPU, with an average training time of 2.17 h. This miniaturization is facilitated by the absence of direct interaction with image data, as the model relies on pre-extracted features from solar observations. While BaLNet includes information from the solar disk, it does so using lower-dimensional feature vectors, reducing the complexity of the model.

We found that removal of the variational component from the architecture results in a loss of predictive power, in exchange for a significant reduction in computational demands. Numerical results were still an improvement over those obtained with state-of-the-art models found in the literature, further highlighting the effectiveness of the Bayesian approach in solar wind forecasting.

In conclusion, BaLNet demonstrates that a Bayesian framework, combined with a lightweight architecture and carefully selected features, can significantly improve solar wind speed forecasting while maintaining computational efficiency. The model’s ability to provide well-calibrated uncertainty estimates and to handle high-dimensional input data without directly processing complex images opens new avenues for operational forecasting.

In future, the BaLNet model will be validated in an operational environment to ensure that the results presented in this paper hold during the remainder of the current solar cycle. Additionally, to facilitate the usage of BaLNet as a source of information for models predicting effects on Earth (e.g., Collado-Villaverde et al., 2024), the temporal resolution will be enhanced by reducing the target timestep to 5 min, while maintaining the current five days lead time. Moreover, in this same direction, BaLNet’s applicability to forecasting additional solar wind parameters, such as density ρ and T, will be investigated.

Acknowledgments

The authors are grateful to all the members of the Space Weather community making data readily available for usage. We acknowledge the use of ACE SWEPAM and MAG data retrieved from the OMNIWeb database (http://omniweb.gsfc.nasa.gov/). We also acknowledge the use of SDO AIA and HMI samples retrieved from the Joint Science Operations Center (http://jsoc.stanford.edu/). WSA-ENLIL runs were retrieved from the NASA Community Coordinated Modeling Center Runs-on-Request service (https://ccmc.gsfc.nasa.gov/tools/runs-on-request/). The research presented in this paper utilizes the following Python libraries: Astropy (Price-Whelan et al., 2018), Sunpy (The SunPy Community et al., 2020), Numpy (Harris et al., 2020), SciPy (Virtanen et al., 2020), Pandas (McKinney, 2010; jbrockmendel et al., 2020), Tensorflow (Abadi et al., 2016), Keras2, and Tensorflow Probability (Dillon et al., 2017). The editor thanks two anonymous reviewers for their assistance in evaluating this paper.

Data availability statement

The data and code required to train BaLNet is available in the following GitHub repository: https://github.com/R012/BaLNet. Level 2 solar wind data is provided along with all preprocessing steps required for reproducibility. Coronal hole data measurements are provided after discarding samples marked as corrupt by their FITS headers.

Appendix

A1 Full model characterization

This Appendix provides a complete characterization of the presented architecture for different configurations of the experiment. In the following table, H denotes the lead time of the forecast, and C the time (in days) taken as an approximation of a Carrington rotation. CI refers to the 95% confidence interval extracted from the predicted normal distribution generated by the model.

Table A1.

Full modeling results. H denotes the lead time of the forecast. C refers to the time taken as an approximation of a Carrington rotation. CI refers to the 95% confidence interval extracted from the predicted normal distribution generated by the model.


References

Cite this article as: Cobos-Maestre M, Flores-Soriano M & Barrero D 2025. BaLNet: A probabilistic Bayesian neural network for solar wind speed forecasting. J. Space Weather Space Clim. 15, 34. https://doi.org/10.1051/swsc/2025029.

All Tables

Table 1

Division of the dataset between training, validation, and testing subsets.

Table 2

Weight, mean value (μ) and standard deviation (σ) of Vt0$ {V}_{{t}_0}$ for each multivariate normal used in the Gaussian mixture prior. The rest of the variates, totalling 56, are omitted in the interest of brevity.

Table 3

Comparison of numerical results with a lead time of one day. Models denoted as “pers.” refer to persistence models. The best numerical results per metric are bolded.

Table 4

Same as Table 3 but for a three days lead time.

Table 5

Same as Table 3 but for a five days lead time.

Table 6

PICP of BaLNet’s 95% CI with respect to lead time.

Table 7

Maximum and minimum predicted μβ(x), and properties of the distributions of aleatoric and epistemic uncertinaty presented in Figures 6, 7 and 8. All measurements are provided in km s −1.

Table 8

Comparison of results obtained in the cross-validation experiment with results obtained in the sequential experiment for a lead time of five days. Each cross-validation fold is labeled using its start and end dates.

Table 9

Comparison of performance results obtained using different values of HCH. The best numerical results for each combination of H and measurement are bolded.

Table 10

Summary of ablation study results. Models are labeled using the component of BaLNet that is ablated. MAE and R2 display the same tendencies as RMSE and CC, respectively. The best values for each measurement evaluated are bolded.

Table A1.

Full modeling results. H denotes the lead time of the forecast. C refers to the time taken as an approximation of a Carrington rotation. CI refers to the 95% confidence interval extracted from the predicted normal distribution generated by the model.

All Figures

thumbnail Figure 1

Comparison of CH identification using the ASSA algorithm (left panel) and our algorithm (right panel). The sample shows a large, centered CH during 2019-09-25.

In the text
thumbnail Figure 2

Schematic representation of the 5-fold cross-validation procedure. The time series is split into sequential folds to preserve causality. For each step of the experiment, the fold marked in orange is used as the test set, and the fold marked in blue is used as the validation set. The remainder of the data is used for training.

In the text
thumbnail Figure 3

Schematic representation of the CH information vectorization process. First, CHs are identified on available AIA and HMI samples. Then, measurements are taken and stored in tabular form. Finally, CHs are ordered from left to right, and the first four are taken and linearized into a single vector.

In the text
thumbnail Figure 4

Schematic representation of the temporal alignment scheme. Two time windows of size W are selected and treated as parallel streams of features. The first time window is offset C days from the target forecast date. The second time window is offset some time H with respect to the forecasted time window. In this visual example, a single set of CH measurements is displayed in the interest of brevity, with W = 5, H = 5, and C = 27.26.

In the text
thumbnail Figure 5

Summary of the proposed architecture. The weights of the variational layer are structured as a Bayesian inference process that generates a normal distribution for each weight. The resulting distribution is then sampled each time the model is run, leading to stochastic outputs.

In the text
thumbnail Figure 6

Top panel presents the comparison of forecasted values and observation with a one day lead time. The dashed line corresponds to the mean of the distribution forecasted by the model. The area shaded in orange corresponds to the 95% confidence interval. The bottom panel presents the corresponding uncertainty disentanglement. The area shaded in grey in both panels denotes a CME arrival with a peak speed of 770 km s−1.

In the text
thumbnail Figure 7

Same as Figure 6, but for a three days lead time.

In the text
thumbnail Figure 8

Same as Figure 6, but for a five days lead time.

In the text
thumbnail Figure 9

Top panel shows the comparison of forecasted value and observation with a five days lead time for the period between May and early August 2014. The area shaded in orange represents the 95% confidence interval. The bottom panel presents the corresponding uncertainty disentanglement. The area shaded in grey in both panels denotes a CME arrival with peak speed of 610 km s−1.

In the text

Current usage metrics show cumulative count of Article Views (full-text article views including HTML views, PDF and ePub downloads, according to the available data) and Abstracts Views on Vision4Press platform.

Data correspond to usage on the plateform after 2015. The current usage metrics is available 48-96 hours after online publication and is updated daily on week days.

Initial download of the metrics may take a while.