Explanation of Dynamic Physical Field Predictions using WassersteinGrad: Application to Autoregressive Weather Forecasting

TL;DR

WassersteinGrad explains dynamic physical field predictions by computing the entropic Wasserstein barycenter, enhancing autoregressive weather forecasting model interpretability.

stat.ML 🔴 Advanced 2026-04-24 22 views
Younes Essafouri Laure Raynaud Luciano Drozda Laurent Risser
WassersteinGrad weather forecasting autoregressive models explainable AI dynamic physical fields

Key Findings

Methodology

This paper introduces a novel method called WassersteinGrad for explaining predictions of dynamic physical fields. The method addresses the geometric displacement issue found in traditional gradient smoothing methods by computing the entropic Wasserstein barycenter of perturbed attribution maps. Specifically, WassersteinGrad maps perturbed attribution maps into the space of spatial probability measures and utilizes the entropic Wasserstein barycenter to extract a geometric consensus, achieving more interpretable predictions in high-dimensional inputs.

Key Results

  • On regional weather datasets, WassersteinGrad outperforms gradient-based baseline methods in interpretability metrics. In both single-step and autoregressive forecasting settings, WassersteinGrad demonstrates superior interpretability, particularly in spatial consistency and physical localization.
  • Experimental validation shows that WassersteinGrad significantly reduces geometric displacement in attribution maps while maintaining predictive model performance. Specific data indicate that at noise levels σ<0.4, model performance degrades by less than 1%, yet attribution centroids advect by an average of 10-15 km.
  • In autoregressive forecasting, WassersteinGrad absorbs geometric distortion accumulation through geometric consensus, exhibiting robustness superior to other baseline methods.

Significance

The introduction of WassersteinGrad holds significant implications for both academia and industry. It addresses the longstanding challenge of interpretability in dynamic physical field predictions and provides a more reliable explanation tool for AI applications in high-stakes environments. By enhancing the transparency of model predictions, WassersteinGrad helps build trust in AI systems, particularly in high-risk areas such as weather forecasting. Additionally, the geometric consensus mechanism offers new insights for explainable AI research in other domains.

Technical Contribution

WassersteinGrad differs significantly from existing state-of-the-art methods in several technical aspects. Firstly, it introduces the entropic Wasserstein barycenter to address geometric displacement, providing a new theoretical guarantee in explainable AI. Secondly, the method offers new engineering possibilities through its geometric consensus mechanism, enabling more precise attribution in high-dimensional dynamic inputs. Lastly, WassersteinGrad excels in interpretability, robustness, and sparsity metrics, showcasing its potential in practical applications.

Novelty

WassersteinGrad is the first method to apply the entropic Wasserstein barycenter to explain predictions of dynamic physical fields. Compared to existing methods like SmoothGrad, it not only addresses geometric displacement but also achieves significant improvements in interpretability and robustness. This innovation opens new pathways for the application of explainable AI in dynamic physical fields.

Limitations

  • WassersteinGrad's computational complexity on high-dimensional grids is relatively high, despite optimization through entropic regularization and Sinkhorn iterations. Further improvements in computational efficiency are needed.
  • The generalizability of this method across different predictive models and spatial domains remains to be validated, particularly in other physical fields with complex dynamics.
  • The current noise injection strategy still uses white Gaussian noise, lacking physical information. Future work should explore more physically meaningful perturbation strategies.

Future Work

Future research directions include: 1) exploring more efficient computational methods to reduce the computational cost of WassersteinGrad on high-dimensional grids; 2) validating the method's generalizability and effectiveness in other dynamic physical fields; 3) developing more physically meaningful noise injection strategies to enhance the physical plausibility of explanations and model robustness.

AI Executive Summary

As the demand for integrating artificial intelligence into high-stakes environments continues to grow, explaining the reasoning behind neural network predictions has shifted from a theoretical curiosity to a strict operational requirement. In autoregressive neural predictions on dynamic physical fields, such as weather forecasting, interpretability is particularly crucial. Traditional gradient-based feature attribution methods are widely used due to their scalability to high-dimensional inputs. However, these methods exhibit a fundamental failure mode in dynamic physical fields: stochastic input perturbations cause geometric displacement in attribution maps rather than stationary amplitude noise. This displacement results in pointwise averaging blurring these spatially misaligned features.

To address this issue, this paper introduces a method called WassersteinGrad, which extracts a geometric consensus of perturbed attribution maps by computing their entropic Wasserstein barycenter. The results, obtained on regional weather data and a meteorologist-validated neural model, demonstrate promising explainability properties of WassersteinGrad over gradient-based baselines across both single-step and autoregressive forecasting settings.

The core technical principle of WassersteinGrad is to use the entropic Wasserstein barycenter to address the geometric displacement issue. By mapping perturbed attribution maps into the space of spatial probability measures and utilizing the entropic regularization of the Wasserstein barycenter, WassersteinGrad extracts a geometric consensus, achieving more interpretable predictions in high-dimensional inputs. This innovation not only enhances the transparency of model predictions but also provides a more reliable explanation tool for AI applications in high-stakes environments.

Experimental results show that WassersteinGrad significantly reduces geometric displacement in attribution maps while maintaining predictive model performance. At noise levels σ<0.4, model performance degrades by less than 1%, yet attribution centroids advect by an average of 10-15 km. Additionally, in autoregressive forecasting, WassersteinGrad absorbs geometric distortion accumulation through geometric consensus, exhibiting robustness superior to other baseline methods.

The introduction of WassersteinGrad holds significant implications for both academia and industry. It addresses the longstanding challenge of interpretability in dynamic physical field predictions and provides a more reliable explanation tool for AI applications in high-stakes environments. By enhancing the transparency of model predictions, WassersteinGrad helps build trust in AI systems, particularly in high-risk areas such as weather forecasting. Additionally, the geometric consensus mechanism offers new insights for explainable AI research in other domains.

However, WassersteinGrad's computational complexity on high-dimensional grids is relatively high, despite optimization through entropic regularization and Sinkhorn iterations. Further improvements in computational efficiency are needed. Additionally, the generalizability of this method across different predictive models and spatial domains remains to be validated, particularly in other physical fields with complex dynamics. Future research directions include exploring more efficient computational methods to reduce computational costs, validating the method's generalizability, and developing more physically meaningful noise injection strategies.

Deep Analysis

Background

As deep learning models achieve remarkable complexity and performance across various tasks, their inherent opacity remains a critical barrier to operational trust in high-stakes applications. This opacity is particularly problematic in weather forecasting, where erroneous predictions can have severe physical or societal consequences. In recent years, deep learning has increasingly competed with traditional physical solvers in weather forecasting, becoming a new predictive tool. However, as these systems approach deployment in high-stakes decision pipelines, explaining their predictions becomes a pressing issue. Among existing explainable AI strategies, gradient-based feature attribution methods have become standard due to their scalability to high-dimensional spatial domains. However, these methods have a fundamental limitation: they often appear visually noisy and locally shattered due to the highly non-linear nature of deep networks. To reduce this noise, widely adopted smoothing techniques like SmoothGrad first sample several noisy neighbors of the input observation with spatially independent Gaussian noise and then average the gradients explaining the predictions obtained using the perturbed inputs. Although these strategies have proven effective on image data, we argue that reliance on pointwise averaging leads to poorly localized explanations when used to forecast the state of dynamic physical phenomena.

Core Problem

In predicting dynamic physical fields, traditional gradient-based feature attribution methods exhibit a fundamental failure mode: stochastic input perturbations cause geometric displacement in attribution maps rather than stationary amplitude noise. This displacement results in pointwise averaging blurring these spatially misaligned features, affecting the interpretability of predictions. Specifically, in meteorological phenomena, this geometric displacement manifests as attribution masses migrating from their true spatial location to nearby physically incorrect locations. This phenomenon is formally recognized as phase error in meteorological verification, predicting the right thing but in the wrong location. Additionally, when the prediction model is used in an autoregressive manner, this geometric displacement phenomenon is reinforced. Each autoregressive step likely introduces an independent geometric distortion, and when backpropagating gradients from final lead times to explain noisy inputs, these spatial misalignments accumulate. Therefore, developing novel explainable AI solutions that take into account the intrinsic properties of modern autoregressive forecast models is particularly interesting.

Innovation

The core innovation of this paper lies in introducing a transport-based aggregation framework called WassersteinGrad to address the geometric displacement issue in dynamic physical field predictions. Specifically, WassersteinGrad extracts a geometric consensus of perturbed attribution maps by computing their entropic Wasserstein barycenter. This innovation not only enhances the transparency of model predictions but also provides a more reliable explanation tool for AI applications in high-stakes environments. Compared to existing methods like SmoothGrad, WassersteinGrad not only addresses the geometric displacement issue but also achieves significant improvements in interpretability and robustness. By mapping perturbed attribution maps into the space of spatial probability measures and utilizing the entropic regularization of the Wasserstein barycenter, WassersteinGrad extracts a geometric consensus, achieving more interpretable predictions in high-dimensional inputs. This innovation opens new pathways for the application of explainable AI in dynamic physical fields.

Methodology

  • �� Input Channel Selection: Select an input channel cin on which spatial attributions will be represented (e.g., select a spatial slice).

  • �� Output Channel Selection: Select an output channel cout of the predicted field to explain (e.g., select surface precipitation).

  • �� Region of Interest (ROI): Define a region of interest B and compute a scalar target over this region.

  • �� Attribution Computation: Compute the gradient of the target with respect to the selected input channel.

  • �� Perturbed Input Construction: Perturb the input with Gaussian noise to construct channel-specific perturbed inputs.

  • �� Attribution Map Calculation: Compute the gradient of the target with respect to the perturbed input slice.

  • �� Attribution Map Mapping: Map each attribution map to a discrete spatial probability distribution.

  • �� Wasserstein Barycenter Calculation: Extract geometric consensus by computing the entropic Wasserstein barycenter of the perturbed attribution maps.

Experiments

The experimental design includes using the AROME high-resolution meteorological benchmark dataset, derived from the AROME limited-area model at Météo-France, providing kilometer-scale analyses over Western Europe. A subdomain over France is used as the test split, spanning January to December 2023. The forecasting backbone is a pretrained hybrid convolutional-attention U-Net trained using Py4cast, predicting meteorological states at 1-hour lead time. In the experiments, total surface precipitation is chosen as the prediction target, and zonal wind at 250 hPa is used as the input channel. All stochastic methods use N perturbed samples with noise variance σ.2x t)x t)), WGBary and WGBary×Grad use Sinkhorn λ=0.001, selected via a faithfulness-sparsity-robustness tradeoff.

Results

Experimental results show that WassersteinGrad outperforms gradient-based baseline methods in interpretability metrics. In both single-step and autoregressive forecasting settings, WassersteinGrad demonstrates superior interpretability, particularly in spatial consistency and physical localization. Specific data indicate that at noise levels σ<0.4, model performance degrades by less than 1%, yet attribution centroids advect by an average of 10-15 km. Additionally, in autoregressive forecasting, WassersteinGrad absorbs geometric distortion accumulation through geometric consensus, exhibiting robustness superior to other baseline methods.

Applications

WassersteinGrad has direct application potential in high-stakes environments such as weather forecasting. By enhancing the transparency of model predictions, WassersteinGrad helps build trust in AI systems, particularly in high-risk areas such as weather forecasting. Additionally, the geometric consensus mechanism offers new insights for explainable AI research in other domains. In the future, this method could also be applied to predictions in other dynamic physical fields, such as oceanography and atmospheric sciences, helping scientists better understand and explain complex natural phenomena.

Limitations & Outlook

WassersteinGrad's computational complexity on high-dimensional grids is relatively high, despite optimization through entropic regularization and Sinkhorn iterations. Further improvements in computational efficiency are needed. Additionally, the generalizability of this method across different predictive models and spatial domains remains to be validated, particularly in other physical fields with complex dynamics. The current noise injection strategy still uses white Gaussian noise, lacking physical information. Future work should explore more physically meaningful perturbation strategies.

Plain Language Accessible to non-experts

Imagine you're in a kitchen cooking a meal. You have a bunch of ingredients (input data), and you want to know how to turn them into a delicious dish (prediction result). Traditional methods are like following a fixed recipe, regardless of the freshness or seasonal changes of the ingredients. WassersteinGrad, on the other hand, is like a smart chef who adjusts the cooking method based on the changes in the ingredients, ensuring that every dish is tasty and consistent. This smart chef observes the changes in the ingredients (input perturbations) and finds the best cooking method (geometric consensus) to ensure the quality and consistency of each dish (prediction result). This way, even if the ingredients change a bit (input noise), the chef can still make a delicious dish (interpretable prediction).

ELI14 Explained like you're 14

Hey there, friends! Do you know how weather forecasts are made? Scientists use supercomputers to predict the weather, just like you use a computer to calculate every move in a game. But sometimes these predictions are hard to understand, like when you don't know why your game character suddenly runs off course. To make these predictions easier to understand, scientists invented a method called WassersteinGrad. Imagine you're playing a maze game, and WassersteinGrad is like a smart guide who tells you where to go at every step, even if the maze walls are moving (input perturbations). This way, you can better understand the game's rules (prediction results) and find the right exit (interpretable prediction). Isn't that cool?

Glossary

WassersteinGrad

A novel method for explaining predictions of dynamic physical fields by computing the entropic Wasserstein barycenter of perturbed attribution maps.

Used to address the geometric displacement issue found in traditional gradient smoothing methods in dynamic physical fields.

Autoregressive Model

A predictive model where the current prediction result is used as input for the next step's prediction.

Used in weather forecasting for continuous time-step predictions.

SmoothGrad

An explainable AI method that reduces noise by perturbing the input multiple times and averaging the gradients.

Used to enhance the visual clarity of gradient attribution maps.

Entropy Wasserstein Barycenter

A method for computing the geometric consensus of multiple probability distributions, made efficient through entropic regularization.

Used to extract geometric consensus from perturbed attribution maps.

Gradient Attribution

A method for explaining model predictions by computing the gradient of the model output with respect to the input.

Used to identify parts of the input that most influence the prediction result.

Phase Error

A type of prediction error where the predicted phenomenon is spatially misplaced.

In meteorological verification, used to describe predicting the right phenomenon but in the wrong location.

Optimal Transport

A mathematical tool for computing the minimal transport cost between two probability distributions.

Used in WassersteinGrad to compute the geometric consensus of perturbed attribution maps.

Sinkhorn Iteration

An efficient algorithm for solving entropic regularization optimal transport problems.

Used in WassersteinGrad to compute the entropic Wasserstein barycenter.

Gaussian Noise

A commonly used type of random noise with normal distribution characteristics.

Used in SmoothGrad and WassersteinGrad for input perturbation.

Attention Mechanism

A technique used to enhance a model's focus on important parts of the input, commonly used in deep learning models.

Used in Transformer models to dynamically adjust the weights of input features.

Open Questions Unanswered questions from this research

  • 1 How can the computational efficiency of WassersteinGrad be further improved on high-dimensional grids? Despite optimization through entropic regularization and Sinkhorn iterations, computational costs remain high in practical applications, necessitating the development of more efficient computational methods.
  • 2 What is the generalizability of WassersteinGrad across different predictive models and spatial domains? Particularly in other physical fields with complex dynamics, the effectiveness of this method remains to be validated.
  • 3 How can more physically meaningful noise injection strategies be developed? The current noise injection strategy still uses white Gaussian noise, lacking physical information, and future work should explore more physically meaningful perturbation strategies.
  • 4 Is the geometric consensus mechanism of WassersteinGrad equally effective in other dynamic physical fields? The successful application of this mechanism in weather forecasting raises the question of its applicability to other domains, such as oceanography and atmospheric sciences.
  • 5 How can the interpretability and robustness of WassersteinGrad be further improved? Although the method has achieved significant improvements in interpretability and robustness, further exploration of new methods is needed to enhance its performance.

Applications

Immediate Applications

Weather Forecasting

WassersteinGrad can be directly applied to weather forecasting, enhancing the interpretability and transparency of prediction results, helping meteorologists better understand and explain weather changes.

Atmospheric Science Research

By enhancing the transparency of model predictions, WassersteinGrad can help atmospheric scientists better understand and explain complex atmospheric phenomena, advancing scientific research.

Oceanography Applications

In oceanography, WassersteinGrad can be used to explain predictions of ocean dynamics models, helping scientists better understand ocean changes.

Long-term Vision

Climate Change Research

WassersteinGrad can be used in climate change research, helping scientists better understand and explain predictions of climate models, advancing climate change research.

Smart City Planning

By improving the accuracy and interpretability of weather forecasts, WassersteinGrad can provide more reliable data support for smart city planning, helping city managers make more informed decisions.

Abstract

As the demand to integrate Artificial Intelligence into high-stakes environments continues to grow, explaining the reasoning behind neural-network predictions has shifted from a theoretical curiosity to a strict operational requirement. Our work is motivated by the explanations of autoregressive neural predictions on dynamic physical fields, as in weather forecasting. Gradient-based feature attribution methods are widely used to explain the predictions on such data, in particular due to their scalability to high-dimensional inputs. It is also interesting to remark that gradient-based techniques such as SmoothGrad are now standard on images to robustify the explanations using pointwise averages of the attribution maps obtained from several noised inputs. Our goal is to efficiently adapt this aggregation strategy to dynamic physical fields. To do so, our first contribution is to identify a fundamental failure mode when averaging perturbed attribution maps on dynamic physical fields: stochastic input perturbations do not induce stationary amplitude noise in attribution maps, but instead cause a geometric displacement of the attributions. Consequently, pointwise averaging blurs these spatially misaligned features. To tackle this issue, we introduce WassersteinGrad, which extracts a geometric consensus of perturbed attribution maps by computing their entropic Wasserstein barycenter. The results, obtained on regional weather data and a meteorologist-validated neural model, demonstrate promising explainability properties of WassersteinGrad over gradient-based baselines across both single-step and autoregressive forecasting settings.

stat.ML cs.LG

References (20)

for the Meteo-France

Pierre Brousseau, V. Vogt, É. Arbogast et al.

75 citations ⭐ Influential

A multiscale analysis of mean-field transformers in the moderate interaction regime

Giuseppe Bruno, Federico Pasqualotto, Andrea Agazzi

2025 10 citations ⭐ Influential View Analysis →

SmoothGrad: removing noise by adding noise

D. Smilkov, Nikhil Thorat, Been Kim et al.

2017 2579 citations ⭐ Influential View Analysis →

The Mean-Field Dynamics of Transformers

Philippe Rigollet

2025 8 citations ⭐ Influential View Analysis →

Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps

K. Simonyan, A. Vedaldi, Andrew Zisserman

2013 8163 citations ⭐ Influential View Analysis →

A mathematical perspective on Transformers

Borjan Geshkovski, Cyril Letrouit, Yury Polyanskiy et al.

2023 132 citations ⭐ Influential View Analysis →

Fuzzy verification of high‐resolution gridded forecasts: a review and proposed framework

E. Ebert

2008 476 citations

Certified Adversarial Robustness via Randomized Smoothing

Jeremy M. Cohen, Elan Rosenfeld, J. Kolter

2019 2438 citations View Analysis →

Decoupled Weight Decay Regularization

I. Loshchilov, F. Hutter

2017 33038 citations

The Shattered Gradients Problem: If resnets are the answer, then what is the question?

David Balduzzi, Marcus Frean, Lennox Leary et al.

2017 445 citations View Analysis →

Towards better understanding of gradient-based attribution methods for Deep Neural Networks

Marco Ancona, Enea Ceolini, Cengiz Öztireli et al.

2017 991 citations

Axiomatic Attribution for Deep Networks

Mukund Sundararajan, Ankur Taly, Qiqi Yan

2017 7682 citations View Analysis →

On the Robustness of Interpretability Methods

David Alvarez-Melis, T. Jaakkola

2018 629 citations View Analysis →

Fast Discrete Distribution Clustering Using Wasserstein Barycenter With Sparse Support

Jianbo Ye, Panruo Wu, J. Z. Wang et al.

2015 136 citations View Analysis →

Solutions of stationary McKean–Vlasov equation on a high-dimensional sphere and other Riemannian manifolds

Anna Shalova, André Schlichting

2024 11 citations View Analysis →

A Displacement-Based Error Measure Applied in a Regional Ensemble Forecasting System

C. Keil, G. Craig

2007 100 citations

Backpropagation Applied to Handwritten Zip Code Recognition

Yann LeCun, B. Boser, J. Denker et al.

1989 12008 citations

Finding the right XAI method - A Guide for the Evaluation and Ranking of Explainable AI Methods in Climate Science

P. Bommer, M. Kretschmer, Anna Hedström et al.

2023 64 citations View Analysis →

Are artificial neural networks black boxes?

J. M. Benítez, J. Castro, I. Requena

1997 560 citations

On Spectral Properties of Gradient-Based Explanation Methods

Amir Mehrpanah, Erik Englesson, Hossein Azizpour

2025 2 citations View Analysis →