Jatan Buch

and 4 more

The annual area burned due to wildfires in the western United States (WUS) increased by more than 300% between 1984 and 2020. However, accounting for the nonlinear, spatially heterogeneous interactions between climate, vegetation, and human predictors driving the trends in fire frequency and sizes at different spatial scales remains a challenging problem for statistical fire models. Here we introduce a novel stochastic machine learning (ML) framework to model observed fire frequencies and sizes in 12 km x 12 km grid cells across the WUS. This framework is implemented using Mixture Density Networks trained on a wide suite of input predictors. The modeled WUS fire frequency corresponds well with observations at both monthly (r= 0.94) and annual (r= 0.85) timescales, as do the monthly (r= 0.90) and annual (r= 0.88) area burned. Moreover, the annual time series of both fire variables exhibit strong correlations (r >= 0.6) in 16 out of 18 ecoregions. Our ML model captures the interannual variability and the distinct multidecade increases in annual area burned for both forested and non-forested ecoregions. Evaluating predictor importance with Shapley additive explanations, we find that fire month vapor pressure deficit (VPD) is the dominant driver of fire frequencies and sizes across the WUS, followed by 1000-hour dead fuel moisture (FM1000), total monthly precipitation (Prec), mean daily maximum temperature (Tmax), and fraction of grassland cover in a grid cell. Our findings serve as a promising use case of ML techniques for wildfire prediction in particular and extreme event modeling more broadly.

Sara Shamekh

and 1 more

Accurately representing vertical turbulent fluxes in the planetary boundary layer is vital for moisture and energy transport. Nonetheless, the parameterization of the boundary layer remains a major source of inaccuracy in climate models. Recently, machine learning techniques have gained popularity for representing oceanic and atmospheric processes, yet their high dimensionality limits interpretability. This study introduces a new neural network architecture employing non-linear dimensionality reduction to predict vertical turbulent fluxes in a dry convective boundary layer. Our method utilizes turbulent kinetic energy and scalar profiles as input to extract a physically constrained two-dimensional latent space, providing the necessary yet minimal information for accurate flux prediction. We obtained data by coarse-graining Large Eddy Simulations covering a broad spectrum of boundary layer conditions, from weakly to strongly unstable. These regimes are employed to constrain the latent space disentanglement, enhancing interpretability. By applying this constraint, we decompose the vertical turbulent flux of various scalars into two main modes of variability: wind shear and convective transport. Our data-driven parameterization accurately predicts vertical turbulent fluxes (heat and passive scalars) across turbulent regimes, surpassing state-of-the-art schemes like the eddy-diffusivity mass flux scheme. By projecting each variability mode onto its associated scalar gradient, we estimate the diffusive flux and learn the eddy diffusivity. The diffusive flux is found to be significant only in the surface layer for both modes and becomes negligible in the mixed layer. The retrieved eddy diffusivity is considerably smaller than previous estimates used in conventional parameterizations, highlighting the predominant non-diffusive nature of transport.

Arthur Grundner

and 3 more

A promising method for improving the representation of clouds in climate models, and hence climate projections, is to develop machine learning-based parameterizations using output from global storm-resolving models. While neural networks can achieve state-of-the-art performance, they are typically climate model-specific, require post-hoc tools for interpretation, and struggle to predict outside of their training distribution. To avoid these limitations, we combine symbolic regression, sequential feature selection, and physical constraints in a hierarchical modeling framework. This framework allows us to discover new equations diagnosing cloud cover from coarse-grained variables of global storm-resolving model simulations. These analytical equations are interpretable by construction and easily transferable to other grids or climate models. Our best equation balances performance and complexity, achieving a performance comparable to that of neural networks ($R^2=0.94$) while remaining simple (with only 13 trainable parameters). It reproduces cloud cover distributions more accurately than the Xu-Randall scheme across all cloud regimes (Hellinger distances $<0.09$), and matches neural networks in condensate-rich regimes. When applied and fine-tuned to the ERA5 reanalysis, the equation exhibits superior transferability to new data compared to all other optimal cloud cover schemes. Our findings demonstrate the effectiveness of symbolic regression in discovering interpretable, physically-consistent, and nonlinear equations to parameterize cloud cover.

Reda ElGhawi

and 6 more

The process of evapotranspiration transfers water vapour from vegetation and soil surfaces to the atmosphere, the so-called latent heat flux (𝑄 LE), and thus crucially modulates Earth’s energy, water, and carbon cycles. Vegetation controls 𝑄 LE through regulating the leaf stomata (i.e., surface resistance π‘Ÿ s) and through altering surface roughness (aerodynamic resistance π‘Ÿ a). Estimating π‘Ÿ s and π‘Ÿ a across different vegetation types proves to be a key challenge in predicting 𝑄 LE. Here, we propose a hybrid modeling approach (i.e., combining mechanistic modeling and machine learning) for 𝑄 LE where neural networks independently learn the resistances from observations as intermediate variables. In our hybrid modeling setup, we make use of the Penman-Monteith equation based on the Big Leaf theory in conjunction with multi-year flux measurements across different forest and grassland sites from the FLUXNET database. We follow two conceptually different strategies to constrain the hybrid model to control for equifinality arising when estimating the two resistances simultaneously. One strategy is to impose an a priori constraint on π‘Ÿ a based on our mechanistic understanding (theory-driven strategy), while the other strategy makes use of more observational data and adds a constraint in predicting π‘Ÿ a through multi-task learning of the latent as well as the sensible heat flux (𝑄 H ; data-driven strategy). Our results show that all hybrid models exhibit a fairly high predictive skill for the target variables with 𝑅 2 = 0.82-0.89 for grasslands and 𝑅 2 = 0.70-0.80 for forests sites at the mean diurnal scale. The predictions of π‘Ÿ s and π‘Ÿ a show physical consistency across the two regularized hybrid models, but are physically implausible in the under-constrained hybrid model. The hybrid models are robust in reproducing consistent results for energy fluxes and resistances across different scales (diurnal, seasonal, interannual), reflecting their ability to learn the physical dependence of the target variables on the meteorological inputs. As a next step, we propose to test these heavily observation-informed parameterizations derived through hybrid modeling as a substitute for overly simple ad hoc formulations in Earth system models.

Sara Shamekh

and 3 more

Accurate prediction of precipitation intensity is of crucial importance for both human and natural systems, especially in a warming climate more prone to extreme precipitation. Yet climate models fail to accurately predict precipitation intensity, particularly extremes. One missing piece of information in traditional climate model parameterizations is sub-grid scale cloud structure and organization, which affects precipitation intensity and stochasticity at the grid scale. Here we show, using storm-resolving climate simulations and machine learning, that by implicitly learning sub-grid organization, we can accurately predict precipitation variability and stochasticity with a low dimensional set of variables. Using a neural network to parameterize coarse-grained precipitation, we find mean precipitation is predictable from large scale quantities only; however, the neural network cannot predict the variability of precipitation (R 2 ∼ 0.4) and underestimates precipitation extremes. Performance is significantly improved when the network is informed by our novel organization metric, correctly predicting precipitation extremes and spatial variability (R 2 ∼ 0.95). The organization metric is implicitly learned by training the algorithm on high-resolution precipitable water, encoding organization degree and humidity amount at the subgrid-scale. The organization metric shows large hysteresis, emphasizing the role of memory created by sub-grid scale structures. We demonstrate this organization metric can be predicted as a simple memory process from information available at the previous time steps. These findings stress the role of organization and memory in accurate prediction of precipitation intensity and extremes and the necessity of parameterizing sub-grid scale convective organization in climate models to better project future changes in the water cycle and extremes.

Xuan Xi

and 3 more

Arthur Grundner

and 5 more

A promising approach to improve cloud parameterizations within climate models and thus climate projections is to use deep learning in combination with training data from storm-resolving model (SRM) simulations. The Icosahedral Non-Hydrostatic (ICON) modeling framework permits simulations ranging from numerical weather prediction to climate projections, making it an ideal target to develop neural network (NN) based parameterizations for sub-grid scale processes. Within the ICON framework, we train NN based cloud cover parameterizations with coarse-grained data based on realistic regional and global ICON SRM simulations. We set up three different types of NNs that differ in the degree of vertical locality they assume for diagnosing cloud cover from coarse-grained atmospheric state variables. The NNs accurately estimate sub-grid scale cloud cover from coarse-grained data that has similar geographical characteristics as their training data. Additionally, globally trained NNs can reproduce sub-grid scale cloud cover of the regional SRM simulation. Using the game-theory based interpretability library SHapley Additive exPlanations, we identify an overemphasis on specific humidity and cloud ice as the reason why our column-based NN cannot perfectly generalize from the global to the regional coarse-grained SRM data. The interpretability tool also helps visualize similarities and differences in feature importance between regionally and globally trained column-based NNs, and reveals a local relationship between their cloud cover predictions and the thermodynamic environment. Our results show the potential of deep learning to derive accurate yet interpretable cloud cover parameterizations from global SRMs, and suggest that neighborhood-based models may be a good compromise between accuracy and generalizability.