WxFormer: Weather-to-Climate Scale Atmosphere Emulation
WxFormer is a weather to climate scale UNET crossformer model used to autoregressively predict the next state of the atmosphere at hourly time resolution. This model is developed within our Community Runnable Earth Digital Intelligence Twin (CREDIT) framework here at NCAR (which John Schreck and I are the lead developers).
Our aim with CREDIT is to create a platform that democratizes access to the weather emulation space, enabling the broader community to engage with and contribute to advanced atmospheric modeling. By providing an accessible and robust framework, we hope to foster collaboration, innovation, and a deeper understanding of atmospheric processes on both weather and climate scales.
In developing WxFormer, we are stress testing NCAR’s Derecho supercomputer. Utilizing Full Sharded Data Parallel (FSDP) training across 64 GPUs, we have established a high-performance system that the community can also leverage. This setup not only advances our predictive capabilities but also empowers researchers, educators, and enthusiasts by providing the tools and data necessary to explore and improve atmospheric science.
We believe that this open and collaborative approach will drive significant advancements in the field and help address some of the most pressing challenges in climate and weather prediction.
Leveraging Machine Learning and Nudging Increments to Improve the MJO in CESM
This project aims to improve the representation of the Madden-Julian Oscillation (MJO) within the Community Earth System Model (CESM) by integrating a machine learning correction to the online model state. This is accomplished by leveraging tendencies learned from a simple data assimilation system. Due to physics-based and numerical deficiencies (e.g., subgrid parameterization approximations) climate model simulations contain innate biases and uncertainties which can ultimately hamper decision making. A measure of the quality of a modeling system is the analysis increment from a data assimilation system, or the augmentation applied to the initial state of the atmosphere to determine the analysis state. In a perfect model with perfect observations the analysis increment will always be zero. However, a clear indication of bias in the forward model is the presence of systematic features in the DA analysis increments, such as persistent values in the increment mean (after appropriate temporal averaging) or regularly recurring/flow-dependent spatial patterns (Dee, 2005). In this study we utilize the ERA5 reanalysis to linearly relax model simulations back to an observed trajectory and collect the model increments. These increments are then fed into a suite of machine learning model frameworks which learn state dependent corrections to the MJO in online model runs. As the MJO has global effects the hope is that by correcting these tropical signals we achieve a more robust model globally.
Connecting CNNs with Fortran-based CESM presented significant challenges due to the contrasting paradigms and computational demands. We overcame these hurdles using FTORCH and Forpy, enabling seamless interactions between machine learning models and the atmospheric science codebase. This integration marks a significant technological advancement in the field, merging cutting-edge machine learning techniques with traditional numerical weather prediction models. I invite interested parties to collaborate or seek guidance on integrating machine learning parameterizations into CESM/CAM, leveraging NCAR’s computational resources. Our team is ready to provide support and share insights gained from this pioneering project.
This project not only advances our understanding and prediction capabilities for the MJO but also sets a precedent for future enhancements in climate modeling through interdisciplinary collaborations. Below I am showing an animation of the online tendency adjustments to the zonal component of the wind at the surface as the model moves between different MJO states. A CNN is responsible for making this full predicted tendency change (left is the anomaly increment and right is climatological increment + this anomaly)
A Framework for a CAM5 / CAM6 Supermodel
GitHub Repository | Link to Full Paper |
The modeling of weather and climate has been a remarkable success, with forecast accuracy steadily improving and model biases decreasing. Traditionally, combining outputs from multiple models post-simulation has enhanced forecast skill and reduced biases. However, are we fully exploiting the capabilities of state-of-the-art models?
Introduction to Supermodeling
Supermodeling represents a significant advancement in the multimodel ensemble approach. Unlike traditional methods, supermodeling allows individual models to exchange state information in real-time, influencing each other’s behavior during simulations. By optimizing parameters based on past observations, supermodeling reduces errors early, preventing them from propagating and affecting larger scales and variables. This process leads to a synchronized solution that remains closer to observed evolutions, effectively creating a new dynamical system—a supermodel. This innovative approach can significantly enhance current weather forecasts and climate predictions.
Supermodeling Framework
Our framework introduces the first atmosphere-connected supermodel using versions of the Community Atmosphere Model (CAM) 5 and 6. These models exchange information interactively in real-time, allowing them to correct each other’s systematic errors. This integration is flexible and efficient, facilitating forward progress in modeling.
In our study, we examine an untrained supermodel where each version is equally weighted in creating pseudo-observations. The models successfully synchronize in storm track regions across multiple time scales and variables, even those not directly exchanged. While synchronization is less in the tropics, model variability is only reduced in regions with very low synchronization. Additionally, the low-frequency modes of variability, such as NAO/PNA, are not degraded compared to the base models, and biases in some variables are reduced compared to control simulations and non-interactive ensembles.
Supermodeling is not limited to weather and climate modeling but can be applied to any complex system with multiple models, enhancing prediction capabilities. This framework marks a promising step toward more accurate and reliable weather and climate forecasts.
William E. Chapman¹, Francine Schevenhoven², Judith Berner¹, Noel Keenlyside², Ingo Bethke², Ping-Gin Chiu², Alok Kumar Gupta², and Jesse Nusbaumer¹
¹National Center for Atmospheric Research, Boulder, CO, USA
²Geophysical Institute and Bjerknes Centre for Climate Research, University of Bergen
Leveraging ML in CAM to correct Model error
In the second phase of our project [see the method described above], we have advanced our machine learning (ML) algorithms to now learn comprehensive corrections to the model’s state using data assimilation (DA) increments. Weather and climate models, such as the Community Atmosphere Model (CAM) and the Community Earth System Model (CESM), often exhibit biases due to limited resolution and suboptimal physical parameterizations. To address this, our approach utilizes vision-based machine learning techniques to derive corrective tendencies from a hindcast simulation that is linearly nudged towards observational analysis. These techniques have proven effective in predicting nudging tendencies directly from the model state.
We train the ML algorithm using only the model state to correct the systematic errors inherent in the model. This correction leads to significant improvements in the background climate state. Furthermore, modes of low variability, which are often misrepresented without these corrections, now exhibit substantial enhancements. Below, we present an animation illustrating the application of surface wind field corrections at various time steps, showcasing the effectiveness of our ML-driven approach in enhancing model accuracy.
Linking CNNs and Fortran is hard and it was a major technological jump to get these things going, we use FTORCH and Forpy as our solutions. Please let me know if you would like help getting ML parameterizations into CESM/CAM on the NCAR computer systems!
Leveraging DART/Nudging Increments to Correct Model Bias in the Community Atmosphere Model
Due to physics-based and numerical deficiencies (e.g., subgrid parameterization approximations) climate model simulations contain innate biases and uncertainties which can ultimately hamper decision making. A measure of the quality of a modeling system is the analysis increment from a data assimilation system, or the augmentation applied to the initial state of the atmosphere to determine the analysis state. In a perfect model with perfect observations the analysis increment will always be zero. However, a clear indication of bias in the forward model is the presence of systematic features in the analysis increments, such as persistent values in the increment mean (after appropriate temporal averaging) or regularly recurring/flow-dependent spatial patterns (Dee, 2005). In this work we develop and compare model-error representation schemes derived from data assimilation increments and nudging tendencies in multi-decadal simulations of the community atmosphere model, version 6. Each scheme applies a bias correction during simulation run-time to the zonal and meridional winds. We quantify to which extent such online adjustment schemes improve the model climatology and variability on daily to seasonal timescales. Generally, we observe a ca. 30% improvement to annual upper-level zonal winds, with largest improvements in boreal spring (ca. 35%) and winter (ca. 47%). Despite only adjusting the wind fields, we additionally observe a ca. 20% improvement to annual precipitation over land, with the largest improvements in boreal fall (ca. 36%) and winter (ca. 25%), and a ca. 50% improvement to annual sea level pressure, globally. With mean state adjustments alone, the dominant pattern of boreal low-frequency variability over the Atlantic (the North Atlantic Oscillation) is significantly improved. Additional stochasticity further increases the modal explained variances, which brings it closer to the observed value. A streamfunction tendency decomposition reveals that the improvement is due to an adjustment to the high- and low-frequency eddy-eddy interaction terms. In the Pacific, the mean state adjustment alone led to an erroneous deepening of the Aleutian low, but this was remedied with the addition of stochastically selected tendencies. Finally, from a practical standpoint, we discuss the performance of using data assimilation increments versus nudging tendencies for an online model-error representation. link to paper
Project Lead: Will Chapman
Exploring the Relative and Combined Contribution of the MJO and ENSO to Midlatitude Subseasonal Predictability with an Interpretable Neural Network
Here we explore the relative contribution of the Madden-Julian Oscillation (MJO) and El Niño Southern Oscillation (ENSO) to midlatitude subseasonal predictive skill of upper atmospheric circulation over the North Pacific, using an inherently interpretable neural network applied to pre-industrial control runs of the Community Earth System Model version 2. We find that this interpretable network generally favors the state of ENSO, rather than the MJO, to make correct predictions on a range of subseasonal lead times and predictand averaging windows. Moreover, the predictability of positive circulation anomalies over the North Pacific is comparatively lower than that of their negative counterparts, especially evident when the ENSO state is important. However, when ENSO is in a neutral state, our findings indicate that the MJO provides some predictive information, particularly for positive anomalies. We identify three distinct evolutions of these MJO states, offering fresh insights into opportune forecasting windows for MJO teleconnections.
Project Leads: Will Chapman, Kirsten Mayer
Distilling Systematic Model Error from DA/Nudging Tendencies Using Machine Learning
Due to physics-based and numerical deficiencies (e.g., subgrid parameterization approximations) climate model simulations contain innate biases and uncertainties which can ultimately hamper decision making. A measure of the quality of a modeling system is the analysis increment from a data assimilation system, or the augmentation applied to the initial state of the atmosphere to determine the analysis state. In a perfect model with perfect observations the analysis increment will always be zero. However, a clear indication of bias in the forward model is the presence of systematic features in the analysis increments, such as persistent values in the increment mean (after appropriate temporal averaging) or regularly recurring/flow-dependent spatial patterns (Dee, 2005). In this work we leverage a perfect modeling framework to distill true model error from systematic DA and Linear relaxation analysis increments. A machine learning based equation discovery method (via the PYSR python package) is used to seperate tendencies not associated with model error from those which represent a sysematic model drift.
Project Lead: Will Chapman
Probabilistic Weather Prediction With Neural Networks
Most dynamic ensembles are underdispersive on synoptic time scales, meaning that they are giving us less reliable probabilistic information than we hope for. Modern post-processing methods have been developed to address this issue and calibrate models. However, dynamically generated ensembles are extremely computationally expensive. Using integrated vapor transport as the variable of interests, we show here that on weather time scales, we can use deep learning to generate probabilistic models from deterministic systems, that either outperform or compete with modern ensemble methods (even when they have been calibrated). link to paper
Project Lead: Will Chapman
Collaborators: Luca Delle Monache, Aneesh Subramanian, Stefano Alessandrini, Negin Hayatbini, Shang-Ping Xie, Marty Ralph
Phase-Dependent Forecast Skill of the Madden Julian Oscillation (MJO) Teleconnection in Early and Late Winter.
Using a coupled ensemble hindcast of the 20th century (period 1901–2010), the phase-dependent Madden Julian Oscillation (MJO) teleconnection variability in the midlatitudes was investigated with November and February model initializations. The February initialized hindcasts show enhanced teleconnection anomalies and forecast accuracy when compared with their November counterparts.
The phase-dependent forecast skill of the MJO teleconnection was examined by partitioning ensemble members initialized during active MJO phases 3 and 4 (MJO34) and during active MJO phases 7 and 8 (MJO78). We show that MJO78 forecasts have significantly higher forecast skill over the Pacific for the February initializations when compared with their MJO34 counterparts. The potential role of transient eddies was assessed, which supported the evidence that the transient eddies in MJO38 (MJO78) forecasts act to diminish (maintain) the midlatitude circulation anomalies.
Finally, we investigated the spatiotemporal evolution of MJO forced geopotential height anomalies in the ensemble spread with week-reliant singular-value decomposition. Significant phase-dependent differences in the forecast uncertainty of the late-season MJO34 and MJO78 teleconnections exist. The uncertainty growth is linked to two sources 1) chaotic growth of the uncertainty in the midlatitude atmosphere due to internal variability 2) tropically derived uncertainty owed to the growth of the leading mode in the ensemble spread of upper-level tropical divergence, which manifests as independent realizations of the MJO itself. The MJO78 forecast teleconnections are shown to be inherently more predictable than MJO34 forecasts by ~5 forecast days.
Project Lead: Will Chapman
Collaborators: Aneesh Subramanian, Shang-Ping Xie, Antje Weisheimer
Assessing the Potential Predictability of North Pacific Winter IVT and Precipitation Extremes in Subseasonal to Seasonal Forecasts
Chaos within the atmosphere causes the predictability of weather at a single instant in time to range from a few days to a few weeks depending on existing circulation patterns (Lorenz 1965). However, recent studies have shown that there are “windows of opportunity” that allow skillful forecasts to be extended into the subseasonal to seasonal (S2S) range (Robertson et al. 2015, Vitart et al. 2017, White et al. 2017, Mariotti 2020). Despite the limit of predictability at singular moments, some processes can create signals in the predictability of broader windows of time that are stronger than the noise of uncertainty caused by chaos. There is still considerable uncertainty over the differences in predictability amongst the characteristics that describe our atmosphere. Lavers et al. (2016) demonstrated that integrated vapor transport (IVT), which plays a key role in driving atmospheric rivers (ARs) (Shields et al. 2018, Ralph et al. 2019) and severe US west coast precipitation (Waliser and Guan 2017, Ricciotti and Cordeira 2022), has potential predictive skill at longer lead times than precipitation itself in medium-range forecasts. Determining what useful information can be extracted from S2S forecasts can have meaningful impacts on water management decisions on the US west coast, a region that is prone to drought and flooding (Das et al. 2013, Mann and Gleick 2015, Williams et al. 2015, Corringham et al. 2019). There is still little known of the discrepancies between the predictability of IVT compared to predictability of precipitation at S2S lead times. In this study, we explore the differences between the potential predictability of IVT and precipitation in S2S forecasts. We will present results showing that the overall significant skill for both precipitation and IVT drops below an Anomaly Correlation Coefficient (ACC) value of 0.6 in almost all spatial locations after 2 weeks. The Pacific North America pattern (PNA) has been shown to be associated with forecast skill in S2S forecasts and can have serious implications on US west coast precipitation (Baxter and Nigam 2013). We will show that there is an area of persistent skill within a region that experiences frequent impactful AR genesis activity (Prince et al. 2021) when forecasts are conditioned the PNA and various weather regimes.
Project Lead: Tim Higgins
Collaborators: Aneesh Subramanian, Will Chapman, Andrew Winters, David Lavers
Interpretable Machine Learning applied to Seasonal Forecasting of Western US Precipitation
Seasonal forecasting of precipitation across the Western United States remains a major scientific challenge. Improvements to the existing forecast skill would be highly valuable for stakeholders and decision makers for planning around drought and floods. Relatively little research has been directed towards testing machine learning for seasonal forecasting. A major barrier is the limited amount of data to train machine learning models at the seasonal time resolution. To address this issue, we test the feasibility of training machine learning on large initial condition climate model simulations. These simulations span several thousand years, providing a large amount of data to train on. link to paper
Project Lead: Peter Gibson
Collaborators: Will Chapman, Alphan Altinok, Mike Deflorio, Luca Delle Monche, Duane Waliser
Parameterizing subgrid-scale eddy effects using deep learning
Most eddy-permitting models presently employ some kind of hyper-viscosity, which is shown to cause a significant amount of energy dissipation. However, comparison to higher resolution simulations shows that only enstrophy, but almost no energy, should be dissipated below the grid-scale. As a result of the artificial energy sink associated with viscous parameterizations, the eddy fields in eddy permitting models are generally not energetic enough. - Jansen and Held, 2014
Here a new approach for sub-grid eddy parameterization in eddy-permitting ocean models is explored by using deep learning. We test this in idealized QG models and show substantial improvements in coarse models.
Project Leads: Will Chapman, Nick Lutsko, and Tom Beucler
Potential Increase in MJO Predictability Due to Global Warming
The Madden-Julian Oscillation (MJO) is the leading source of predictability in our climate system on the subseasonal time scale. In this study, we explore and explain the increasing MJO predictability during the past century. We use RMMI to represent MJO. First, we will show the increasing MJO predictability trend we observed from model ensemble forecasts and reanalysis data. Following the traditional method of using model ensemble forecasts and evaluating with the bivariate anomaly correlation coefficient, we obtained a significant positive trend in MJO predictability for the past century. We then analyzed the MJO in ECMWF coupled climate reanalysis for the 20th century (CERA-20C) using the Weighted Permutation Entropy (WPE) method, which has been proven as a useful tool in analyzing predictability. The higher the WPE, the lower the predictability. We witnessed a consistent decreasing trend in WPE among all 10 CERA-20C ensemble members, which reflects a robust, increasing trend in the MJO predictability.
Then, we will present the MJO predictability change in CESM2 and CESM2-WACCM historical runs using the WPE method. Most historical runs are with a WPE changing trend within the spread of the trends estimated from the control run; however, the distribution of the WPE trends in historical runs shifts to the negative side compared to the distribution calculated from the control run. This suggests that the increasing MJO predictability we observed in the past century is likely caused by the internal climate variability and the external forcing (the global warming).
Next, we will present the MJO predictability change in CESM2 and CESM2-WACCM future projections under the ssp585 scenario. With a much stronger global warming forcing, the distribution of the WPE trends shifts even more to the negative side than the distribution calculated from the historical runs, which further supports the assumption that global warming can increase the MJO predictability.
Finally, we will explain why there is such an increase in MJO predictability. In both reanalysis data and CESM2/CESM2-WACCM ssp585 future projection, we noticed that, within a range of 10 days, the sequential amplifying/weakening of RMM1, RMM2 and MJO amplitude, and the organized eastward propagation occur more and more frequently. These regular patterns make the MJO more predictable.
Project Lead: Danni Du
Collaborators: Aneesh Subramanian, Will Chapman, Weiqing Han
Monthly Modulation of ENSO Teleconnections: Implications for North American Potential Predictability
Using a high-resolution atmospheric general circulation model simulation of unprecedented ensemble size, we examine potential predictability of monthly anomalies under El Niño Southern Oscillation (ENSO) forcing and background internal variability. This study reveals the pronounced month-to-month evolution of both the ENSO forcing signal and internal variability. Internal variance in upper-level geopotential height decreases ($\sim10\%$) over the North Pacific during El Niño as the westerly jet extends eastward, allowing forced signals to account for a greater fraction of the total variability, and leading to increased potential predictability. We identify March of El Niño years as the most predictable month followed closely by February using a signal-to-noise anaylsis. In contrast, December, a month typically included in teleconnection studies, shows little-to-no potential predictability. We show that the seasonal evolution of SST forcing and variability leads to significant signal-to-noise relationships that can be directly linked to both upper-level and surface variable predictability for a given month. The stark changes in forced response, internal variability, and thus signal-to-noise across an ENSO season indicate that subseasonal fields should be used to diagnose potential predictability over North America associated with ENSO teleconnections. Using surface air temperature and precipitation as examples, this study provides motivation to pursue ‘windows of forecast opportunity’, in which statistical skill can be developed, tested, and leveraged to determine times and regions in which this skill may be elevated. link to paper
Project Lead: Will Chapman
Collaborators: Aneesh Subramanian, Mike Sierks, Shang-Ping Xie, Marty Ralph, Youichi Kamae
Assessing Vulnerability and Adaptive Management Under Climate Change Scenarios: Lessons from California’s Largest Reservoir
Climate change is exacerbating the long-standing tensions between water supply and flood-risk mitigation across the Western US and beyond. As springtime snowmelt declines in the face of warming trends, reducing opportunities to refill reservoirs after wintertime flood risks subside, water managers face the decision whether to continue operations designed for a bygone era or to pursue adaptation measures. Differences in factors such as climate, hydrology, and reservoir operations between basins require that impacts of climate change and proposed adaptation strategies be examined on a case-by-case basis. This study investigates projected climate change impacts on California’s Lake Shasta and identifies specific variables that govern its vulnerability. Using a newly developed, highly flexible model, we analyze coming threats to water supply and flood risk under existing operations and several forms of adaptive responses to climate change. Compared to the historical period, we simulate 27% declines in carryover storage at the end of the 21st century, under the more severe of two warming scenarios, if operations are left unchanged. Compounding the direct impacts due to decreased snowpack, we find existing reservoir operating procedures are responsible for one-third of average losses. Both operational and infrastructural adaptive measures were explored by altering rule curve and increasing reservoir storage capacity. Despite many interventions favoring water supply over flood risk, historical levels of carryover storage were irretrievable at the end of the century under the warmer of the two warming scenarios examined in this study. link to paper
Collaborators: Mike Dettinger, Will Chapman, Marty Ralph
Hawaii Lee Wind Reconstruction Using Deep Learning for Satellite Ambiguity Selection
Satellite scatterometer retrievals provide the only regular vector wind observations over vast swaths of the global oceans and are therefore vital for climate study (Chelton & Xie, 2010; Xie, 2004) and forecasting applications (Atlas et al., 2001; Chelton et al., 2006). However, satellite scatterometer winds have been identified as often errant in regions where coastal orography interacts with oceanic surface winds (Kilpatrick et al 2019). These errors are especially prevalent in Hawaii’s lee wake in the summertime easterly trade wind regime, where upstream winds force two orographically tied vortices which have been well documented (Patzert 1969; Nickerson and Dias, 1981; Smith and Grubišic´ 1993) and affect local precipitation patterns and mesoscale ocean circulation (Yang et al. 2008). Here we test comparitive empirical methods for spatial reconstruction of satellite wind for correcting inaccuracies in Hawaii’s Lee Wind Wake. Methods: Convolutional Neural Networks “inpainting”, Maximum Covariance Analysis, and Canonical Correlates.
Project Lead: Will Chapman
Collaborators: Tom Kilpatrick, Shang-Ping Xie, David John Gagne
Improving Atmospheric River Prediction with Machine Learning
This study tests the utility of convolutional neural networks as a postprocessing framework for improving the National Center for Environmental Prediction’s Global Forecast System’s integrated vapor transport forecast field in the Eastern Pacific and western United States. Integrated vapor transport is the characteristic field of atmospheric rivers, which provide over 65% of yearly precipitation at some western U.S. locations. The method reduces full‐field root‐mean‐square error (RMSE) at forecast leads from 3 hours to seven days (9–17% reduction), while increasing correlation between observations and predictions (0.5–12% increase). This represents an an approximately one‐ to two‐day lead time improvement in RMSE. Decomposing RMSE shows that random error and conditional biases are predominantly reduced. Systematic error is reduced up to five‐day forecast lead, but accounts for a smaller portion of RMSE. This work demonstrates convolutional neural network’s potential to improve forecast skill out to seven days for precipitation events affecting the western United States. link to paper
Project Lead: Will Chapman
Collaborators: Aneesh Subramanian, Luca Delle Monache, Shang-Ping Xie, Marty Ralph