The use of machine learning (ML) for the online correction of coarse-resolution atmospheric models has proven effective in reducing biases in near-surface temperature and precipitation rate. However, this often introduces biases in the upper atmosphere and improvements are not always reliable across ML-corrective models trained with different random seeds. Furthermore, ML corrections can feed back on the baseline physics of the atmospheric model and produce profiles that are outside the distribution of samples used in training, leading to low confidence in the predicted corrections. This study introduces the use of a novelty detector to mask the predicted corrections when the atmospheric state is deemed out-of-sample. The novelty detector is trained on profiles of temperature and specific humidity in a semi-supervised fashion using samples from the coarsened reference fine-resolution simulation. Offline, the novelty detector determines more columns to be out-of-sample in simulations which are known, using simple metrics like mean bias, to drift further from the reference simulation. Without novelty detection, corrective ML leads to the development of undesirably large climate biases for some ML random seeds but not others. Novelty detection deems about 21% of columns to be novelties in year-long simulations. The spread in the root mean square error (RMSE) of time-mean spatial patterns of surface temperature and precipitation rate across a random seed ensemble is sharply reduced when using novelty detection. In particular, the random seed with the worst RMSE is improved by up to 60% (depending on the variable) while the best seed maintains its low RMSE.

W. Andre Perkins

and 3 more

We present a machine learning based emulator of a microphysics scheme for condensation and precipitation processes (Zhao-Carr) used operationally in a global atmospheric forecast model (FV3GFS). Our tailored emulator architecture achieves high skill (≥94%) in predicting condensate and precipitation amounts and maintains low global-average bias (≤4%) for 1 year of continuous simulation when replacing the Fortran scheme. The stability and success of this emulator stems from key design decisions. By separating the emulation of condensation and precipitation processes, we can better enforce physical priors such as mass conservation and locality of condensation, and the vertical dependence of precipitation falling downward, using specific network architectures. An activity classifier for condensation imitates the discrete-continuous nature of the Fortran microphysics outputs (i.e., tendencies are identically zero where the scheme is inactive, and condensate is zero where clouds are fully evaporated). A temperature-scaled conditional loss function ensures accurate condensate adjustments for a high dynamic range of cloud types (e.g., cold, low-condensate cirrus clouds or warm, condensate-rich clouds). Despite excellent overall performance, the emulator exhibits some deficiencies in the uppermost model levels, leading to biases in the stratosphere. The emulator also has short episodic skill dropouts in isolated grid columns and is computationally slower than the original Fortran scheme. Nonetheless, our challenges and strategies should be applicable to the emulation of other microphysical schemes. More broadly, our work demonstrates that with suitable physically motivated architectural choices, ML techniques can accurately emulate complex human-designed parameterizations of fast physical processes central to weather and climate models.

Isabel Louise McCoy

and 4 more

Aerosol-cloud-precipitation interactions are a leading source of uncertainty in estimating climate sensitivity. Remote marine boundary layers where accumulation mode (~100-400 nm diameter) aerosol concentrations are relatively low are very susceptible to aerosol changes. These regions also experience heightened Aitken mode aerosol (~10-100 nm) concentrations associated with ocean biology. Aitken aerosols may significantly influence cloud properties and evolution by replenishing cloud condensation nuclei and droplet number lost through precipitation (i.e., Aitken buffering). We use a large-eddy simulation with an Aitken-mode enabled microphysics scheme to examine the role of Aitken buffering in a mid-latitude decoupled boundary layer cloud regime observed on July 15, 2017 during the ACE-ENA flight campaign: cumulus rising into stratocumulus under elevated Aitken concentrations (~100-200 mg-1). In situ measurements are used to constrain and evaluate this case study. Our simulation accurately captures observed aerosol-cloud-precipitation interactions and reveals time-evolving processes driving regime development and evolution. Aitken activation into the accumulation mode occurs primarily in the cumulus layer, providing a reservoir for turbulence and convection to carry accumulation aerosols into the drizzling stratocumulus layer above. Thus, the cloud regime is buffered against precipitation removal, reducing cloud break-up and associated increases in heterogeneity. We examine cloud evolution sensitivity to initial aerosol conditions. With halved accumulation number, Aitken aerosols restore accumulation concentrations, maintain droplet number similar to original values, and prevent cloud break-up. Without Aitken aerosols, precipitation-driven cloud break-up occurs rapidly. In this regime, mesoscale and synoptic-scale uplift enhance cloud condensate and brightness, but Aitken buffering sustains brighter, more homogeneous clouds for longer.

Brian Henn

and 7 more

Coarse-grid weather and climate models rely particularly on parameterizations of cloud fields, and coarse-grained cloud fields from a fine-grid reference model are a natural target for a machine-learned parameterization. We machine-learn the coarsened-fine cloud properties as a function of coarse-grid model state in each grid cell of NOAA’s FV3GFS global atmosphere model with 200 km grid spacing, trained using a 3 km fine-grid reference simulation with a modified version of FV3GFS. The ML outputs are coarsened fine fractional cloud cover and liquid and ice cloud condensate mixing ratios, and the inputs are coarse model temperature, pressure, relative humidity, and ice cloud condensate. The predicted fields are skillful and unbiased, but somewhat under-dispersed, resulting in too many partially-cloudy model columns. When the predicted fields are applied diagnostically (offline) in FV3GFS’s radiation scheme, they lead to small biases in global-mean top-of-atmosphere (TOA) and surface radiative fluxes. An unbiased global mean TOA net radiative flux is obtained by setting to zero any predicted cloud with grid cell mean cloud fraction less than a threshold of 6.5%; this does not significantly degrade the ML prediction of cloud properties. The diagnostic, ML-derived radiative fluxes are  far more accurate than those obtained with the existing cloud parameterization in the nudged coarse-grid model, as they leverage the accuracy of the fine-grid reference simulation’s cloud properties. 

Matthew Wyant

and 4 more