The Problem with Traditional Weather Prediction
Weather forecasting is one of the most computationally expensive scientific endeavors on Earth. Every day, national meteorological agencies around the world run numerical weather prediction (NWP) models that divide the atmosphere into millions of grid cells, then solve the Navier-Stokes equations of fluid dynamics at each cell for every time step. The European Centre for Medium-Range Weather Forecasts (ECMWF) runs its flagship HRES model on a supercomputer with tens of thousands of processors, and a single 10-day global forecast takes roughly one hour to compute.
This approach has been the backbone of operational weather prediction since the 1950s. It works remarkably well — modern 5-day forecasts are as accurate as 3-day forecasts were 20 years ago. But NWP faces fundamental limitations. The computational cost scales exponentially with resolution, making it difficult to increase grid detail without massive hardware investments. Ensemble forecasting, where you run dozens of slightly varied predictions to quantify uncertainty, is constrained by compute budgets. And the physics-based approach requires hand-tuned parameterizations for processes like cloud formation and convection that occur at scales smaller than the grid.
Enter AI Weather Models
In 2022 and 2023, a wave of AI weather models demonstrated that machine learning could match or exceed traditional NWP accuracy while running orders of magnitude faster. Three models stand out: Google DeepMind's GraphCast, Huawei's Pangu-Weather, and NVIDIA's FourCastNet. Each takes a fundamentally different approach to the same problem, but they share a common strategy: learn the dynamics of the atmosphere directly from decades of observational data.
GraphCast
GraphCast uses a graph neural network (GNN) architecture that represents the Earth's atmosphere as a mesh of interconnected nodes. Given two consecutive atmospheric states (6 hours apart), GraphCast predicts the next state by passing messages between nodes on a multi-scale icosahedral mesh. It was trained on 39 years of ERA5 reanalysis data and operates on a 0.25-degree grid with 37 pressure levels — the same resolution as ECMWF HRES. For a deeper dive, see our GraphCast explainer.
Pangu-Weather
Huawei's Pangu-Weather uses a 3D vision transformer architecture rather than a graph network. It treats the atmospheric state as a 3D volume (latitude x longitude x pressure levels) and applies a modified Swin Transformer to learn spatial relationships. Pangu-Weather trains separate models for 1-hour, 3-hour, 6-hour, and 24-hour prediction horizons, then chains them together for longer forecasts. This multi-timescale approach helps it maintain accuracy over extended forecast periods.
FourCastNet
NVIDIA's FourCastNet takes yet another approach, using Adaptive Fourier Neural Operators (AFNOs) that perform learning in the frequency domain. By operating on Fourier transforms of the atmospheric data rather than the raw spatial data, FourCastNet can efficiently capture global-scale patterns while maintaining computational efficiency. It was one of the first AI weather models to demonstrate NWP-competitive accuracy.
How They're Trained: The ERA5 Foundation
All three models rely on ERA5, the fifth-generation atmospheric reanalysis produced by ECMWF. ERA5 combines historical weather observations (from satellites, weather stations, radiosondes, aircraft, and ships) with a physics-based model to produce a globally consistent dataset of atmospheric conditions from 1940 to the present. It provides hourly data at 0.25-degree resolution across 37 pressure levels, covering variables like temperature, wind, humidity, and geopotential height.
Training an AI weather model on ERA5 is conceptually straightforward: given the atmospheric state at time T, predict the state at time T+6h (or T+1h, depending on the model). The model learns the mapping between consecutive states by seeing millions of examples from decades of weather history. What makes this work is that the atmosphere, despite its complexity, follows consistent physical patterns that neural networks can learn to approximate.
import requests
API_KEY = "sk-sci-your-api-key"
BASE = "https://api.scirouter.ai/v1"
HEADERS = {"Authorization": f"Bearer {API_KEY}"}
# Get a 7-day forecast for a location
response = requests.post(f"{BASE}/climate/forecast",
headers=HEADERS,
json={
"latitude": 40.7128,
"longitude": -74.0060,
"forecast_days": 7,
"variables": ["temperature_2m", "precipitation",
"wind_speed_10m", "humidity"]
})
forecast = response.json()
for day in forecast["daily"]:
print(f"{day['date']}: {day['temp_max']:.1f}C / "
f"{day['temp_min']:.1f}C, "
f"Precip: {day['precipitation_mm']:.1f}mm")Why AI Is Faster: Seconds vs Hours
Traditional NWP models must numerically integrate differential equations forward in time, step by step. Each time step requires solving equations at every grid cell, and the time step size is constrained by numerical stability (the Courant-Friedrichs-Lewy condition). This means a 10-day forecast requires thousands of sequential computation steps.
AI weather models sidestep this entirely. GraphCast generates a 6-hour forecast in a single forward pass through the network. A 10-day forecast requires 40 autoregressive steps (10 days divided by 6-hour increments), and each step takes less than a second on a modern GPU. The total time is under 60 seconds, compared to roughly one hour on a supercomputer for ECMWF HRES.
This speed advantage has profound implications. It makes it practical to run hundreds of ensemble members instead of the 50 that ECMWF typically runs, providing much better uncertainty quantification. It enables real-time reforecasting as new observations come in. And it democratizes weather prediction — a researcher with a single GPU can generate forecasts that previously required supercomputer access.
Accuracy Benchmarks
In head-to-head comparisons, GraphCast outperformed ECMWF HRES on over 90% of the 1,380 verification targets in the WeatherBench2 benchmark. These targets cover different atmospheric variables (temperature, wind, humidity, geopotential) at different pressure levels and forecast lead times. The improvements are particularly notable at 5-10 day lead times, where GraphCast's predictions degrade more slowly than HRES.
Pangu-Weather showed similar competitiveness, with particular strength in upper-atmospheric variables. FourCastNet, while slightly less accurate than GraphCast and Pangu-Weather on average, demonstrated impressive efficiency and was one of the first to show that AI could genuinely compete with operational NWP.
- GraphCast: Beats ECMWF HRES on 90%+ of WeatherBench2 targets, strongest at 5-10 day range
- Pangu-Weather: Competitive with HRES, particularly strong for upper-level geopotential and wind fields
- FourCastNet: Slightly lower accuracy but extremely fast, pioneered the Fourier operator approach
- GenCast: Google DeepMind's diffusion-based model generates calibrated probabilistic forecasts, outperforming ENS
The Future of Weather Forecasting
Weather agencies are rapidly adopting AI models into operational workflows. ECMWF launched its AI Forecasting Engine (AIFE) to integrate machine learning alongside traditional NWP. National weather services in the US, UK, Japan, and China are all experimenting with AI forecast products. The trend is toward hybrid systems that combine the physical consistency of NWP with the speed and pattern recognition of AI.
Key areas of active development include:
- Probabilistic forecasting: Models like GenCast use diffusion architectures to generate full probability distributions rather than single deterministic predictions
- Higher resolution: Current AI models operate at 25 km resolution but regional models at 1-5 km are under development
- Extreme event prediction: Specialized models for hurricanes, severe convection, and atmospheric rivers where standard training data is insufficient
- Longer horizons: Extending skillful forecasts from 10 days toward subseasonal-to-seasonal timescales (2-6 weeks)
Accessing AI Weather Forecasts via API
SciRouter's climate endpoints will bring AI-powered weather forecasting to developers through a simple REST API. Instead of running your own model or parsing complex GRIB files from ECMWF, you can query forecasts programmatically with latitude, longitude, and desired variables. The API handles model inference, data formatting, and caching.
To explore what AI weather forecasting looks like in practice:
- GraphCast Explained – deep technical dive into the GNN architecture
- Historical Climate Data API – access ERA5, NASA POWER, and NOAA datasets
- Climate Change and Your City – use historical data to measure local climate trends
Ready to see AI weather prediction in action? Open the Weather Forecaster Studio or get a free API key to start building with climate data.