Extreme Weather Bench
Builds on the successful work of WeatherBench and introduces a set of high-impact weather events, spanning across multiple spatial and temporal scales and different parts of the weather spectrum.
https://github.com/brightbandtech/extremeweatherbench
Category: Atmosphere
Sub Category: Meteorological Observation and Forecast
Keywords
benchmarking meteorology
Last synced: about 16 hours ago
JSON representation
Repository metadata
Benchmarking of machine learning and numerical weather prediction (MLWP & NWP) models, with a focus on extreme events.
- Host: GitHub
- URL: https://github.com/brightbandtech/extremeweatherbench
- Owner: brightbandtech
- License: mit
- Created: 2024-08-15T15:33:50.000Z (over 1 year ago)
- Default Branch: main
- Last Pushed: 2025-12-12T03:38:27.000Z (15 days ago)
- Last Synced: 2025-12-15T10:15:46.042Z (11 days ago)
- Topics: benchmarking, meteorology
- Language: Python
- Homepage: https://www.brightband.com/benchmarks
- Size: 1.04 GB
- Stars: 87
- Watchers: 4
- Forks: 5
- Open Issues: 31
- Releases: 1
-
Metadata Files:
- Readme: README.md
- License: LICENSE
README.md
Extreme Weather Bench (EWB)
EWB is currently in limited pre-release. Bugs are likely to occur for now.
v1.0 to be published alongside EWB preprint.
As AI weather models are growing in popularity, we need a standardized set of community driven tests that evaluate the models across a wide variety of high-impact hazards. Extreme Weather Bench (EWB) builds on the successful work of WeatherBench and introduces a set of high-impact weather events, spanning across multiple spatial and temporal scales and different parts of the weather spectrum. We provide data to use for testing, standard metrics for evaluation by forecasters worldwide for each of the phenomena, as well as impact-based metrics. EWB is a community system and will be adding additional phenomena, test cases and metrics in collaboration with the worldwide weather and forecast verification community.
Events
EWB has cases broken down by multiple event types within src/extremeweatherbench/data/events.yaml between 2020 and 2024. EWB case studies are documented here.
Available:
| Event Type | Number of Cases |
|---|---|
| 🌇 Heat Waves | 46 |
| 🧊 Freezes | 14 |
| 🌀 Tropical Cyclones | 106 |
| ☔️ Atmospheric Rivers | 56 |
| 🌪️ Severe Convection | 115 |
| Total Cases | 337 |
EWB paper and talks
- AMS 2025 talk: https://ams.confex.com/ams/105ANNUAL/meetingapp.cgi/Paper/451220
- EWB paper is in preparation and will be submitted in late 2025
How do I suggest new data, metrics, or otherwise get involved?
We welcome your involvement! The success of a benchmark suite rests on community involvement and feedback. There are several ways to get involved:
- Get involved in community discussion using the discussion board
- Submit new code requests using the issues
- Send us email at hello@brightband.com
Installing EWB
Currently, the easiest way to install EWB is using the pip command:
$ pip install git+https://github.com/brightbandtech/ExtremeWeatherBench.git
It is highly recommend to use uv if possible:
$ git clone https://github.com/brightbandtech/ExtremeWeatherBench.git
$ cd ExtremeWeatherBench
$ uv sync
How to Run EWB
Running EWB on sample data (included) is straightforward.
Using command line initialization:
$ ewb --default
Note: this will run every event type, case, target source, and metric for the individual event type as they become available for GFS initialized FourCastNetv2. It is expected a full evaluation will take some time, even on a large VM.
Using Jupyter Notebook or a Script:
from extremeweatherbench import cases, inputs, metrics, evaluate, utils
# Select model
model = 'FOUR_v200_GFS'
# Set up path to directory of file - zarr or kerchunk/virtualizarr json/parquet
forecast_dir = f'gs://extremeweatherbench/{model}.parq'
# Preprocessing function exclusive to handling the CIRA parquets
def preprocess_bb_cira_forecast_dataset(ds: xr.Dataset) -> xr.Dataset:
"""Preprocess CIRA kerchunk (parquet) data in the ExtremeWeatherBench bucket.
A preprocess function that renames the time coordinate to lead_time,
creates a valid_time coordinate, and sets the lead time range and resolution not
present in the original dataset.
Args:
ds: The forecast dataset to rename.
Returns:
The renamed forecast dataset.
"""
ds = ds.rename({"time": "lead_time"})
# The evaluation configuration is used to set the lead time range and resolution.
ds["lead_time"] = np.array(
[i for i in range(0, 241, 6)], dtype="timedelta64[h]"
).astype("timedelta64[ns]")
return ds
# Define a forecast object; in this case, a KerchunkForecast
fcnv2_forecast = inputs.KerchunkForecast(
name="fcnv2_forecast", # identifier for this forecast in results
source=forecast_dir, # source path
variables=["surface_air_temperature"], # variables to use in the evaluation
variable_mapping=inputs.CIRA_metadata_variable_mapping, # mapping to use for variables in forecast dataset to EWB variable names
storage_options={"remote_protocol": "s3", "remote_options": {"anon": True}}, # storage options for access
preprocess=preprocess_bb_cira_forecast_dataset # required preprocessing function for CIRA references
)
# Load in ERA5; source defaults to the ARCO ERA5 dataset from Google and variable mapping is provided by default as well
era5_heatwave_target = inputs.ERA5(
variables=["surface_air_temperature"], # variable to use in the evaluation
storage_options={"remote_options": {"anon": True}}, # storage options for access
chunks=None, # define chunks for the ERA5 data
)
# EvaluationObjects are used to evaluate a single forecast source against a single target source with a defined event type. Event types are declared with each case. One or more metrics can be evaluated with each EvaluationObject.
heatwave_evaluation_list = [
inputs.EvaluationObject(
event_type="heat_wave",
metric_list=[
metrics.MaximumMeanAbsoluteError(),
metrics.RootMeanSquaredError(),
metrics.MaximumLowestMeanAbsoluteError(),
],
target=era5_heatwave_target,
forecast=fcnv2_forecast,
),
]
# Load in the EWB default list of event cases
case_metadata = cases.load_ewb_events_yaml_into_case_collection()
# Create the evaluation class, with cases and evaluation objects declared
ewb_instance = evaluate.ExtremeWeatherBench(
case_metadata=case_metadata,
evaluation_objects=heatwave_evaluation_list,
)
# Execute a parallel run and return the evaluation results as a pandas DataFrame
heatwave_outputs = ewb_instance.run(
parallel_config={'backend':'loky','n_jobs':16} # Uses 16 jobs with the loky backend
)
# Save the results
heatwave_outputs.to_csv('heatwave_evaluation_results.csv')
Owner metadata
- Name: Brightband
- Login: brightbandtech
- Email:
- Kind: organization
- Description: Brightband is making weather and climate predictable for all, to help humanity adapt to increasingly extreme weather.
- Website: brightband.com
- Location: United States of America
- Twitter: brightbandtech
- Company:
- Icon url: https://avatars.githubusercontent.com/u/159490539?v=4
- Repositories: 1
- Last ynced at: 2024-09-18T00:32:39.957Z
- Profile URL: https://github.com/brightbandtech
GitHub Events
Total
- Create event: 134
- Release event: 1
- Issues event: 71
- Watch event: 57
- Delete event: 115
- Issue comment event: 91
- Public event: 1
- Push event: 1168
- Pull request review event: 138
- Pull request review comment event: 207
- Pull request event: 166
- Fork event: 3
Last Year
- Create event: 135
- Release event: 1
- Issues event: 73
- Watch event: 58
- Delete event: 117
- Issue comment event: 95
- Public event: 1
- Push event: 1189
- Pull request review event: 138
- Pull request review comment event: 207
- Pull request event: 167
- Fork event: 3
Committers metadata
Last synced: 5 days ago
Total Commits: 596
Total Committers: 4
Avg Commits per committer: 149.0
Development Distribution Score (DDS): 0.029
Commits in past year: 530
Committers in past year: 4
Avg Commits per committer in past year: 132.5
Development Distribution Score (DDS) in past year: 0.026
| Name | Commits | |
|---|---|---|
| aaTman | m****r@g****m | 579 |
| Amy McGovern | a****n@o****u | 12 |
| Daniel Rothenberg | d****l@d****m | 3 |
| Ummara Ali Syeda | 8****2 | 2 |
Committer domains:
- danielrothenberg.com: 1
- ou.edu: 1
Issue and Pull Request metadata
Last synced: 6 days ago
Total issues: 63
Total pull requests: 194
Average time to close issues: about 1 month
Average time to close pull requests: 4 days
Total issue authors: 5
Total pull request authors: 5
Average comments per issue: 1.1
Average comments per pull request: 0.48
Merged pull request: 137
Bot issues: 0
Bot pull requests: 0
Past year issues: 62
Past year pull requests: 189
Past year average time to close issues: about 1 month
Past year average time to close pull requests: 4 days
Past year issue authors: 5
Past year pull request authors: 4
Past year average comments per issue: 1.11
Past year average comments per pull request: 0.46
Past year merged pull request: 133
Past year bot issues: 0
Past year bot pull requests: 0
Top Issue Authors
- aaTman (48)
- amymcgovern (8)
- darothen (3)
- hansmohrmann (2)
- alxmrs (2)
Top Pull Request Authors
- aaTman (179)
- amymcgovern (8)
- darothen (3)
- ummaraali2 (3)
- gideonite (1)
Top Issue Labels
- Improvement (6)
- documentation (3)
- bug (3)
- enhancement (3)
- good first issue (2)
- Feature (2)
- v1 (1)
Top Pull Request Labels
Package metadata
- Total packages: 2
- Total downloads: unknown
- Total dependent packages: 0 (may contain duplicates)
- Total dependent repositories: 0 (may contain duplicates)
- Total versions: 4
proxy.golang.org: github.com/brightbandtech/ExtremeWeatherBench
- Homepage:
- Documentation: https://pkg.go.dev/github.com/brightbandtech/ExtremeWeatherBench#section-documentation
- Licenses: mit
- Latest release: v0.2.0 (published 24 days ago)
- Last Synced: 2025-12-19T22:08:35.639Z (7 days ago)
- Versions: 2
- Dependent Packages: 0
- Dependent Repositories: 0
-
Rankings:
- Dependent packages count: 5.401%
- Average: 5.583%
- Dependent repos count: 5.764%
proxy.golang.org: github.com/brightbandtech/extremeweatherbench
- Homepage:
- Documentation: https://pkg.go.dev/github.com/brightbandtech/extremeweatherbench#section-documentation
- Licenses: mit
- Latest release: v0.2.0 (published 23 days ago)
- Last Synced: 2025-12-19T22:08:33.660Z (7 days ago)
- Versions: 2
- Dependent Packages: 0
- Dependent Repositories: 0
-
Rankings:
- Dependent packages count: 5.401%
- Average: 5.583%
- Dependent repos count: 5.764%
Dependencies
- actions/checkout v4 composite
- actions/checkout v3 composite
- actions/setup-python v5 composite
- actions/setup-python v3 composite
- astral-sh/setup-uv v4 composite
- pre-commit/action v3.0.1 composite
- cartopy >=0.24.1
- cftime >=1.6.4.post1
- dacite >=1.8.1
- dask [complete]>=2024.12.1
- fastparquet >=2024.11.0
- gcsfs >=2024.12.0
- geopandas >=1.0.1
- h5py >=3.12.1
- ipywidgets >=8.1.5
- kerchunk >=0.2.7
- numpy >=2.2.0
- pandas >=2.2.3
- pyyaml >=6.0.2
- regionmask >=0.13.0
- rioxarray >=0.18.1
- s3fs >=2024.12.0
- scikit-learn >=1.6.0
- scores >=2.0.0
- seaborn >=0.13.2
- shapely >=2.0.6
- tqdm >=4.67.1
- ujson >=5.10.0
- virtualizarr >=1.2.0
- xarray >=2024.11.0
- zarr >=2.18.4
- 157 dependencies
Score: -Infinity