|
43 | 43 | T_NetcdfEngine = Literal["netcdf4", "scipy", "h5netcdf"]
|
44 | 44 | T_Engine = Union[
|
45 | 45 | T_NetcdfEngine,
|
46 |
| - Literal["pydap", "pynio", "pseudonetcdf", "cfgrib", "zarr"], |
| 46 | + Literal["pydap", "pynio", "pseudonetcdf", "zarr"], |
47 | 47 | type[BackendEntrypoint],
|
48 | 48 | str, # no nice typing support for custom backends
|
49 | 49 | None,
|
|
64 | 64 | "h5netcdf": backends.H5NetCDFStore.open,
|
65 | 65 | "pynio": backends.NioDataStore,
|
66 | 66 | "pseudonetcdf": backends.PseudoNetCDFDataStore.open,
|
67 |
| - "cfgrib": backends.CfGribDataStore, |
68 | 67 | "zarr": backends.ZarrStore.open_group,
|
69 | 68 | }
|
70 | 69 |
|
@@ -387,7 +386,7 @@ def open_dataset(
|
387 | 386 | ends with .gz, in which case the file is gunzipped and opened with
|
388 | 387 | scipy.io.netcdf (only netCDF3 supported). Byte-strings or file-like
|
389 | 388 | objects are opened by scipy.io.netcdf (netCDF3) or h5py (netCDF4/HDF).
|
390 |
| - engine : {"netcdf4", "scipy", "pydap", "h5netcdf", "pynio", "cfgrib", \ |
| 389 | + engine : {"netcdf4", "scipy", "pydap", "h5netcdf", "pynio", \ |
391 | 390 | "pseudonetcdf", "zarr", None}, installed backend \
|
392 | 391 | or subclass of xarray.backends.BackendEntrypoint, optional
|
393 | 392 | Engine to use when reading files. If not provided, the default engine
|
@@ -479,7 +478,7 @@ def open_dataset(
|
479 | 478 | relevant when using dask or another form of parallelism. By default,
|
480 | 479 | appropriate locks are chosen to safely read and write files with the
|
481 | 480 | currently active dask scheduler. Supported by "netcdf4", "h5netcdf",
|
482 |
| - "scipy", "pynio", "pseudonetcdf", "cfgrib". |
| 481 | + "scipy", "pynio", "pseudonetcdf". |
483 | 482 |
|
484 | 483 | See engine open function for kwargs accepted by each specific engine.
|
485 | 484 |
|
@@ -576,7 +575,7 @@ def open_dataarray(
|
576 | 575 | ends with .gz, in which case the file is gunzipped and opened with
|
577 | 576 | scipy.io.netcdf (only netCDF3 supported). Byte-strings or file-like
|
578 | 577 | objects are opened by scipy.io.netcdf (netCDF3) or h5py (netCDF4/HDF).
|
579 |
| - engine : {"netcdf4", "scipy", "pydap", "h5netcdf", "pynio", "cfgrib", \ |
| 578 | + engine : {"netcdf4", "scipy", "pydap", "h5netcdf", "pynio", \ |
580 | 579 | "pseudonetcdf", "zarr", None}, installed backend \
|
581 | 580 | or subclass of xarray.backends.BackendEntrypoint, optional
|
582 | 581 | Engine to use when reading files. If not provided, the default engine
|
@@ -666,7 +665,7 @@ def open_dataarray(
|
666 | 665 | relevant when using dask or another form of parallelism. By default,
|
667 | 666 | appropriate locks are chosen to safely read and write files with the
|
668 | 667 | currently active dask scheduler. Supported by "netcdf4", "h5netcdf",
|
669 |
| - "scipy", "pynio", "pseudonetcdf", "cfgrib". |
| 668 | + "scipy", "pynio", "pseudonetcdf". |
670 | 669 |
|
671 | 670 | See engine open function for kwargs accepted by each specific engine.
|
672 | 671 |
|
@@ -803,7 +802,7 @@ def open_mfdataset(
|
803 | 802 | If provided, call this function on each dataset prior to concatenation.
|
804 | 803 | You can find the file-name from which each dataset was loaded in
|
805 | 804 | ``ds.encoding["source"]``.
|
806 |
| - engine : {"netcdf4", "scipy", "pydap", "h5netcdf", "pynio", "cfgrib", \ |
| 805 | + engine : {"netcdf4", "scipy", "pydap", "h5netcdf", "pynio", \ |
807 | 806 | "pseudonetcdf", "zarr", None}, installed backend \
|
808 | 807 | or subclass of xarray.backends.BackendEntrypoint, optional
|
809 | 808 | Engine to use when reading files. If not provided, the default engine
|
|
0 commit comments