@@ -791,13 +791,15 @@ def open_dataarray(
791
791
def open_mfdataset (
792
792
paths : str | NestedSequence [str | os .PathLike ],
793
793
chunks : T_Chunks | None = None ,
794
- concat_dim : str
795
- | DataArray
796
- | Index
797
- | Sequence [str ]
798
- | Sequence [DataArray ]
799
- | Sequence [Index ]
800
- | None = None ,
794
+ concat_dim : (
795
+ str
796
+ | DataArray
797
+ | Index
798
+ | Sequence [str ]
799
+ | Sequence [DataArray ]
800
+ | Sequence [Index ]
801
+ | None
802
+ ) = None ,
801
803
compat : CompatOptions = "no_conflicts" ,
802
804
preprocess : Callable [[Dataset ], Dataset ] | None = None ,
803
805
engine : T_Engine | None = None ,
@@ -1101,8 +1103,7 @@ def to_netcdf(
1101
1103
* ,
1102
1104
multifile : Literal [True ],
1103
1105
invalid_netcdf : bool = False ,
1104
- ) -> tuple [ArrayWriter , AbstractDataStore ]:
1105
- ...
1106
+ ) -> tuple [ArrayWriter , AbstractDataStore ]: ...
1106
1107
1107
1108
1108
1109
# path=None writes to bytes
@@ -1119,8 +1120,7 @@ def to_netcdf(
1119
1120
compute : bool = True ,
1120
1121
multifile : Literal [False ] = False ,
1121
1122
invalid_netcdf : bool = False ,
1122
- ) -> bytes :
1123
- ...
1123
+ ) -> bytes : ...
1124
1124
1125
1125
1126
1126
# compute=False returns dask.Delayed
@@ -1138,8 +1138,7 @@ def to_netcdf(
1138
1138
compute : Literal [False ],
1139
1139
multifile : Literal [False ] = False ,
1140
1140
invalid_netcdf : bool = False ,
1141
- ) -> Delayed :
1142
- ...
1141
+ ) -> Delayed : ...
1143
1142
1144
1143
1145
1144
# default return None
@@ -1156,8 +1155,7 @@ def to_netcdf(
1156
1155
compute : Literal [True ] = True ,
1157
1156
multifile : Literal [False ] = False ,
1158
1157
invalid_netcdf : bool = False ,
1159
- ) -> None :
1160
- ...
1158
+ ) -> None : ...
1161
1159
1162
1160
1163
1161
# if compute cannot be evaluated at type check time
@@ -1175,8 +1173,7 @@ def to_netcdf(
1175
1173
compute : bool = False ,
1176
1174
multifile : Literal [False ] = False ,
1177
1175
invalid_netcdf : bool = False ,
1178
- ) -> Delayed | None :
1179
- ...
1176
+ ) -> Delayed | None : ...
1180
1177
1181
1178
1182
1179
# if multifile cannot be evaluated at type check time
@@ -1194,8 +1191,7 @@ def to_netcdf(
1194
1191
compute : bool = False ,
1195
1192
multifile : bool = False ,
1196
1193
invalid_netcdf : bool = False ,
1197
- ) -> tuple [ArrayWriter , AbstractDataStore ] | Delayed | None :
1198
- ...
1194
+ ) -> tuple [ArrayWriter , AbstractDataStore ] | Delayed | None : ...
1199
1195
1200
1196
1201
1197
# Any
@@ -1212,8 +1208,7 @@ def to_netcdf(
1212
1208
compute : bool = False ,
1213
1209
multifile : bool = False ,
1214
1210
invalid_netcdf : bool = False ,
1215
- ) -> tuple [ArrayWriter , AbstractDataStore ] | bytes | Delayed | None :
1216
- ...
1211
+ ) -> tuple [ArrayWriter , AbstractDataStore ] | bytes | Delayed | None : ...
1217
1212
1218
1213
1219
1214
def to_netcdf (
@@ -1647,8 +1642,7 @@ def to_zarr(
1647
1642
zarr_version : int | None = None ,
1648
1643
write_empty_chunks : bool | None = None ,
1649
1644
chunkmanager_store_kwargs : dict [str , Any ] | None = None ,
1650
- ) -> backends .ZarrStore :
1651
- ...
1645
+ ) -> backends .ZarrStore : ...
1652
1646
1653
1647
1654
1648
# compute=False returns dask.Delayed
@@ -1671,8 +1665,7 @@ def to_zarr(
1671
1665
zarr_version : int | None = None ,
1672
1666
write_empty_chunks : bool | None = None ,
1673
1667
chunkmanager_store_kwargs : dict [str , Any ] | None = None ,
1674
- ) -> Delayed :
1675
- ...
1668
+ ) -> Delayed : ...
1676
1669
1677
1670
1678
1671
def to_zarr (
0 commit comments