Skip to content

Commit 0e0a907

Browse files
author
Darshan Prajapati
committed
Fix Tests
1 parent eda169e commit 0e0a907

File tree

4 files changed

+6
-4
lines changed

4 files changed

+6
-4
lines changed

weather_mv/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -205,7 +205,7 @@ weather-mv bq --uris "gs://your-bucket/*.grib" \
205205
--output_table $PROJECT.$DATASET_ID.$TABLE_ID \
206206
--xarray_open_dataset_kwargs '{"engine": "cfgrib", "indexpath": "", "backend_kwargs": {"filter_by_keys": {"typeOfLevel": "surface", "edition": 1}}}' \
207207
--temp_location "gs://$BUCKET/tmp" \
208-
--input_chunks '{ "time": 1 }'
208+
--input_chunks '{ "time": 1 }' \
209209
--direct_num_workers 2
210210
```
211211

weather_mv/loader_pipeline/bq.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -327,7 +327,7 @@ def df_to_rows(self, rows: pd.DataFrame, ds: xr.Dataset, uri: str) -> t.Iterator
327327
yield row
328328

329329
def chunks_to_rows(self, _, ds: xr.Dataset) -> t.Iterator[t.Dict]:
330-
logger.info(f"Processing for time: {ds['time'].values} and level: {ds['level'].values}")
330+
logger.info(f"Processing for time: {ds['time'].values}")
331331
uri = ds.attrs.get(DATA_URI_COLUMN, '')
332332
# Re-calculate import time for streaming extractions.
333333
if not self.import_time or self.zarr:

weather_mv/loader_pipeline/bq_test.py

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -205,7 +205,7 @@ class ExtractRowsTestBase(TestDataBase):
205205
def extract(self, data_path, *, variables=None, area=None, open_dataset_kwargs=None,
206206
import_time=DEFAULT_IMPORT_TIME, disable_grib_schema_normalization=False,
207207
tif_metadata_for_start_time=None, tif_metadata_for_end_time=None, zarr: bool = False, zarr_kwargs=None,
208-
skip_creating_polygon: bool = False) -> t.Iterator[t.Dict]:
208+
skip_creating_polygon: bool = False, input_chunks={ 'time': 1}) -> t.Iterator[t.Dict]:
209209
if zarr_kwargs is None:
210210
zarr_kwargs = {}
211211
op = ToBigQuery.from_kwargs(
@@ -215,7 +215,7 @@ def extract(self, data_path, *, variables=None, area=None, open_dataset_kwargs=N
215215
tif_metadata_for_start_time=tif_metadata_for_start_time,
216216
tif_metadata_for_end_time=tif_metadata_for_end_time, skip_region_validation=True,
217217
disable_grib_schema_normalization=disable_grib_schema_normalization, coordinate_chunk_size=1000,
218-
skip_creating_polygon=skip_creating_polygon)
218+
skip_creating_polygon=skip_creating_polygon, input_chunks=input_chunks)
219219
coords = op.prepare_coordinates(data_path)
220220
for uri, chunk in coords:
221221
yield from op.extract_rows(uri, chunk)
@@ -792,6 +792,7 @@ def test_extracts_rows(self):
792792
variables=list(), area=list(), xarray_open_dataset_kwargs=dict(), import_time=None, infer_schema=False,
793793
tif_metadata_for_start_time=None, tif_metadata_for_end_time=None, skip_region_validation=True,
794794
disable_grib_schema_normalization=False,
795+
input_chunks={'time': 1}
795796
)
796797

797798
with TestPipeline() as p:

weather_mv/loader_pipeline/pipeline_test.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -69,6 +69,7 @@ def setUp(self) -> None:
6969
'log_level': 2,
7070
'use_local_code': False,
7171
'skip_creating_polygon': False,
72+
'input_chunks': { 'time': 1 }
7273
}
7374

7475

0 commit comments

Comments
 (0)