@@ -139,7 +139,7 @@ def _get_90_days_filenames_for_dates(self, dates, max_days=90):
139
139
)
140
140
continue
141
141
142
- # Check next 5 days
142
+ # Check next 90 days
143
143
found = False
144
144
for days in range (1 , max_days + 1 ):
145
145
future_date = target + timedelta (days = days )
@@ -154,7 +154,7 @@ def _get_90_days_filenames_for_dates(self, dates, max_days=90):
154
154
break
155
155
156
156
if not found :
157
- self .logger .warning (
157
+ self .logger .error (
158
158
f"No available data within { max_days } days of { target } "
159
159
)
160
160
@@ -249,10 +249,11 @@ def process_historical_zipped_data(self, zipped_filepaths, dates):
249
249
250
250
for file in unzipped_files :
251
251
date = get_datetime_from_filename (file [0 ])
252
- self .logger .info (f"Processing historical { SFED } data from { date } " )
252
+ date_pretty = date .strftime ("%Y-%m-%d" )
253
+ self .logger .info (f"Processing historical { SFED } data from { date_pretty } " )
253
254
sfed_da = self .process_data (file [0 ], band_type = SFED )
254
255
255
- self .logger .info (f"Processing historical { MFED } data from { date } " )
256
+ self .logger .info (f"Processing historical { MFED } data from { date_pretty } " )
256
257
mfed_da = self .process_data (file [1 ], band_type = MFED )
257
258
self ._combine_bands (sfed_da , mfed_da , date = date )
258
259
@@ -329,7 +330,8 @@ def query_api(self, date):
329
330
return sfed_unzipped , mfed_unzipped , latest_date
330
331
331
332
def process_historical_data (self , filepath , date , band_type ):
332
- self .logger .info (f"Processing historical { band_type } data from { date } " )
333
+ date_pretty = date .strftime ("%Y-%m-%d" )
334
+ self .logger .info (f"Processing historical { band_type } data from { date_pretty } " )
333
335
334
336
with xr .open_dataset (filepath ) as ds :
335
337
ds = ds .transpose ("time" , "lat" , "lon" )
0 commit comments