Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Mean clim patch #920

Merged
merged 19 commits into from
Apr 16, 2023
Merged
Show file tree
Hide file tree
Changes from 3 commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 9 additions & 1 deletion pcmdi_metrics/mean_climate/lib/create_mean_climate_parser.py
Original file line number Diff line number Diff line change
Expand Up @@ -24,6 +24,14 @@ def create_mean_climate_parser():
required=False,
)

parser.add_argument(
"--varname_in_test_data",
type=ast.literal_eval,
dest="varname_in_test_data",
help="Variable name in input model file",
required=False,
)

parser.add_argument(
"--regions",
type=ast.literal_eval,
Expand Down Expand Up @@ -256,4 +264,4 @@ def create_mean_climate_parser():
required=False,
)

return parser
return parser
17 changes: 12 additions & 5 deletions pcmdi_metrics/mean_climate/lib/load_and_regrid.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,12 +3,13 @@
import xcdat as xc
import numpy as np

def load_and_regrid(data_path, varname, level=None, t_grid=None, decode_times=True, regrid_tool='regrid2', debug=False):
def load_and_regrid(data_path, varname, varname_in_file=None, level=None, t_grid=None, decode_times=True, regrid_tool='regrid2', debug=False):
"""Load data and regrid to target grid

Args:
data_path (str): full data path for nc or xml file
varname (str): variable name
varname_in_file (str): variable name if data array named differently
level (float): level to extract (unit in hPa)
t_grid (xarray.core.dataset.Dataset): target grid to regrid
decode_times (bool): Default is True. decode_times=False will be removed once obs4MIP written using xcdat
Expand All @@ -17,9 +18,12 @@ def load_and_regrid(data_path, varname, level=None, t_grid=None, decode_times=Tr
"""
if debug:
print('load_and_regrid start')


if varname_in_file is None:
varname_in_file = varname

# load data
ds = xcdat_open(data_path, data_var=varname, decode_times=decode_times) # NOTE: decode_times=False will be removed once obs4MIP written using xcdat
ds = xcdat_open(data_path, data_var=varname_in_file, decode_times=decode_times) # NOTE: decode_times=False will be removed once obs4MIP written using xcdat

# calendar quality check
if "calendar" in list(ds.time.attrs.keys()):
Expand Down Expand Up @@ -51,11 +55,14 @@ def load_and_regrid(data_path, varname, level=None, t_grid=None, decode_times=Tr

# regrid
if regrid_tool == 'regrid2':
ds_regridded = ds.regridder.horizontal(varname, t_grid, tool=regrid_tool)
ds_regridded = ds.regridder.horizontal(varname_in_file, t_grid, tool=regrid_tool)
elif regrid_tool in ['esmf', 'xesmf']:
regrid_tool = 'xesmf'
regrid_method = 'bilinear'
ds_regridded = ds.regridder.horizontal(varname, t_grid, tool=regrid_tool, method=regrid_method)
ds_regridded = ds.regridder.horizontal(varname_in_file, t_grid, tool=regrid_tool, method=regrid_method)

if varname != varname_in_file:
ds_regridded[varname] = ds_regridded[varname_in_file]

# preserve units
try:
Expand Down
23 changes: 18 additions & 5 deletions pcmdi_metrics/mean_climate/mean_climate_driver.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,6 +30,7 @@
test_data_set = parameter.test_data_set
realization = parameter.realization
vars = parameter.vars
varname_in_test_data = parameter.varname_in_test_data
reference_data_set = parameter.reference_data_set
target_grid = parameter.target_grid
regrid_tool = parameter.regrid_tool
Expand Down Expand Up @@ -63,6 +64,7 @@
elif isinstance(realization, str):
if realization.lower() in ["all", "*"]:
find_all_realizations = True
realizations = "Search for all realizations!!"
else:
realizations = [realization]

Expand All @@ -78,6 +80,7 @@
'test_data_set:', test_data_set, '\n',
'realization:', realization, '\n',
'vars:', vars, '\n',
'varname_in_test_data:', varname_in_test_data, '\n',
'reference_data_set:', reference_data_set, '\n',
'target_grid:', target_grid, '\n',
'regrid_tool:', regrid_tool, '\n',
Expand Down Expand Up @@ -146,6 +149,11 @@
print('varname:', varname)
print('level:', level)

if varname_in_test_data is not None:
varname_testdata = varname_in_test_data[varname]
else:
varname_testdata = varname

# set dictionary for .json record
result_dict = tree()

Expand All @@ -172,7 +180,7 @@
obs_dict[varname][ref_dataset_name]["template"])
print('ref_data_full_path:', ref_data_full_path)
# load data and regrid
ds_ref = load_and_regrid(ref_data_full_path, varname, level, t_grid, decode_times=False, regrid_tool=regrid_tool, debug=debug)
ds_ref = load_and_regrid(data_path=ref_data_full_path, varname=varname, level=level, t_grid=t_grid, decode_times=False, regrid_tool=regrid_tool, debug=debug)
ds_ref_dict = OrderedDict()
# for record in output json
result_dict['References'][ref] = obs_dict[varname][ref_dataset_name]
Expand All @@ -182,18 +190,21 @@
# ----------
for model in test_data_set:

print('=================================')
print('model, runs, find_all_realizations:', model, realizations, find_all_realizations)

result_dict["RESULTS"][model][ref]["source"] = ref_dataset_name

if find_all_realizations:
test_data_full_path = os.path.join(
test_data_path,
filename_template).replace('%(variable)', varname).replace('%(model)', model).replace('%(model_version)', model).replace('%(realization)', '*')
ncfiles = glob.glob(test_data_full_path)
print('test_data_full_path: ', test_data_full_path)
ncfiles = sorted(glob.glob(test_data_full_path))
realizations = []
for ncfile in ncfiles:
realizations.append(ncfile.split('/')[-1].split('.')[3])
print('=================================')
print('model, runs:', model, realizations)
print('realizations (after search): ', realizations)

for run in realizations:
# identify data to load (annual cycle (AC) data is loading in)
Expand All @@ -208,7 +219,7 @@
ds_test_dict = OrderedDict()

# load data and regrid
ds_test = load_and_regrid(test_data_full_path, varname, level, t_grid, decode_times=True, regrid_tool=regrid_tool, debug=debug)
ds_test = load_and_regrid(data_path=test_data_full_path, varname=varname, varname_in_file=varname_testdata, level=level, t_grid=t_grid, decode_times=True, regrid_tool=regrid_tool, debug=debug)
print('load and regrid done')
result_dict["RESULTS"][model]["units"] = ds_test[varname].units
result_dict["RESULTS"][model][ref][run]["InputClimatologyFileName"] = test_data_full_path.split('/')[-1]
Expand Down Expand Up @@ -279,6 +290,8 @@
)

except Exception as e:
if debug:
raise
print('error occured for ', model, run)
print(e)

Expand Down