use benedict to make loops easier

This commit is contained in:
Justus Kuhlmann 2023-10-09 16:06:13 +00:00
parent eb58920e34
commit fa3987479b

View file

@ -4,6 +4,8 @@ import re
import numpy as np # Thinly-wrapped numpy
from ..obs import Obs
from .utils import sort_names, check_idl
from benedict import benedict
from copy import deepcopy
def read_sfcf(path, prefix, name, quarks='.*', corr_type="bi", noffset=0, wf=0, wf2=0, version="1.0c", cfg_separator="n", silent=False, **kwargs):
@ -65,7 +67,9 @@ def read_sfcf(path, prefix, name, quarks='.*', corr_type="bi", noffset=0, wf=0,
list of Observables with length T, observable per timeslice.
bb-type correlators have length 1.
"""
ret = read_sfcf_multi(path, prefix, [name], quarks_list=[quarks], corr_type_list=[corr_type], noffset_list=[noffset], wf_list=[wf], wf2_list=[wf2], version=version, cfg_separator=cfg_separator, silent=silent, nice_output=True, **kwargs)
ret = read_sfcf_multi(path, prefix, [name], quarks_list=[quarks], corr_type_list=[corr_type],
noffset_list=[noffset], wf_list=[wf], wf2_list=[wf2], version=version,
cfg_separator=cfg_separator, silent=silent, nice_output=True, **kwargs)
return ret
@ -195,32 +199,11 @@ def read_sfcf_multi(path, prefix, name_list, quarks_list=['.*'], corr_type_list=
idl = []
# setup dect structures
intern = {}
for name, corr_type in zip(name_list, corr_type_list):
intern[name] = {}
b2b, single = _extract_corr_type(corr_type)
intern[name]["b2b"] = b2b
intern[name]["single"] = single
intern[name]["spec"] = {}
for quarks in quarks_list:
intern[name]["spec"][quarks] = {}
for off in noffset_list:
intern[name]["spec"][quarks][str(off)] = {}
for w in wf_list:
intern[name]["spec"][quarks][str(off)][str(w)] = {}
for w2 in wf2_list:
intern[name]["spec"][quarks][str(off)][str(w)][str(w2)] = {}
# setup dict structures
intern = _setup_intern_dict(name_list, corr_type_list, quarks_list, noffset_list, wf_list, wf2_list)
internal_ret_dict = {}
for name in name_list:
internal_ret_dict[name] = {}
for quarks in quarks_list:
internal_ret_dict[name][quarks] = {}
for off in noffset_list:
internal_ret_dict[name][quarks][str(off)] = {}
for w in wf_list:
internal_ret_dict[name][quarks][str(off)][str(w)] = {}
ret_key_paths, internal_ret_dict = _setup_ret_dict(name_list, quarks_list, noffset_list, wf_list, wf2_list)
dict_template = deepcopy(internal_ret_dict)
if not appended:
for i, item in enumerate(ls):
@ -249,20 +232,21 @@ def read_sfcf_multi(path, prefix, name_list, quarks_list=['.*'], corr_type_list=
if i == 0:
if version != "0.0" and compact:
file = path + '/' + item + '/' + sub_ls[0]
for name in name_list:
for k in ret_key_paths:
ori_keys = k.split("/")
name = ori_keys[0]
quarks = ori_keys[1]
off = ori_keys[2]
w = ori_keys[3]
w2 = ori_keys[4]
if version == "0.0" or not compact:
file = path + '/' + item + '/' + sub_ls[0] + '/' + name
for quarks in quarks_list:
for off in noffset_list:
for w in wf_list:
for w2 in wf2_list:
# here, we want to find the place within the file,
# where the correlator we need is stored.
# to do so, the pattern needed is put together
# from the input values
intern[name]["spec"][quarks][str(off)][str(w)][str(w2)]["pattern"] = _make_pattern(version, name, off, w, w2, intern[name]['b2b'], quarks)
start_read, T = _find_correlator(file, version, intern[name]["spec"][quarks][str(off)][str(w)][str(w2)]["pattern"], intern[name]['b2b'], silent=silent)
start_read, T = _find_correlator(file, version, intern[name]["spec"][quarks][off][w][w2]["pattern"], intern[name]['b2b'], silent=silent)
intern[name]["spec"][quarks][str(off)][str(w)][str(w2)]["start"] = start_read
intern[name]["T"] = T
# preparing the datastructure
@ -270,39 +254,31 @@ def read_sfcf_multi(path, prefix, name_list, quarks_list=['.*'], corr_type_list=
deltas = []
for j in range(intern[name]["T"]):
deltas.append([])
internal_ret_dict[name][quarks][str(off)][str(w)][str(w2)] = deltas
internal_ret_dict[k] = deltas
if compact:
rep_deltas = _read_compact_rep(path, item, sub_ls, intern, im)
for name in name_list:
for quarks in quarks_list:
for off in noffset_list:
for w in wf_list:
for w2 in wf2_list:
rep_deltas = _read_compact_rep(path, item, sub_ls, intern, ret_key_paths, im, dict_template)
for k in ret_key_paths:
name = k.split("/")[0]
for t in range(intern[name]["T"]):
internal_ret_dict[name][quarks][str(off)][str(w)][str(w2)][t].append(rep_deltas[name][quarks][str(off)][str(w)][str(w2)][t])
internal_ret_dict[k][t].append(rep_deltas[k][t])
else:
for name in name_list:
for name in internal_ret_dict:
rep_data = []
for cnfg, subitem in enumerate(sub_ls):
for quarks in quarks_list:
for off in noffset_list:
for w in wf_list:
for w2 in wf2_list:
for k in _get_deep_paths(internal_ret_dict.subset(name)):
cfg_path = path + '/' + item + '/' + subitem
file_data = _read_o_file(cfg_path, name, intern, version, im)
file_data = _read_o_file(cfg_path, name, intern, version, im, dict_template)
rep_data.append(file_data)
for quarks in quarks_list:
for off in noffset_list:
for w in wf_list:
for w2 in wf2_list:
for k in _get_deep_paths(internal_ret_dict.subset(name)[name]):
for t in range(intern[name]["T"]):
internal_ret_dict[name][quarks][str(off)][str(w)][str(w2)][t].append([])
internal_ret_dict[name][k][t].append([])
for cfg in range(no_cfg):
internal_ret_dict[name][quarks][str(off)][str(w)][str(w2)][t][i].append(rep_data[cfg][quarks][str(off)][str(w)][str(w2)][t])
internal_ret_dict[name][k][t][i].append(rep_data[cfg][k][t])
else:
for name in name_list:
for name in internal_ret_dict:
if "files" in kwargs:
ls = kwargs.get("files")
else:
@ -310,11 +286,14 @@ def read_sfcf_multi(path, prefix, name_list, quarks_list=['.*'], corr_type_list=
for exc in name_ls:
if not fnmatch.fnmatch(exc, prefix + '*.' + name):
name_ls = list(set(name_ls) - set([exc]))
name_ls = sort_names(name_ls)
for quarks in quarks_list:
for off in noffset_list:
for w in wf_list:
for w2 in wf2_list:
for k in _get_deep_paths(internal_ret_dict.subset(name)[name]):
ori_keys = k.split("/")
quarks = ori_keys[0]
off = ori_keys[1]
w = ori_keys[2]
w2 = ori_keys[3]
pattern = _make_pattern(version, name, off, w, w2, intern[name]['b2b'], quarks)
deltas = []
for rep, file in enumerate(name_ls):
@ -327,7 +306,7 @@ def read_sfcf_multi(path, prefix, name_list, quarks_list=['.*'], corr_type_list=
deltas.append([])
for t in range(intern[name]['T']):
deltas[t].append(rep_data[t])
internal_ret_dict[name][quarks][str(off)][str(w)][str(w2)] = deltas
internal_ret_dict[name][k] = deltas
if name == name_list[0]:
idl.append(rep_idl)
@ -344,37 +323,68 @@ def read_sfcf_multi(path, prefix, name_list, quarks_list=['.*'], corr_type_list=
if not silent:
print("Done")
result_dict = {}
for name in name_list:
result_dict[name] = {}
for quarks in quarks_list:
result_dict[name][quarks] = {}
for off in noffset_list:
result_dict[name][quarks][str(off)] = {}
for w in wf_list:
result_dict[name][quarks][str(off)][str(w)] = {}
for w2 in wf2_list:
result_dict = deepcopy(dict_template)
for k in ret_key_paths:
name = k.split("/")[0]
result = []
for t in range(intern[name]["T"]):
result.append(Obs(internal_ret_dict[name][quarks][str(off)][str(w)][str(w2)][t], new_names, idl=idl))
result_dict[name][quarks][str(off)][str(w)][str(w2)] = result
result.append(Obs(internal_ret_dict[k][t], new_names, idl=idl))
print(result)
result_dict[k] = result
if nice_output:
result_dict = _reduce_dict(result_dict)
return result_dict
def _read_o_file(cfg_path, name, intern, version, im):
def _setup_ret_dict(name_list, quarks_list, noffset_list, wf_list, wf2_list):
internal_ret_dict = {}
for name in name_list:
internal_ret_dict[name] = {}
for quarks in quarks_list:
internal_ret_dict[name][quarks] = {}
for off in noffset_list:
internal_ret_dict[name][quarks][str(off)] = {}
for w in wf_list:
internal_ret_dict[name][quarks][str(off)][str(w)] = {}
for w2 in wf2_list:
internal_ret_dict[name][quarks][str(off)][str(w)][str(w2)] = []
internal_ret_dict = benedict(internal_ret_dict, keypath_separator="/")
ret_key_paths = _get_deep_paths(internal_ret_dict)
return ret_key_paths, internal_ret_dict
def _setup_intern_dict(name_list, corr_type_list, quarks_list, noffset_list, wf_list, wf2_list):
intern = {}
for name, corr_type in zip(name_list, corr_type_list):
intern[name] = {}
b2b, single = _extract_corr_type(corr_type)
intern[name]["b2b"] = b2b
intern[name]["single"] = single
intern[name]["spec"] = {}
for quarks in quarks_list:
intern[name]["spec"][quarks] = {}
for off in noffset_list:
intern[name]["spec"][quarks][str(off)] = {}
for w in wf_list:
intern[name]["spec"][quarks][str(off)][str(w)] = {}
for w2 in wf2_list:
intern[name]["spec"][quarks][str(off)][str(w)][str(w2)] = {}
return intern
def _read_o_file(cfg_path, name, intern, version, im, template):
file = cfg_path + '/' + name
return_vals = {}
return_vals = deepcopy(template)
return_vals = return_vals.subset(name)[name]
print(return_vals)
with open(file) as fp:
lines = fp.readlines()
for quarks in intern[name]["spec"].keys():
return_vals[quarks] = {}
for off in intern[name]["spec"][quarks].keys():
return_vals[quarks][off] = {}
for w in intern[name]["spec"][quarks][off].keys():
return_vals[quarks][off][w] = {}
for w2 in intern[name]["spec"][quarks][off][w].keys():
for k in _get_deep_paths(return_vals):
ori_keys = k.split("/")
quarks = ori_keys[0]
off = ori_keys[1]
w = ori_keys[2]
w2 = ori_keys[3]
T = intern[name]["T"]
start_read = intern[name]["spec"][quarks][off][w][w2]["start"]
deltas = []
@ -468,19 +478,17 @@ def _find_correlator(file_name, version, pattern, b2b, silent=False):
return start_read, T
def _read_compact_file(rep_path, cfg_file, intern, im):
return_vals = {}
def _read_compact_file(rep_path, cfg_file, intern, im, template):
return_vals = deepcopy(template)
with open(rep_path + cfg_file) as fp:
lines = fp.readlines()
for name in intern.keys():
return_vals[name] = {}
for quarks in intern[name]["spec"].keys():
return_vals[name][quarks] = {}
for off in intern[name]["spec"][quarks].keys():
return_vals[name][quarks][off] = {}
for w in intern[name]["spec"][quarks][off].keys():
return_vals[name][quarks][off][w] = {}
for w2 in intern[name]["spec"][quarks][off][w].keys():
for key in _get_deep_paths(return_vals):
ori_keys = key.split("/")
name = ori_keys[0]
quarks = ori_keys[1]
off = ori_keys[2]
w = ori_keys[3]
w2 = ori_keys[4]
T = intern[name]["T"]
start_read = intern[name]["spec"][quarks][off][w][w2]["start"]
# check, if the correlator is in fact
@ -496,39 +504,29 @@ def _read_compact_file(rep_path, cfg_file, intern, im):
for k in range(6, T + 6):
floats = list(map(float, corr_lines[k].split()))
t_vals.append(floats[-2:][im])
return_vals[name][quarks][off][w][w2] = t_vals
return_vals[key] = t_vals
return return_vals
def _read_compact_rep(path, rep, sub_ls, intern, im_list):
def _read_compact_rep(path, rep, sub_ls, intern, keys, im_list, template):
rep_path = path + '/' + rep + '/'
no_cfg = len(sub_ls)
return_vals = {}
for name in intern.keys():
return_vals[name] = {}
for quarks in intern[name]["spec"].keys():
return_vals[name][quarks] = {}
for off in intern[name]["spec"][quarks].keys():
return_vals[name][quarks][off] = {}
for w in intern[name]["spec"][quarks][off].keys():
return_vals[name][quarks][off][w] = {}
for w2 in intern[name]["spec"][quarks][off][w].keys():
print(template)
return_vals = deepcopy(template)
for k in keys:
name = k.split("/")[0]
deltas = []
for t in range(intern[name]["T"]):
deltas.append(np.zeros(no_cfg))
return_vals[name][quarks][off][w][w2] = deltas
return_vals[k] = deltas
for cfg in range(no_cfg):
cfg_file = sub_ls[cfg]
cfg_data = _read_compact_file(rep_path, cfg_file, intern, im_list)
for name in intern.keys():
for quarks in intern[name]["spec"].keys():
for off in intern[name]["spec"][quarks].keys():
for w in intern[name]["spec"][quarks][off].keys():
for w2 in intern[name]["spec"][quarks][off][w].keys():
cfg_data = _read_compact_file(rep_path, cfg_file, intern, im_list, template)
for k in keys:
name = k.split("/")[0]
for t in range(intern[name]["T"]):
return_vals[name][quarks][off][w][w2][t][cfg] = cfg_data[name][quarks][off][w][w2][t]
return_vals[k][t][cfg] = cfg_data[k][t]
print(template)
return return_vals
@ -641,3 +639,13 @@ def _reduce_dict(d):
if len(list(d.keys())) == 1:
ret = d[list(d.keys())[0]]
return ret
def _get_deep_paths(f, sep='/'):
max_count = max([k.count(sep) for k in f.keypaths()])
deep_paths = []
for k in f.keypaths():
if k.count(sep) == max_count:
deep_paths.append(k)
return deep_paths